
Evaluation of Parametric and Nonparametric Machine‐Learning Techniques for Prediction of Saturated and Near‐Saturated Hydraulic Conductivity
Author(s) -
Kotlar Ali Mehmandoost,
Iversen Bo V.,
Jong van Lier Quirijn
Publication year - 2019
Publication title -
vadose zone journal
Language(s) - English
Resource type - Journals
SCImago Journal Rank - 1.036
H-Index - 81
ISSN - 1539-1663
DOI - 10.2136/vzj2018.07.0141
Subject(s) - pedotransfer function , support vector machine , machine learning , hydraulic conductivity , mathematics , artificial intelligence , soil science , lasso (programming language) , bootstrapping (finance) , nonparametric statistics , linear regression , soil water , statistics , computer science , environmental science , econometrics , world wide web
Core Ideas Accurate Ks and K10 were obtained using water content data at several matric potentials. Robust K s and K 10 prediction by machine‐learning methods confirmed by bootstrapping. Gaussian process regression predicted K s and K 10 with minimum number of predictors. Parametric and nonparametric supervised machine learning techniques were used to estimate saturated and near‐saturated hydraulic conductivities ( K s and K 10 , respectively) from easily measurable soil properties including the name of the pedological horizon (HOR), soil texture (sand, silt, and clay), organic matter (OM), bulk density (BD), and water contents (θ pF1 , θ pF2 , θ pF3 , and θ pF4.2 ) measured at four different matric heads (−10, −100, −1000, and −15,848 cm, respectively). Using a stepwise linear model (SWLM) and the Lasso regression as parametric methods with 316 data in training and 135 data in the testing phase, four pedotransfer functions (PTFs) were obtained in which water contents for both methods play an important role compared with other variables. The SWLM showed better performance than Lasso in the testing phase for log( K s ) and log( K 10 ) prediction, with RMSE values of 0.666 and 0.551 cm d −1 and R 2 of 0.26 and 0.65. Nonparametric supervised machine learning methods trained and tested with a similar data set significantly improved the accuracy of K s prediction, with R 2 of 0.52, 0.36, and 0.53 for Gaussian process regression (GPR), support vector machine (SVM), and ensemble (ENS) methods in the testing stage. These methods also described 74.9, 66.7, and 72.5% of the variation of log( K 10 ). Bootstrapping validated the strong performance of nonparametric techniques. The feature selection capability of GPR determined that instead of using a model with all predictors, HOR, silt, θ pF1 , and θ pF3 are sufficient for the prediction of log( K s ), while HOR, silt, and OM can predict log( K 10 ) as accurate as the comprehensive model with all variables.