Application of Developing Artificial Intelligence (AI) Techniques to Model Pan Evaporation Trends in Slovak River Sub-Basins


The geographical positioning of Slovakia contributes to a transitional climate that integrates both maritime and continental elements, although altitude significantly influences the climate profile [37]. Among the primary climatic factors, air temperature, alongside atmospheric precipitation, plays a crucial role in determining the climatic conditions of a particular area. The Danubian Lowland emerges as the warmest region, characterized by an average annual air temperature of nearly 10 °C, as indicated by extensive long-term temperature records. Conversely, the average air temperature in the East Slovak Plain region is slightly lower in comparison. The average annual air temperature in river basins and valleys that are interconnected with the lowlands, such as Považie, Ponitrie, and Pohronie, typically falls within the range of 7 to 9 °C. In contrast, the highest basins, including Popradská and Oravská Kotlina, as well as northern Spiš, record average annual temperatures of less than 6 °C. It is observed that as altitude increases, there is a corresponding decrease in the average annual air temperature. This trend becomes particularly evident in locations situated at approximately 2000 m a. s. l. [37]. The geographic position, altitude, wind direction, and leeward aspect of mountains significantly influence atmospheric precipitation in the region. In Slovakia, the average annual precipitation exhibits considerable variation, ranging from approximately 2000 mm in the High Tatras to less than 500 mm in areas such as Galanta, Senec, and the eastern portion of Žitný Island. The phenomenon known as the precipitation shadow created by the mountains results in relatively low rainfall totals in these regions. As a result, the Spiš basins experience a notable decrease in moisture, being shielded from moist air masses originating from the south by the Slovak Ore Mountains and from the southwest to the northwest by the High and Low Tatras. On average, the region receives less than 600 mm of precipitation annually. In Slovakia, there is a noted trend of increasing precipitation with elevation. The mountains situated in the northwestern and northern parts of Slovakia typically experience higher levels of atmospheric precipitation compared to those found in the central, southern, and eastern regions. This phenomenon can be attributed to their increased exposure to prevailing north-westerly wind patterns. Furthermore, elevated atmospheric precipitation levels may also occur in the windward areas of mountains located further south during southern cyclonic conditions, a situation that is particularly prevalent in the eastern Slovak region of Vihorlat. The climate characteristics referred to in this context are based on data from the period 1961 to 2010, as detailed in the Climatic Conditions of Slovak Republic (2022) [38]. The criteria for the zoning of the Slovak Republic are elaborated upon in the study by Novotná et al. (2022) [29].

2.3. Statistics and Machine Learning Toolbox Application

The data analysis conducted in this study utilized MATLAB, version 2024a, Statistics and Machine Learning Toolbox [40]. Because of its robust computational capabilities, large library, and easy-to-use interface, MATLAB is a great option for artificial intelligence (AI) modeling. MATLAB is accessible to both novice and seasoned developers, enabling users to build and train AI models with a few lines of code or through low-code apps. To ensure thorough investigation and choose the optimal model for accuracy outcomes, we used all 28 ML and DL models that were accessible in MATLAB in the study. With this method, users can experiment with different modeling approaches and choose the one that best suits their needs.

At the same time, MATLAB is recognized as a high-performance programming language and interactive environment that is widely employed for numerical calculations, data analysis, algorithm development, and data visualization. The MATLAB R2024a version offers users an extensive range of machine learning models that are applicable for data analysis and predictive modeling. There are several benefits of simulating pan evaporation patterns using MATLAB in combination with machine learning (ML) and deep learning (DL) approaches, especially in environmental research. With a wealth of built-in functions and visualization capabilities, MATLAB offers an interactive, high-level environment that makes it easier to develop intricate algorithms and analyze big datasets. When ML and DL are combined, non-linear correlations between evaporation rates and meteorological variables can be modelled, improving forecast accuracy over conventional techniques. These cutting-edge methods are flexible, scalable for handling large amounts of data across time, and able to pinpoint important variables. When combined, MATLAB and these AI techniques improve water resource management’s real-time monitoring and decision making, resulting in more environmentally friendly responses to climatic unpredictability.

These models used in the study are as follows:

1. (LR): LR is a fundamental statistical technique employed to model the relationship between a dependent variable and one or more independent variables using a linear equation. This model forecasts outcomes based on a linear fit.

2. Interaction LR: This variant of linear regression incorporates interaction terms among variables. It evaluates not only the direct effects of predictor variables but also examines how the influence of one predictor variable varies with the level of another predictor.

3. Robust LR: This approach modifies traditional linear regression to minimize the influence of outliers within the dataset. Techniques such as Huber loss can be applied to enhance the model’s resistance to extreme values.

4. Stepwise LR: Stepwise linear regression represents an automated technique for selecting a subset of predictors by methodically adding or removing variables based on established criteria, such as the Akaike information criterion (AIC) or the Bayesian information criterion (BIC). The term “stepwise” denotes the iterative process of refining predictors to determine the most suitable model complexity.

5. Fine Tree: A fine tree is a type of decision tree characterized by its deep branching structure, which may lead to overfitting. While it excels at capturing intricate details within the dataset, its ability to generalize new data is often limited.

6. Medium Tree: The medium tree represents a decision tree that effectively balances depth and breadth. It demonstrates improved resistance to overfitting in comparison to a fine tree while still being capable of identifying significant patterns within the data.

7. Coarse Tree: A coarse tree is a simplified decision tree model featuring fewer splits and branches. This design enhances its generalizability and reduces the likelihood of overfitting when compared to both fine and medium trees.

8. Linear SVM: The linear support vector machine is designed to identify a linear hyperplane that effectively separates distinct classes within the feature space. This approach is particularly suitable for datasets that exhibit linear separability.

9. Quadratic SVM: This variant of support vector machine employs a quadratic kernel function to establish a decision boundary capable of addressing situations where the classes are not linearly separable in the original feature space.

10. Cubic SVM: The cubic support vector machine, akin to the quadratic variant, utilizes a cubic kernel function, thereby facilitating the formation of more intricate decision boundaries.

11. Fine Gaussian SVM: This model leverages a Gaussian radial basis function (RBF) kernel to manage high-dimensional data. The term “fine” denotes that the model has been meticulously calibrated to capture complex interrelationships within the data.

12. Medium Gaussian SVM: The medium Gaussian support vector machine achieves a balance between detail and generalization, effectively mitigating the risks of overfitting and underfitting.

13. Coarse Gaussian SVM: This model features a Gaussian support vector machine with a broader decision boundary, which reduces the likelihood of overfitting; however, it may also fail to detect subtle patterns present within the data.

14. EL Least Squares: This methodology represents an advanced optimization of linear regression, specifically designed to manage larger datasets with greater effectiveness by employing sophisticated numerical techniques for least squares fitting.

15. EL SVM: This term denotes an enhanced implementation of support vector machines (SVM) that emphasizes rapid convergence and computational efficiency, thereby rendering it well-suited for the analysis of large datasets.

16. Ensemble: Boosted Trees: An ensemble method that combines weak learners (typically, shallow trees) in a sequential manner, where each tree corrects the errors of the previous ones. This often leads to a powerful predictive model.

17. Ensemble: Bagged Trees: Utilizes bootstrap aggregating (bagging) to build multiple decision trees from random samples of the data. The final prediction is usually made by averaging or majority voting from all trees, enhancing robustness and reducing overfitting.

18. Squared Exponential Gaussian Process Regression: This approach to Gaussian process regression utilizes a squared exponential kernel to effectively model smooth functions. It is particularly adept at generating continuous and smooth predictions.

19. Matern 5/2 Gaussian Process Regression: This variant of Gaussian process regression employs the Matern kernel with a smoothness parameter of 5/2, providing enhanced flexibility in modeling functions that display varying levels of smoothness.

20. Exponential Gaussian Process Regression: This methodology incorporates an exponential kernel within Gaussian process regression, rendering it suitable for modeling functions that may exhibit limited smoothness but demonstrate exponential decay behavior.

21. Rational Quadratic Gaussian Process Regression: This Gaussian process utilizes the rational quadratic kernel, which integrates features from both squared exponential and linear kernels. This combination allows for a high degree of flexibility in capturing patterns with diverse smoothness characteristics.

22. Narrow NN: This neural network configuration comprises fewer neurons within each layer. Although this structure may limit its ability to capture intricate relationships, it offers advantages such as expedited training processes and a reduced likelihood of overfitting.

23. Medium NN: This architecture incorporates a moderate number of neurons, aiming to establish a balance between model complexity and the dynamics of training.

24. Wide NN: A wide neural network features a larger number of neurons in one or more layers, enabling it to identify complex patterns. However, this complexity may lead to challenges, particularly concerning overfitting.

25. Bilayered NN: This architecture consists of two primary layers, generally encompassing one hidden layer succeeded by an output layer. Its simplicity often facilitates easier interpretation.

26. Trilayered NN: This more elaborate structure includes three distinct layers: input, hidden, and output. The inclusion of an additional layer enhances the network’s capacity to learn more comprehensive representations of the input data.

27. SVM Kernel: The SVM Kernel refers to the kernel function employed in support vector machines, which facilitates the transformation of input data into a higher-dimensional space. This transformation enhances the separation of classes, thereby improving classification performance. Common types of kernels utilized in this context include linear, polynomial, and radial basis function (RBF) kernels.

28. Least Squares Regression Kernel: The least squares regression kernel is utilized in scenarios involving least squares fitting within kernelized feature spaces. This approach allows for the efficient management of regression tasks in high-dimensional environments, significantly enhancing computational efficacy and accuracy.

In the MATLAB R2024a environment, a diverse array of models is available, allowing users to select optimal methodologies tailored to their specific data characteristics and analytical objectives. Each model type presents distinct advantages and applies to various predictive tasks across different domains. To fulfil the objectives of this study, all the previously mentioned machine learning models have been employed. The evaluation of individual artificial intelligence methods will be conducted using the minimum redundancy maximum relevance (mRMR) method. This prominent feature selection technique in machine learning is particularly effective for enhancing model performance by identifying a subset of features that are relevant to the target variable while exhibiting minimal redundancy among themselves. This approach proves valuable in the context of high-dimensional datasets, where irrelevant or redundant features can diminish model accuracy and elevate computational complexity [41]. Various model settings were tested, but no better results were achieved than in automatic mode. The input data have been filtered. Fault data were deleted, and no interpolation was used. After analyzing the measured data, it was found that it is not possible to interpolate the measured data. Daily measurement averages and minimum and maximum values differed significantly according to the weather. Stations with a longer failure of measurements of monitored variables were excluded from the work. The input data were selected based on availability in the given basins. The setting parameters of AI models are shown in Appendix B.
Choosing the appropriate performance indicators is crucial since every indicator has its own properties. Since each performance indicator has unique characteristics, selecting the right ones is essential. Additionally, understanding each statistical measure’s capabilities might help one better appreciate how well the model performs. Therefore, many well-known statistical variables were used in this study to assess the prediction performance of the model. The model accuracy was assessed according to the RMSE (root mean square error), which measures the average error between predicted and actual values. Lower values indicate better model performance. The MSE (mean squared error) is a measure of the average squared error. R2 represents the proportion of variance in the dependent variable that can be predicted from the independent variables. The MSE ranges from 0 to 1, with higher values indicating better fit. The MAE (mean absolute error) measures the average absolute errors. Lower values are better. There are four basic metrics used to evaluate regression models. Each of them offers a unique insight into the accuracy and reliability of the model. The MSE evaluates the total error of the model. It emphasizes large errors (due to quadratic penalty). Quadratic penalty is useful when you want to emphasize large errors more. It has advantageous mathematical properties (e.g., derivatives for optimization). Sensitivity to outliers can significantly increase the MSE. The RMSE is the square root of the MSE and allows the error to be interpreted in the same units as the target variable. It is a more practical interpretation compared to the MSE because the resulting value is not quadratic. The MAE penalizes all errors equally. It is less sensitive to outliers than the MSE and RMSE. The indicators are defined in the study Abed et al. (2022) [31].



Source link

Beáta Novotná www.mdpi.com