Lab
Data Science & Big Data Lab
Institution: Pablo de Olavide University
About the lab
Featured research (7)
Quantum computing holds great promise for enhancing ma- chine learning algorithms, particularly by integrating classical and quan- tum techniques. This study compares two prominent quantum develop- ment frameworks, Qiskit and Pennylane, focusing on their suitability for hybrid quantum-classical support vector machines with quantum kernels. Our analysis reveals that Qiskit requires less theoretical information to be used, while Pennylane demonstrates superior performance in terms of execution time. Although both frameworks exhibit variances, our ex- periments reveal that Qiskit consistently yields superior classification accuracy compared to Pennylane when training classifiers with quantum kernels. Additionally, our results suggest that the performance of both frameworks remains stable for up to 20 qubits, indicating their suitability for practical applications. Overall, our findings provide valuable insights into the strengths and limitations of Qiskit and Pennylane for hybrid quantum-classical machine learning.
Water consumption forecasting is an essential tool for water management, as it allows for efficient planning and allocation of water resources, an undervalued but indispensable resource for all living beings. With the increasing demand for accurate and timely water forecasting, traditional forecasting methods are proving to be insufficient. Deep learning techniques, which have shown remarkable performance in a wide range of applications, offer a promising approach to address the challenges of water consumption forecasting. In this work, the use of deep learning models for medium-term water consumption forecasting of residential areas is explored. A deep feed-forward neural network is developed to predict water consumption of a company's customers for the next quarter. First, customers are grouped according to their consumption as these customers include both household consumers and special consumers such as public swimming pools, sports halls or small industries. Then, a deep feed-forward neural network is designed for household customers by obtaining the optimal values for those hyperparameters that have a great influence on the network performance. Results are reported using a real-world dataset composed of the water consumption from 1999 to 2015 on a quarterly basis, corresponding to 3262 clients of a water supply company. Finally, the proposed algorithm is evaluated by comparing it with other reference algorithms including an LSTM network.
Diabetic retinopathy is an eye complication of a widespread disease named diabetes mellitus. The most widely used method for diagnosing diabetic retinopathy is the analysis of retinal fundus images obtained by retinography. Deep Learning-based methods have shown promising results as a diagnostic tool for diabetic retinopathy, achieving, in some cases, performance close to the human inspection of images. However, the performance of these methods relies heavily on fine-tuning the algorithm hyperparameters and big data sets. In this work, we propose training a Deep Learning network with evolutionary algorithms to classify three stages of Diabetic Retinopathy: i) no sign of diabetic retinopathy, ii) Non-proliferative dia-betic retinopathy, and iii) proliferative diabetic retinopathy. We propose a neuroevolution methodology for selecting the most efficient Deep Learning model. The results of the neuroevolution methodology were improved by including Simulated Annealing strategies, Population Reinitialization, and ensembles. With high accuracy, sensitivity, specificity, and kappa index rates of 0.889, 0.889, 0.951, and 0.822, respectively, in the best case found, the experiments show that our neuroevolution methodology for selecting the Deep Learning model hyperparameters is a competitive alternative for training deep neural networks to classify three stages of diabetic retinopathy even with a small data set.
This paper proposes an application of the Automated Deep Learning model to predict the presence of olive flies in crops. Compared to baseline algorithms such as Random Forest or K-Nearest Neighbor, our Automated Deep Learning model demonstrates superior performance. Explainable Artificial Intelligence techniques such as Local Interpretable Model-Agnostic Explanations and Shapley Additive explanations are applied to interpret the results, revealing solar radiation as a key predictor for the presence of olive fly. This study enhances deep learning for agriculture, showcasing Automated Deep Learning superiority, and providing interpretable insights for effective pest management.Keywordsdeep learningexplainable artificial intelligenceagriculture
From an engineering point of view, non‐linear systems are essential to the operation of control systems, because all systems actually have a non‐linear state in nature. In reality, there are many different kinds of non‐linear systems hidden by this negative definition. For successful analysis and control, the identification of non‐linear systems using unknown models is typically necessary. Till now, numerous approaches are developed for identifying non‐linear systems, but it cannot be employed with a large number of components. Moreover, system identification is typically restricted to output and input signals alone, also such systems are rarely used in reality. This is the primary justification for using non‐linear systems in this research. So, this research proposed a non‐linear model of system identification for large‐scale systems under the consideration of two systems: bilinear system and Volterra system. Therefore, a novel algorithm named Self Adaptive Penguin Search Optimization (SAPeSO) is introduced to attain the system characteristics properly and minimize the output variation. Finally, the effectiveness of the proposed work is compared with existing works in terms of various error measures. This research mainly focuses on the application‐oriented engineering problems. In particular, the Mean Absolute Error (MAE) of the proposed work for the Volterra system at 4000 samples is 18.83%, 14.05%, 8.88%, 29.72%, 19.91%, and 6.70% which is better than the existing bald eagle search (BES), arithmetic optimization algorithm (AOA), whale optimization algorithm (WOA), nonlinear autoregressive moving average with exogenous inputs‐ frequency response function + principal component analysis (NARMAX‐FRF+PCA), Global Gravitational Search Algorithm‐Assisted Kalman Filter (CGS‐KF), and sparse regression and separable least squares method (SR‐SLSM) methods, respectively. Finally, the error is minimum for the proposed model when compared with the other traditional approaches.