Javascript is required
Search
Volume 4, Issue 3, 2025
Open Access
Research article
Development of a Machine Learning-Driven Web Platform for Automated Identification of Rice Insect Pests
samuel n. john ,
nasiru a. musa ,
joshua s. mommoh ,
etinosa noma-osaghe ,
ukeme i. udioko ,
james l. obetta
|
Available online: 05-22-2025

Abstract

Full Text|PDF|XML

An advanced machine learning (ML)-driven web platform was developed and deployed to automate the identification of rice insect pests, addressing limitations associated with traditional pest detection methods and conventional ML algorithms. Historically, pest identification in rice cultivation has relied on expert evaluation of pest species and their associated crop damage, a process that is labor-intensive, time-consuming, and prone to inaccuracies, particularly in the misclassification of pest species. In this study, a subset of the publicly available IP102 benchmark dataset, consisting of 7,736 images across 12 rice pest categories, was curated for model training and evaluation. Two classification models—a Support Vector Machine (SVM) and a deep Convolutional Neural Network (CNN) based on the Inception_ResNetV2 architecture—were implemented and assessed using standard performance metrics. Experimental results demonstrated that the Inception_ResNetV2 model significantly outperformed SVM, achieving an accuracy of 99.97%, a precision of 99.46%, a recall of 99.81%, and an F1-score of 99.53%. Owing to its superior performance, the Inception_ResNetV2 model was integrated into a web-based application designed for real-time pest identification. The deployed system exhibited an average response time of 5.70 seconds, representing a notable improvement in operational efficiency and usability over previous implementations. The results underscore the potential of artificial intelligence in transforming agricultural practices by enabling accurate, scalable, and timely pest diagnostics, thereby enhancing pest management strategies, mitigating crop losses, and supporting global food security initiatives.

Open Access
Research article
Comparative Analysis of Machine Learning Models for Predicting Indonesia's GDP Growth
rossi passarella ,
muhammad ikhsan setiawan ,
zaqqi yamani
|
Available online: 07-03-2025

Abstract

Full Text|PDF|XML

Accurate forecasting of Gross Domestic Product (GDP) growth remains essential for supporting strategic economic policy development, particularly in emerging economies such as Indonesia. In this study, a hybrid predictive framework was constructed by integrating fuzzy logic representations with machine learning algorithms to improve the accuracy and interpretability of GDP growth estimation. Annual macroeconomic data from 1970 to 2023 were utilised, and 19 input features were engineered by combining numerical economic indicators with fuzzy-based linguistic variables, along with a forecast label generated via the Non-Stationary Fuzzy Time Series (NSFTS) method. Six supervised learning models were comparatively assessed, including Random Forest (RF), Support Vector Regression (SVR), eXtreme Gradient Boosting (XGBoost), Huber Regressor, Decision Tree (DT), and Multilayer Perceptron (MLP). Model performance was evaluated using Mean Absolute Error (MAE) and accuracy metrics. Among the tested models, the RF algorithm demonstrated superior performance, achieving the lowest MAE and an accuracy of 99.45% in forecasting GDP growth for 2023. Its robustness in capturing non-linear patterns and short-term economic fluctuations was particularly evident when compared to other models. These findings underscore the RF model's capability to serve as a reliable tool for economic forecasting in data-limited and volatile macroeconomic environments. By enabling more precise GDP growth predictions, the proposed hybrid framework offers a valuable decision-support mechanism for policymakers in Indonesia, contributing to more informed resource allocation, proactive economic intervention, and long-term development planning. The methodological innovation of integrating NSFTS with machine learning extends the frontier of data-driven macroeconomic modelling and provides a replicable template for forecasting applications in other emerging markets.

Abstract

Full Text|PDF|XML

The integration of artificial intelligence (AI) in precision agriculture has facilitated significant advancements in crop health monitoring, particularly in the early identification and classification of foliar diseases. Accurate and timely diagnosis of plant diseases is critical for minimizing crop loss and enhancing agricultural sustainability. In this study, an interpretable deep learning model—referred to as the Multi-Crop Leaf Disease (MCLD) framework—was developed based on a Convolutional Neural Network (CNN) architecture, tailored for the classification of tomato and grapevine leaf diseases. The model architecture was derived from the Visual Geometry Group Network (VGGNet), optimized to improve computational efficiency while maintaining classification accuracy. Leaf image datasets comprising healthy and diseased samples were employed to train and evaluate the model. Performance was assessed using multiple statistical metrics, including classification accuracy, sensitivity, specificity, precision, recall, and F1-score. The proposed MCLD framework achieved a detection accuracy of 98.40% for grapevine leaf diseases and a classification accuracy of 95.71% for tomato leaf conditions. Despite these promising results, further research is required to address limitations such as generalizability across variable environmental conditions and the integration of field-acquired images. The implementation of such interpretable AI-based systems is expected to substantially enhance precision agriculture by supporting rapid and accurate disease management strategies.

Abstract

Full Text|PDF|XML
Electroencephalography (EEG) provides a non-invasive approach for capturing brain dynamics and has become a cornerstone in clinical diagnostics, cognitive neuroscience, and neuroengineering. The inherent complexity, low signal-to-noise ratio, and variability of EEG signals have historically posed substantial challenges for interpretation. In recent years, artificial intelligence (AI), encompassing both classical machine learning (ML) and advanced deep learning (DL) methodologies, has transformed EEG analysis by enabling automatic feature extraction, robust classification, regression-based state estimation, and synthetic data generation. This survey synthesizes developments up to 2025, structured along three dimensions. The first dimension is task category, e.g., classification, regression, generation and augmentation, clustering and anomaly detection. The second dimension is the methodological framework, e.g., shallow learners, Convolutional Neural Networks (CNNs), Recurrent Neural Networks (RNNs), Transformers, Graph Neural Networks (GNNs), and hybrid approaches. The third dimension is application domain, e.g., neurological disease diagnosis, brain-computer interfaces (BCIs), affective computing, cognitive workload monitoring, and specialized tasks such as sleep staging and artifact removal. Publicly available EEG datasets and benchmarking initiatives that have catalyzed progress were reviewed in this study. The strengths and limitations of current AI models were critically evaluated, including constraints related to data scarcity, inter-subject variability, noise sensitivity, limited interpretability, and challenges of real-world deployment. Future research directions were highlighted, including federated learning (FL) and privacy-preserving learning, self-supervised pretraining of Transformer-based architectures, explainable artificial intelligence (XAI) tailored to neurophysiological signals, multimodal fusion with complementary biosignals, and the integration of lightweight on-device AI for continuous monitoring. By bridging historical foundations with cutting-edge innovations, this survey aims to provide a comprehensive reference for advancing the development of accurate, robust, and transparent AI-driven EEG systems.
Open Access
Research article
Application of Artificial Intelligence on MNIST Dataset for Handwritten Digit Classification for Evaluation of Deep Learning Models
jide ebenezer taiwo akinsola ,
micheal adeolu olatunbosun ,
Ifeoluwa Michael Olaniyi ,
moruf adedeji adeagbo ,
emmanuel ajayi olajubu ,
ganiyu adesola aderounmu
|
Available online: 09-18-2025

Abstract

Full Text|PDF|XML

Handwritten digit classification represents a foundational task in computer vision and has been widely adopted in applications ranging from Optical Character Recognition (OCR) to biometric authentication. Despite the availability of large benchmark datasets, the development of models that achieve both high accuracy and computational efficiency remains a central challenge. In this study, the performance of three representative machine learning paradigms—Chi-Squared Automatic Interaction Detection (CHAID), Generative Adversarial Networks (GANs), and Feedforward Deep Neural Networks (FFDNNs)—was systematically evaluated on the Modified National Institute of Standards and Technology (MNIST) dataset. The assessment was conducted with a focus on classification accuracy, computational efficiency, and interpretability. Experimental results demonstrated that deep learning approaches substantially outperformed traditional Decision Tree (DT) methods. GANs and FFDNNs achieved classification accuracies of approximately 97%, indicating strong robustness and generalization capability for handwritten digit recognition tasks. In contrast, CHAID achieved only 29.61% accuracy, highlighting the limited suitability of DT models for high-dimensional image data. It was further observed that, despite the computational demand of adversarial training, GANs required less time per epoch than FFDNNs when executed on modern GPU architectures, thereby underscoring their potential scalability. These findings reinforce the importance of model selection in practical deployment, particularly where accuracy, computational efficiency, and interpretability must be jointly considered. The study contributes to the ongoing discourse on the role of artificial intelligence (AI) in pattern recognition by providing a comparative analysis of classical machine learning and deep learning approaches, thereby offering guidance for the development of reliable and efficient digit recognition systems suitable for real-world applications.

- no more data -