Journal of Electronics, Electromedical Engineering, and Medical Informatics
http://jeeemi.org/index.php/jeeemi
<p>The Journal of Electronics, Electromedical Engineering, and Medical Informatics, (JEEEMI), is a peer-reviewed periodical scientific journal aimed at publishing research results of the Journal focus areas. The Journal is published by the Department of Electromedical Engineering, Health Polytechnic of Surabaya, Ministry of Health, Indonesia. The role of the Journal is to facilitate contacts between research centers and the industry. The aspiration of the Editors is to publish high-quality scientific professional papers presenting works of significant scientific teams, experienced and well-established authors as well as postgraduate students and beginning researchers. All articles are subject to anonymous review processes by at least two independent expert reviewers prior to publishing on the International Journal of Electronics, Electromedical Engineering, and Medical Informatics website.</p>Department of Electromedical Engineering, POLTEKKES KEMENKES SURABAYAen-US Journal of Electronics, Electromedical Engineering, and Medical Informatics2656-8632<p><strong>Authors who publish with this journal agree to the following terms:</strong></p> <ol> <li class="show">Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution-ShareAlikel 4.0 International <a title="CC BY SA" href="https://creativecommons.org/licenses/by-sa/4.0/" target="_blank" rel="noopener">(CC BY-SA 4.0)</a> that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.</li> <li class="show">Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.</li> <li class="show">Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See <a href="http://opcit.eprints.org/oacitation-biblio.html" target="_new">The Effect of Open Access</a>).</li> </ol>Predicting Construction Costs with Machine Learning: A Comparative Study on Ensemble and Linear Models
http://jeeemi.org/index.php/jeeemi/article/view/799
<p><strong>Accurate prediction of construction costs plays a pivotal role in ensuring successful project delivery, influencing budget formulation, resource allocation, and financial risk management. However, traditional estimation methods often struggle to handle complex, nonlinear relationships inherent in construction datasets. This study proposes a process innovation by systematically evaluating six machine learning (ML) models, i.e., Ridge Regression, Lasso Regression, Elastic Net, K-Nearest Neighbors (KNN), XGBoost, and CatBoost, on a standardized RSMeans dataset comprising 4,477 real-world construction data points. The primary aim is to benchmark the predictive performance, generalizability, and stability of both linear and ensemble models in construction cost forecasting. Each model is subjected to rigorous hyperparameter tuning using grid search with 5-fold cross-validation. Performance is assessed using <em>R</em>² (coefficient of determination), RMSE (root mean squared error), and MBE (mean bias error), while confidence intervals are computed to quantify predictive uncertainty. Results indicate that linear models achieve modest accuracy (<em>R</em>² ≈ 0.83), but struggle to model nonlinear interactions. In contrast, ensemble-based models outperform significantly, i.e., XGBoost and CatBoost achieve R² values of 0.988 and 0.987, respectively, RMSE values below 0.5, and near-zero MBE. Moreover, confidence interval visualization and feature importance analysis provide transparency and interpretability, enhancing the models’ practical applicability. Unlike prior studies that compare models in isolation, this work introduces a unified, interpretable framework and highlights the trade-offs between accuracy, overfitting, and deployment readiness. The findings have real-world implications for contractors, project managers, and cost engineers seeking reliable, data-driven decision support systems. In summary, this study introduces a scalable and robust ML-based framework that offers process innovation in construction cost estimation, paving the way for more intelligent, efficient, and risk-aware construction project management.</strong></p>Lifei ChenSew Sun TiangKim Soon ChongAbhishek SharmaTarek BerghoutWei Hong Lim
Copyright (c) 2025 Lifei Chen, Sew Sun Tiang, Kim Soon Chong, Abhishek Sharma, Tarek Berghout, Wei Hong Lim
https://creativecommons.org/licenses/by-sa/4.0
2025-05-202025-05-207362464510.35882/jeeemi.v7i3.799Power-Efficient 8-Bit ALU Design Using Squirrel Search and Swarm Intelligence Algorithms
http://jeeemi.org/index.php/jeeemi/article/view/822
<p><strong>The Arithmetic Logic Unit (ALU) serves as a core digital computing element which performs arithmetic functions along with logic operations. The normal operation of ALU designs leads to increased power consumption because of signal redundancy and continuous operation when new data inputs are unavailable. The research implements the Squirrel Search Algorithm (SSA) combined with Swarm Intelligence Algorithm (SIA) for 8-bit ALU optimization to achieve maximum resource efficiency alongside computational accuracy. The optimization properties of SSA and SIA make them ideal choices for digital circuit design applications because they yielded successful results in power-aware systems. The proposed method utilizes SSA-based conditional execution paired with SIA-based transition minimization to direct operations to execute only during fluctuating input data conditions thus eliminating undesired calculations. Studies confirm SSA and SIA function more effectively than distributed clock gating for power saving because they enable runtime-dependent optimization without creating significant computational overhead. The experimental Xilinx Vivado tests executed on an AMD Spartan-7 FPGA (XC7S50FGGA484) running at 100 MHz frequency established that SSA eliminates power consumption from 6 mW to 2 mW, and SIA achieves a power level of 4 mW. The SSA algorithm generates worst negative slack (WNS) values of 8.740 ns which SIA produces as 6.531 ns improving system timing performance. SSA-optimized ALU requires the same number of LUTs as the unoptimized design at 42 LUTs yet SIA uses 50 LUTs because of added logical elements. We observe no changes in flip-flop use during SSA where nine FFs remain yet SIA shows an increase in its usage up to 29 FFs due to input tracking. The study proves that bio-inspired methods create energy-efficient platforms which make them ideal for implementing ALU designs with FPGAs. Research studies demonstrate that hybrid swarm intelligence techniques represent an unexplored potential to optimize power-efficient architectures thus reinforcing their significance for future high-performance energy-efficient digital systems.</strong></p>Ashish PasayaSarman HadiaKiritkumar Bhatt
Copyright (c) 2025 Sarman Hadia, Ashish Pasaya, Kiritkumar Bhatt
https://creativecommons.org/licenses/by-sa/4.0
2025-05-282025-05-287366367910.35882/jeeemi.v7i3.822Advanced Traffic Flow Optimization Using Hybrid Machine Learning and Deep Learning Techniques
http://jeeemi.org/index.php/jeeemi/article/view/948
<p><strong>Road traffic congestion remains a persistent and critical challenge in modern urban environments, adversely affecting travel times, fuel consumption, air quality, and overall urban livability. To address this issue, this study proposes a hybrid ensemble learning framework for accurate short-term traffic flow prediction across signalized urban intersections. The model integrates Random Forest, Gradient Boosting, and Multi-Layer Perceptron within a weighted voting ensemble mechanism, wherein model contributions are dynamically scaled based on individual validation performance. Benchmarking is performed against traditional and advanced baselines, including Linear Regression, Support Vector Regression, and Long Short-Term Memory (LSTM) networks. A real-world traffic dataset, comprising 56 consecutive days of readings from six intersections, is utilized to validate the approach. A robust preprocessing pipeline is implemented, encompassing anomaly detection, temporal feature engineering especially time-of-day and day-of-week normalization, and sliding window encoding to preserve temporal dependencies. Experimental evaluations on 4-intersection and 6-intersection scenarios reveal that the ensemble consistently outperforms all baselines, achieving a peak R² of 0.954 and an RMSE of 0.045. Statistical significance testing using Welch’s t-test confirms the reliability of these improvements. Furthermore, SHAP-based interpretability analysis reveals the dominant influence of temporal features during high-variance periods. While computational overhead and data sparsity during rare events remain limitations, the framework demonstrates strong applicability for deployment in smart traffic systems. Its predictive accuracy and model transparency make it a viable candidate for adaptive signal control, congestion mitigation, and urban mobility planning. Future work may explore real-time streaming adaptation, external event integration, and generalization across heterogeneous urban networks.</strong></p>Mohammed El Kaim BillahAbdelfettah Mabrouk
Copyright (c) 2025 Mohammed El Kaim Billah, Abdelfettah Mabrouk
https://creativecommons.org/licenses/by-sa/4.0
2025-06-272025-06-277381783410.35882/jeeemi.v7i3.948Grad-CAM based Visualization for Interpretable Lung Cancer Categorization using Deep CNN Models
http://jeeemi.org/index.php/jeeemi/article/view/690
<p>The Grad-CAM (Gradient-weighted Class Activation Mapping) technique has loomed as a crucial tool for elucidating deep learning models, particularly convolutional neural networks (CNNs), by visually accentuating the regions of input images that accord most to a model's predictions. In the context of lung cancer histopathological image classification, this approach provides discernment into the decision-making process of models like InceptionV3, XceptionNet, and VGG19. These CNN architectures, renowned for their high performance in image categorization tasks, can be leveraged for automated diagnosis of lung cancer from histopathological images. By applying Grad-CAM to these models, heatmaps can be generated that divulge the areas of the tissue samples most influential in categorizing the images as lung adenocarcinomas, squamous cell carcinoma, and benign patches. This technique allows for the visualization of the network's focus on specific regions, such as cancerous cells or abnormal tissue structures, which may otherwise be difficult to explicate. Using pre-trained models fine-tuned for the task, the Grad-CAM method assesses the gradients of the target class concerning the final convolutional layer, generating a heatmap that can be overlaid on the input image. The results of Grad-CAM for InceptionV3, XceptionNet, and VGG19 offer distinct insights, as each model has unique characteristics. InceptionV3 pivots on multi-scale features, XceptionNet apprehend deeper patterns with separable convolutions, and VGG19 emphasizes simpler, more global attributes. By justaposing the heatmaps generated by each architecture, one can assess the model’s focus areas, facilitating better comprehension and certainty in the model's prophecy, crucial for clinical applications. Ultimately, the Grad-CAM approach not only intensify model transparency but also aids in ameliorating the interpretability of lung cancer diagnosis in histopathological image categorization.</p>Rashmi MothkurPullagura SoubhagyalakshmiSwetha C. B.
Copyright (c) 2025 Rashmi Mothkur, Pullagura Soubhagyalakshmi, Swetha C. B.
https://creativecommons.org/licenses/by-sa/4.0
2025-05-032025-05-037356758010.35882/jeeemi.v7i3.690Applied Machine Learning in EEG data Classification to Classify Major Depressive Disorder by Critical Channels
http://jeeemi.org/index.php/jeeemi/article/view/719
<p><strong>The electroencephalogram (EEG) stands out as a promising non-invasive tool for assessing depression. However, the efficient selection of channels is crucial for pinpointing key channels that can differentiate between different stages of depression within the vast dataset. This study outcome a comprehensive strategy for optimizing EEG channels to classify Major Depressive Disorder (MDD) using machine learning (ML) and deep learning (DL) approaches, and monitor effect of central lobe channels. A thorough review underscores the vital significance of EEG channel selection in the analysis of mental disorders. Neglecting this optimization step could result in heightened computational expenses, squandered resources, and potentially inaccurate classification results. Our assessment encompassed a range of techniques, such as Asymmetric Variance Ratio (AVR), Amplitude Asymmetry Ratio (AAR), Entropy-based selection employing Probability Mass Function (PMF), and Recursive Feature Elimination (RFE) where, RFE exhibited superior performance, particularly in pinpointing the most pertinent EEG channels while including central lobe channels like Fz, Cz, and Pz. With this accuracy between 97 to 99% is recorded by Electroencephalography Neural Network (EEGNet). Our experimental findings indicate that, models using RFE achieved enhancement in accuracy to classifying depressive disorders across diverse classifiers: EEGNet (96%), Random Forest (95%), Long Short-Term Memory (LSTM: 97.4%), 1D-CNN with 95%, and Multi-Layer Perceptron (98%) irrespective of central lobe incorporation. A pivotal contribution of this research is the development of a robust Multilayer Perceptron (MLP) model trained on EEG data from 382 participants, achieved accuracy of 98.7%, with a perfect precision score of 1.00, F1-Score of 0.983, and a Recall-Score of 0.966, to make it an enhanced technique for depression classification. Significant channels identified include Fp1, Fp2, F7, F4, F8, T3, C3, Cz, T4, T5, and P3, offering critical insights about depression. Our findings shows that, optimized EEG channel selection via RFE enhances depression classification accuracy in the field of brain-computer interface.</strong></p>Sudhir DhekaneAnand Khandare
Copyright (c) 2025 SUDHIR DHEKANE, Anand Khandare
https://creativecommons.org/licenses/by-sa/4.0
2025-05-052025-05-057358159610.35882/jeeemi.v7i3.719Computational Analysis of Medical Image Generation Using Generative Adversarial Networks (GANs)
http://jeeemi.org/index.php/jeeemi/article/view/784
<p>The limited availability of diverse, high-quality medical images constitutes a significant obstacle to training reliable deep-learning models that can be used in clinical settings. The traditional methods used for data augmentation generate inadequate medical photos that result in poor model performance and a low rate of successful generalization. This research studies the effectiveness of DCGAN cGAN CycleGAN and SRGAN GAN architectures through performance testing in five essential medical imaging datasets, including Diabetic Retinopathy, Pneumonia and Brain Tumor and Skin Cancer and Leukemia. The main achievement of this research was to perform an extensive evaluation of these GAN models through three key metrics: generation results, training loss metrics, and computational resource utilization. DCGAN generated stable high-quality synthetic images, whereas its generator produced losses from 0.59 (Pneumonia) to 6.24 (Skin Cancer), and its discriminator output losses between 0.29 and 6.25. CycleGAN showed the best convergence potential for Diabetic Retinopathy with generator and discriminator losses of 2.403 and 2.02 and Leukemia with losses at 3.325 and 3.129. The SRGAN network produced high-definition images at a generator loss of 6.253 and discriminator loss of 6.119 for the Skin Cancer dataset. Still, it failed to maintain crucial medical characteristics in grayscale images. GCN exhibited stable performance across all loss metrics and datasets. The DCGAN model required the lowest computing resources for 4 to 7 hours, using 0.9M and 1.4M parameters. The framework of SRGAN consumed between 7 and 10 hours and needed 1.7M to 2.3M parameters for its operation, and CycleGAN required identical computational resources. DCGAN was determined as the ideal model for synthetic medical image generation since it presented an optimal combination of quality output and resource efficiency. The research indicates that using DCGAN-generated images to increase medical datasets serves as a solution for boosting AI-based diagnostic system capabilities within healthcare.</p>Shrina PatelAshwin Makwana
Copyright (c) 2025 Shrina Patel, Ashwin Makwana
https://creativecommons.org/licenses/by-sa/4.0
2025-05-082025-05-087359761010.35882/jeeemi.v7i3.784Breast Cancer Classification on Ultrasound Images Using DenseNet Framework with Attention Mechanism
http://jeeemi.org/index.php/jeeemi/article/view/779
<p><strong>Breast cancer is one of </strong><strong>the most prevalent and life-threatening diseases</strong><strong> among women worldwide. Early detection of breast cancer being critical for increasing survival rates. </strong><strong>Ultrasound image is commonly used for breast cancer screening</strong><strong> due to its non-invasive, safe, and cost-effective. However, ultrasound images are often of low quality and have significant noise, which can hinder the effectiveness of classification models. This study proposes an enhanced breast cancer classification model that leverages transfer learning in combination with attention mechanisms to improve diagnostic performance. The main contribution of this research is the introduction of Dense-SASE, a novel architecture that combines DenseNet-121 with two powerful attention modules: Scaled-Dot Product Attention and Squeeze-and-Excitation (SE) Block. These mechanisms are integrated to improve feature representation and allow the model to focus on the most relevant regions of the ultrasound images. The proposed method was evaluated on a publicly available breast ultrasound image dataset, with classification performed across three categories: normal, benign, and malignant. Experimental results demonstrate that the Dense-SASE model achieves an accuracy of 98.29%, a precision of 97.97%, a recall of 98.98%, and an F1-score of 98.44%</strong><strong>.</strong><strong> Additionally, Grad-CAM visualizations demonstrated the model's capability to localize lesion areas effectively, avoiding non-informative regions, and confirming the model's interpretability. In conclusion, the Dense-SASE model significantly improves the accuracy and reliability of breast cancer classification in ultrasound images. By effectively learning and focusing on clinically relevant features, this approach offers a promising solution for computer-aided diagnosis (CAD) systems and has the potential to assist radiologists in early and accurate breast cancer detection.</strong></p>Hanina Nafisa AzkaWiharto WihartoEsti Suryani
Copyright (c) 2025 Wiharto, Hanina Nafisa Azka, Esti Suryani
https://creativecommons.org/licenses/by-sa/4.0
2025-05-092025-05-097361162310.35882/jeeemi.v7i3.779Classification of Cervical Cell Types Based on Machine Learning Approach: A Comparative Study
http://jeeemi.org/index.php/jeeemi/article/view/829
<p>Cervical cancer remains a major global health issue and is the second most common cancer affecting women worldwide. Early detection is crucial for effective treatment but remains challenging due to the asymptomatic nature of the disease and the visual complexity of cervical cell structures, which are often affected by inconsistent staining, poor contrast, and overlapping cells. This study aims to classify cervical cell images using Artificial Intelligence (AI) techniques by comparing the performance of Convolutional Neural Networks (CNN), Support Vector Machine (SVM), and K-Nearest Neighbors (KNN). The Herlev Pap smear image dataset was used for experimentation. In the preprocessing phase, images were resized to 100 × 100 pixels and enhanced through grayscale conversion, Gaussian smoothing for noise reduction, contrast stretching, and intensity normalization. Segmentation was performed using region-growing and active contour methods to accurately isolate cell nuclei. All classifiers were implemented using MATLAB. Experimental results show that CNN achieved the highest performance, with an accuracy of 85%, precision of 86.7%, and sensitivity of 83%, outperforming both SVM and KNN. These findings indicate that CNN is the most effective approach for cervical cell classification in this study. However, limitations such as class imbalance and occasional segmentation inconsistencies impacted overall performance, particularly in detecting abnormal cells. Future work will focus on improving classification accuracy especially for abnormal samples by exploring data augmentation techniques such as Generative Adversarial Networks (GANs) and implementing ensemble learning strategies. Additionally, integrating the proposed system into a real-time diagnostic platform using a graphical user interface (GUI) could support clinical decision-making and enhance cervical cancer screening programs.</p>Wan Azani Wan MustafaKhalis KhiruddinKhairur Rijal JamaludinFirdaus Yuslan KhusairiShahrina Ismail
Copyright (c) 2025 Wan Azani Mustafa, Khalis Khiruddin, Khairur Rijal Jamaludin, Firdaus Yuslan Khusairi and Shahrina Ismail
https://creativecommons.org/licenses/by-sa/4.0
2025-05-202025-05-207364666210.35882/jeeemi.v7i3.829Dual Attention and Channel Atrous Spatial Pyramid Pooling Half-UNet for Polyp Segmentation
http://jeeemi.org/index.php/jeeemi/article/view/893
<p><strong>Colorectal cancer (CRC) is a leading cause of cancer-related deaths, with two million cases detected in 2020 and causing one million deaths annually. Approximately 95% of CRC cases originate from colorectal adenomatous polyps. Early detection through accurate polyp segmentation is crucial for preventing and treating CRC effectively. While colonoscopy screening remains the primary detection method, its limitations have prompted the development of Computer-Aided Diagnostic (CAD) systems enhanced by deep learning models. This study proposes a novel neural network architecture called Dual Attention and Channel Atrous Spatial Pyramid Pooling Half-UNet (DACHalf-UNet) for medical polyp image segmentation that balances optimal performance with computational efficiency. The proposed model builds upon the U-Net framework by integrating Double Squeeze-and-Excitation (DSE) blocks in the encoder after the Ghost Module, Channel Atrous Spatial Pyramid Pooling (CASPP) in the bottleneck and decoder, and Attention Gate (AG) mechanisms within the architecture. DACHalf-UNet was trained and evaluated on the CVC-ClinicDB and Kvasir-SEG datasets for 70 epochs. Evaluations demonstrated superior performance with F1-Score and IoU values of 94.23% and 89.28% on CVC-ClinicDB, and 88.40% and 81.47% on Kvasir-SEG, respectively. Comparative analysis showed that DACHalf-UNet outperforms existing architectures including U-Net, U-Net++, ResU-Net, AGU-Net, CSAP-UNet, PRCNet, UNeXt, and UNeSt. Notably, the model achieves this performance with only 0.56 million trainable parameters and 30.29 GFLOPs, significantly reducing computational complexity compared to previous methods. These results demonstrate that DACHalf-UNet effectively addresses the need for accurate and efficient polyp segmentation, potentially enhancing CAD systems and contributing to improved CRC detection and treatment outcomes.</strong></p>Beatrix Datu SariraHeri Prasetyo
Copyright (c) 2025 Beatrix Datu Sarira, Heri Prasetyo
https://creativecommons.org/licenses/by-sa/4.0
2025-05-282025-05-287368069110.35882/jeeemi.v7i3.893Performance Evaluation of Classification Algorithms for Parkinson’s Disease Diagnosis: A Comparative Study
http://jeeemi.org/index.php/jeeemi/article/view/713
<p>Selection and implementation of classification algorithms along with proper preprocessing methods are important for the accuracy of predictive models. This paper compares some well-known and frequently used algorithms for classification tasks and performs in depth analysis. In this study we analyzed four most frequently used algorithm viz random forest (RF), decision tree (DT), logistic regression (LR) and support vector machine (SVM). To conduct the study on the well-known Oxford Parkinson’s disease Detection dataset obtained from the UCI Machine Learning Repository. We evaluated the algorithms' performance using six distinct approaches. Firstly, we used the classifiers where we didn’t used any method to enhance the performance of the classifier. Secondly, we applied Principal Component Analysis (PCA) to minimize the dimensionality of the dataset. Thirdly, we used collinearity-based feature elimination (CFE) method where we applied correlation among the features and if the correlation between a pair of features exceeds the threshold of 0.9, we eliminated one from the pair. Fourthly, we adopt synthetic minority oversampling technique (SMOTE) to synthetically increase the instances of the minority class. Fifth, we combined PCA+SMOTE and on sixth method, we combined CFE + SMOTE. The study demonstrates that SVM is highly effective for Parkinson’s disease classification. SVM maintained high accuracy, precision, recall and F1-score across various preprocessing techniques including PCA, CFE and SMOTE, making it robust and reliable for clinical applications. RF showed improved results with SMOTE. However, it experienced reduced performance with PCA and CFE, indicating its dependence on original feature interactions. DT benefited from PCA, while LR showed limited improvements and sensitivity to oversampling. These findings emphasize the importance of selecting appropriate preprocessing techniques to enhance model performance.</p>Dhiraj BaruahRizwan RehmanPranjal Kumar BoraPriyakshi MahantaKankana DuttaPinakshi Konwar
Copyright (c) 2025 Dhiraj Baruah, Rizwan Rehman, Pranjal Kumar Bora, Priyakshi Mahanta, Kankana Dutta, Pinakshi Konwar
https://creativecommons.org/licenses/by-sa/4.0
2025-05-302025-05-307369271210.35882/jeeemi.v7i3.713Performance Comparison of Extreme Learning Machine (ELM) and Hierarchical Extreme Learning Machine (H-ELM) Methods for Heart Failure Classification on Clinical Health Datasets
http://jeeemi.org/index.php/jeeemi/article/view/904
<p>Heart failure is one of the leading causes of death worldwide and requires accurate and timely diagnosis to improve patient outcomes. However, early detection remains a significant challenge due to the complexity of clinical data, high dimensionality of features, and variability in patient conditions. Traditional clinical methods often fall short in identifying subtle patterns that indicate early stages of heart failure, motivating the need for intelligent computational techniques to support diagnostic decisions. This study aims to enhance predictive modeling for heart failure classification by comparing two supervised machine learning approaches: Extreme Learning Machine (ELM) and Hierarchical Extreme Learning Machine (HELM). The main contribution of this research is the empirical evaluation of HELM's performance improvements over conventional ELM using 10-fold cross-validation on a publicly available clinical dataset. Unlike traditional neural networks, ELM offers fast training by randomly assigning weights and analytically computing output connections, while HELM extends this with a multi-layer structure that allows for more complex feature representation and improved generalization. Both models were assessed based on classification accuracy and Area Under the Curve (AUC), two critical metrics in medical classification tasks. The ELM model achieved an accuracy of 73.95% ± 8.07 and an AUC of 0.7614 ± 0.093, whereas the HELM model obtained a comparable accuracy of 73.55% ± 7.85 but with a higher AUC of 0.7776 ± 0.085. In several validation folds, HELM outperformed ELM, notably reaching 90% accuracy and 0.9250 AUC in specific cases. In conclusion, HELM demonstrates improved robustness and discriminatory capability in identifying heart failure cases. These findings suggest that HELM is a promising candidate for implementation in clinical decision support systems. Future research may incorporate feature selection, hyperparameter optimization, and evaluation across multi-center datasets to improve generalizability and real-world applicability.</p>Ichwan Dwi NugrahaTriando Hamonangan SaragihIrwan BudimanDwi KartiniFatma IndrianiWahyu Caesarendra
Copyright (c) 2025 Ichwan Dwi Nugraha, Triando Hamonangan Saragih, Irwan Budiman, Dwi Kartini, Fatma Indriani, Wahyu Caesarendra
https://creativecommons.org/licenses/by-sa/4.0
2025-05-312025-05-317371372810.35882/jeeemi.v7i3.904Advancement of Lung Cancer Diagnosis with Transfer Learning: Insights from VGG16 Implementation
http://jeeemi.org/index.php/jeeemi/article/view/704
<p><strong>Lung cancer continues to be one of the leading causes of cancer-related mortality globally, largely due to the challenges associated with its early and accurate detection. Timely diagnosis is critical for improving survival rates, and advances in artificial intelligence (AI), particularly deep learning, are proving to be valuable tools in this area. This study introduces an enhanced deep learning-based approach for lung cancer classification using the VGG16 neural network architecture. While previous research has demonstrated the effectiveness of ResNet-50 in this domain, the proposed method leverages the strengths of VGG16 particularly its deep architecture and robust feature extraction capabilities to improve diagnostic performance. To address the limitations posed by scarce labelled medical imaging data, the model incorporates transfer learning and fine-tuning techniques. It was trained and validated on a well-curated dataset of lung CT images. The VGG16 model achieved a high training accuracy of 99.09% and a strong validation accuracy of 95.41%, indicating its ability to generalize well across diverse image samples. These results reflect the model’s capacity to capture intricate patterns and subtle features within medical imagery, which are often critical for accurate disease classification. A comparative evaluation between VGG16 and ResNet-50 reveals that VGG16 outperforms its predecessor in terms of both accuracy and reliability. The improved performance underscores the potential of the proposed approach as a reliable and scalable AI-driven diagnostic solution. Overall, this research highlights the growing role of deep learning in enhancing clinical decision-making, offering a promising path toward earlier detection of lung cancer and ultimately contributing to better patient outcomes</strong>.</p>Vedavrath LakideV. Ganesan
Copyright (c) 2025 Vedavrath Lakide and V. Ganesan
https://creativecommons.org/licenses/by-sa/4.0
2025-06-012025-06-017372973910.35882/jeeemi.v7i3.704Exploring Dataset Variability in Diabetic Retinopathy Classification Using Transfer Learning Approaches
http://jeeemi.org/index.php/jeeemi/article/view/838
<p>Diabetic retinopathy (DR) stands as a primary international cause of vision impairment that needs effective and swift diagnostic services to protect eye structures from advancing deterioration. The variations of imaging data that appear between sources create major obstacles for achieving consistent performance from models. The elimination of performance fluctuation problems during DR classifications across two benchmark datasets EYE-PACS and APTOS is examined through systematic transfer learning analysis using different high-performing CNN architectures including VGG16, VGG19, ResNet50, Xception, InceptionV3, MobileNetV2, and InceptionResNetV2. The research evaluates how data heterogeneity affects and how augmentation approaches impact the accuracy while stabilizing robustness in deep learning models. The research provides new insights through its extensive investigation of generalization performance based on dataset changes which utilize modified data augmentation methods for retinal images. A collection of data transformations such as rotation, flipping, zooming and brightness modifications create simulated realistic scenarios to handle imbalanced data classes. Academic research involved CNN pre-training followed by transfer learning on both databases while researchers evaluated the models through both untreated source data and augmented image testing procedures. InceptionResNetV2 outperformed its counterparts with 96.2% accuracy and Xception delivered 95.7% accuracy in APTOS evaluation and both models scored 95.9% and 95.4% respectively on EYE-PACS testing. When augmentation was applied it increased the performance level by 3% to 5% across all running models. The experimental outcomes demonstrate how adequate variable training allows these models to recognize datasets regardless of their heterogeneity. This analysis confirms that combining reliable deep learning structures with purposeful data enhancement techniques substantially enhances DR diagnosis reliability to build scalable future diagnostic solutions for ophthalmology practice.</p>Kinjal PatniShruti YagnikPratik Patel
Copyright (c) 2025 Kinjal Patni Patni, Shruti Yagnik, Pratik Patel
https://creativecommons.org/licenses/by-sa/4.0
2025-06-072025-06-077376377710.35882/jeeemi.v7i3.838BHMI: A Multi-Sensor Biomechanical Human Model Interface for Quantifying Ergonomic Stress in Armored Vehicle
http://jeeemi.org/index.php/jeeemi/article/view/877
<p>Ergonomic stress inside armored military vehicles presents a critical yet often overlooked risk to soldier safety, operational effectiveness, and long-term health. Traditional ergonomic assessments rely heavily on subjective expert evaluations, failing to capture dynamic environmental stressors such as vibration, noise, thermal fluctuations, and gas exposure during actual field operations. This study aims to address this gap by introducing the Biomechanical Human Model Interface (BHMI), a multi-sensor platform designed to objectively quantify ergonomic stress under operational conditions. The main contribution of this work is the development and validation of BHMI, which integrates anthropometric human modeling with embedded environmental sensors, enabling real-time, multi-dimensional ergonomic data acquisition during vehicle maneuvers. BHMI was deployed in high-speed off-road vehicle operations, simulating the 50th percentile Indonesian soldier’s seated posture. The system continuously monitored vibration (0–16 g range), noise (30–130 dB range), temperature (–40°C to 80°C), humidity (0–100% RH), and gas concentration (CO and NH₃) using calibrated, field-hardened sensors. Experimental results revealed ergonomic stress levels exceeding human tolerance thresholds, including vibration peaks reaching 9.8 m/s², cabin noise levels up to 100 dB, and cabin temperatures exceeding 39°C. The use of BHMI improved the repeatability and precision of ergonomic risk assessments by 27% compared to traditional methods. Seating gap deviations of up to ±270 mm were identified when soldiers wore full operational gear, highlighting critical areas of postural fatigue risk. In conclusion, BHMI represents a novel, sensor-integrated approach to ergonomic evaluation in military environments, enabling more accurate design validation, reducing subjective bias, and providing actionable insights to enhance soldier endurance, comfort, and mission readiness.</p>Giva Andriana MutiaraHardy AdiluhungPeriyadi PeriyadiMuhammad Rizqy AlfarisiLisda Meisaroh
Copyright (c) 2025 Giva Andriana Mutiara, Hardy Adiluhung, Periyadi Periyadi, Muhammad Rizqy Alfarisi, Lisda Meisaroh
https://creativecommons.org/licenses/by-sa/4.0
2025-06-092025-06-097377879910.35882/jeeemi.v7i3.877Addressing Intrinsic Data Characteristics Issues of Imbalance Medical Data Using Nature Inspired Percolation Clustering
http://jeeemi.org/index.php/jeeemi/article/view/835
<p><strong>Data on diseases are generally skewed towards either positive or negative cases, depending on their prevalence. The problem of imbalance can significantly impact the performance of classification models, resulting in biased predictions and reduced model accuracy for the underrepresented class. Other factors that affect the performance of classifiers include intrinsic data characteristics, such as noise, outliers, and within-class imbalance, which complicate the learning task. Contemporary imbalance handling techniques employ clustering with SMOTE (Synthetic Minority Oversampling Technique) to generate realistic synthetic data that preserves the underlying data distribution, generalizes unseen data and mitigates overfitting to noisy points. Centroid-based clustering methods (e.g., K-means) often produce synthetic samples that are too clustered or poorly spaced. At the same time, density-based methods (e.g., DBSCAN) may fail to generate sufficient meaningful synthetic samples in sparse regions. The work aims to develop nature-inspired clustering that, combined with SMOTE, generates synthetic samples that adhere to the underlying data distribution and maintain sparsity among the data points that enhance performance of classifier. We propose PC-SMOTE, which leverages Percolation Clustering (PC), a novel clustering algorithm inspired by percolation theory. The methodology of PC utilizes a connectivity-driven framework to effectively handle irregular cluster shapes, varying densities, and sparse minority instances. The experiment was designed using a hybrid approach to assess PC-SMOTE using synthetically generated data with variable spread and other parameters; second, the algorithm was evaluated on eight sets of real medical datasets. The results show that the PC-SMOTE method works excellently for the Breast cancer dataset, Parkinson's dataset, and Cervical cancer dataset, where AUC is in the range of 96% to 99%, which is high compared to the other two methods. This demonstrates the effectiveness of the PC-SMOTE algorithm in handling datasets with both low and high imbalance ratios and often demonstrates competitive or superior performance compared to K-means and DBSCAN combined with SMOTE in terms of AUC, F1-score, G-mean, and PR-AUC.</strong></p>Kaikashan SiddavatamSubhash Shinde
Copyright (c) 2025 Kaikashan Siddavatam, Subhash Shinde
https://creativecommons.org/licenses/by-sa/4.0
2025-06-052025-06-057374076210.35882/jeeemi.v7i3.835Automated ICD Medical Code Generation for Radiology Reports using BioClinicalBERT with Multi-Head Attention Network
http://jeeemi.org/index.php/jeeemi/article/view/775
<p><strong>International Classification of Diseases (ICD) coding plays a pivotal role in healthcare systems with its provision of a standard method for classifying medical diagnoses, treatments, and procedures. However, the process of manually applying ICD codes to clinical records is both time-consuming and error-prone, particularly considering the large magnitude of medical terminologies and the periodic changes to the coding system. This work introduces a Hierarchical Multi-Head Attention Network (HMHAN) that aims to automate ICD coding using domain-related embeddings with an attention mechanism. The proposed method uses BioClinicalBERT for feature extraction from clinical text and then a two-level attention mechanism to learn hierarchical dependencies between labels. BioClinicalBERT is pre-trained on large biomedical and clinical corpora that enable it to capture complex contextual relationships specific to medical language more effectively. The multi-head attention mechanism enables the model to focus on different parts of the input text simultaneously, learning intricate associations between medical terms and corresponding ICD codes at various levels. This method uses SMOTE (Synthetic Minority Oversampling Technique) based multi-label resampling to solve class imbalance. SMOTE generates synthetic examples for underrepresented classes, allowing the model to learn better from imbalanced data without overfitting. For this work, MIMIC-IV dataset of de-identified radiology reports and corresponding ICD codes are used. The performance of the model is assessed with F1 score, Hamming loss, and ROC-AUC metrics. Results obtained from the model with an F1 score of 0.91, Hamming loss of 0.07, and ROC-AUC of 0.92 show promising research directions to automate the ICD coding process. This system will improve the effectiveness of healthcare workflows by automating ICD code generation for advanced clinical care.</strong></p>Sasikala D.Sarrvesh N.Sabarinath J.Theetchenya S.Kalavathi S.
Copyright (c) 2025 Sasikala D., Sarrvesh N., Sabarinath J., Theetchenya S., Kalavathi S.
https://creativecommons.org/licenses/by-sa/4.0
2025-06-132025-06-137380081610.35882/jeeemi.v7i3.775