Breast Cancer Classification Using z-score Thresholding and Machine Learning
Abstract
Image processing and machine learning are being used in biomedical purposes as a supporting tool in the detection and diagnosis of certain diseases. Breast cancer is one of these diseases which the researchers have put great effort into for decades. To accomplish this task, image and feature-based public datasets are available to be used. Due to several reasons such as hardware or preprocessing, images can get noisy. The noise in images which can lead to anormal / outliers in the dataset may decrease the detection accuracy and can mislead the medical staff during diagnosis stage. Therefore, this study aims to present the effect of removing the outliers from dataset on the detection accuracy of breast cancer. The method removes the outliers detected by z-score analysis. The remaining data is normalized, and classification accuracy of 10 methods are obtained by direct implementation. The methods are XGBoost, Neural Network, CNN, RNN, AdaBoost, LSTM, GRU, Random Forest, SVM and Logistic Regression. A public dataset Wisconsin diagnosis breast cancer (WDBC) was used in this study. Ablation study was conducted by fine-tuning the threshold value of z-score method. The result showed that the best accuracy was obtained when the threshold value is 3. Also, comparison was made between the results made on the entire dataset and dataset after its outliers were removed. The results showed that the average accuracy of all the classifiers is 98.08%. As a conclusion, the results indicate that removal of the outliers from the dataset increases the overall accuracy of breast cancer detection.
Downloads
References
H. Sung et al., “Global Cancer Statistics 2020: GLOBOCAN Estimates of Incidence and Mortality Worldwide for 36 Cancers in 185 Countries,” CA Cancer J Clin, vol. 71, no. 3, pp. 209–249, May 2021, doi: 10.3322/caac.21660.
N. Harbeck et al., “Breast cancer,” Nat Rev Dis Primers, vol. 5, no. 1, p. 66, Sep. 2019, doi: 10.1038/s41572-019-0111-2.
G. Litjens et al., “A survey on deep learning in medical image analysis,” Med Image Anal, vol. 42, pp. 60–88, Dec. 2017, doi: 10.1016/j.media.2017.07.005.
A. Esteva et al., “Dermatologist-level classification of skin cancer with deep neural networks,” Nature, vol. 542, no. 7639, pp. 115–118, Feb. 2017, doi: 10.1038/nature21056.
W. Wolberg, O. Mangasarian, N. Street, and W. Street. "Breast Cancer Wisconsin (Diagnostic)," UCI Machine Learning Repository, 1993. [Online]. Available: https://doi.org/10.24432/C5DW2B.
W. H. Wolberg and O. L. Mangasarian, “Multisurface method of pattern separation for medical diagnosis applied to breast cytology.,” Proceedings of the National Academy of Sciences, vol. 87, no. 23, pp. 9193–9196, Dec. 1990, doi: 10.1073/pnas.87.23.9193.
M. Elter. "Mammographic Mass," UCI Machine Learning Repository, 2007. [Online]. Available: https://doi.org/10.24432/C53K6Z.
M. Elter, R. Schulz‐Wendtland, and T. Wittenberg, “The prediction of breast cancer biopsy outcomes using two CAD approaches that both emphasize an intelligible decision process,” Med Phys, vol. 34, no. 11, pp. 4164–4172, Nov. 2007, doi: 10.1118/1.2786864.
Z. Khandezamin, M. Naderan, and M. J. Rashti, “Detection and classification of breast cancer using logistic regression feature selection and GMDH classifier,” J Biomed Inform, vol. 111, p. 103591, Nov. 2020, doi: 10.1016/j.jbi.2020.103591.
A. S. Yaro, F. Maly, P. Prazak, and K. Malý, “Outlier Detection Performance of a Modified Z-Score Method in Time-Series RSS Observation With Hybrid Scale Estimators,” IEEE Access, vol. 12, pp. 12785–12796, 2024, doi: 10.1109/ACCESS.2024.3356731.
C. C. Aggarwal, Outlier Analysis. Cham: Springer International Publishing, 2017. doi: 10.1007/978-3-319-47578-3.
A. Zimek, E. Schubert, and H. Kriegel, “A survey on unsupervised outlier detection in high‐dimensional numerical data,” Statistical Analysis and Data Mining: The ASA Data Science Journal, vol. 5, no. 5, pp. 363–387, Oct. 2012, doi: 10.1002/sam.11161.
I. Guyon, J. Weston, S. Barnhill, and V. Vapnik, “Gene Selection for Cancer Classification using Support Vector Machines,” Mach Learn, vol. 46, no. 1–3, pp. 389–422, Jan. 2002, doi: 10.1023/A:1012487302797.
S. Yadav and S. Shukla, “Analysis of k-Fold Cross-Validation over Hold-Out Validation on Colossal Datasets for Quality Classification,” in 2016 IEEE 6th International Conference on Advanced Computing (IACC), IEEE, Feb. 2016, pp. 78–83. doi: 10.1109/IACC.2016.25.
L. Breiman, “Random Forests,” Mach Learn, vol. 45, no. 1, pp. 5–32, Oct. 2001, doi: 10.1023/A:1010933404324.
S. Lundberg and S.-I. Lee, “A Unified Approach to Interpreting Model Predictions,” Nov. 2017.
E. Genc, M. E. Yildirim, and Y. B. Salman, “Human activity recognition with fine-tuned CNN-LSTM,” Journal of Electrical Engineering, vol. 75, no. 1, pp. 8–13, Feb. 2024, doi: 10.2478/jee-2024-0002.
V. Fascianelli, A. Battista, F. Stefanini, S. Tsujimoto, A. Genovesio, and S. Fusi, “Neural representational geometries reflect behavioral differences in monkeys and recurrent neural networks,” Nat Commun, vol. 15, no. 1, p. 6479, Aug. 2024, doi: 10.1038/s41467-024-50503-w.
Y. Zhang, S. Xu, L. Zhang, W. Jiang, S. Alam, and D. Xue, “Short-term multi-step-ahead sector-based traffic flow prediction based on the attention-enhanced graph convolutional LSTM network (AGC-LSTM),” Neural Comput Appl, vol. 37, no. 20, pp. 14869–14888, Jul. 2025, doi: 10.1007/s00521-024-09827-3.
L. Zhang, J. Zhang, W. Gao, F. Bai, N. Li, and N. Ghadimi, “A deep learning outline aimed at prompt skin cancer detection utilizing gated recurrent unit networks and improved orca predation algorithm,” Biomed Signal Process Control, vol. 90, p. 105858, Apr. 2024, doi: 10.1016/j.bspc.2023.105858.
Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, May 2015, doi: 10.1038/nature14539.
G. Litjens et al., “A survey on deep learning in medical image analysis,” Med Image Anal, vol. 42, pp. 60–88, Dec. 2017, doi: 10.1016/j.media.2017.07.005.
Ö. F. İnce, I. F. Ince, M. E. Yıldırım, J. S. Park, J. K. Song, and B. W. Yoon, “Human activity recognition with analysis of angles between skeletal joints using a RGB‐depth sensor,” ETRI Journal, vol. 42, no. 1, pp. 78–89, Feb. 2020, doi: 10.4218/etrij.2018-0577.
S. M. Kasongo, “A deep learning technique for intrusion detection system using a Recurrent Neural Networks based framework,” Comput Commun, vol. 199, pp. 113–125, Feb. 2023, doi: 10.1016/j.comcom.2022.12.010.
M. Abdar et al., “A review of uncertainty quantification in deep learning: Techniques, applications and challenges,” Information Fusion, vol. 76, pp. 243–297, Dec. 2021, doi: 10.1016/j.inffus.2021.05.008.
N. Munir, J. Huang, C.-N. Wong, and S.-J. Song, “Machine learning based eddy current testing: A review,” Results in Engineering, vol. 25, p. 103724, Mar. 2025, doi: 10.1016/j.rineng.2024.103724.
S. Lundberg and S.-I. Lee, “A Unified Approach to Interpreting Model Predictions,” Nov. 2017.
E. Tjoa and C. Guan, “A Survey on Explainable Artificial Intelligence (XAI): Towards Medical XAI,” Aug. 2020, doi: 10.1109/TNNLS.2020.3027314.
M. B. Lopes, A. Veríssimo, E. Carrasquinha, S. Casimiro, N. Beerenwinkel, and S. Vinga, “Ensemble outlier detection and gene selection in triple-negative breast cancer data,” BMC Bioinformatics, vol. 19, no. 1, p. 168, Dec. 2018, doi: 10.1186/s12859-018-2149-7.
S. H. Ali and M. Shehata, “A New Breast Cancer Discovery Strategy: A Combined Outlier Rejection Technique and an Ensemble Classification Method,” Bioengineering, vol. 11, no. 11, p. 1148, Nov. 2024, doi: 10.3390/bioengineering11111148.
A. Alloqmani, Y. B. Abushark, and A. I. Khan, “Anomaly Detection of Breast Cancer Using Deep Learning,” Arab J Sci Eng, vol. 48, no. 8, pp. 10977–11002, Aug. 2023, doi: 10.1007/s13369-023-07945-z.
J. Lv and L. Wang, “Hybrid modeling of adsorption process using mass transfer and machine learning techniques for concentration prediction,” Journal of Saudi Chemical Society, vol. 29, no. 4, p. 12, Sep. 2025, doi: 10.1007/s44442-025-00016-y.
D. Wu, X. Ma, and D. L. Olson, “Financial distress prediction using integrated Z-score and multilayer perceptron neural networks,” Decis Support Syst, vol. 159, p. 113814, Aug. 2022, doi: 10.1016/j.dss.2022.113814.
A. M. Sharifnia, D. E. Kpormegbey, D. K. Thapa, and M. Cleary, “A Primer of Data Cleaning in Quantitative Research: Handling Missing Values and Outliers,” J Adv Nurs, Mar. 2025, doi: 10.1111/jan.16908.
W. Li, Y. Yin, X. Quan, and H. Zhang, “Gene Expression Value Prediction Based on XGBoost Algorithm,” Front Genet, vol. 10, Nov. 2019, doi: 10.3389/fgene.2019.01077.
Martin. Anthony, Neural Network Learning : Theoretical Foundations. Cambridge University Press, 2022.
E. Genc, M. E. Yildirim, and Y. B. Salman, “Human activity recognition with fine-tuned CNN-LSTM,” Journal of Electrical Engineering, vol. 75, no. 1, pp. 8–13, Feb. 2024, doi: 10.2478/jee-2024-0002.
A. Sherstinsky, “Fundamentals of Recurrent Neural Network (RNN) and Long Short-Term Memory (LSTM) network,” Physica D, vol. 404, p. 132306, Mar. 2020, doi: 10.1016/j.physd.2019.132306.
X. Li, L. Wang, and E. Sung, “AdaBoost with SVM-based component classifiers,” Eng Appl Artif Intell, vol. 21, no. 5, pp. 785–795, Aug. 2008, doi: 10.1016/j.engappai.2007.07.001.
M. A. Chandra and S. S. Bedi, “Survey on SVM and their application in image classification,” International Journal of Information Technology, vol. 13, no. 5, pp. 1–11, Oct. 2021, doi: 10.1007/s41870-017-0080-1.
T. Perumal, N. Mustapha, R. Mohamed, and F. M. Shiri, “A Comprehensive Overview and Comparative Analysis on Deep Learning Models,” Journal on Artificial Intelligence, vol. 6, no. 1, pp. 301–360, 2024, doi: 10.32604/jai.2024.054314.
V. F. Rodriguez-Galiano, B. Ghimire, J. Rogan, M. Chica-Olmo, and J. P. Rigol-Sanchez, “An assessment of the effectiveness of a random forest classifier for land-cover classification,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 67, pp. 93–104, Jan. 2012, doi: 10.1016/j.isprsjprs.2011.11.002.
S. Sperandei, “Understanding logistic regression analysis,” Biochem Med (Zagreb), pp. 12–18, 2014, doi: 10.11613/BM.2014.003.
A. Masbakhah, U. Sa’adah, and M. Muslikh, “Heart Disease Classification Using Random Forest and Fox Algorithm as Hyperparameter Tuning,” Journal of Electronics, Electromedical Engineering, and Medical Informatics, vol. 7, no. 4, pp. 964–976, Aug. 2025, doi: 10.35882/jeeemi.v7i4.932.
L. De Coster et al., “On the optimal z-score threshold for SISCOM analysis to localize the ictal onset zone,” EJNMMI Res, vol. 8, no. 1, p. 34, Dec. 2018, doi: 10.1186/s13550-018-0381-9.
A. Curtis, T. Smith, B. Ziganshin, and J. Elefteriades, “The Mystery of the Z-Score,” AORTA, vol. 04, no. 04, pp. 124–130, Aug. 2016, doi: 10.12945/j.aorta.2016.16.014.
S. Aamir et al., “Predicting Breast Cancer Leveraging Supervised Machine Learning Techniques,” Comput Math Methods Med, vol. 2022, pp. 1–13, Aug. 2022, doi: 10.1155/2022/5869529.
Z. Mushtaq, A. Yaqub, S. Sani, and A. Khalid, “Effective K-nearest neighbor classifications for Wisconsin breast cancer data sets,” Journal of the Chinese Institute of Engineers, vol. 43, no. 1, pp. 80–92, Jan. 2020, doi: 10.1080/02533839.2019.1676658.
H. Rajaguru and S. C. S R, “Analysis of Decision Tree and K-Nearest Neighbor Algorithm in the Classification of Breast Cancer,” Asian Pacific Journal of Cancer Prevention, vol. 20, no. 12, pp. 3777–3781, Dec. 2019, doi: 10.31557/APJCP.2019.20.12.3777.
F. Khan et al., “Cloud-Based Breast Cancer Prediction Empowered with Soft Computing Approaches,” J Healthc Eng, vol. 2020, pp. 1–16, May 2020, doi: 10.1155/2020/8017496.
N. Al-Azzam and I. Shatnawi, “Comparing supervised and semi-supervised Machine Learning Models on Diagnosing Breast Cancer,” Annals of Medicine and Surgery, vol. 62, pp. 53–64, Feb. 2021, doi: 10.1016/j.amsu.2020.12.043.
A. Rasool, C. Bunterngchit, L. Tiejian, Md. R. Islam, Q. Qu, and Q. Jiang, “Improved Machine Learning-Based Predictive Models for Breast Cancer Diagnosis,” Int J Environ Res Public Health, vol. 19, no. 6, p. 3211, Mar. 2022, doi: 10.3390/ijerph19063211.
S. Zhou, C. Hu, S. Wei, and X. Yan, “Breast Cancer Prediction Based on Multiple Machine Learning Algorithms,” Technol Cancer Res Treat, vol. 23, Jan. 2024, doi: 10.1177/15330338241234791.
P. Ghosh and D. Chatterjee, “Comparative Analysis of Machine Learning Algorithms for Breast Cancer Classification: SVM Outperforms XGBoost, CNN, RNN, and Others,” Apr. 2024, doi: 10.1101/2024.04.22.590658.
Copyright (c) 2025 Mustafa Eren Yildirim, Yucel B. Salman

This work is licensed under a Creative Commons Attribution-ShareAlike 4.0 International License.
Authors who publish with this journal agree to the following terms:
- Authors retain copyright and grant the journal right of first publication with the work simultaneously licensed under a Creative Commons Attribution-ShareAlikel 4.0 International (CC BY-SA 4.0) that allows others to share the work with an acknowledgement of the work's authorship and initial publication in this journal.
- Authors are able to enter into separate, additional contractual arrangements for the non-exclusive distribution of the journal's published version of the work (e.g., post it to an institutional repository or publish it in a book), with an acknowledgement of its initial publication in this journal.
- Authors are permitted and encouraged to post their work online (e.g., in institutional repositories or on their website) prior to and during the submission process, as it can lead to productive exchanges, as well as earlier and greater citation of published work (See The Effect of Open Access).
 
							 
							



 
 .png)
.png)
.png)
.png)
.png)
