A Framework for Integrating Domain Knowledge into Machine Learning Models for Improved Interpretability

Authors

Asier Beñat Ignacio
Explainable AI (XAI) Engineer, Spain.
Eneko Santi Borja
Postdoctoral Fellow in Explainable AI, Spain.

Keywords:

interpretability, domain knowledge, machine learning, hybrid modeling, explainable AI

Synopsis

Modern machine learning (ML) models achieve impressive predictive performance but often lack interpretability, limiting trust and adoption in high-stakes domains such as healthcare, finance, and engineering. This paper proposes a structured framework for integrating domain knowledge directly into machine learning models to improve interpretability without sacrificing performance. We present systematic techniques combining expert rules, constraint-based learning, and feature engineering, and evaluate their impact across synthetic and real-world datasets. Results indicate that domain knowledge integration enhances model transparency, reduces uncertainty, and fosters actionable insights.

References

[1] Breiman, L.: Random forests. Machine Learning 45(1), 5–32 (2001)

[2] Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo (1993)

[3] Domingos, P.: A few useful things to know about machine learning. Communications of the ACM 55(10), 78–87 (2012)

[4] Ribeiro, M.T., Singh, S., Guestrin, C.: “Why should I trust you?” Explaining the predictions of any classifier. In: Proceedings of the ACM SIGKDD Conference on Knowledge Discovery and Data Mining, pp. 1135–1144 (2016)

[5] Doshi-Velez, F., Kim, B.: Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608 (2017)

[6] Rudin, C.: Stop explaining black box machine learning models for high stakes decisions and use interpretable models instead. Nature Machine Intelligence 1(5), 206–215 (2019)

[7] Holzinger, A., Biemann, C., Pattichis, C.S., Kell, D.B.: What do we need to build explainable AI systems for the medical domain? Artificial Intelligence in Medicine 101, 101709 (2019)

[8] Uppuluri, V. (2018). The Future of Business Intelligence in Value-Based Care Models. Journal of Artificial Intelligence, Machine Learning & Data Science, 1(1), 3009–3015. https://doi.org/10.51219/JAIMLD/vijitha-uppuluri/623

[9] Guidotti, R., Monreale, A., Ruggieri, S., Turini, F., Pedreschi, D., Giannotti, F.: A survey of methods for explaining black box models. ACM Computing Surveys 51(5), 1–42 (2018)

[10] Carvalho, D.V., Pereira, E.M., Cardoso, J.S.: Machine learning interpretability: A survey on methods and metrics. Electronics 8(8), 832 (2019)

[11] Letham, B., Rudin, C., McCormick, T.H., Madigan, D.: Interpretable classifiers using rules and Bayesian analysis. The Annals of Applied Statistics 9(3), 1350–1371 (2015)

[12] Gunning, D., Aha, D.: DARPA’s explainable artificial intelligence (XAI) program. AI Magazine 40(2), 44–58 (2019)

[13] Murdoch, W.J., Singh, C., Kumbier, K., Abbasi-Asl, R., Yu, B.: Definitions, methods, and applications in interpretable machine learning. Proceedings of the National Academy of Sciences 116(44), 22071–22080 (2019)

[14] Zhang, Q., Yang, Y., Wu, Y., Hoi, S.C.H.: Interpreting CNNs via decision trees. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 6261–6270 (2019)

[15] Potla, R.B. (2021). Blueprinting a Manufacturing Data Lakehouse: Harmonizing BOM, Routing, and Serialization Data for Advanced Analytics. International Journal of Innovative Research in Engineering & Multidisciplinary Physical Sciences, 9(1), 1–12. https://doi.org/10.37082/IJIRMPS.v9.i1.232841

[16] Tonekaboni, S., Joshi, S., McCradden, M.D., Goldenberg, A.: What clinicians want: Contextualizing explainable machine learning for clinical end use. In: Proceedings of the Machine Learning for Healthcare Conference, pp. 359–380 (2019)

[17] Lundberg, S.M., Lee, S.I.: A unified approach to interpreting model predictions. In: Advances in Neural Information Processing Systems, pp. 4765–4774 (2017)

[18] Molnar, C.: Interpretable Machine Learning. Lulu Press, Morrisville (2020)

[19] Zhang, Y., Chen, X., Zhou, D.: Incorporating domain knowledge into deep learning for medical image analysis. IEEE Transactions on Medical Imaging 39(6), 1934–1945 (2020)

[20] Holzinger, A., Langs, G., Denk, H., Zatloukal, K., Müller, H.: Causability and explainability of artificial intelligence in medicine. Wiley Interdisciplinary Reviews: Data Mining and Knowledge Discovery 9(4), e1312 (2019)

IJAIML

Published

April 15, 2022