Developing an Efficient Predictive Model Based on ML and DL Approaches to Detect Diabetes

Said Gadri

Abstract


During the last decade, some important progress on machine learning ML area have been made, especially with the apparition of a new subfield called deep learning DL and CNN networks (Convolutional Neural Networks). This new tendency is used to perform much more sophisticated algorithms allowing high performance in many disciplines such as: pattern recognition, image classification, computer vision, as well as other supervised and unsupervised classification tasks. In this work, we have developed an automatic classifier that permits to classify a number of diabetic patients based on some blood characteristics by using ML and DL approaches. Initially, we have proceeded to the classification task using many ML algorithms. Then we proposed a simple CNN model composed of many layers. Finally, we established a comparison between ML and DL algorithms. For programming task, we have used Python, Tensorflow and Keras which are the most used in the field.

Full Text:

PDF

References


Lee H., Grosse R., Ranganath R. , and Ng A.Y(2009). Convolutional deep belief networks for scalable unsupervised learning of hierarchical representations. In Proceedings of the 26th Annual International Conference on Machine Learning, pages 609–616. ACM.

Pinto N., Doukhan D., DiCarlo J.J. , and Cox D.D. (2009) A high-throughput screening approach to discovering good forms of biologically inspired visual representation. PLoS computational biology, 5(11):e1000579.

Turaga S.C., Murray J.F., Jain V. Roth F., Helmstaedter M., Briggman K., Denk W., and Seung H.S. (2010). Convolutional networks can learn to generate affinity graphs for image segmentation. Neural Computation, 22(2):511–538.

Abadi B., Agarwal M., Barham A, P, Brevdo E, Chen Z, Citro C, Corrado GS, Davis A, Dean J, Devin M, Ghemawat S, Goodfellow I, Harp A, Irving G, Isard M, Jia Y, Jozefowicz R, Kaiser L, Kudlur M, Levenberg J, Mané D, Monga R, Moore S, Murray D, Olah C, Schuster M, Shlens J, Steiner B, Sutskever I, Talwar K, Tucker P, Vanhoucke V, Vasudevan V, Viégas F, Vinyals O, Warden P, Wattenberg M, Wicke M, Yu Y, Zheng X. (2018) TensorFlow: large-scale machine learning on heterogeneous systems 2015. http://tensorflow.org/. Accessed 1 Nov 2018.

Theano Development Team. (2016). Theano: a Python framework for fast computation of mathematical expressions. arXiv e-prints arXiv:1605.02688.

Chollet F, et al (2018). Keras. 2015. https://keras.io. Accessed 1 Nov 2018.

Paszke A, Gross S, Chintala S, Chanan G, Yang E, DeVito Z, Lin Z, Desmaison A, Antiga L, Lerer A (2017). Automatic differentiation in pytorch. In: NIPS-W.

Chetlur S, Woolley C, Vandermersch P, Cohen J, Tran J, Catanzaro B, Shelhamer, E. cudnn (2014). Efficient primitives for deep learning.

Krizhevsky A, Sutskever I, Hinton GE (2012). Imagenet classification with deep convolutional neural networks. In: Neural information processing systems. p. 25.

Fukushima K. (1980). Neocognitron: a self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position. Biol Cybern. 1980;36(4):193–202.

LeCun Y, Bottou L, Bengio Y, Haffner P (1998). Gradient-based learning applied to document recognition. Proc IEEE. 86(11):2278–324.

Witten IH, Frank E, Hall MA, Pal CJ (2016). Data mining, Fourth Edition: Practical machine learning tools and techniques. 4th ed. San Francisco: Morgan Kaufmann Publishers Inc.

Goodfellow I, Bengio Y, Courville A (2016). Deep learning. Cambridge: The MIT Press; 2016.

Minar MR, Naher J. (2018). Recent advances in deep learning: an overview. arXiv:1807.08169.

LeCun Y, Bengio Y, Hinton G (2015). Deep learning. Nature;521:436.

Schmidhuber J. (2015). Deep learning in neural networks: an overview. Neural Net;61:85–117.

Rumelhart DE, Hinton GE, Williams RJ. (1986). Learning representations by back-propagating errors. Nature;323:533.

Y. Le Cun,, Y., Boser B., Denker J.S., Henderson D., Howard R.E., Hubbard W., Jackel L.D. (1990) :Handwritten digit recognition with a back-propagation network. In Advances in neural information processing systems.

LeCun Y, Boser B, Denker JS, Henderson D, Howard RE, Hubbard W, Jackel LD. (1989). Backpropagation applied to handwritten zip code recognition. Neural Comput;1(4):541–51.

Hinton GE, Osindero S, Teh Y-W. (2006). A fast learning algorithm for deep belief nets. Neural Comput;18(7):1527–54. https://doi.org/10.1162/neco.2006.18.7.1527.

Bengio Y, Lamblin P, Popovici D, Larochelle H. (2006) Greedy layer-wise training of deep networks. In: Proceedings of the 19th international conference on neural information processing systems. NIPS’06. MIT Press, Cambridge, MA, USA. p. 153–60.

Russakovsky O, Deng J, Su H, Krause J, Satheesh S, Ma S, Huang Z, Karpathy A, Khosla A, Bernstein M, Berg AC, Fei-Fei L. (2015). ImageNet large scale visual recognition challenge. Int J Comput Vision (IJCV);115(3):211–52. https://doi.org/10.1007/s11263-015-0816-y.

Kumar M. (2016). An incorporation of artificial intelligence capabilities in cloud computing. Int J Eng Comput Sci. https://doi.org/10.18535/ijecs/v5i11.63.

Saiyeda A, Mir MA. (2017). Cloud computing for deep learning analytics: a survey of current trends and challenges. Int J Adv Res Comput Sci;8(2):68–72.

Dumbill E. (2012). What is big data?: an introduction to the big data landscape. http://radar.oreilly.com/2012/01/what-is-big-data.html

Najafabadi MM, Villanustre F, Khoshgoftaar TM, Seliya N, Wald R, Muharemagic E. (2015). Deep learning applications and challenges in big data analytics. J Big Data; 2(1):1. https://doi.org/10.1186/s40537-014-0007-7.

Hinton G, Salakhutdinov R. (2011). Discovering binary codes for documents by learning deep generative models. Top Cogn Sci.;3(1):74–91.

Salakhutdinov R, Hinton G. (2009). Semantic hashing. Int J Approx Reason;50(7):969–78.




DOI: https://doi.org/10.31449/inf.v45i3.3041

Creative Commons License
This work is licensed under a Creative Commons Attribution 3.0 License.