Hostname: page-component-cd9895bd7-gvvz8 Total loading time: 0 Render date: 2024-12-24T01:20:43.109Z Has data issue: false hasContentIssue false

An efficient semisupervised feedforward neural network clustering

Published online by Cambridge University Press:  02 December 2014

Roya Asadi*
Affiliation:
Department of Artificial Intelligence, Faculty of Computer Science and Information Technology, University of Malaya, Kuala Lumpur, Malaysia
Mitra Asadi
Affiliation:
Department of Research, Iranian Blood Transfusion Organization, Tehran, Iran
Sameem Abdul Kareem
Affiliation:
Department of Artificial Intelligence, Faculty of Computer Science and Information Technology, University of Malaya, Kuala Lumpur, Malaysia
*
Reprint requests to: Roya Asadi, Department of Artificial Intelligence, Faculty of Computer Science and Information Technology, University of Malaya, Kuala Lumpur, 60503, Selangor, Malaysia. E-mail: [email protected]

Abstract

We developed an efficient semisupervised feedforward neural network clustering model with one epoch training and data dimensionality reduction ability to solve the problems of low training speed, accuracy, and high memory complexity of clustering. During training, a codebook of nonrandom weights is learned through input data directly. A standard weight vector is extracted from the codebook, and the exclusive threshold of each input instance is calculated based on the standard weight vector. The input instances are clustered based on their exclusive thresholds. The model assigns a class label to each input instance through the training set. The class label of each unlabeled input instance is predicted by considering a linear activation function and the exclusive threshold. Finally, the number of clusters and the density of each cluster are updated. The accuracy of the proposed model was measured through the number of clusters and the quantity of correctly classified nodes, which was 99.85%, 100%, and 99.91% of the Breast Cancer, Iris, and Spam data sets from the University of California at Irvine Machine Learning Repository, respectively, and the superior F measure results between 98.29% and 100% accuracies for the breast cancer data set from the University of Malaya Medical Center to predict the survival time.

Type
Regular Articles
Copyright
Copyright © Cambridge University Press 2014 

Access options

Get access to the full version of this content by using one of the access options below. (Log in options will check for institutional or personal access. Content may require purchase if you do not have access.)

References

REFERENCES

Alippi, C., Piuri, V., & Sami, M. (1995). Sensitivity to errors in artificial neural networks: a behavioral approach. IEEE Transactions on Circuits and Systems I: Fundamental Theory and Applications 42(6), 358361.CrossRefGoogle Scholar
Andonie, R., & Kovalerchuk, B. (2007). Neural Networks for Data Mining: Constraints and Open Problems. Ellensburg, WA: Central Washington University, Computer Science Department.Google Scholar
Asadi, R., & Kareem, S.A. (2013). Review of feedforward neural network classification preprocessing techniques. Proc. 3rd Int. Conf. Mathematical Sciences (ICMS3), pp. 567–573, Kuala Lumpur, Malaysia.Google Scholar
Asadi, R., & Kareem, S.A. (2014). An unsupervised feedforward neural network model for efficient clustering. Manuscript submitted for publication.Google Scholar
Asadi, R., Sabah Hasan, H., & Abdul Kareem, S. (2013). Review of current online dynamic unsupervised feedforward neural network classification. Proc. Computer Science and Electronics Engineering (CSEE—ISI/Scopus) Conf., Kuala Lumpur, Malaysia.Google Scholar
Asadi, R., Sabah Hasan, H., & Abdul Kareem, S. (2014). Review of current online dynamic unsupervised feedforward neural network classification. International Journal of Artificial Intelligence and Neural Networks 4(2), 12.Google Scholar
Asuncion, A., & Newman, D. (2007). UCI Machine Learning Repository. Irvine, CA: University of California, School of Information and Computer Science. Accessed at http://www.ics.uci.edu/~mlearn/MLRepositoryGoogle Scholar
Bengio, Y. (2000). 1M. Zurada. Introduction to the Special Issue on neural networks for data mining and knowledge discovery. IEEE Transactions on Neural Networks 100(3), 545549.Google Scholar
Bengio, Y., Buhmann, J., Embrechts, M., & Zurada, J. (2000). Neural networks for data mining and knowledge discovery [Special Issue]. IEEE Transactions on Neural Networks 11(2).Google Scholar
Bose, N.K., & Liang, P. (1996). Neural Network Fundamentals With Graphs, Algorithms, and Applications. New York: McGraw–Hill.Google Scholar
Bouchachia, A., Gabrys, B., & Sahel, Z. (2007). Overview of some incremental learning algorithms. Proc. Fuzzy Systems Conf. Fuzz-IEEE, pp. 1–16, London, July 23–26.CrossRefGoogle Scholar
Camastra, F., & Verri, A. (2005). A novel kernel method for clustering. IEEE Transactions on Pattern Analysis and Machine Intelligence 27(5), 801805.CrossRefGoogle ScholarPubMed
Chattopadhyay, M., Pranab, K., & Mazumdar, S. (2011). Principal component analysis and self-organizing map for visual clustering of machine-part cell formation in cellular manufacturing system. Systems Research Forum 5(1), 2551.CrossRefGoogle Scholar
Costa, J.A.F., & Oliveira, R.S. (2007). Cluster analysis using growing neural gas and graph partitioning. Proc. Int. Joint Conf. Neural Networks, Orlando, FL, August 12–17.CrossRefGoogle Scholar
Craven, M.W., & Shavlik, J.W. (1997). Using neural networks for data mining. Future Generation Computer Systems 13(2), 211229.CrossRefGoogle Scholar
Daffertshofer, A., Lamoth, C.J.C., Meijer, O.G., & Beek, P.J. (2004). PCA in studying coordination and variability: a tutorial. Clinical Biomechanics 19(4), 415428.CrossRefGoogle ScholarPubMed
Dasarathy, B.V. (1990). Nearest Neighbor Pattern Classification Techniques. Los Alamitos, CA: IEEE Computer Society Press.Google Scholar
Demuth, H., Beale, M., & Hagan, M. (2008). Neural Network Toolbox TM 6: User's Guide. Natick, MA: Math Works.Google Scholar
Deng, D., & Kasabov, N. (2003). On-line pattern analysis by evolving self-organizing maps. Neurocomputing 51, 87103.CrossRefGoogle Scholar
Fisher, R. (1950). The Use of Multiple Measurements in Taxonomic Problems: Contributions to Mathematical Statistics. New York: Wiley. (Original work published 1936)Google Scholar
Fritzke, B. (1995). A growing neural gas network learns topologies. Advances in Neural Information Processing Systems 7, 625632.Google Scholar
Fritzke, B. (1997). Some Competitive Learning Methods. Dresden: Dresden University of Technology, Artificial Intelligence Institute.Google Scholar
Furao, S., Ogura, T., & Hasegawa, O. (2007). An enhanced self-organizing incremental neural network for online unsupervised learning. Neural Networks 20(8), 893903.CrossRefGoogle ScholarPubMed
Germano, T. (1999). Self-organizing maps. Accessed at http://davis.wpi.edu/~matt/courses/somsGoogle Scholar
Goebel, M., & Gruenwald, L. (1999). A survey of data mining and knowledge discovery software tools. ACM SIGKDD Explorations Newsletter 1(1), 2033.CrossRefGoogle Scholar
Gui, V., Vasiu, R., & Bojković, Z. (2001). A new operator for image enhancement. Facta Universitatis-Series: Electronics and Energetics 14(1), 109117.Google Scholar
Hamker, F.H. (2001). Life-long learning cell structures—continuously learning without catastrophic interference. Neural Networks 14(4–5), 551573.CrossRefGoogle ScholarPubMed
Han, J., & Kamber, M. (2006). Data Mining, Southeast Asia Edition: Concepts and Techniques. San Francisco, CA: Morgan Kaufmann.Google Scholar
Hazlina, H., Sameem, A., NurAishah, M., & Yip, C. (2004). Back propagation neural network for the prognosis of breast cancer: comparison on different training algorithms. Proc. 2nd. Int. Conf. Artificial Intelligence in Engineering & Technology (ICAIET), pp. 445–449.Google Scholar
Hebb, D.O. (1949). The Organization of Behavior: A Neuropsychological Approach. New York: Wiley.Google Scholar
Hinton, G.E. (1989). Deterministic Boltzmann learning performs steepest descent in weight space. Neural Computation 1(1), 143150.CrossRefGoogle Scholar
Hebboul, A., Hacini, M., & Hachouf, F. (2011). An incremental parallel neural network for unsupervised classification. Proc. 7th Int. Workshop on Systems, Signal Processing Systems and Their Applications (WOSSPA), pp. 400403, Tipaza, Algeria, May 9–11.Google Scholar
Hegland, M. (2003). Data Mining—Challenges, Models, Methods and Algorithms. Canberra, Australia: Australia National University, ANU Data Mining Group.Google Scholar
Hinton, G.E., & Salakhutdinov, R.R. (2006). Reducing the dimensionality of data with neural networks. Science 313(5786), 504.CrossRefGoogle ScholarPubMed
Honkela, T. (1998). Description of Kohonen's self-organizing map. Accessed at http://www.cis.hut.fi/~tho/thesisGoogle Scholar
Jacquier, E., Kane, A., & Marcus, A.J. (2003). Geometric or arithmetic mean: a reconsideration. Financial Analysts Journal 59(6), 4653.CrossRefGoogle Scholar
Jean, J.S., & Wang, J. (1994). Weight smoothing to improve network generalization. IEEE Transactions on Neural Networks 5(5), 752763.CrossRefGoogle ScholarPubMed
Jolliffe, I. (1986). Principal Component Analysis (pp. 17). New York: Springer.CrossRefGoogle Scholar
Jolliffe, I.T. (2002). Principal Component Analysis (pp. 19). New York: Springer–Verlag.Google Scholar
Kamiya, Y., Ishii, T., Furao, S., & Hasegawa, O. (2007). An online semisupervised clustering algorithm based on a self-organizing incremental neural network. Proc. Int. Joint Conf. Neural Networks (IJCNN), pp. 1061–1066.CrossRefGoogle Scholar
Kantardzic, M. (2011). Data Mining: Concepts, Models, Methods, and Algorithms. New York: Wiley–Interscience.CrossRefGoogle Scholar
Kasabov, N.K. (1998). ECOS: evolving connectionist systems and the ECO learning paradigm. Proc. 5th Int. Conf. Neural Information Processing, ICONIP’98, pp. 123–128.Google Scholar
Kemp, R.A., MacAulay, C., & Palcic, B. (1997). Detection of malignancy associated changes in cervical cell nuclei using feed-forward neural networks. Journal of the European Society for Analytical Cellular Pathology 14(1), 3140.CrossRefGoogle ScholarPubMed
Kohonen, T. (1997). Self-Organizing Maps (Springer Series in Information Sciences, Vol. 30, pp. 22–25). Berlin: Springer–Verlag.CrossRefGoogle Scholar
Kohonen, T. (2000). Self-Organization Maps (3rd ed.). Berlin: Springer–Verlag.Google Scholar
Larochelle, H., Mandel, M., Pascanu, R., & Bengio, Y. (2012). Learning algorithms for the classification restricted Boltzmann machine. Journal of Machine Learning Research 13, 643669.Google Scholar
Linde, Y., Buzo, A., & Gray, R. (1980). An algorithm for vector quantizer design. IEEE Transactions on Communications 28(1), 8495.CrossRefGoogle Scholar
Lindsay, R.S., Funahashi, T., Hanson, R.L., Matsuzawa, Y., Tanaka, S., Tataranni, P.A., et al. (2002). Adiponectin and development of type 2 diabetes in the Pima Indian population. Lancet 360(9326), 5758.CrossRefGoogle ScholarPubMed
Martinetz, T.M., Berkovich, S.G., & Schulten, K.J. (1993). Neural-gas network for vector quantization and its application to time-series prediction. IEEE Transactions on Neural Networks 4(4), 558569.CrossRefGoogle Scholar
McClelland, J.L., Thomas, A.G., McCandliss, B.D., & Fiez, J.A. (1999). Understanding failures of learning: Hebbian learning, competition for representational space, and some preliminary experimental data. Progress in Brain Research 121, 7580.CrossRefGoogle ScholarPubMed
McCloskey, S. (2000). Neural networks and machine learning, p. 755. Accessed at http://www.cim.mcgill.ca/~scott/RIT/research_project.htmlGoogle Scholar
Melek, W.W., & Sadeghian, A. (2009). A theoretic framework for intelligent expert systems in medical encounter evaluation. Expert Systems 26(1), 8299.CrossRefGoogle Scholar
Oh, M., & Park, H.M. (2011). Preprocessing of independent vector analysis using feed-forward network for robust speech recognition. Proc. Neural Information Processing Conf., pp. 366–373.CrossRefGoogle Scholar
Özbay, Y., Ceylan, R., & Karlik, B. (2006). A fuzzy clustering neural network architecture for classification of ECG arrhythmias. Computers in Biology and Medicine 36(4), 376388.CrossRefGoogle ScholarPubMed
Pavel, B. (2002). Survey of Clustering Data Mining Techniques. San Jose, CA: Accrue Software.Google Scholar
Peng, J.-M., & Lin, Z. (1999). A non-interior continuation method for generalized linear complementarity problems. Mathematical Programming 86(3), 533563.CrossRefGoogle Scholar
Prudent, Y., & Ennaji, A. (2005). An incremental growing neural gas learns topologies. Proc. IEEE Int. Joint Conf. Neural Networks, IJCNN'05, pp. 1211–1216.CrossRefGoogle Scholar
Rougier, N., & Boniface, Y. (2011). Dynamic self-organising map. Neurocomputing 74(11), 18401847.CrossRefGoogle Scholar
Shen, F., Yu, H., Sakurai, K., & Hasegawa, O. (2011). An incremental online semisupervised active learning algorithm based on self-organizing incremental neural network. Neural Computing and Applications 20(7), 10611074.CrossRefGoogle Scholar
Tong, X., Qi, L., Wu, F., & Zhou, H. (2010). A smoothing method for solving portfolio optimization with CVaR and applications in allocation of generation asset. Applied Mathematics and Computation 216(6), 17231740.CrossRefGoogle Scholar
Ultsch, A., & Siemon, H.P. (1990). Kohonen's self organizing feature maps for exploratory data analysis. Proc. Int. Neural Networks Conf., pp. 305–308.Google Scholar
Van der Maaten, L.J., Postma, E.O., & Van den Herik, H.J. (2009). Dimensionality reduction: a comparative review. Journal of Machine Learning Research 10(1), 6671.Google Scholar
Vandesompele, J., De Preter, K., Pattyn, F., Poppe, B., Van Roy, N., De Paepe, A., et al. (2002). Accurate normalization of real-time quantitative RT-PCR data by geometric averaging of multiple internal control genes. Genome Biology 3(7).CrossRefGoogle ScholarPubMed
Werbos, P. (1974). Beyond regression: new tools for prediction and analysis in the behavioral sciences. PhD Thesis. Harvard University.Google Scholar
Wolberg, W.H., & Mangasarian, O.L. (1990). Multisurface method of pattern separation for medical diagnosis applied to breast cytology. Proceedings of the National Academy of Sciences 87(23), 91939196.CrossRefGoogle ScholarPubMed
Ziegel, E.R. (2002). Statistical inference. Technometrics 44(4).CrossRefGoogle Scholar