Partial Histogram Bayes Learning Algorithm for Classification Applications
 Abstract
 Keywords
 References

Abstract
This paper presents a proposed supervised classification technique namely partial histogram Bayes (PHBayes) learning algorithm. Conventional classifier based on Gaussian function has limitation when dealing with different probability distribution functions and requires large memory for large number of instance. Alternatively, histogram based classifiers are flexible for different probability density function. The aims of PHBayes are to handle large number of instances in datasets with lesser memory requirement, and fast in training and testing phases. The PHBayes depends on portion of the observed histogram that is similar to the probability density function. PHBayes was analyzed using synthetic and real data. Several factors affecting classification accuracy were considered. The PHBayes was compared with other established classifiers and demonstrated higher accurate classification, lesser memory even when dealing with large number of instance, and faster in training and testing phases.

Keywords
Classification; Histogram noise estimation and reduction; Histogram probability distribution; Naïve Bayes; Supervised learning.

References
[1] Freedman, S. T. and Adams, J. A. Filtering data based on humaninspired forgetting. IEEE Trans. Syst. Man Cybern. B, Cybern. 2011, 41(6):15441555.
[2] Mohammad, M. M., Jing, G., Latifur, K., Jiawei, H. and Bhavani, T. Classification and novel class detection in conceptdrifting data streams under time constraints. IEEE Trans. Knowl. Data Eng. 2011, 23(6):859874.
[3] Bogdan, W. M. Neural network architectures and learning algorithms. IEEE Ind. Electron. Mag. 2009, 3(4):5663.
[4] French, R.M. Catastrophic forgetting in connectionist networks. Trends Cogn. Sci. 1999, 3(4):128135.
[5] Mensink, T., Verbeek, J., Perronnin, F. and Csurka, G. Distancebased image classification: Generalizing to new classes at nearzero cost. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35(11):26242637.
[6] Mathur, A. and Foody, G. M. Multiclass and binary SVM classification: Implications for training and classification users. IEEE Geosci. Remote Sens. Lett. 2008, 5(2):241245.
[7] ChihWei, H. and ChihJen, L. A comparison of methods for multiclass support vector machines. IEEE Trans. Neural Netw. 2002, 13(2):415425.
[8] Ravikumar, B., Thukaram, D. and Khincha, H. P. Comparison of multiclass SVM classification methods to use in a supportive system for distance relay coordination. IEEE Trans. Power Del. 2010, 25(3):12961305.
[9] Chaudhuri, P., Ghosh, A. K. and Oja, H. Classification based on hybridization of parametric and nonparametric classifiers. IEEE Trans. Pattern Anal. Mach. Intell. 2009, 31(7):11531164.
[10] [10] Veenman, C. J. and Reinders, M. J. T. The nearest subclass classifier: A compromise between the nearest mean and nearest neighbor classifier. IEEE Trans. Pattern Anal. Mach. Intell. 2005, 27(9):14171429.
[11] Viswanath, P. and Sarma, T. H. An improvement to knearest neighbor classifier. Proceedings of the Recent Advances in Intelligent Computational Syst., pp. 227231, 2011.
[12] Yiguang, L., Sam, G. S., Chunguang, L. and Zhisheng, Y. kNS: A classifier by the distance to the nearest subspace. IEEE Trans. Neural Netw. 2011, 22(8):12561268.
[13] Hui, W. Neighborhood counting measure and minimum risk metric. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 32(4):766768.
[14] Xueyi, W. A fast exact knearest neighbors algorithm for high dimensional search using kmeans clustering and triangle inequality. Proceedings of the Int. Joint Conf. on Neural Networks, pp. 12931299, 2011.
[15] Hui, W. Nearest neighbors by neighborhood counting. IEEE Trans. Pattern Anal. Mach. Intell. 2006, 28(6):942953.
[16] Yoonho, H., Bohyung, H. and HeeKap, A. A fast nearest neighbor search algorithm by nonlinear embedding. IEEE Conf. on Computer Vision and Pattern Recognition, pp. 30533060, Providence, 2012.
[17] Moutafis, P., Leng, M., and Kakadiaris, I. A. An overview and empirical comparison of distance metric learning methods. IEEE Trans. Cybern. 2017, 47(3):612625.
[18] Yu, J., Bai, M., Wang, G. and Shi, X. Fault diagnosis of planetary gearbox with incomplete information using assignment reduction and flexible naive Bayesian classifier. J. of Mechanical Science and Technology, 2018, 32(1):3747.
[19] Yonghong, H., Englehart, K. B., Hudgins, B. and Chan, A. D. C. A Gaussian mixture model based classification scheme for myoelectric control of powered upper limb prostheses. IEEE Trans. Biomed. Eng. 2005, 52(11):18011811.
[20] Karthikeyan, G., Rajendra, A. U., Chua Kuang, C., Choo Min L. and Thomas, A. K. Oneclass classification of mammograms using trace transform functionals. IEEE Trans. Instrum. Meas. 2014, 63(2):304311.
[21] Wei, L., Saurabh, P., James, F. E. and Mann, B. L. Localitypreserving dimensionality reduction and classification for hyperspectral image analysis. IEEE Trans. Geosci. Remote Sens. 2012, 50(4):11851198.
[22] Saurabh, P., Minshan, C., Wei, L. and James, F. E. Segmented mixtureofGaussian classification for hyperspectral image analysis. IEEE Geosci. Remote Sens. Lett. 2014, 11(1):138142.
[23] Wei, L., Saurabh, P. and James, F. E. Hyperspectral image classification using Gaussian mixture models and Markov random fields. IEEE Geosci. Remote Sens. Lett. 2014, 11(1):153157.
[24] XiZhao, W., YuLin, H. and Wang, D. D. Nonnaive Bayesian classifiers for classification problems with continuous attributes. IEEE Trans. Cybern. 2014, 44(1):2139.
[25] John, G. H. and Langley, P. Estimating continuous distributions in Bayesian classifiers. Proceedings of the Conf. on Uncertain. Artif. Intell. pp. 338–345, Montréal, 1995.
[26] Liu, J. N. K., YuLin, H., XiZhao, W. and YanXing, H. A comparative study among different kernel functions in flexible naive Bayesian classification. Proceedings of the Int. Conf. on Mach. Learning and Cybern., pp. 638643, 2011.
[27] Lawend, H. O.; and Muad, A. M. A nonparametric partial histogram Bayes learning algorithm for classification applications. Proceedings of the IEEE Int. Conf. on Control Sys., Computing and Eng., pp. 3539, 2014.
[28] HungJu, H. and ChunNan, H. Bayesian classification for data from the same unknown class. IEEE Trans. Syst. Man Cybern. B, Cybern. 2002, 32(2):137145.
[29] Soria, D., Garibaldi, J. M., Biganzoli, E. and Ellis, I. O. A comparison of three different methods for classification of breast cancer data. Proceedings of the Seventh Int. Conf. on Mach. Learning and Appl , pp. 619624, 2008.
[30] YanShi, D. and KeSong, H. A comparison of several ensemble methods for text categorization. Proceedings of the Services Computing, pp. 419422, 2004.
[31] Dilrukshi, I. and Zoysa, K. D. Twitter news classification: Theoretical and practical comparison of SVM against naive Bayes algorithms. Proceedings of the Int. Conf. on Advances in ICT for Emerging Regions, pp. 278278, 2013.
[32] Shameem, F. and Nisar, H. Comparison of classification techniques Proceedings of the SVM and Naives Bayes to predict the arboviral diseasedengue. IEEE Int. Conf. on Bioinformatics and Biomedicine Workshops, pp. 538539. 2011.
[33] Perez, A., Larranaga, P. and Inza, I. Bayesian classifiers based on kernel density estimation: Flexible classifiers. Int. J. of Approximate Reasoning. 2009, 50(2):341362.
[34] Papoulis, A. Probability, random variables, and stochastic processes. McGrawHill. 1991.
[35] Asuncion, A. and Newman, D.J. UCI machine learning repository. 2007, http://www.ics.uci.edu/~mlearn/MLRepository.html.

View 
Download  Article ID: 20787 
