[1] J. M. Keller, M. R. Gray and J. A. Givens, “A fuzzy k-nearest neighbor algorithm,” In IEEE transactions on systems, pp. 580-585, 1985.
[2] T. Denoeux, “A k-nearest neighbor classification rule based on Dempster-Shafer theory,” In IEEE Transactions on Systems, Man, and Cybernetics, vol. 25, no. 5, pp. 804-813, May 1995.
[3] K. Fukunaga and P. M. Narendra, "A Branch and Bound Algorithm for Computing k-Nearest Neighbors," In IEEE Transactions on Computers, vol. C-24, no. 7, pp. 750-753, July 1975.
[4] S. A. Dudani, “The Distance-Weighted k-Nearest-Neighbor Rule,” In IEEE Transactions on Systems, Man, and Cybernetics, vol. SMC-6, no. 4, pp. 325-327, April 1976.
[5] J. R. Quinlan, “Induction of decision trees,” Machine Learning, vol. 1, pp. 81–106, 1986.
[6] H. Schmid, “Probabilistic part of speech tagging using decision trees,” In Proceedings of the International Conference on New Methods in Language Processing, pp. 44-49, 1994.
[7] C. Z. Janikow, "Fuzzy decision trees: issues and methods," In IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics), vol. 28, no. 1, pp. 1-14, Feb. 1998.
[8] S. Tsang, B. Kao, K. Y. Yip, W. Ho and S. D. Lee, "Decision Trees for Uncertain Data," In IEEE Transactions on Knowledge and Data Engineering, vol. 23, no. 1, pp. 64-78, Jan. 2011.
[9] L. Breiman, “Random Forests,” Machine Learning, vol. 45, no. 1, pp. 5–32, 2001.
[10] A. Bosch, A. Zisserman and X. Munoz, "Image Classification using Random Forests and Ferns," IEEE International Conference on Computer Vision, Rio de Janeiro, pp. 1-8, 2007.
[11] P. O. Gislason, J. A. Benediktsson, and J. R. Sveinsson, “Random forests for land cover classification,” Pattern Recognit. Lett., vol. 27, no. 4, pp. 294–300, Mar. 2006.
[12] C. Cortes and V. Vapnik, “Support-Vector Network,” Machine Learning, pp. 273–297, 1995.
[13] T. Joachims, “Text categorization with support vector machines: Learning with many relevant features,” In European Conference on Machine Learning, pp. 137–142, 1998.
[14] W. M. Campbell, D. E. Sturim and D. A. Reynolds, “Support vector machines using GMM supervectors for speaker verification,” In IEEE Signal Processing Letters, vol. 13, no. 5, pp. 308-311, May 2006.
[15] K. P. Bennett and A. Demiriz, “Semi-supervised support vector machines,” In Advances in Neural Information processing systems, pp. 368-374, 1999.
[16] Y. LeCun, Y. Bengio, and G. Hinton, “Deep learning,” Nature, vol. 521, no. 7553, pp. 436–444, 2015.
[17] A. Krizhevsky, I. Sutskever, and G. E. Hinton, “Imagenet classification with deep convolutional neural networks,” Advances in Neural Information Processing Systems, pp. 1097-1105, 2012.
[18] T. Le, S. Huang and D. Jaw, "Cross-Resolution Feature Fusion for Fast Hand Detection in Intelligent Homecare Systems," In IEEE Sensors Journal, vol. 19, no. 12, pp. 4696-4704, 15 June15, 2019.
[19] Y. Liu, D. Jaw, S. Huang and J. Hwang, "DesnowNet: Context-Aware Deep Network for Snow Removal," In IEEE Transactions on Image Processing, vol. 27, no. 6, pp. 3064-3073, June 2018.
[20] J. A. Hartigan and M. A. Wong, "A K-Means Clustering Algorithm", Applied Statistics, vol. 28, no. 1, pp. 100-108, 1979.
[21] A. K. Jain, “Data clustering: 50 years beyond K-means,” Pattern Recognit. Lett., vol. 31, no. 8, pp. 651–666, Jan. 2010.
[22] T. Kanungo et al., “An efficient k-means clustering algorithm: Analysis and implementation,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 24, no. 7, pp. 881–892, Jul. 2002.
[23] S. Z. Selim and M. A. Ismail, "K-Means-Type Algorithms: A Generalized Convergence Theorem and Characterization of Local Optimality," In IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. PAMI-6, no. 1, pp. 81-87, Jan. 1984.
[24] J. F. Navarro, C. S. Frenk, and S. D. M. White, ‘‘A universal density profile from hierarchical clustering,’’ Astrophysical J., vol. 490, no. 2, pp. 493–508, Dec. 1997.
[25] P. Bajcsy and N. Ahuja, "Location- and density-based hierarchical clustering using similarity analysis," In IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 20, no. 9, pp. 1011-1015, Sept. 1998.
[26] X. Tang and P. Zhu, "Hierarchical Clustering Problems and Analysis of Fuzzy Proximity Relation on Granular Space," In IEEE Transactions on Fuzzy Systems, vol. 21, no. 5, pp. 814-824, Oct. 2013.
[27] S. Wold, K. Esbensen, and P. Geladi, “Principal component analysis,” Chemometr. Intell. Lab. Syst., vol. 2, pp. 37–52, 1987.
[28] I. Joliffe, Principal Component Analysis. New York: Springer-Verlag, 1986.
[29] B. Moore, "Principal component analysis in linear systems: Controllability, observability, and model reduction," In IEEE Transactions on Automatic Control, vol. 26, no. 1, pp. 17-32, February 1981.
[30] M. E. Tipping, “Sparse kernel principal component analysis,” In Neural Information Processing Systems, T. K. Leen, T. G. Dietterich, and V. Tresp, Eds. Cambridge: MIT Press, pp. 633–639, 2000.
[31] S. Mika, B. Scholkopf, A. Smola, K. M uller, M. Scholz, and G. R atsch, “Kernel PCA and de-noising in feature spaces,” Advances in Neural Information Processing Systems, vol. 11, no. 1, pp. 536–542, 1999.
[32] Chengjun Liu, "Gabor-based kernel PCA with fractional power polynomial models for face recognition," In IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 26, no. 5, pp. 572-581, May 2004.
[33] G. Hinton and R. Salakhutdinov, “Reducing the dimensionality of data with neural networks,” Science, vol. 313, no. 5786, pp. 504–507, 2006.
[34] W. Wang, Y. Huang, Y. Wang, and L. Wang, “Generalized autoencoder: A neural network framework for dimensionality reduction,” Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, pp. 490–497, Jun. 2014.
[35] D. Silver et al., “Mastering the game of Go with deep neural networks and tree search,” Nature, vol. 529, no. 7587, pp. 484–489, Jan. 2016.
[36] D. Silver et al., “Mastering the game of go without human knowledge,” Nature, vol. 550, no. 7676, pp. 354–359, 2017.