Fast and Accuracy Control Chart Pattern Recognition using a New cluster-k-Nearest Neighbor
Creators
Description
By taking advantage of both k-NN which is highly accurate and K-means cluster which is able to reduce the time of classification, we can introduce Cluster-k-Nearest Neighbor as "variable k"-NN dealing with the centroid or mean point of all subclasses generated by clustering algorithm. In general the algorithm of K-means cluster is not stable, in term of accuracy, for that reason we develop another algorithm for clustering our space which gives a higher accuracy than K-means cluster, less subclass number, stability and bounded time of classification with respect to the variable data size. We find between 96% and 99.7 % of accuracy in the lassification of 6 different types of Time series by using K-means cluster algorithm and we find 99.7% by using the new clustering algorithm.
Files
9707.pdf
Files
(216.5 kB)
Name | Size | Download all |
---|---|---|
md5:2949a1b5c1a32af6099a676deb0a1643
|
216.5 kB | Preview Download |
Additional details
References
- B. V. Dasarsthy, Ed., Nearest Neighbor (NN) Norms: NN Pattern classification techniques. Los Alamitos, CA:IEEE Computer Society Press,1990.
- P. J. Hardin, Parametric and Nearest Neighbor methods for hybrid classification: A comparaison of pixel assignment accuracy, photogramm Eng. Remote Sens., Vol. 60 pp. 1439-1448, Dec. 1994 .
- T. P. Yunck, A technique to identify NN, IEEE Trans. Syst., Man, Cybern,. Vol. SMC-6, no. 10, pp. 678-683, 1976.
- X. Jia and J. A. Richards, Cluster-space representation for hyperspectral data classification, IEEE Trans, Geosci. Remote Sens., vol. 40, no. 3, pp. 593-598, Mar. 2002.
- X. Jia and J. A. richards, Fast k-NN classification Using the Cluster-Space Approach, IEEE Trans., Geosci. Remote Sens., vol. 2, no. 2, April 2005.
- K. wagstaff and S. Rogers, Constrained k-means Clustering with Background Knowledge, Proc. of the 8th International conference on Machine learning, p. 577-584, 2001
- Pentakalos, O., Menasce, D., and Yesha, Y., Automated clustering-based workload characterization, Joint Sixth NASA Goddard Space Flight Center Conference on Mass Storage Systems and Technologies and Fifth IEEE Symposium on Mass Storage Systems, College Park, MD, March 23-26, 1998.
- Jain, A., Murty, M., Flynn, P., Data Clustering: A Review, http://scgwiki.iam.unibe.ch:8080/SCG/uploads/596/p264-jain.pdf
- Eisen, Michael. Cluster 3.0 Manual. http://bonsai.ims.u-tokyo.ac.jp/ mdehoon/ software/cluster/cluster3.pdf. [10] Hale. Introductory Statistics, http://espse.ed.psu.edu/statistics. [11] R. O. Duda, P. E. Hart, and D. G. Stork, Pattern Classification, 2nd ed. New York: Wiley, 2000. [12] I. T. Nabney, Netlab, Algorithms for Pattern Recognition. London, U.K.: Springer-Verlag, 2003. [13] D. J. C. MacKay. (1997). Gaussian processesA replacement for supervised neural networks, in Lecture Notes TutorialNIPS . [Online]. Available: http://www.inference.phy.cam.ac.uk/mackay/gp.pdf [14] D. M. Titterington, A. F. M. Smith, and U. E. Makov, Statistical Analysis of Finite Mixture Distributions. John Wiley, New York, 1985. [15] C. M. Bishop, Neural Networks for Pattern Recognition. Clarendon Press, Oxford, 1995. [16] S. Brahim-Belhouari and A. Bermak, Pattern Recognition Letters. Elsevier Science Publisher, 2005. [17] S. Brahim-Belhouari,Modified k-means cluster. University technology of PETRONAS, 2008.