Abstract
We mathematically and experimentally evaluate the validity of dimension-reduction methods for the computation of similarity in image pattern recognition. Image pattern recognition identifies instances of particular objects and distinguishes differences among images. This recognition uses pattern recognition techniques for the classification and categorisation of images. In numerical image pattern recognition techniques, images are sampled using an array of pixels. This sampling procedure derives vectors in a higher-dimensional metric space from image patterns. To ensure the accuracy of pattern recognition techniques, the dimension reduction of the vectors is an essential methodology since the time and space complexities of processing depend on the dimension of the data. Dimension reduction causes information loss of topological and geometrical features of image patterns. Through both theoretical and experimental comparisons, we clarify that dimension-reduction methodologies that preserve the topology and geometry in the image pattern space are essential for linear pattern recognition. For the practical application of methods of dimension reduction, the random projection works well compared with downsampling, the pyramid transform, the two-dimensional random projection, the two-dimensional discrete cosine transform and nonlinear multidimensional scaling if we have no a priori information on the input data.
Similar content being viewed by others
Notes
In this paper, the 2DDCT is applied to an image without partitioning, while the JPEG and MPEG compression algorithms divide an image into blocks of \(8 \times 8\) pixels before applying the 2DDCT.
This generalised principal component analysis is a different method to the GPCA [17] although they bear the same name.
Note that we use the two-dimensional DCT-II without dividing an image into blocks, while the JPEG and MPEG compression algorithms use the two-dimensional DCT-II by partitioning an image into \(N \times N\) blocks of \(8\times 8\) pixels.
The MDS embeds data into a low-dimensional space. This embedding is a nonlinear dimension-reduction method, while the kernel method uses a linear dimension-reduction method in a high-dimensional space.
References
Turk, M., Pentland, A.: Eigenfaces for recognition. J. Cogn. Neurosci. 3(1), 71–86 (1991)
Maeda, K.: From the Subspace Methods to the Mutual Subspace Method. Computer Vision, vol. 285, pp. 135–156. Springer, Berlin (2010)
Murase, H., Nayar, S.K.: Illumination planning for object recognition using parametric eigenspace. IEEE Trans. Pattern Anal. Mach. Intell. 16, 1219–1227 (1994)
Park, C.H., Park, H.: Fingerprint classification using fast Fourier transform and nonlinear discriminant analysis. Pattern Recognit. 38, 495–503 (2005)
Park, H.A., Park, K.R.: Iris recognition based on score level fusion by using SVM. Pattern Recognit. Lett. 28, 2019–2028 (2007)
Csurka, G., Dance, C., Fan, L., Willamowski, J., Bray, C.: Visual categorization with bags of keypoints. In: Proceedings of the ECCV Workshop on Statistical Learning in Computer Vision, pp. 1–22 (2004)
Van der Maaten, L.J. P., Postma, E.O., Vanden Herik, H.J.: Dimensionality Reduction: A Comparative Review. Technical report, Tilburg University (2009)
Burt, P.J., Adelson, E.H.: The Laplacian pyramid as a compact image code. IEEE Trans. Commun. 31, 532–540 (1983)
Karlsson, A.: Nonexpanding maps, Busemann functions, and multiplicative ergodic theory. In: Rigidity in Dynamics and Geometry, pp. 283–294. Springer (2002)
Borgefors, G., Ramella, G., di Baja, G.S.: Shape and topology preserving multi-valued image pyramids for multi-resolution skeletonization. Pattern Recognit. Lett. 22, 741–751 (2001)
Kropatsch, W.G., Haxhimusa, Y., Pizlo, Z., Langs, G.: Vision pyramids that do not grow too high. Pattern Recognit. Lett. 26, 319–337 (2005)
Lu, H., Plataniotis, K.N., Venetsanopoulos, A.N.: A survey of multilinear subspace learning for tensor data. Pattern Recognit. 44, 1540–1551 (2011)
Yang, J., Zhang, D., Frangi, A.F., Yang, J.-Y.: Two-dimensional PCA: a new approach to appearance-based face representation and recognition. IEEE Trans. Pattern Anal. Mach. Intell. 26, 131–137 (2004)
Otsu, N.: Mathematical Studies on Feature Extraction in Pattern Recognition. PhD thesis, Electrotechnical Laboratory (1981)
Aase, S.O., Husoy, J.H., Waldemar, P.: A critique of SVD-based image coding systems. Proc. IEEE Int. Symp. Circuits Syst. 4, 13–16 (1999)
Ding, C., Ye, J.: Two-dimensional singular value decomposition (2DSVD) for 2D maps and images. In: Proceedings of the SIAM International Conference on Data Mining, pp. 32–43 (2005)
Ye, J., Janardan, R., Qi, L.: GPCA: An efficient dimension reduction scheme for image compression and retrieval. In: Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, pp. 354–363 (2004)
Moore, J.B., Mahony, R.E., Helmke, U.: Numerical gradient algorithms for eigenvalue and singular value calculations. SIAM J. Matrix Anal. Appl. 15, 881–902 (1994)
Lu, H., Plataniotis, K.N., Venetsanopoulos, A.N.: MPCA: Multilinear principal component analysis of tensor objects. IEEE Trans. Neural Netw. 19(1), 18–39 (2008)
Lathauwer, L.D., Moor, BDe, Vandewalle, J.: A multilinear singular value decomposition. SIAM J. Matrix Anal. Appl. 21(4), 1253–1278 (2000)
Wang, H., Ahuja, N.: Compact representation of multidimensional data using tensor randk-one decomposition. In: Proceedings of the International Conference on Pattern Recognition, vol. 1, pp. 44–47 (2004)
Lu, H., Plataniotis, K.N., Venetsanopoulos, A.N.: Uncorrelated multilinear principal component analysis for unsupervised multilinear subspace learning. IEEE Trans. Neural Netw. 20(11), 1820–1836 (2009)
Allen, G.I.: Sparse higher-order principal components analysis. In: Proceedings of the International Conference on Artificial Intelligence and Statistics, pp. 27–36 (2012)
Johnson, W., Lindenstrauss, J.: Extensions of Lipschitz maps into a Hilbert space. Contemp. Math. 26, 189–206 (1984)
Arya, S., Mount, D.M., Netanyahu, N.S., Silverman, R., Wu, A.Y.: An optimal algorithm for approximate nearest neighbor searching in fixed dimensions. In: Proceedings of the ACM-SIAM Symposium on Discrete Algorithms, pp. 573–582 (1994)
Achlioptas, D., McSherry, F.: Fast computation of low-rank matrix approximations. J. ACM 54(2), 9 (2007)
Sakai, T., Imiya, A.: Practical algorithms of spectral clustering: toward large-scale vision-based motion analysis. In: Machine Learning for Vision-Based Motion Analysis, pp. 3–26. Springer (2011)
Baraniuk, R.G., Wakin, M.B.: Random projections of smooth manifolds. Found. Comput. Math. 9, 51–77 (2009)
Bingham, E., Mannila, H.: Random projection in dimensionality reduction: Applications to image and text data. In: Proceedings of the International Conference on Knowledge Discovery and Data Mining, pp. 245–250 (2001)
Achlioptas, D.: Database-friendly random projections: Johnson–Lindenstrauss with binary coins. J. Comput. Syst. Sci. 66, 671–687 (2003)
Watanabe, T., Takimoto, E., Amano, K., Maruoka, A.: Random projection and its application to learning. In: Proceedings of the Workshop on Randomness and Computation, pp. 3–4 (2005)
Matousek, J.: On variants of the Johnson–Lindenstrauss lemma. Random Struct. Algorithms 33, 142–156 (2008)
Ailon, N., Liberty, E.: Almost optimal unrestricted fast Johnson–Lindenstrauss transform. ACM Trans. Algorithms 9, 21:1–211:2 (2013)
Schölkopf, B., Smola, A., Müller, K.-R.: Nonlinear component analysis as a kernel eigenvalue problem. Neural Comput. 10, 1299–1319 (1998)
Borg, I., Groenen, P.: Modern Multidimensional Scaling: Theory and Applications, 2nd edn. Springer, New York (2005)
Williams, C.K.I.: On a connection between kernel PCA and metric multidimensional scaling. Mach. Learn. 46, 11–19 (2002)
Tenenbaum, J.B., de Silva, V., Langford, J.C.: A global geometric framework for nonlinear dimensionality reduction. Science 290, 2319–2323 (2000)
Roweis, S.T., Saul, L.K.: Nonlinear dimensionality reduction by locally linear embedding. Science 290, 2323–2326 (2000)
Venna, J., Kaski, S.: Local multidimensional scaling. Neural Netw. 19, 889–899 (2006)
Vidal, R., Yi, M., Sastry, S.: Generalized principal component analysis (GPCA). IEEE Trans. Pattern Anal. Mach. Intell. 27, 1945–1959 (2005)
Goh, A., Vidal, R.: Clustering and dimensionality reduction on Riemannian manifolds. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pp. 1–7 (2008)
Harandi, M.T., Salzmann, M., Hartley, R.: From manifold to manifold: geometry-aware dimensionality reduction for PD matrices. Proc. Eur. Conf. Comput. Vis. 8690, 17–32 (2014)
Fisher, R.A.: The use of multiple measurements in taxonomic problems. Ann. Eugen. 7, 179–188 (1936)
Vapnik, V., Lerner, A.: Pattern recognition using generalized portrait method. Autom. Remote Control 24, 774–780 (1963)
Iijima, T.: Theory of pattern recognition. Electronics and Communications in Japan, pp. 123–134 (1963)
Watanabe, S.: Karhunen–Loeve expansion and factor analysis. In: Proceedings of the Transactions of the Fourth Prague Conference on Information Theory, Statistical Decision Functions, Random Processes, pp. 635–660 (1967)
Itoh, H., Sakai, T., Kawamoto, K., Imiya, A.: Topology-preserving dimension-reduction methods for image pattern recognition. In: Proceedings of the Scandinavian Conference on Image Analysis, pp. 195–204 (2013)
Fukui, K., Maki, A.: Difference subspace and its generalization for subspace-based methods. IEEE Trans. Pattern Anal. Mach. Intell., pp. 1–1, In print (2015)
Cock, K.D., Moor, B.D.: Subspace angles between ARMA models. Syst. Control Lett. 46, 265–270 (2002)
Hamm, J., Lee, D. D.: Grassmann discriminant analysis: a unifying view on subspace-based learning. In: Proceedings of the International Conference on Machine Learning, pp. 376–383 (2008)
Boser, E., Guyon, I., Vapnik, V.: A training algorithm for optimal margin classifiers. In: Proceedings of the Workshop on Computational Learning Theory, pp. 144–152 (1992)
Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20(3), 273–297 (1995)
Hsu, C.-W., Lin, C.-J.: A comparison of methods for multiclass support vector machines. IEEE Trans. Neural Netw. 13, 415–425 (2002)
Sivic, J., Zisserman, A.: Efficient visual search of videos cast as text retrieval. IEEE Trans. Pattern Anal. Mach. Intell. 31, 591–606 (2009)
Jégou, H., Perronnin, F., Douze, M., Sanchez, J., Perez, P., Schmid, C.: Aggregating local image descriptors into compact codes. IEEE Trans. Pattern Anal. Mach. Intell. 34, 1704–1716 (2012)
Vempala, S.S.: The Random Projection Method. American Mathematical Society, Providence (2004)
Magen, A.: Dimensionality reductions that preserve volumes and distance to affine spaces, and their algorithmic applications. Proc. Randomization and Approximation Techniques in Computer. Science 2483, 239–253 (2002)
Agarwal, P. K., Har-Peled, S., Yu, H.: Embeddings of surfaces, curves, and moving points in Euclidean space. In: Proceedings of the Annual Symposium on Computational Geometry, pp. 381–389 (2007)
Dasgupta, S., Gupta, A.: An elementary proof of the Johnson–Lindenstrauss lemma. Technical report, UC Berkeley (1996)
Helmke, U., Moore, J.B.: Singular-value decomposition via gradient and self-equivalent flows. Linear Algebra Appl. 169, 223–248 (1992)
Liang, Z., Shi, P.: An analytical algorithm for generalized low-rank approximations of matrices. Pattern Recognit. 38, 2213–2216 (2005)
Itoh, H., Sakai, T., Kawamoto, K., Imiya, A.: Dimension reduction methods for image pattern recognition. In: Proceedings of the International Workshop on Similarity-Based, Pattern Recognition, pp. 26–42 (2013)
Björck, Å., Golub, G .H.: Numerical methods for computing angles between linear subspaces. Math. Comput. 27, 579–594 (1975)
Golub, G.H., Van Loan, C.F.: Matrix Computations. The Johns Hopkins University Press, Baltimore (1996)
Georghiades, A.S., Belhumeur, P.N., Kriegman, D.J.: From few to many: Illumination cone models for face recognition under variable lighting and pose. IEEE Trans. Pattern Anal. Mach. Intell. 23, 643–660 (2001)
Samaria, F., Harter, A.: Parameterisation of a stochastic model for human face identification. In: Proceedings of the IEEE Workshop on Applications of Computer Vision (1994)
Leibe, B., Schiele, B.: Analyzing appearance and contour based methods for object categorization. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition vol. 2, pp. 409–415 (2003)
Mobahi, H., Collobert, R., Weston, J.: Deep learning from temporal coherence in video. In: Proceedings of the International Conference on Machine Learning (2009)
LeCun, Y., Bottou, L., Bengio, Y., Haffner, P.: Gradient-based learning applied to document recognition. Proc. IEEE 86, 2278–2324 (1998)
Saito, T., Yamada, H., Yamada, K.: On the data base ETL9 of handprinted characters in JIS Chinese characters and its analysis. IEICE Trans. J68–D, 757–764 (1985)
Fei-Fei, L., Fergus, R., Perona, P.: Learning generative visual models from few training examples: an incremental Bayesian approach tested on 101 object categories. Comput. Vis. Image Underst. 106, 59–70 (2007)
Everingham, M., Van Gool, L., Williams, C. K. I., Winn, J., Zisserman, A.: The PASCAL Visual Object Classes Challenge 2010 (VOC2010) Results. http://www.pascal-network.org/challenges/VOC/voc2010/workshop/index. html
Wright, J., Yang, A.Y., Ganesh, A., Sastry, S.S., Ma, Y.: Robust face recognition via sparse representation. IEEE Trans. Pattern Anal. Mach. Intell. 31, 210–227 (2009)
Kim, T.-K., Kittler, J., Cipolla, R.: Discriminative learning and recognition of image set classes using canonical correlations. IEEE Trans. Pattern Anal. Mach. Intell. 29, 1005–1018 (2007)
Liu, H., Ding, X.: Handwritten character recognition using gradient feature and quadratic classifier with multiple discrimination schemes. In: Proceedings of the International Conference on Document Analysis and Recognition, pp. 19–23 (2005)
He, K., Zhang, Z., Ren, S., Jian, S.: Spatial pyramid pooling in deep convolutional networks for visual recognition. In: Proceedings of the European Conference on Computer Vision, pp. 346–361 (2014)
Everingham, M., Eslami, S.M.A., Gool, L.V., Williams, C.K.I., Winn, J., Zisserman, A.: The Pascal Visual Object Classes challenge: a retrospective. Int. J. Comput. Vis. 111, 98–136 (2015)
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Itoh, H., Imiya, A. & Sakai, T. Dimension Reduction and Construction of Feature Space for Image Pattern Recognition. J Math Imaging Vis 56, 1–31 (2016). https://doi.org/10.1007/s10851-015-0629-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10851-015-0629-1