Texture and Color Features based Color Image Retrieval using Canonical Correlation

Table of contents

1. I. Introduction

he proliferation of digitizing the functions of private and public sectors, governmental organizations impetus to increase the repository capacity of the devices as well as the speed of accessing and retrieving the data through the internet. Owing to awareness and usage of the Internet and World Wide Web (WWW), a number of Internet users increased exponentially. Perhaps, a number of approaches have been developed with a consideration of the internet users and the difficulties in accessing the data available on the Internet and the WWW, they are not completely fulfilling the users' requirements. Specifically, images demand large amount of space and consume more time to retrieve from the image repository.

At the earlier stage, the text annotation based image retrieval method has been developed which is time consuming (Yap and Wu, 2015), and also it is not effective and efficient because it retrieves the images based on the text annotated manually by the user. Then, the content-based image retrieval (CBIR) method has been developed, which attracted a number of researchers. Most of the techniques have been developed based on the contents of the images, namely low-level global visual features, viz. colour properties, shape, texture, spatial orientation, etc., which are used as a query for the retrieval process (Huang et al., 1997;Stricker and Orengo, 1995;Pentland et al., 1994;Fuh et al., 2000). Liu, et. al. (2011) propose micro-structure Author ?: Division of Computer and Information Sciences, Faculty of Science, Annamalai University, Annamalainagar -608 002, Tamil Nadu, INDIA. e-mail: [email protected] Author ?: B.Tech (ECE), Pondicherry Engineering College, Pondicherry-605014. e-mail: [email protected] descriptor, which extracts and integrates colour, texture, shape and colour layout information for image retrieval. The authors claim that it has only 72 dimensions for full colour image. But this is too large compared to those of the methods proposed by (Seetharaman, 2015;Krishnamoorthi and Sathiya Devi, 2013). The work proposed in (Murala et al., 2012) encodes the relationship between reference pixel and its surrounding neighbors by comparing gray-level values, and it extracts the edge information on local extreme in different angles. Though, a number of works have been developed based on the statistical distributional approaches and parametric tests of hypothesis, most of them are not effective and efficient, since there is no guarantee that all the images are distributed as independent and identical to Gaussian random process. The statistical parametric tests can be applied on an image, if it is distributed to Gaussian process. Otherwise, it does not lead to a good result. Generally, the parametric tests can be used if the data are quantitative and are distributed as independent and identical to Gaussian process. This motivated to develop the proposed method.

This paper adopts the idea of Canonical correlation analysis (CCA) for image retrieval. It is an effective and efficient technique for both textured and structured images, since it does not strictly rely on the distributional process. The Correlation analysis is dependent on the co-ordinate system, in which the variables are described, so even if there is a very strong linear relationship between two sets of multidimensional variables, this relationship might not be visible as a correlation. The CCA seeks a pair of linear transformation for each set of variables such that when the set of variables are transformed, the corresponding co-ordinates are maximally correlated. It is one of the valuable multi-data processing methods (Sun, et al. 1994;Zhang, et al. 1999). In recent years, CCA has been applied to several fields such as signal processing, computer vision, neural network and speech recognition.

In recent years, there has been a vast increase in the amount of multimedia content available both in off-line and online, though it is unable to access or make use of this data unless it is organized in such a way as to allow efficient browsing. To enable content based image retrieval with no reference to labeling, this paper attempts to learn the semantic representation of the images, and their associated text. This paper presents a general approach using KCCA that can be used for content based (David and John, 2003) as well as region based retrieval (Fyfe and Lai, 2001;David and John, 2003). In both cases, the KCCA approach and the Generalized Vector Space Model (GVSM) are compared, which aims at capturing some term to term correlations by looking at co-occurrence information.

The canonical correlation is the most general multivariate form, because multiple regression, discriminant function analysis and MANOVA are the special cases of the Canonical correlation analysis. Thus, it yields better results. The one more advantage of using canonical correlation is that there is no specific requirement for the data that should follow the normality conditions, if it is a descriptive analysis.

2. II. Proposed Retrieval Method

m m 2 2 1 1 q x x x I f f f + + + = ? (1)

and the linear combination of the feature vectors i y of the target image is

n n 2 2 1 1 t y y y I f f f + + + = ?(2)

The first canonical correlation is the maximum among the correlation coefficients between the variates q I and t I that is considered as the canonical correlation coefficient between the query and target images.

The function to be maximized is

) I S , I (S max t y q x I , I t q corr = ? t y q x t y q x I , I I S I S I S , I S max t q = ? If [ ] ) y , x ( f ? denote the empirical expectation of the function, ) y , x ( f , then [ ] ? = = ? N 1 i i i ) y , x ( N 1 ) y , x ( f f

Now, the correlation expression can be rewritten as in equation ( 3)

[ ] ? ? ? ? ? ? = ? ? ? ? 2 t 2 q t q t q y , I E x , I E y , I x , I E ) I , I max( (3) [ ] [ ] [ ] t ' t q ' q t ' q t q I ' yy I E I ' xx I E I ' xy I E ) I , I max( ? ? ? = ? (4) Follows that [ ] [ ] [ ] t ' t q ' q t ' q t q I ' yy E I I ' xx E I I ' xy E I ) I , I max( ? ? ? = ? (5) C Cyy Cyx Cxy Cxx y x y x E ) y , x ( C ' = ? ? ? ? ? ? = ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? ? = ? (6)

The total covariance matrix C is a block matrix, where the within-sets covariance matrices are xx C and , yy C and the between-sets covariance matrices are

' yx xx C C = .

Thus, we can rewrite the function, ? as

t yy ' t q xx ' q t xy ' q t q I C I I C I I C I ) I , I max( = ? (7)

The maximum canonical correlation is the maximum of ? with respect to q I and t I . Statistical hypothesis:

0 ) I , I ( : H 0 ) I , I ( : H t q t q 0 ? ? = ? ? ) 1 ( ) 1 n m ( 2 1 1 p 2 2 j ) n , m min( 1 j ? = ? ? ? ? ? ? ? ? ? + + ? ? ? = ? ln (8)

Which is asymptotically distributed as a Chisquared with

3. IV. Measure of Performance

In order to measure the performance of the proposed method, the F ? -measure is adopted, which is a harmonic mean of precision (P) and recall (R) values, and that are given in equations ( 9), (10) and (11). The F ? , measures the effectiveness of retrieval with respect to a user who attaches ? times as much importance to recall as precision (Van Rijsbergen, 1979). The ? value is decided with respect to weight which is given to the P ? is considered as 0.5, since more weight is given to the precision. Thus, equation ( 9) is rewritten as in equation (12).

( ) ( ) ) R ( P ) R ( P ) 1 ( measure F 2 2 + × ? ? × ? + = ? (9)

Where, In the context of the proposed research, the ? is considered as 0.5 since more weight is given to the precision. Thus, equation ( 9) can be written as in equation ( 12).

4. { } { }

Note: Images

5. V. Experiments and Results

A colour image is given as input key image to the system, and the shapes in the image are segmented; the segmented image is modelled to HSV colour space. The texture features are extracted from the V component of the HSV colour space. The colour features, such as H, S and V, and the texture feature are considered as observation and used in the CCA method expressed in equation (7). For a sample, due to space constraints, a shape segmented colour image is presented in Fig. 1. In order to experiment the CCA method discussed in section 2, the expression in equation ( 7) is implemented with the image and their feature databases constructed and discussed in section 4 that measures the correlation between both query and target images. Due to space constraints, for a sample, some of the images considered from the image database are presented in this paper that have been included during the experiments.

The target images are identified and marked in the image database, based on the results obtained by experiments. According to the users' requirements, the user can fix the level of significance of the tests by which the user can retrieve a number of same or similar images. The proposed system matches and retrieves the same image (target and query images are the same) from the image feature database at the level of significance 0% or 0.001 (100% accuracy -same images); almost the same images for 1% to 2% (99% to 98% accuracy) level of significance; similar images for the level of significance from 3% to 5%; related images for the level of significance from 6% to 8%. The selected images are indexed (ranked) from lowest to highest, i.e. in an ascending order based on the test statistic values obtained. The image which corresponds to the first value in the indexed list is marked as target image and it is retrieved from the image database. In order to validate the proposed system, the image in column 1 of Figure 1(a) is given as input query image, for which the system retrieves the images in row against them. The retrieved output images show that the proposed CCA method is robust for scaling and rotation for both textured and structured images. Since the proposed system serves same as the distributional approach, the proposed system acts as invariant for rotation and noise.

The experiment is conducted at various levels of significance on the input query Wall Street Bull image given in Figure 2(a). The images in column 2 of the Figure 2(b) are retrieved at 2% level of significance; images in columns 3, 2 and 1 are retrieved at 5% level of significance; at 8% and 12% level of significance, the system retrieves the images in columns 4, 3, 2, 1 and 7 respectively; at 15% level of significance, all the images presented in Figure 2(b) are retrieved. The number of shapes in both query and target images are compared, if it matches, then it is emphasized that both the images are same or similar. The obtained output of the experiments, based on the canonical correlation, is tabulated in Table 1; it can be observed from the output that the proposed CCA yields 0.9982 for the images Wall Street Bull-1 versus Wall Street Bull -1, i.e. the query and target images are same. Similarly, the canonical correlation coefficients are obtained for different combinations of the Bull images are presented in Table 1. The computed canonical correlation coefficient matrix is a symmetric. The main advantage of the proposed system is that it facilitates the user to retrieve only the required number of images by fixing the level of significance, whereas the existing methods retrieve a number of similar images. Thus, in the existing methods, the user has to expend time to select the required image from the set of retrieved same or similar images. In addition to that the proposed system retrieves a set of similar images by fixing the level of significance at a desired level. Furthermore, to prove the efficacy of the proposed method, a number of structured images are considered for the experiment. For a sample, a few of them are presented in Figure 3. The image given in Figure 3(a) is fed as input query to the system, and the level of significance is fixed at 0.08 (i.e. 8 percent); for which the system retrieves the images presented in row 1. The images in row 2 are retrieved, while the level of significance is fixed at 0.12 or lesser (i.e. 12 percent or lesser), whereas it retrieves the images in rows 3 and 4 at the level of significance 0.15 The obtained results reveal that the proposed method outperforms the existing systems. But, this paper takes account of texture and structure images. Averages of the precision and recall values are calculated for various methods: OP, BD and MD measures, and the results are presented in Table 2. Bar chart is plotted for the observed results, and is presented in Figures 5.

6. Time Complexity

The performance of the proposed system is compared with the existing systems: OP method, BD and MD distance measures in terms of computational time complexity. The system is implemented using Java SE 7 compiler with the system specification, Intel Core i5-4440 processor based PC with 4GB DDR3 RAM. The time consumed by the proposed system for feature extraction, matching and retrieving the images is measured in terms of seconds after a rigorous experimentation, and the obtained results are presented in Table 3. The proposed system demands lesser time compared to that of the existing systems, and also the proposed system yields better retrieval results.

7. VI. Conclusion

In this paper a novel technique, Canonical correlation analysis and the Chi-square test are employed. The Chi-square test tests the correlation coefficient between the query and target images that whether it is significant or not. The canonical correlation coefficient extracts the spatial relationship among the pixels in the image, by which the spatial orientation of the texture properties are extracted; and the properties between the query and target images are tested. If it is significant, then it is concluded that the input query and target images are same or similar; otherwise, it is inferred that the two images differ significantly. It is observed from the results that the proposed technique yields better results than the existing techniques J e XV Issue VI Version I

Figure 1. Je
XV Issue VI Version I Texture and Color Features based Color Image Retrieval using Canonical Correlation Where, ' I represents the transpose of a vector or matrix I. The covariance matrix of (x, y) is
Figure 2. © 20 15 3 2015 J
1532015Global Journals Inc. (US) Global Journal of Researches in Engineering ( ) Volume XV Issue VI Version I Year Texture and Color Features based Color Image Retrieval using Canonical Correlation and R. In the context of the proposed research work, the
Figure 3. Figure 1 :
1Figure 1: Segmented Shapes: (a): actual image; (b): red component shape; (c): green component shape; (d): blue component shape; (e): combination of green and blue; (f): combination of red and green.
Figure 4. I
The canonical correlation coefficient ? is computed, based on the equation(8), on the input query image. The query and target images are considered in various combinations, and the computed ? values are tested using the Chi-squared test statistic given in equation (9) of the section 2.© 2015 Global Journals Inc. (US) Texture and Color Features based Color Image Retrieval using Canonical Correlation
Figure 5. Figure 3 :
3Figure 3 : Structure images downloaded from Internet -(a): input key image; (b): retrieved output imagesMoreover, to emphasis that the system works well for the textured or semi-structured images, a number of images are considered from the Corel-10K and 5K image databases; and the obtained retrieved images are present in Figure4. The system retrieves the images in columns 2-6 for the input key images in column 1. Images in in column 2 are retrieved at the level of significance 0.01 (i.e. 1%); the images in column 3 are retrieved while the level of significance is fixed at 0.08 or lesser (8% or lesser); while the level of significance is at 0.15 or lesser (i.e. 15% or lesser), the images in columns 4 and 5 are retrieved; images in column 7 are retrieved at the level of significance, 0.18 (i.e. 18% or lesser).
Figure 6. Figure 4 :
4Figure 4 : Corel 5K and 10K images -Texture or Semistructured images. Column 1: input query images; columns 2-5: retrieved output images A comparative study is performed to measure the efficiency of the proposed method with the existing methods: Orthogonal Polynomial (OP) (Krishnamoorthi and Sathiya Devi, 2013), Bhattacharyya distance (BD) (Yang and Newsam, 2013), and Mahalanobis distance (MD) (Hoi, et al., 2010) in terms of average of the precision and recall values of the retrieved results. The average of the precision and recall values are obtained at the level of significance (?) 0.12 percent and above.
Figure 7. Figure 2 :
2Figure 2 : Wall Street Bull -(a): actual image with size 75 × 100; row 2 of (b): actual image with size 96 × 128; row 3 of (b): images in row 2 are rotated clockwise through 90 degrees; row 4 of (b): images in row 2 are rotated clockwise through 180 degrees.
Figure 8. Figure 5 :
5Figure 5 : Bar Chart: Comparison of Retrieval Rates of Different Methods
Figure 9.
(m - i + 1)(n - i + 1) degrees freedom for
large sample size, N. If ( ? I q , I t ) < ? 2 ? , then the two
images belong to the same class; otherwise, the two images are different, where ? is the level of
significance (Kanti et al., 1979).
Therefore, totally there are [1932 (actual image) + ((16
× (695+682)) + (4676 + 3336 + 558 + 537)) × 3
(rotated through 90 o , 180 o and 270 o ) +30,289 (scaled)
+ 626 (satellite)] = 1, 28, 482 images. Based on these
image collections, an image database and FV database
are constructed.
Figure 10. Table 1 :
1
Bull images
Vs. Bull-1 Bull-2 Bull-3 Bull-4 Bull-5
Bull-1 0.9999 0.9881 0.7506 0.7457 0.6152
Bull-2 0.9881 0.9998 0.7561 0.7389 0.5896
Bull-3 0.7506 0.7561 0.9999 0.7931 0.6856
Bull-4 0.7457 0.7389 0.7931 0.9998 0.7953
Bull-5 0.6152 0.5896 0.6856 0.7953 1.0000
Figure 11. Table 2 :
2
Image Database Proposed system (? is at 0.12 and above) OP BD MD
R P F 0.5 R P F 0.5 R P F 0.5 R P F 0.5
CalTech 0.9585 0.7952 0.8233 0.9015 0.7631 0.7873 0.8523 0.6937 0.7205 0.8252 0.6817 0.7063
Corel 0.9098 0.8585 0.8683 0.9255 0.8039 0.8256 0.7829 0.7736 0.7754 0.8095 0.6912 0.7120
Structure 0.9681 0.7852 0.8160 0.8762 0.7729 0.7916 0.8139 0.6483 0.6758 0.8069 0.6275 0.6567
Images
Figure 12. Table 3 :
3
Time Consumption Proposed System OP BD MD
Feature
extraction time 0.485 sec 0.798 sec 0.975 sec 0.852 sec
Searching time 0.0581 sec 0.072 sec 0.070 sec 0.069 sec
1

Appendix A

  1. Evaluation of shape descriptors for shape-based image retrieval. A Amanatiadis , V G Kaburlasos , A Gasteratos , S E Papadakis . IET Image Processing 2011. 5 (5) p. .
  2. Photobook: Content-based manipulation of image databases. SPIE Storage and Retrieval for Image and Video Databases II, A Pentland , R Picard , Sclaroff . 1994. p. .
  3. Hierarchical colour image region segmentation for contentbased image retrieval system. C-S Fuh , S-W Cho , K Essig . IEEE Transactions on Image Processing 2000. 9 (1) p. .
  4. Kernel and nonlinear canonical correlation analysis. C Fyfe , P L Lai . International Journal of Neural Systems 2001.
  5. Ch Srinivasa Rao , S Srinivas Kumar , B N Chatterji . Content based image retrieval using Contourlet Transform, 2007. 7 p. .
  6. A novel region-based image retrieval method using relevance feedback. F Jing , B Zhang , F Z Lin , Y W Ma , H J Zhang . Proc. 3 rd ACM Int. Workshop on Multimedia Information Retrieval, (3 rd ACM Int. Workshop on Multimedia Information Retrieval) 2001. MIR2001.
  7. An efficient and effective region-based image retrieval framework. F Jing , M Li , H J Zhang , B Zhang . IEEE Transactions on Image Processing 2004. 13 (5) p. .
  8. Handwritten numerical recognition based on multiple algorithm. F Kimura , M Shridhar . Pattern Recognition 1991. 24 (10) p. .
  9. Image retrieval based on micro-structure descriptor. G H Liu , Z-Y Lib , L Zhangc , Y Xud . Pattern Recognition 2011. 44 (9) p. .
  10. Multiple classifiers for colour flag and trademark image retrieval. I-S Hsieh , H C Fan . IEEE Transactions on Image Processing 2001. 10 (6) p. .
  11. Spatial template extraction for image retrieval by region matching. J-W Hsieh , W Eric , L Grimson . IEEE Transactions on Image Processing 2003. 12 (11) p. .
  12. Image indexing using colour correlogram. J Huang , S R Kunar , M Mitra , W J Zhu , R Zabih . Proc. IEEE Comp. Soc. Conf. Comp. Vis. and Patt. Recog, (IEEE Comp. Soc. Conf. Comp. Vis. and Patt. Recog) 1997. 1997. p. .
  13. Image classification and querying using composite region templates. J R Smith , C S Li . Journal of Computer Vision and Image Understanding 1999. 75 (12) p. .
  14. Scalable integrated region-based image retrieval using IRM and statistical clustering. J Z Wang , Y P Du . Proc. ACM and IEEE Joint Conference on Digital Libraries, (ACM and IEEE Joint Conference on Digital Libraries) 2001. 2001. VA. p. .
  15. A soft relevance framework in content-based image retrieval systems. K-H Yap , K Wu . IEEE Transactions on Circuits and Systems 2005. 15 (2) p. .
  16. Statistical framework for image retrieval based on multiresolution features and similarity method. K Seetharaman , M Kamarasan . Multimedia Tools and Applications 2014. 73 (3) p. .
  17. Statistical Distributional Approach for Scale and Rotation Invariant Colour Image Retrieval Using Multivariate Parametric tests and Orthogonality Condition. K Seetharaman , M Jaikarthic . Journal of Visual Communication and Image Representation 2014. 25 (5) p. .
  18. Image retrieval based on micro-level spatial structure features and content analysis using Full Range Gaussian Markov Random Field model. K Seetharaman . Engineering Applications of Artificial Intelligence 2015. 40 p. .
  19. Storage and Retrieval for Image and Video Databases. M Stricker , M Orengo . Proc. (SPIE 2420), (SPIE 2420) 1995. p. . (Similarity of colour images)
  20. Kcca for different level precision in content-based image retrieval. R David , John Hardoon , Shawe-Taylor . Submitted to Third International Workshop on Content-Based Multimedia Indexing, (Rennes, France
    ) 2003.
  21. A simple computational model for image retrieval with weighted multifeatures based on orthogonal polynomials and genetic algorithm. R Krishnamoorthi , S Sathiya Devi . Neurocomputing 2013. 116 p. .
  22. Recognition of images in large databases using colour and texture. S Belongie , C Carson , H Greenspan , J Malik . IEEE Transactions on Pattern Analysis and Machine Intelligence 2002. 24 (8) p. .
  23. Semi-Supervised Distance Metric Learning for Collaborative Image Retrieval and Clustering. S C H Hoi , W Liu , S F Chang . 10.1145/1823746.1823752. ACM Transactions on Multimedia Computing, Communications, and Applications 2010. 6 (3) .
  24. Directional local extrema patterns: a new descriptor for content based image retrieval. S Murala , R P Maheshwari , R Balasubramanian . International journal of Multimedia Information Retrieval 2012. 1 (3) p. .
  25. Decision combination in multiple classifier systems. T K Ho , J J Hull , S N Srihari . IEEE Transactions on Pattern Analysis and Machine Intelligence 1994. 16 (1) p. .
  26. Multivariate Statistical Analysis. W S Sun , L X Chen . Information Retrieval 1994. 1979. Higher Education Press. 26 (2) . (C. J.. nd ed.)
  27. Multivariate Statistical Introduction, X T Zhang , K K Fang . 1999. Beijing: Sciences Press.
  28. An efficient lowdimensional colour indexing scheme for regionbased image retrieval. Y Deng , B S Manjunath . Proc. IEEE Int. Conf. ASSP, (IEEE Int. Conf. ASSP) 1999. 6 p. .
  29. Geographic Image Retrieval Using Local Invariant Features. Y Yang , S Newsam . IEEE Transactions on Geoscience and remote Sensing 2013. 51 (2) p. .
Notes
1
© 20 15 Global Journals Inc. (US) Global Journal of Researches in Engineering ( ) Volume XV Issue VI Version I
Date: 2015-03-15