Feature extraction using wavelet and fractal

https://doi.org/10.1016/S0167-8655(01)00003-4Get rights and content

Abstract

In this paper, we are investigating the utility of several emerging techniques to extract features. A novel method of feature extraction is proposed, which includes utilizing the central projection transformation (CPT) to describe the shape, the wavelet transformation to aid in the boundary identification, and the fractal features to enhance image discrimination. It reduces the dimensionality of a two-dimensional pattern by way of a central projection approach, and thereafter, performs Daubechies' wavelet transform on the derived one-dimensional pattern to generate a set of wavelet transform sub-patterns, namely, curves that are non-self-intersecting. The divider dimensions are computed from these curves with a modified box-counting approach. These divider dimensions constitute a new feature vector for the original two-dimensional pattern, defined over the curve's fractal dimensions. We have conducted several experiments in which a set of printed Chinese characters, English letters of varying fonts and other images were classified. Based on the Euclidean distance between the different feature vectors, the experiments have satisfying results.

Introduction

Feature extraction is the heart of a pattern recognition system. In pattern recognition, features are utilized to identify one class of pattern from another. The goal of feature extraction is to find a transformation from an n-dimensional observation space X to a smaller m-dimensional feature space XiT=(xi1,xi2,…,xin), i=1,…,m that retains most of the information needed for pattern classification. The coordinate axes that define the feature space XiT are called features. There are two main reasons for performing feature extraction. First, the computational complexity for pattern classification is reduced by dealing with the data in a lower dimensional space. Secondly, for a given number of training samples, one can generally obtain more accurate estimates of the class-conditional density functions and thus formulate a more reliable decision rule. Whether or not this decision rule actually performs better than the one applied in the observation space depends on how much information was lost in the feature transformation. In some cases, it is possible to derive features that sacrifice none of the information needed for classification.

In this paper, one may think that the best description of a pattern is itself. So, why do we need the feature extraction step? In other words, why not use the pattern themselves as features? Classification may be seen as a function C from the feature space XiT to the pattern space X(C:X→XiT), where C is the set of all possible classes. If X has a big dimension, and this will be the case if bitmaps are used as features, the application of C may require a lot of CPU time. By extracting features from the pattern we decrease the dimensionality of XiT and therefore speed up the classification. The second important reason is that normalization is faster and more accurate in feature space than in pattern space.

The “source” features can be extracted from the bitmaps. This is not the only possible source. Other commonly used methods can be described below (Jain et al., 1996; Sazaklis, 1997):

1. Extraction from the boundary curves of the pattern. A boundary following algorithm can be run on the bitmap in order to obtain the coordinates [x(t),y(t)] of the contour of the pattern. This curve is closed and therefore the functions x(t) and y(t) are periodic. Feature extraction from the boundaries can, however, easily be normalized to translation, scaling and rotation. In addition, they have low sensitivity to noise. Of course, methods based on these features cannot be used if noisy gaps “open” the characters, or if the characters are not connected.

2. Extraction from the HV-projections. From the bitmap b(x,y), 0⩽xW−1, 0⩽yH−1, the following two functions are computed:h(y)=∑i=0W−1b(i,y)andv(x)=∑j=0H−1b(x,j),h and v are, respectively, called the horizontal and vertical projections of b. Features computed from h and v cannot be normalized to rotation nor can the shapes of the characters be reconstructed from h(y) and v(x).

3. Pattern profiles. A symbol can be described in terms of its four profiles: left L(y), right R(y), top T(x), and bottom B(x). The profiles give a measure of the variations of the shape on each side of the character. For the bitmap b(x,y), 0⩽xW−1, 0⩽yH−1, and the left profile L(y) may be defined byL(y):xifb(x,y)isthefirstsetofpixelsontherowy,Wifthereisnosetofpixelontherowy.Fig. 1 shows the left and right profiles of the character `R'. The profile functions are very sensitive to noise, especially if it is present all over the bitmap. The contours cannot keep the information about the inside patterns.

4. Direct extraction from the bitmap. No special remark needs to be done about this source, all depend on the features that are extracted.

5. Extraction from gray-level rasters. The four previous schemes are used in bitmaps. Recently some methods have been described that use gray-level rasters instead. Variations in the gray-levels are used either directly as features or for generating functions from which features are computed.

6. Structural features. Structural features aim at capturing the essential shape information of the characters, generally from their skeletons, and sometimes from their contours. The features include: loops, junctions, crossing and end points, concavity and convexity, arcs and strokes.

Pattern recognition requires the extraction of features from regions of the image, and the processing of these features with a pattern classification algorithm. Many of the features used in our applications tend to be local in nature, which means their calculation requires a connected region of the image over which an average or other statistics is extracted. In this paper, we present a novel approach to extracting features in pattern recognition that utilizes a central projection transformation which combines the wavelet and fractal theories. In particular, this approach reduces the dimensionality of a two-dimensional pattern by way of a central projection method, and thereafter, performs Daubechies' wavelet transform on the derived one-dimensional pattern to generate a set of wavelet transformation sub-patterns, namely, curves that are non-self-intersecting. The divider dimensions are readily computed from the resulting non-self-intersecting curve. These divider dimensions constitute a new feature vector for the original two-dimensional pattern, defined using the curves' fractal dimensions. Once these feature vectors have been captured, we can compare them to the training set by calculating the Euclidean distance between the different feature vectors. The smallest distance was considered the match. Fig. 2 illustrates the overall approach.

Section snippets

Dimensionality reduction based on central projection

Projection is a very basic and common operation that is used in pattern recognition and image processing. Projection refers to the mapping of a two-dimensional region of an image into a waveform whose values are the sums of the values of the image points. It is obtained by determining the number of black pixels that fall onto a projection axis. Projection profiles represent a global feature of a character (Jain et al., 1996). They play a very important role in character recognition.

Multiresolution analysis and wavelet decomposition

In the preceding section, we have shown how to effectively transform two-dimensional patterns into one-dimensional ones with central projection transformation. In the present section, we shall apply wavelet transform and multiresolution analysis to derived one-dimensional patterns to produce wavelet transform sub-patterns. The wavelet transform is also to aid in the boundary identification. Multiresolution analysis (MRA) was first published in 1989 by Mallat, and the advanced research and

Computing divider dimension of one-dimensional patterns

The fractal dimension is a useful method to quantify the complexity of feature details present in an image. In this section, we shall discuss the problem of computing the divider dimension of those curves, and thereafter, use the computed divider dimension to construct a feature vector for the original two-dimensional pattern in question for pattern recognition.

Until today, there is no common definition of what is a fractal. But it is clear that a fractal has many differences from Euclidean

Experiments

This section presents the procedure as well as the results of our experiments that aim at recognizing a set of two-dimensional patterns.

Conclusions

In this paper, the notion of extraction feature with wavelet and fractal theories is presented as a powerful technique in pattern recognition. We have investigated the utility of several emerging techniques to extract features. This novel method of feature extraction includes utilizing the CPT to describe the shape, the wavelet transformation to aid in the boundary identification, and the fractal features to enhance image discrimination. Its essential advantage is that it can be used to

Acknowledgements

This work was supported by research grants received from the Research Grant Council (RGC) of Hong Kong and a Faculty Research Grant (FRG) from Hong Kong Baptist University.

References (13)

  • C.K. Chui

    Wavelets – A Mathematical Tool for Signal Analysis

    (1997)
  • T. Crimmins

    A complete set of Fourier descriptors for two dimensional shapes

    IEEE Trans. Syst., Man Cybernet

    (1982)
  • I. Daubechies et al.

    Introduction to the special issue on wavelet transforms and multiresolution signal analysis

    IEEE Trans. Inform. Theory

    (1992)
  • G.A. Edgar

    Measure, Topology and Fractal Geometry

    (1990)
  • K.L. Falconer

    The Geometry of Fractal Sets

    (1985)
  • K. Falconer

    Fractal Geometry: Mathematical Foundation and Applications

    (1990)
There are more references available in the full text version of this article.

Cited by (55)

  • An intelligent approach for variable size segmentation of non-stationary signals

    2015, Journal of Advanced Research
    Citation Excerpt :

    DWT represents the signal variation in frequency with respect to time. After decomposing the signal, fractal dimension (FD) is employed as a relevant tool to detect the transients in a signal [13]. FD can be used as a feature for adaptive signal segmentation because FD can indicate changes not only in amplitude but also in frequency.

  • Dynamic unbalance detection of Cardan shaft in high-speed train applying double decomposition and double reconstruction method

    2015, Measurement: Journal of the International Measurement Confederation
    Citation Excerpt :

    While effective for small quantities, wavelet decomposition of large data volumes often leads to difficulty in judging the data processing result in practice. Therefore, post-processing methods of wavelet decomposition such as wavelet scale energy statistical analysis [6–8], wavelet fractal analysis [9–12], and wavelet singular value decomposition [13–16] have been fully developed. Because wavelet singular value decomposition has unique advantages in de-noising and eliminating correlations of signals, it has widely been used in fault diagnosis and image processing.

  • Spike detection approaches for noisy neuronal data: Assessment and comparison

    2014, Neurocomputing
    Citation Excerpt :

    At any point, the mean value of the two envelopes defined respectively by local maxima and local minima must be zero. FD is a measure of nonlinear dynamics of time series and has recently become popular in analysis of biomedical signals such as EEG [33,34]. If the amplitude and frequency of a signal change, the dimension of FD will change the same as Fig. 2.

  • A hybrid evolutionary approach to segmentation of non-stationary signals

    2013, Digital Signal Processing: A Review Journal
  • Fractal pursuit for compressive sensing signal recovery

    2013, Computers and Electrical Engineering
    Citation Excerpt :

    Fractal dimension has been used in many fields of study. Tang et al. [15,16] select fractal feature to recognize the printed Chinese characters, English letters and handwritten signatures; Lee and Hsieh [17] propose a more robust and efficient improvement of BCD and apply it to the classification of images; Ebrahimi and Vrscay [18] research the fractal method of projections onto convex sets; Lin [19] utilizes fractal to solve the inverse problem in the signal denoising. On the basis of the above mentioned fractal measure theory, we will use fractal recognition to extract the prior sparsity from signal sensing data, and use fractal minimization to realize the CS reconstruction.

View all citing articles on Scopus
View full text