Nischal verma future image frame generation
-
Upload
nakul-surana -
Category
Documents
-
view
126 -
download
3
Transcript of Nischal verma future image frame generation
Future Image Frame Generation Using Artificial Neural Network with Selected Features
Nishchal K. Verma
Department of Electrical Engineering Indian Institute of Technology
Kanpur, India 208016 [email protected]
Abstract— This paper presents a novel approach for the generation of Future image frames using Artificial Neural Network (ANN) on spatiotemporal framework. The input to this network are hyper-dimensional color and spatiotemporal features of every pixel of an image in an image sequence. Principal Component Analysis, Mutual Information, Interaction Information and Bhattacharyya Distance measure based feature selection techniques have been used to reduce the dimensionality of the feature set. The pixel values of an image frame are predicted using a simple ANN back propagation algorithm. The ANN network is trained for R, G and B values for each and every pixel in an image frame. The resulting model is successfully applied on an image sequence of a landing fighter plane. As Mentioned above four feature selection techniques are used to compare the performance of the proposed ANN model. The quality of the generated future image frames is assessed using, Canny edge detection based Image Comparison Metric(CIM) and Mean Structural Similarity Index Measure(MSSIM) image quality measures. The proposed approach is found to have generated six future image frames successfully with acceptable quality of images.
Keywords- Future Image Frame generation, Spatiotemporal framework, Feature selection, Artificial Neural Network.
I. INTRODUCTION Prediction of future data from its past samples has always
been an area of great interest. This paper proposes an efficient method to predict future image frames of an image sequence. Future image frames of the image sequence of an event using the past information of the image sequence. To generate an image frame, the R, G and B components of each and every pixel of it has to be generated. This has motivated to model an image sequence in such a manner that it can capture the color components associated with the space as well as the time component Many earlier attempts have successfully made use of the spatiotemporal representation for an image sequence [1],[2]. The challenges in the generation of future images are to utilize the spatiotemporal representation along with a suitable ANN modeling scheme with proper feature selection techniques. Some preliminary work has already been done in this direction[2],[3],[4] using additive GFM [5]. In an image sequence modeling, we often encounter with poor representation of objects causing many uncertainties and contains irrelevant data that unnecessarily increases the dimension of the input features. Research literatures suggest [2],[6] that ANN theory along with feature selection could be useful in dealing with computational complexities involved in
future frame generation. For future image sequence modeling initiatives, ANN modeling approach claim for their strong candidature. Utilizing these concepts and related formulations motivated to develop a useful ANN modeling scheme along with feature selection for generation of landing fighter plane images. In this framework, space and time are treated uniformly in the spatiotemporal domain called unified feature space[1]. The extracted space-time regions allow us for the generation of spatiotemporal events. In this paper, images of a fighter aircraft landing on an aircraft carrier are used for the implementation of the model. The ANN model can generate, with a good accuracy, future image frames of the fighter planes. Also to evaluate the quality of the generated future image frames, the structural similarity of corresponding actual image frame is computed using CIM and MSSIM indices. This work may also be helpful in early detection and target tracking and to take defense measures in advance. The rest of the paper is organized as follows: In Section-II, the feature extraction and spatiotemporal representation of image sequence is discussed. Feature selection techniques are discussed in Section-III. Section-IV explains the formulation of the ANN model for image frames generation. The step-wise algorithm is given in Section-V. In Section-VI, the assessment of the structural quality of generated future satellite image is discussed. Next, actual images of the fighter, are used for validation of the system in Section-VII. Finally, the conclusions are drawn in Section-VIII.
II. SPATIOTEMPORAL DOMAIN AND FEATURE EXTRACTION
For the generation of future image frames, both spatial and temporal information are needed to be considered in an integrated manner. An image sequence contains consistent visual contents that change over time. The image sequence is to be seen as a single entity, as opposed to a sequence of separate image frames. Space and time need to be treated uniformly. An image sequence has both spatial and temporal features. Feature selection reduces the amount of data to be processed and helps in getting better output as only relevant information to be used. Of the various schemes [7] available in the literatures, the spatiotemporal feature space, i.e., representation of an image sequence in spatiotemporal domain is found to be more suitable for the problem at hand. For RGB color scheme features like color components R, G and B of every pixel and their location coordinate in x and y along with the time of occurrence of the frame at tth instant, in the image sequence are extracted and a six dimensional unified feature space, (R, G, B,
x, y, t) is used in [1],[2],[3],[4],[5] for representation of spatiotemporal images and in this way all the pixels in each frame of an image sequence are uniquely defined.
III. FEATURE SELECTION Feature selection is normally known as variable selection,
feature reduction, attribute selection or variable subset selection, Feature selection is a process of selecting a subset of relevant features for building robust learning models. Large number of features increase computational complexities. By removing irrelevant and redundant features from the data,
feature selection process helps in improving the performance of models. Enhancing generalization capability, speeding up learning process and improving model ability to interpret. Most of the time it is not possible to identify usefulness of features by intuition. Hence, the need arises for an automatic feature selection process to select optimal feature set. While developing models for classification, the features form a feature space where they are also called as variables or dimensions of feature space. Optimal feature set contains all relevant features and ignores irrelevant and redundant features. However, no single algorithm guarantees this in general to generate an optimal set of features. Selection process is very much dependent to a specific problem in hand. There are many feature selection techniques available depending upon the feature space and size of the feature set. Pixels in an image are highly correlated to the pixels in their neighborhood. In this paper Principal Component Analysis(PCA)[8], the Bhattacharyya Distance measure [9],[12],[13],[14] (derived from Bhattacharyya coefficient [10]) and the Interaction Information[11],[15],[16] are used to extract relevant features.
IV. ANN MODEL FOR FUTURE IMAGE FRAME GENERATION
In this section, a simple separate ANN models for future image frame generation are derived using R,G and B color components of each pixel with the assumption of independence of features or attributes. A Simple architecture of ANN Model for R-component with n-inputs is shown in Figure 1, having one hidden layer with m neurons and one output layer neuron. The training of ANN is mainly undertaken using the Back-propagation learning algorithm [6], [17]. Back-propagation algorithm is one of the training algorithms for feed-forward
neural networks,
which uses two parameters in conjunction with the gradient descent learning algorithm. The first parameter is the learning rate η , which is essentially a parameter that determines how directly the gradient descent should be applied to the weight matrix and threshold values. The gradient is multiplied by the η and then added to the weight matrix or threshold value. This will slowly optimize the weights to values that will produce a lower error. The second parameter is called momentum, which specifies, to what degree, the weight changes from the previous iteration should be applied to the current iteration. This helps back propagation algorithm to come out of local minima.
From the fig 1. Rja can be computed:
( ), ,01
1n
R R R Rj j i i j
i
a w x w=
= +∑
and total output of the thj hidden layer neuron can be written as:
( ) ( )1 21
Rj
R Rj j a
Z ae λ
σ−
= =+
∑
∑
∑
( )1Raσ
( )2Raσ
( )Rmaσ
∑ ( )Rcγ
2,1
Rw,1
R
mw
,2
R
mw
1,
R
nw
,
R
m nw
2,
R
nw
2,2
Rw
1,1
Rw
0Rv
1,0Rw
2,0Rw
1Ra
2Ra
Rma
Rmz
1Rz
1Rv
2Rv
Rmv
Rc Output
Input Neuron LayerHidden Neuron Layer
,0Rmw
Output Layer Neuron
1R
x
2R
x
Rnx
nIn
puts
Fig. 1: ANN Architecture with one hidden layer for R-component
0Rz
0Rx
0Rx
0Rx
2RZ
RE R⎡ ⎤⎣ ⎦x
( ), ,01
3n
R R Rj i i j
iw x wσ
=
⎛ ⎞= +⎜ ⎟⎝ ⎠∑
output of the output neuron layer can be written as:
( )01
4m
R R R Rj j
jE R x v z vγ
=
⎛ ⎞⎡ ⎤ = +⎜ ⎟⎣ ⎦ ⎝ ⎠∑
or in a matrix form:
11,1 1,2 1, 1,0
22 ,1 2 ,2 2 , 2 ,0
1 2 0 0
,1 ,2 , ,00
(
RR R R R
RnR R R R
nR R R R R Rm
RnR R R R
m m m n m R
xw w w w xw w w w
E R v v v v zx
w w w wx
γ σ
⎡ ⎤⎧ ⎫⎛ ⎞⎡ ⎤⎢ ⎥⎪ ⎪⎜ ⎟⎡ ⎤ ⎢ ⎥⎢ ⎥⎪ ⎪⎜ ⎟⎢ ⎥ ⎢ ⎥⎪ ⎪⎢ ⎥⎜ ⎟⎢ ⎥ ⎢ ⎥⎡ ⎤ ⎡ ⎤= +⎨ ⎬⎣ ⎦⎢ ⎥⎣ ⎦ ⎜ ⎟⎢ ⎥ ⎢ ⎥⎪ ⎪⎢ ⎥⎜ ⎟⎢ ⎥ ⎢ ⎥⎪ ⎪⎢ ⎥⎢ ⎥⎜ ⎟⎢ ⎥⎣ ⎦⎜ ⎟⎪ ⎪⎢ ⎥⎢ ⎥⎣ ⎦⎝ ⎠⎩ ⎭⎣ ⎦
x 5)
V. ALGORITHM FOR FUTURE IMAGE FRAME GENERATION USING ANN
A. The Algorithm
Assuming that we have L no. of image frames in an image sequence. To derive the suitable ANN model, the image sequence is split into two sets: the set of first F frames be
called rT (training set) and the set of remaining L - F frames be sT (the testing set). Use set rT to develop the ANN model and for its training. Use this model to generate the future frames. Use the set sT to validate the generated frames. It is important to note that for real time application, test set may not be used and whole image sequence data may be used for incremental training of ANN. Step 1: Perform feature extraction on the image sequence and make a unified feature space representation in spatiotemporal domain, i.e. a six-dimensional feature space (R G B x y t) as discussed in section II. This means representing each pixel of
an image in the sequence with a six dimensional feature vector [R G B x y t]. Step 2: Arrange the four-dimensional feature vectors for all color space as: [R, x, y, t], [G, x, y, t] and [B, x, y, t] for each pixel from the six-dimensional feature space. These feature spaces along with neighboring and past pixel values will serve as the final
feature set for the corresponding ANN model. Step 3: Employ an appropriate feature selection technique, to reduce the dimensionality of the feature set, for each color component R, G and B. Step 4: Based on the space and time inputs estimate R, G and B values separately using a model suitable ANN models ANN-R, ANN-G and ANN-B as shown in Figure 1. Step 5: Aggregate R, G and B components of corresponding image pixels to construct the future image frames. Step 6: Apply Canny edge detection algorithm on the test and the generated future image frames for structural comparison based on CIM and MSSIM indices.
Basic Feature Extraction( Pixel
representation in 4-D Spatiotemporal domain
(R/G/B,x,y,t) in an unified feature space
4-Dimensional feature data i.e. R- Component, Spatial(x,y) and
temporal alongwith neighboring and past pixel data
4-Dimensional feature data i.e. G- Component, Spatial(x,y) and
temporal alongwith neighboring and past pixel data
4-Dimensional feature data i.e. B- Component, Spatial(x,y) and
temporal alongwith neighboring and past pixel data
Feature Selection using suitable
algorithm
Feature Selection using
suitable algorithm
Feature Selection using
suitable algorithm
ANN-Model for R-Component
ANN-Model for G-Component
ANN-Model for B-Component
Aggregation of R,G and B
components for every pixel
Future Generated Image Frames
Image
Sequence (A set of image frames)
Fig. 2: Generation of Future images using ANN model along with Feature selection components
B. Implementation
The image sequence(a video clip in general) for which image frames are to be generated is first converted into a set of its constituent image frames. This set is divided into a training and a testing set. The training set image frames are then used to learn to ANN color model Since separate ANN models have to be created for each color component. A four dimensional vectors [R, x, y, t], [G, x, y, t] and [B, x, y, t] from the six dimensional feature space is extracted along with neighboring and past pixel values to train ANN-R, ANN-G and ANN-B. For every pixel separate ANN model is developed for all three color components. For example, if the frames of the image sequence are of size 100 x 100 then the frame to be generated will be of the same size hence we will be having total 10000 x 3 ANN models, one for each pixel for each color component. The pixels in an image frame are strongly correlated to their neighboring pixels in terms of the variation of the color component. This correlation can be used in model as the network can be trained for this variation to increase its prediction capability. These neighboring pixels along with the required one act as feature set to the generation model. To select the best pixel values among all the neighboring ones feature selection is performed on this feature set. After selecting the best set of features, the network is developed depending upon the number of input features, the output and the parameters like number of hidden layer and hidden layer neurons. Each of these ANN networks is trained separately for the variation in the corresponding color for respective pixel over all training set frames. The color component R, G and B of a pixel of future image frame is then generated based on the variations of last two frames of the training set. These generated color components are then aggregated for each pixel to generate the final future image.
VI. IMAGE QUALITY ASSESSMENT Most of the image similarity assessment metrics like peak signal to noise ratio (PSNR), picture quality scale (PQS), noise quality measure (NQM), mean structural similarity (MSSIM), information fidelity criterion (IFC), Canny based image comparison metric (CIM) and visual information fidelity (VIF) perform well on gray scale images. Out of these seven indices, two indices, MSSIM [18] and CIM [19] have been chosen for the evaluation of structural similarities in color images in our image generation applications. The reason for choosing CIM and MSSIM based metric is that CIM index is invariant to rotation, scaling and translation within a reasonable tolerance and MSSIM has been stated to be a better structural descriptor[18] for images. The indices are computed for each color component, R-image, G-image and B-image and then averaged to get a composite index for color RGB images.
VII. SIMULATION AND RESULTS For experimental validation, image sequence of a fighter plane landing on an aircraft carrier is under taken where there is camera motion and zooming effect involved. This special case is considered for simulation as this involves some camera motion(being on a ship) and zooming effect(fighter
approaching the camera). The camera motion causes the background to change and zooming effect will cause the regular change in the object size. This means the network has to be tuned for both motion and size change of the object as well as the changing background. The training set consists of 120 images of size 153 x 116 pixels and the testing set has 6 images of similar size. Note that, the training images and the test images correspond to two non-intersecting time interval. In other words, the test images are the future images that is not seen by the model estimation process. The results constitutes the simulation of the algorithm for four different categories in which one is by using only ANN and the rest three involves ANN with three feature selection techniques Principal Component Analysis (ANN-PCA), Interaction Information(ANN-II) and Bhattacharya Distance(ANNBD). The inputs correspond to a window of two subsequent frames occurring at t - 2 and t -1 and are centered at pixel located at ( ), .x y Generation of images without feature selection trains the ANN using back-propagation with ten input variables:
( )C x, y -1, t - 2 , ( )C x -1, y, t - 2 , ( )C x, y, t - 2 ,
( )C x +1, y, t - 2 , ( )C x, y +1, t - 2 , ( )C x, y -1, t -1 ,
( )C x -1, y, t -1 , ( )C x, y, t -1 , ( )C x +1, y, t -1 , ( )C x, y +1, t -1
and one output variable i.e. ( )C x, y, t . The window covers the immediate vertical and horizontal neighborhood of the pixel at ( ), .x y In case of ANN with feature selection the input to the ANN are chosen from a set of 18 features
( )C x -1, y +1, t - 2 , ( )C x, y -1, t - 2 , ( )C x -1, y, t - 2 ,
( )C x +1, y -1, t - 2 , ( )C x, y, t - 2 , ( )C x -1, y +1, t - 2 ,
( )C x +1, y +1, t - 2 , ( )C x +1, y, t - 2 , ( )C x, y +1, t - 2 ,
( )C x, y -1, t -1 , ( )C x -1, y, t -1 , ( )C x -1, y -1, t - 2 ,
( )C x +1, y -1, t -1 , ( )C x, y, t -1 , ( )C x -1, y +1, t -1 ,
( )C x +1, y, t -1 , ( )C x, y +1, t -1 , ( )C x +1, y +1, t -1 and one
output variable ( )C x, y, t . This window covers the immediate vertical, horizontal and diagonal neighborhood of the pixel at ( ), .x y Out of these 18 features, we select 10 best features, ranked according to the maximum correlation with the output class, using feature selection techniques. ( )C x, y, t represents
the color values R, G and B of the pixel, located at ( ),x y in the frame occurring at time t, for ANN-R, ANN-G and ANN-B respectively. This is to be noted that color information of the previous time instants and spaces are included in the input vector to capture the dynamics as well as to enhance the performance of the system. The images are generated for 1, 3 and 10 hidden layer neurons for different η =0.01,0.05 and 0.09 in each case. The ANN network used has ten input neurons (in the input neuron layer) and one output neuron (in the output neuron layer) with one hidden layer. The activation function of hidden layers is unipolar sigmoid and output layer
is ‘pure-linear’. The momentum is kept Figure 3 shows the actual image and tgenerated image by using ANN-PCA witneurons. It was found from the simulatigenerative methods has best CIM and Mgenerative images for 3 hidden layer neurtables I, II, III, IV, V, VI, VII and VIII. Ittables that CIM and MSSIM indices are highwith η = 0.05.
(a) Test Image 121
(d) Generated Image 121
(a) Test Image 124
(d) Generated Image 124
Fig. 3: Test Images(actua
constant at 0.56. the corresponding th 3 hidden layer ons that all four
MSSIM indices of rons as shown in t is clear from the hest for ANN-PCA
Figure 4 shows the comparisoshows the comparison of MSSby all four generative methods=0.05 and α =0.56. It is clear fmethod is efficient compared tof quality of the generated futu
(b) Test Image 122
(e) Generated Image 122
(b) Test Image 125
(e) Generated Image 125
al images) and Generated Images using ANN-PCA with 3 hidden lay
on of CIM index and figure 5 SIM index of images generated s with 3 hidden layer neuron, η from the graphs that ANN-PCA to other three methods in terms
ure image frames.
yer neurons
(c) Test Image 123
(f) Generated Image 123
(c) Test Image
(f) Generated Image 126
Fig. 4: CIM Inde
Fig. 5: MSSIM Index for 3 hidden layer neurons α = 0.5
ex for 3 hidden layer neurons α=0.56 and η=0.05
56 and η = 0.05
TABLE I: CIM Indices for ANN with 3 hidden layer neuron
Image no.
Learning Rates
0.01 0.05 0.09
121 0.9868 0.9881 0.9829
122 0.9779 0.9878 0.9793
123 0.9772 0.9786 0.9766
124 0.9764 0.9772 0.9743
125 0.9734 0.9745 0.9743
126 0.9677 0.9696 0.9675
TABLE II: CIM Indices for ANN-PCA with 3 hidden layer neurons
Image no.
Learning Rates
0.01 0.05 0.09
121 0.9883 0.9913 0.9859
122 0.9793 0.9888 0.9799
123 0.9776 0.9856 0.9769
124 0.9767 0.9786 0.9763
125 0.9743 0.9751 0.9747
126 0.9678 0.9716 0.9678
TABLE IV: CIM Indices for ANN-BD with 3 hidden layer neurons
Image no.
Learning Rates
0.01 0.05 0.09
121 0.9819 0.9889 0.9817
122 0.9817 0.9850 0.9810
123 0.9779 0.9845 0.9773
124 0.9734 0.9769 0.9704
125 0.9718 0.9742 0.9685
126 0.9582 0.9709 0.9592
TABLE VI: MSSIM Indices for ANN-PCA with 3 hidden layer neurons
Image no.
Learning Rates
0.01 0.05 0.09
121 0.9871 0.9892 0.9810
122 0.9647 0.9663 0.9627
123 0.9342 0.9430 0.9277
124 0.9230 0.9244 0.9149
125 0.8797 0.8869 0.8751
126 0.8437 0.8495 0.8398
TABLE V: MSSIM Indices for ANN with 3 hidden layer neurons
Image no.
Learning Rates
0.01 0.05 0.09
121 0.9853 0.9854 0.9825
122 0.9640 0.9642 0.9604
123 0.9334 0.9355 0.9316
124 0.9224 0.9247 0.9204
125 0.8745 0.8821 0.8799
126 0.8432 0.8457 0.8368
TABLE III: CIM Indices for ANN-II with 3 hidden layer neurons
Image no.
Learning Rates
0.01 0.05 0.09
121 0.9843 0.9856 0.9825
122 0.9714 0.9786 0.9744
123 0.9673 0.9678 0.9635
124 0.9570 0.9632 0.9597
125 0.9563 0.9602 0.9528
126 0.9551 0.9572 0.9510
TABLE VII: MSSIM Indices for ANN-II with 3 hidden layer neurons
Image no.
Learning Rates
0.01 0.05 0.09
121 0.9824 0.9830 0.9817
122 0.9600 0.9613 0.9593
123 0.9289 0.9312 0.9263
124 0.9165 0.9181 0.9145
125 0.8713 0.8718 0.8707
126 0.8370 0.8408 0.8364
TABLE VIII: MSSIM Indices for ANN-BD with 3 hidden layer neurons
Image no.
Learning Rates
0.01 0.05 0.09
121 0.9827 0.9832 0.9821
122 0.9640 0.9643 0.9613
123 0.9323 0.9330 0.9234
124 0.9212 0.9214 0.9124
125 0.8741 0.8769 0.8405
126 0.8430 0.8450 0.8298
VIII. CONCLUSION The performance of the proposed algorithm for an image set of a fighter plane landing on a ship has been studies. Six future images of the image sequence has successfully been generated capturing the motion of the plane successfully. Four variations of the image generation algorithm has been simulated and are analyzed on the basis of quality (CIM and MSSIM indices) of generated image frames. Feature selection methods has been employed to select the optimum features for future image frame generation. It has been deduced that feature selection helps in improving the efficiency of future
image frame generation model. Also from the simulation it is clear that ANN-PCA model with 3 hidden layer neuron for η= 0.05 and α = 0.56 is the optimum future image generation scheme.
ACKNOWLEDGEMENT Author sincerely thanks Department of Science and Technology (DST), New Delhi, India for providing financial support under project no DST/EE/20100272 to carry out this research work.
REFERENCES [1] H. Greenspan, J. Goldberger and A. Mayer, “Probabilistic space-time
video modeling via piecewise GMM,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 26, pp. 384 - 396, 2004.
[2] Nishchal K. Verma, Prateek Tamrakar and Saurabh Agrawal, “On Generating future satellite image frame using artificial neural network”, International Conference on Image and Video Processing and Computer vision(IVPCV-2010), pp.158-164, July 12-14, 2010, Orlando, Florida, USA.
[3] Nishchal K. Verma and N. R. Pal, “Prediction of Satellite Images using Fuzzy Rule based Gaussian Regression,” IEEE 39th Applied Imagery Pattern Recognition workshop, AIPR-2010, pp. 1-8, Washington DC, USA.
[4] Nishchal K. Verma and M. Hanmandlu, Interactive Fuzzy System Using CWM, in Proceedings of IEEE Indicon 2005 Conference, pp.375-378, Chennai, India, 11-13 Dec.2005
[5] Nishchal K. Verma and H. Madasu, “From a Gaussian Mixture Model to Non-additive Fuzzy Systems,” IEEE Transactions on Fuzzy Systems, vol. 15, pp. 809-827, 2007.
[6] R. Parisi, E. D. Di Claudio, G. Orlandi, and B. D. Rao, “A generalized learning paradigm exploiting the structure of feed-forward neural networks,” IEEE Transactions on Neural Networks, vol. 7, pp. 1450-1459, 1996.
[7] Dongfeng Ren and Yuan Ma, “Research on feature extraction from re-mote sensing image,” International Conference on Computer Application and System Modeling (ICCASM), Taiyuan-China, 2010, vol. 1, pp. V1-44
[8] F. Song, Z. Guo and D. Mei, “Feature Selection Using Principal Component Analysis,” International conference on System Science , Engineering Design and Manufacturing Informatization (ICSEM), vol 1, pp. 27- 30, Changchun, China, 2010
[9] A. Bhattacharyya, “A measure of divergence between two multinomial populations,” Sankhy\\bara, The Indian Journal Of Statistics, vol.7, no.4,
pp.401-406, Calcutta,1946. [10] Aherne F., Thacler N. and Rockett P., “The Bhattacharaya metric as an
absloute similarity measure for frequency coded data,” cybernetika, vol. 32(4), pp. 1-7, 1997
[11] Meyer P.E., Schretter C. and Bontempi G. on “Information-Theoretic feature selection in Microarray data using variable complementarity,” IEEE Journal of Selected Topics in Signal Processing, vol 2, Issue 3, pp. 261-274, June 2008.
[12] Bhattacharyya, A., “A measure of divergence between two statistical populations defined by their probability distribution,” Bulletin of the Calcutta Mathematical Society, vol. 35, pp. 99-110, 1943
[13] Euisun Choi and Chulhee Lee on, “Feature extraction based on Bhattacharyya distance for multimodal data,” IEEE 2001 International Geoscience and Remote Sensing Symposium, IGARSS’01, pp. 524-526.
[14] Xuan G., Chai P., Minhui Wu on, “Bhattacharyya distance based feature selection,” proceedings of the 13th International Conference on Pattern Recognition, pp.195 - 199, vol.2, Viena, Austria, 1996.
[15] McGill W J., “Multivariate information transmission,” Psychometrika vol. 19, pp. 97-116, 1954
[16] R. Battiti, “Using Mutual Information for selecting features in supervised neural net learning,” IEEE Transactions on Neural Networks, vol.5, no. 4, pp. 537-550, 1994.
[17] Xinghuo Yu, M. Onder Efe and Okyay Kaynak, “A General Back propagation Algorithm for Feed forward Neural Networks Learning,” IEEE Transactions on Neural Networks, vol. 13 (1), pp. 251-254, 2002.
[18] W. Zhou, A. C. Bovik, H. R. Sheikh and E. P. Simoncelli, “Image quality assessment: from error visibility to structural similarity,” IEEE Transactions on Image Processing, vol. 13, pp. 600-612, 2004.
[19] W. C. Chan, M. V. Le, and P. D. Le, “A wavelet and Canny based image comparison,” IEEE Conference on Cybernetics and Intelligent Systems, vol. 1, pp. 329-333, 2004.
[20] J. Canny, “A computational approach to edge detection,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 8, no. 6, pp. 679 - 698, 1986.
[21] Guyon, “An introduction to variable and feature selection,” Journal of Machine Learning Research, vol. 3, pp. 1157-1182, 2003 .