Wavelet Transform Based Texture Features For Content Based Image Retrieval
Wavelet Transform Based Texture Features For Content Based Image Retrieval
Retrieval
Manesh Kokare, B.N. Chatterji and P.K. Biswas
Electronics and Electrical Communication Engineering Department,
Indian Institute of Technology,
Kharagpur PIN 721 302, India
{mbk, bnc, pkb}@ece.iitkgp.ernet.in
ABSTRACT
The rapid expansion of the Internet and the
wide use of digital data have increased the need
for both efficient image database creation and
retrieval procedure. The challenge in image
retrieval is to develop methods that can capture
the important characteristics of an image,
which makes it unique, and allow its accurate
identification. The focus of this paper is on the
image processing aspects that in particular
using texture information for retrieval. We
present a unique wavelet transform based
texture features for content-based image
retrieval, which is comparable with standard
existing methods. We propose the use of
pyramidal and tree structured wavelet features
using 8-tap Daubechies coefficients for texture
analysis and provide extensive experimental
evaluation. Comparison with various features
using Broadtz texture database indicates that
the combination of energy and standard
deviation of wavelet features provide good
pattern retrieval accuracy for tree structured
wavelet
decomposition
while
standard
deviation alone gives better result in pyramidal
wavelet decomposition. In most of the existing
retrieval methods the most commonly used
Euclidean distance function or measure of
dissimilarity between feature vectors is used,
but we observed that, it is not always the best
metric. We have done the comparison of results
using Euclidean distance and Manhattan
distance for both the tree structured and
pyramidal wavelet decomposition methods and
found that Manhattan distance gives better
result than Euclidean distance metric.
Keywords: Content-based image retrieval
(CBIR), image database, feature database,
similarity, query image, pyramidal wavelet
transform, tree structured wavelet transform.
1. INTRODUCTION
Worldwide networking allows us to
communicate, share, and learn information in
2. WAVELET TRANSFORM
In this section we will discuss pyramidal and
tree structured wavelet transform in brief.
2.1 Pyramidal wavelet transform
The images in a database are likely to be stored
in a compressed form. Superior indexing
performance can therefore be obtained if the
properties of the coding can be exploited in the
indexing technique. Recently, discrete wavelet
transform has become popular in image coding
applications [3, 4]. Wavelets provide
multiresolution capability, good energy
compaction and adaptability to human visual
characteristics.
Wavelet transform represents a
function as a superposition of a family of basis
functions called wavelets. Translating and
dilating the mother wavelet corresponding to a
particular basis can generate a set of basis
functions. The signal is passed through a low
pass and high pass filter, and the filters output
is decimated by two. Thus, wavelet transforms
extract information from signal at different
scales. For reconstruction, the coefficients are
up sampled and passed through another set of
low pass and high pass filters.
The 2-D DWT is generally calculated
using a separable approach. Fig.1 shows a three
level pyramidal wavelet decomposition of an
image S1 of size a b pixels. In the first level of
decomposition, one low pass subimage (S 2 )
and three orientation selective high pass
subimages (W2H ,W2V ,W2D ) are created. In
second level of decomposition, the low pass
subimage is further decomposed in to one low
pass and three high pass subimages
W4H , W4V ,W4D . The process is repeated on the
low pass subimage to form higher level of
wavelet decomposition. In other words, DWT
decomposes an image in to a pyramid structure
of the subimages with various resolutions
corresponding to the different scales. We note
that a three-stage decomposition will create
three low pass subimages and nine (three each
in horizontal, vertical, and diagonal direction)
high pass directional subimages. The low pass
subimages are low-resolution versions of the
original image at different scales. The
horizontal, vertical and diagonal subimages
provide the information about the changes in
the corresponding directions respectively.
b
8
b
a
8
b
4
S8 W H
8
V
8
D
8
W W
Original
Image
a
4
a
2
V
4
b
2
W4H
W2H
D
4
W2V
W2D
3.1
The
texture
database
used in the
experimentation consists of 116 different
texture classes. We have used 108 textures
from Bordatz album [7], seven textures from
USC Database and one artificial texture. Size
of each texture is 512 512 . Each of 512 512
images is divided into sixteen 128 128
nonoverlapping subimages, thus creating a
database of 1,856 patterns in the database.
f qj - f ij
(1)
- f ij )
(2)
j =1
D qiE =
(f
j =1
qj