We Are Intechopen, The World'S Leading Publisher of Open Access Books Built by Scientists, For Scientists
We Are Intechopen, The World'S Leading Publisher of Open Access Books Built by Scientists, For Scientists
6,200
Open access books available
169,000
International authors and editors
185M
Downloads
154
Countries delivered to
TOP 1%
most cited scientists
12.2%
Contributors from top 500 universities
1. Introduction
Image segmentation consists in grouping pixels sharing some common characteristics. In
vision systems, the segmentation layer typically precedes the semantic analysis of an image.
Thus, to be useful for higher-level tasks, segmentation must be adapted to the goal, i.e. able
to effectively segment objects of interest. Our objective is to propose a cognitive vision
approach to the image and video segmentation problem. More precisely, we aim at
introducing learning and adaptability capacities into the segmentation task. Traditionally,
explicit knowledge is used to set up this task in vision systems. This knowledge is mainly
composed of image processing programs (e.g., specialized segmentation algorithms and
post-processing’s) and of program usage knowledge to control segmentation (e.g., algorithm
selection and algorithm parameter settings).
In real world applications, when the context changes, so does the appearance of the images.
It can be due to local changes (e.g., shadows, reflections) and/or global illumination changes
(e.g., due to meteorological conditions). The consequences on segmentation results can be
dramatic. This context adaptation issue emphasizes the need of automatic adaptation
capabilities. Our first objective is to learn the contextual variations of images in order to
discriminate between different segmentation actions. The identification of the contexts will
lead to different segmentation actions as algorithm selection.
When designing a segmentation algorithm, internal parameters (e.g., thresholds or minimal
sizes of regions) are set with default values by the algorithm authors. In practice, it is often
up to an image processing expert to supervise the tuning of these free parameters to get
meaningful results. As seen in Figure 1, it is not clear how to choose the best parameter set
regarding the segmented images: the first one is quite good but several parts of the insect
are missing; the second one is also good, since the insect is well outlined, but too many
meaningless regions are also present. However, complex interactions between free
parameters make the behaviour of the algorithm fairly impossible to predict. Moreover, this
awkward task is tedious and time-consuming. Thus, the algorithm parameter tuning is a
real challenge. To solve this issue, our objective is threefold: first, we want to automate this
task in order to alleviate users’ effort and prevent subjective results. Second, the fitness
function used to assess segmentation quality should be generic (i.e. not application
dependent). Third, no a priori knowledge of segmentation algorithm behaviours is required,
only ground truth data should be provided by users.
www.intechopen.com
266 Tools in Artificial Intelligence
Fig. 2. An example of the segmentation of an image with two different algorithms. The first
algorithm forms regions according to a multi-scale colour criterion while the second uses a
local colour homogeneity criterion.
Once all the algorithms have been optimized, a third issue is to select the best one. However,
when images of the application domain are highly variable, it remains quite impossible to
achieve a good segmentation with only one tuned algorithm. Our objective is to make use of
the extracted knowledge of context variations and parameter tuning to associate a
segmentation action to each identified context.
Finally, in many computer vision systems at the detection layer, the goal is to separate the
object(s) of interest from the image background. When objects of interest and/or image
background are complex (e.g. composed of several subparts), a low-level algorithm cannot
achieve a semantic segmentation, even if optimized. For this reason, a fourth issue is to
refine the image segmentation to provide a semantically meaningful segmentation to higher
vision modules.
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 267
Our final objective in this chapter is to show the potential of our approach through a
segmentation task in a real-world application. The segmentation task we focus on is image
segmentation in a biological application related to early pest detection and counting. This
implies to robustly segment the objects of interest (mature white flies) from the complex
background (rose leaves). Our goal is to demonstrate that the cognitive vision system
coupled with our adaptive segmentation approach achieves a better detection rate of white
flies than tuned with an ad hoc segmentation.
This chapter is structured as follows. Section 2 introduces the reader to image segmentation
in the context of computer vision systems. We propose an overview on topics closely related
to our problem. Section 3 details each step of our learning approach. Section 4 shows how
the learnt segmentation knowledge is used to perform adaptive image segmentation. The
next section is dedicated to the validation of the approach for a real world application: the
segmentation step of a cognitive vision system dedicated to the recognition of biological
organisms in static images. Concluding remarks and suggestions for future work are
discussed in section 6.
2. Related work
In this section, we present some previous work related to image segmentation, segmentation
performance evaluation, algorithm parameter optimization and algorithm selection.
www.intechopen.com
268 Tools in Artificial Intelligence
Segmentation
Assessment Value
A priori Knowledge
Synthetic
Image
Empirical Supervised
OR
Methods Methods
Manual
Segmentation
OR OR
Region-based Edge-based
Segmentation Segmentation
OR
Segmentation
Algorithm
Input
Image
Fig. 3. Segmentation evaluation diagram starting from an input image and returning a
segmentation assessment value
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 269
www.intechopen.com
270 Tools in Artificial Intelligence
application using implicit domain knowledge. Thereby, it has not been proved how the
fitness function can affect the performance of the optimization. Moreover, if authors have
often assessed their optimization methods against default segmentations, they did not make
any quantitative evaluation regarding to other optimization techniques. A comparative
study of optimization algorithms has to be done.
Ground Truth*
Segmentation yes
Image Segmentation Stop?
Evaluation
no
Global
Optimization
Updated Segmentation Algorithm
Parameters
Algorithm Parameter
* in supervised evaluation Space
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 271
expert systems limits their applicability. We can add that the model representation approach
appears to be more realistic in a computing point of view as compared to expert systems.
2.5 Conclusion
We have reviewed the segmentation task in the field of computer vision systems. If
researchers agree that segmentation is one of the fundamental problems in computer vision,
the efforts devoted to cope with this issue since the last four decades have still not led to a
unified solution. Most of the vision systems are application dependent and their
segmentation step is based on heuristic rules for, as example, the tuning of algorithm
parameters. It is, however, well-established that such a priori knowledge is determined by
domain experts from the context in which the segmentation takes place. Hence, the
generalization to other domain of application is strongly limited. Nonetheless, it appears
that the recent cognitive vision approach (ECVISION, 2005) has identified some avenues of
researches to cope with these limitations, as integration of machine learning techniques into
the knowledge acquisition task.
3.1 Overview
Our framework consists of two stages: a learning stage and an adaptive segmentation stage.
The framework relies on training data composed of manual segmentations of the training
images with semantic region annotations. The learning stage extracts the segmentation
knowledge from the training data by means of:
• a data mining module to extract and learn contextual variations,
• an optimization procedure for automatic segmentation parameter tuning,
• a learning module for context adaptation (i.e. to associate a segmentation action to each
identified context),
• a learning module for semantic segmentation; the goal is to train region classifiers with
respect to the annotated manual segmentations of the training images.
www.intechopen.com
272 Tools in Artificial Intelligence
The learning stage is sketched in Figure 5. The module for adaptive image segmentation
relies on the learnt segmentation knowledge. It will be described in section 4. The following
sections details each step of the learning stage.
Segmentation
Knowledge Base
Segmentation
Learning Learnt Algo
Algorithms Module Parameters
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 273
(( ) )
E IA = ρ A I , p A , G I (1)
The value E IA is an assessment value of the matching between the segmentation and the
ground truth. This can be goodness or a discrepancy measure.
The purpose of our optimization procedure is to determine a set of parameter values p̂ IA
which minimizes/maximizes:
(( ) )
pˆ IA = arg min/ Amax ρ A I , p A , G I
p
(2)
The final assessment value Ê IA and the optimal parameter set p̂ IA make a pair sample noted
( )
pˆ IA , Eˆ IA . This pair forms the segmentation knowledge for the image I and the algorithm A.
The set of all collected pairs constitutes the segmentation knowledge S set such that:
(
S = ∪ pˆ IA , Eˆ IA ) (3)
I ∈ℑ
One key-point of this optimization procedure is the definition of the assessment function ρ.
The quality of the final result varies according to this fitness function. The choice of a
segmentation performance evaluation metric is hence fundamental. It is discussed in the
next section.
www.intechopen.com
274 Tools in Artificial Intelligence
false boundary rate e Bf . The former, e mB , specifies the percentage of the points on B IG that are
mistakenly classified as non-boundary points; while the latter, e Bf , indicates the percentage
of the points in B IA that are actually false alarms. Therefore,
T1 and B T2 (4)
e mB = ef = A
B IG BI
where
{ ( ) (
T1 = x | x ∈ B IG ∧ x ∉ B IA )}
{ ( ) (
And T2 = x | x ∈ B IA ∧ x ∉ B IG )} (5)
( )
E IA = ρ B IA , B IG =
1 B
2
(
e m + e Bf ) (6)
result and the ground truth when using algorithm A. The value EIA = 1 indicates that all
pixels are misclassified. However, it is easy to show that this metric comes up against
unsuited response to under-segmented results, as illustrated in Figure 6. Segmentation in
panel (a) shows two regions with a quite good ground truth overlap, only three pixels are
misclassified. In the panel (b), the segmentation shows only one region and
the quality score is logically less than in (a). In the last panel (c), two regions are present but
the centre region badly overlaps the corresponding ground truth centre region. In
opposition with visual assessment, the segmentation quality is worst than in Figure 6(c).
Fig. 6. Limitation of the segmentation evaluation metric when weighting terms ( wmB and w Bf )
are not used.
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 275
The metric is improved by introducing two weighting terms wmB and w Bf which quantify the
average distance between misclassified points to the ground truth boundary such that:
∑ dist ( x, xˆ )
1
wmB = A
I
(7)
T1 x∈T1
∑ dist ( x, xˆ )
1
wBf = G
I
(8)
T2 x∈T2
with x̂ IG the closest pixel to x belonging to B IG ; dist (x1 , x 2 ) is the Euclidean distance
between two pixels x1(u,v) and x2(u,v) in a 4-neighbourhood such that:
Since wmB and w Bf have no fixed upper bounds, the normalization factor is useless and the
segmentation quality measure becomes:
The search of x̂IA (resp. x̂ IG ) is made easier by the use of a distance map (Maurer &
Raghavan, 2003) computed from B IA (resp. B IG ). This operation is exemplified in Figure 7.
By taking back the example in Figure 6 with the new definition of the evaluation metric, the
values of E IA for the cases (a), (b), and (c) are respectively 0.168, 0.75, and 0.679, yielding a
good correlation with a visual assessment.
www.intechopen.com
276 Tools in Artificial Intelligence
Once our performance evaluation metric is defined, the goal is now to minimize the
segmentation error E IA in order to learn optimal segmentation parameters. This is the role of
our closed-loop global optimization procedure.
f : Rd → S (11)
v( I ) ( A, pˆ A )
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 277
1
pA =
ℑA
∑ pˆ
I∈ℑA
A
I
(12)
where ℑA is the subset of training images for which the algorithm A has obtained the best
evaluation results among the other algorithms. Finally, for each training image of the cluster
and each algorithm A tuned with p A , the segmentation quality is computed again. The
algorithm having the best average performance on the training image set is finally selected.
We obtain a discrete function F taking a context identifier θ as input and returning an
algorithm A with a mean parameter setting p A such as:
F :Θ→ S
(13)
θ ( A, p A )
This selection strategy comes to select the robustest algorithm based on objective
comparisons, i.e. the algorithm which can deliver the best results for the cluster with a
globally relevant parameter set. However, this straightforward ranking approach has two
major drawbacks. First, by selecting only one algorithm and averaging its parameters, it
reduces the previously extracted segmentation knowledge amount to one mean case.
Second, even if the selected algorithm over performs the others in most of the cases, the
parameter averaging can have disastrous effects on the algorithm performance.
The principal purpose of this strategy is to overcome the drawbacks of a pure global ranking
strategy by dividing the solution space and by restricting the ranking process onto each
subspace. The main advantage on ranking algorithms inside a subspace is that evaluation
profiles are likely more correlated.
In this section, we have shown that the algorithm selection problem cannot be separated
from the parameter tuning problem. This statement means that a solution to the algorithm
selection issue is composed of both an algorithm and a parameter setting. We have
described our twofold strategy for learning the algorithm selection based on image-content
analysis and algorithm ranking. Starting from a training image set and segmentation
algorithms, our approach first identifies different situations based on image-content
analysis, then select the best algorithm with a mean parameter set for each identified context
based on optimized parameter values. At the end of the learning process, contexts are learnt
( )
with their associated pairs A, p A .
www.intechopen.com
278 Tools in Artificial Intelligence
segmentation algorithms cannot reach a semantic partitioning of the image. Thus, compared
to the ground truth, some regions remain over-segmented, as illustrated in Figure 8. If we
can assign the right label to each region, neighbouring regions with similar labels are
merged and, as a consequence, the residual over-segmentation becomes invisible. This
means to be able to map region features onto a symbolic concept, i.e. a class label. We use
the example-based modelling approach as an implicit representation of the low-level
knowledge. This approach has been applied successfully in many applications such as
detection and segmentation of objects from specific classes e.g., (Schnitman et al., 2006;
Borenstein & Malik, 2006). Starting from representative patch-based samples of objects (e.g.,
fragments), modelling techniques (e.g., mixture of Gaussian, neural networks, naive Bayes
classifiers) are implemented to obtain codebooks or class-specific detectors for the
segmentation of images. Our strategy follows this implicit knowledge representation and
associates it with machine learning techniques to train region classifiers. The following sub-
sections describe this stage in details.
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 279
⎧ max h(r )
⎪ yi | i = arg max H r , if >T
y (r ) = ⎨ r (14)
⎪
⎩ y0 , else
with |r| the number of pixels of the region r, T a threshold, and
H (r ) = {h1 (r ),… , hi (r ),… , hk (r )} the label histogram of the region r such that for a pixel u and
a label yi, hi (r ) = card {u ∈ r | y (u ) = y i }, i ∈ 1, … , k .
If the ratio of the most represented class in the region does not reach the threshold T (here
fixed at 0.8), the region label is set to y 0 ∉ Y . This prevents from labelling badly segmented
region as sketched in Figure 10.
Labelled Ground Truth Segmentation Result
(2 regions) (5 regions)
Fig. 10. Example of the mapping between a labelled ground truth regions and segmented
regions.
www.intechopen.com
280 Tools in Artificial Intelligence
Rℑ = ∪ ∪ {r | y(r ) ≠ y0 } (16)
I∈ℑ r∈RIA
for each region, a feature vector x(r) is extracted and makes with the label a pair sample
noted ( x( r ), y ( r ) ) . The set of all collected pair samples from ℑ constitute the training data set
such as:
Tℑ represents the knowledge of the semantic segmentation task and is composed, at this
time, of raw information. In the following section, we address the problem of knowledge
modelling by statistical analysis.
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 281
equivalent to the Gabor features but are computationally simpler. The basic idea is to
compute the energy of a region as a steerable function. This energy is computed for different
power channel, which are the result of convolving the region pixels with OGD filters of a
specific order. As colour histograms, texture feature vectors depend on the parameter q.
The final feature vector representing a region is a concatenation of the feature vectors
extracted from each cue. The feature extraction process is applied on each region of the
annotated regions set Rℑ so as to build the training data set Tℑ.
Following our cognitive approach of the segmentation problem, we need to avoid manually
selected and tuned algorithms. At the feature selection level, this means to be able to
automatically select and tune the feature extraction algorithm.
Feature Selection
The feature selection is used to reduce the number of features, remove irrelevant,
redundant, or noisy data, and it brings the immediate effects of speeding up and improving
the prediction performance of learning models. Since feature selection is a fertile field of
research, we refer the reader to surveys (Guyon & Elisseeff, 2003; Kohavi & John, 1997;
Blum & Langley, 1997) as good starting literatures. The optimality of a feature subset is
measured by an evaluation criterion. Feature selection algorithms designed with different
evaluation criteria broadly fall into two categories: the filters and the wrappers. Filters select
subsets of features as a pre-processing step, independently of the chosen predictor. Well-
known methods dedicated to this purpose are basic linear transforms of the input features
like Principal Component Analysis (PCA) and Fisher Linear Discriminant Analysis (LDA).
Techniques based on iterative search are also widespread as sequential forward/backward
algorithms (e.g. SFFS, SBS, ReliefF). Wrappers utilize the learning machine of interest (e.g.,
SVM, neural networks) as a black box to score subsets of features according to their
predictive power. Consequently, wrappers are remarkably universal and simple. An
interesting comparative study of such feature selection algorithms can be found in (Molina
et al., 2002).
The feature selection approach we propose is derived from wrappers. Our goal is to find the
best feature extractor configuration which minimizes the joint classification errors of the
class predictors applied on the training data set Tℑ. Unlike classical approaches, we act on
the feature extractor parameters to generate different feature vectors, instead of reducing the
feature vector itself. This approach is sketched in Figure 11. The two free parameters of our
selected feature extractors are the colour space encoder for colour feature extractor, and the
quantization level for both colour and texture feature extractors. The goal is to find the best
combination able to induce the minimum region classification errors. The quality estimation
is conducted via a cross-validation procedure which gives, for each region classifier, the
classification Mean Square Error (MSE). A global MSE is then computed by averaging the
MSE of each region classifiers.
Fig. 11. Feature selection schema based on tuning of the feature extractor parameters.
www.intechopen.com
282 Tools in Artificial Intelligence
We use an iterative search strategy to cover the value spaces of the two parameters q and cs.
This technique guarantees to find a global optimal solution but is computationally
expensive: first, it requires to run M x N x O region classifier training procedures, with M
the number of quantization levels (typically equals to 256), N the number of color spaces,
and O the number of classifiers to train; second, when the value of q increases, so does the
size of the feature vector. So, to avoid an unreasonable computational time, the choice of the
training algorithm must take into account this computational constraint.
Training Algorithm for Class Modelling
After extracting a feature vector for each region of the training data set, the next step is to
model the knowledge in order to produce region classifiers (one classifier per class). For a
feature vector x(r) and a class yi,
with ci (r ) ∈ [0,1] is the estimated probability associated with the hypothesis: ‘’feature vector
x(r) extracted from region r is a representative sample of the class yi’’. The set of these
trained region classifiers is noted C = {c1 , … , c k }.
A variety of techniques have been successfully employed to tackle the problem of
knowledge modelling such as naives Bayes networks, decision trees or support vector
machine (SVM). We propose to use SVM (Burges, 1998) as a template-based approach. SVM
are known to be efficient discriminative strategies for large scale classification problems
such as in image categorization (Chen & Wang, 2004) or object categorization (Huang &
LeCun, 2006). SVM yields also state-of-the-art performance at very low computational cost.
SVM training consists of finding an hyper-surface in the space of possible inputs (i.e. feature
vectors labelled by +1 or -1). This hyper-surface will attempt to split the positive samples
from the negative samples. This split will be chosen to have the largest distance from the
hyper-surface to the nearest of the positive and negative samples.
We adopt a one-vs-rest multi-class scheme with probability information (Wu et al., 2004) to
train one region evaluator per class. We use SVM with radial basis function as region
classifiers. There are two parameters while using RBF kernels: C (penalty parameter of the
error term) and γ (kernel parameter). It is not known beforehand which C and are the best
for one problem; consequently some kind of model selection (parameter search) must be
done. To fit the C and γ parameters, we adopt a grid-search method using 5-fold cross-
validation on training data. Basically, pairs of (C, γ) are tried and the one with the best cross-
validation accuracy is picked. This straightforward model selection efficiently prevents
over-fitting problems. The model selection is wrapped in the feature selection schema with
which it shares the cross-validation step. The training stage ends up when all combinations
of ((q,cs),(C, γ)) have been tested. The one giving the lowest global classification error is
picked and the region classifiers are trained a last time with this configuration.
3.6 Conclusion
In this section, we have presented our learning approach for adaptive image segmentation.
We have detailed each step of the learning module for context adaptation, algorithm
parameter tuning, and semantic image segmentation. The algorithm parameterisation issue
is tackled with a generic optimization procedure based on three independent components.
We have designed our performance evaluation metric to be broadly applicable and with a
low computational cost. It allows assessing a large variety of segmentation algorithms and
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 283
Algorithm Selection
Segmented Image
Fig. 12. Adaptive segmentation of an input image based on algorithm selection, parameter
tuning, and region labelling.
www.intechopen.com
284 Tools in Artificial Intelligence
For a new incoming image I not belonging to the training set, a feature vector is first
extracted then classified into a cluster. The classification is based on the minimization of the
distance between the feature vector and the cluster set {κ i } as follows:
( )
The pair A, p A associated with the detected context θ i is returned.
Once the algorithm is selected and tuned, the image is segmented. For each region, a feature
vector is extracted using the optimized (q, ˆ cs) parameter set and given as input to each
trained region classifiers c i . Classes are scored according to the classifier responses {c i (r )}
and finally, the returned label y(r ) is such as:
When all regions are labelled, neighbouring regions with the same label are merged to form
a semantic partitioning of the image. This final segmentation is returned to the user.
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 285
performance levels are revealed where good performances are achieved for a large range of
the parameter values. However, the global optimum is more difficult to see since the
difference between the good performance level (in blue) and its level is very thin. From
these observations, we can conclude that the evaluation profiles are not always convex hulls
and their granularity can depend on the image. Since the Simplex algorithm does not
guarantee to obtain a global optimum, we divide each parameter space into three sub-spaces
and run an optimization on each sub-space. This means that 3N optimization loops are run
for a segmentation algorithm with N free-parameters. Table 2 present the optimization
results of the five segmentation algorithms in terms of segmentation performance. Globally,
all the algorithms reach a good level except the EGBIS algorithm, as shown in Figure 18.
This result is due to the fact that this algorithm is sensitive to small gradient variations. As
expected, the EGBIS has a big standard deviation (due to the presence of many local optima)
whereas the thresholding one is low (due to its straightforward behaviour). We have also
compared the performances of the optimization algorithms (the Simplex and the GA) with a
systematic search method (third part of Table 2). By systematic, we mean an iterative search
throughout the whole parameter space with a fixed sampling rate. The sampling rate
depends on the dimensionality of the parameter space. The global performances of the three
methods are similar with a very little advantage to the Simplex.
Default
Algorithm Free Parameter Range
Value
CSC (Priese et al., 2002)
t: region merging threshold 5.0-255.0 20.0
Color Structure Code
SRM (Nock & Nielsen, 2004)
Q: coarse-to-fine scale control 1.0-255.0 32.0
Statistical Region Merging
EGBIS (Felzenszwalb &
0.0-1.0
Huttenlocher, 2004) σ: smooth control on input image 0.50
0.0-
Efficient Graph-Based Image k: colour space threshold 500.0
2000.0
Segmentation
Tlow: low threshold 0.0-1.0 -
Hysteresis thresholding
Thigh: high threshold 0.0-1.0 -
m: region merging threshold 0.0-200.0
CWAGM (Alvarado Moya, 2004) 100.0
n: min. region number 1.0-100.0
Color Watershed-Adjacency Graph 10.0
p: min. probability for watershed 0.0-1.0
Merge 0.45
threshold
Table. 1. Components of the segmentation algorithm bank, their names, and parameters to
tune with range and author’s default values.
To decide between the three different methods, we have compared them by considering
their computational cost as described in Table 3. The systematic search is obviously the most
costly method. The Simplex is the fastest method to converge apart from the CWAGM
algorithm. According to the previous performance score tables, the simplex is definitively
the best algorithm to optimize low dimensional parameter spaces in a few numbers of
iterations. For segmentation algorithms with more than two free-parameters, the Genetic
Algorithm should be preferred, requiring less iteration for the same level of performance.
Note that we have limited the number of iterations — mainly for computational cost
www.intechopen.com
286 Tools in Artificial Intelligence
reasons—for the systematic search method to 2550 for the EGBIS algorithm and to 1250 for
the CWAGM algorithm, respectively. These two algorithms are relatively slow compared to
the others and the parameter space to explore is really huge, particularly for the CWAGM.
Fig. 14. Evaluation profiles of the CSC Fig. 15. Evaluation profiles of the SRM
algorithm applied on the four training algorithm applied on the four training
images presented in Figure 13. images presented in Figure 13.
Fig. 16. Different evaluation profiles of the EGBIS algorithm applied on the four training
images presented in Figure 13. t and σ are the two free parameters.
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 287
Fig. 17. Different evaluation profiles of the Hysteresis thresholding algorithm applied on the
four training images presented in Figure 13. Tlow and Thigh are the two free parameters.
www.intechopen.com
288 Tools in Artificial Intelligence
converges in a few numbers of iterations. The same study is done for the GA and the results
are graphically reported in Figure 18 (right). We decide to assess the GA sensitivity to the
initial population size. The number of initial points is here independent of the segmentation
algorithm and varies between 20 and 840. The same conclusions can be drawn. We just can
add that the EGBIS algorithm brings some problem to the GA which falls in many local
optima (peaks of the EGBIS curve in Figure 18 (right)).
Fig. 18. Convergence accuracy of the Simplex algorithm by varying the maxCalls parameter
and convergence accuracy of the GA by varying the initial population size.
Fig. 19. Examples of images for the two identified clusters. Left = cluster 1 (front side of the
leaves), right = cluster 2 (back side of the leaves).
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 289
www.intechopen.com
290 Tools in Artificial Intelligence
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 291
www.intechopen.com
292 Tools in Artificial Intelligence
8. References
Abdul-Karim, M.-A.; Roysam, B.; Dowell-Mesfin, N. M.; Jeromin, A.; Yuksel, M. &
Kalyanaraman, S. (2005). Automatic selection of parameters for vessel/neurite
segmentation algorithms, Transactions on Image Processings, Vol. 14(9), pp. 1338–
1350
Alvarado Moya, J. P. (2004). Segmentation of color images for interactive 3d object retrieval,
Ph.D. thesis, Technical University of Aachen
Bahnu, B.; Lee, S. & Das, S. (1995). Adaptive image segmentation using genetic and hybrid
search methods, Trans. on Aerospace and Electronic Systems, Vol. 31(4), pp. 1268–1291
Blum, A. L. & Langley, P. (1997). Selection of relevant features and examples in machine
learning, Artif. Intell., Vol. 97(1-2), pp. 245–271
Borenstein, E. & Malik, J. (2006). Shape guided object segmentation, in: Proc. of the Int. Conf.
on Computer Vision and Pattern Recognition, Vol. 1, pp. 969–976
Burges, C. J. C. (1998). A tutorial on support vector machines for pattern recognition, Data
Mining and Knowledge Discovery, Vol. 2(2), pp. 121–167
Chen, Y. & Wang, J. Z. (2004). Image categorization by learning and reasoning with regions,
Journal of Machine Learning Research, Vol. 5, pp. 913–939
Cinque, L.; Corzani, F.; Levialdi, S.; Cucchiara, R. & Pignalberi, G. (2002). Improvement in
range segmentation parameters tuning, in: Proc. of the Int. Conf. on Pattern
Recognition, Vol. 1, pp. 10176
ECVISION (2005). A research roadmap of cognitive vision, Technical report, Project IST-2001-
35454
Ester, M.; Kriegel, H.-P.; Sander, J. & Xu, X. (1996). A density-based algorithm for
discovering clusters in large spatial databases with noise, in: Proc. of the Int. Conf. on
Knowledge Discovery and Data Mining, Portland, pp. 226–231
Everingham, M.; Muller, H. & Thomas, B. (2002). Evaluating image segmentation algorithms
using the pareto front, in: Proc. of the Eur. Conf. Computer Vision, Vol. 2353(4), pp.
34–38
Felzenszwalb, P. F. & Huttenlocher, D. P. (2004). Efficient graph-based image segmentation,
Int. Journal on Computer Vision, Vol. 59(2), pp. 167-181
Fowlkes, C. & Martin, D. (2007). The berkeley segmentation dataset and benchmark,
http://www.eecs.berkeley.edu/Research/Projects/CS/vision/grouping/segbenc
h/
Gelasca, E.; Salvador, E. & Ebrahimi, T. (2003). Intuitive strategy for parameter setting in
video segmentation, in: Visual Communications and Image Processing, Vol. 5150, pp.
998–1008
Goldberg, D. E. (1989). Genetic Algorithms in Search, Optimization and Machine Learning,
Addison-Wesley Longman Publishing Co., Boston, MA, USA
www.intechopen.com
A Cognitive Vision Approach to Image Segmentation 293
Guyon, I. & Elisseeff, A. (2003). An introduction to variable and feature selection, J. Mach.
Learn. Res., Vol. 3, pp. 1157–1182
Haralick, R. (1979). Statistical and structural approaches to texture, in: Proc. of the IEEE, Vol.
67
Huang, F. J. & LeCun, Y. (2006). Large-scale learning with svm and convolutional nets for
generic object categorization, in: Proc. of the Int. Conf. on Computer Vision and Pattern
Recognition, pp. 284–291
Jain, A. K. & Farrokhnia, F. (1991). Unsupervised texture segmentation using gabor filters,
Pattern Recognition, Vol. 24(12), pp. 1167–1186
Kohavi, R. & John, G. H. (1997). Wrappers for feature subset selection, Artificial Intelligence,
Vol. 97(1-2), pp. 273–324
Laws, K. (1980). Textured image segmentation, Ph.D. thesis, Univ. Southern California
Lucchese, L. & Mitra, S. (2001). Color image segmentation : A state-of-the-art survey, Vol. 67,
Indian National Science Academy, New Dehli, pp. 207–221
Mao, S. & Kanungo, T. (2000). Automatic training of page segmentation algorithms : An
optimizatin approach, in: Proc. of the Int. Conf. on Pattern Recognition, Barcelona,
Spain, pp. 531–534
Maurer, C. R. J.; Qi, R. & Raghavan, V. (2003). A linear time algorithm for computing exact
euclidean distance transforms of binary images in arbitrary dimensions, Trans. on
Pattern Anlasys and Machine Intelligence., Vol. 25(2), pp. 265–270
Mezaris, V.; Kompatsiaris, I. & Strintzis, M. (2003). Still image objective segmentation
evaluation using ground truth, in:Proc. of the Workshop on Information and Knowledge
Management for Integrated Media Communication, Prague, Czech Republic, pp. 9–14
Molina, L. C.; Belanche, L. & Nebot, A. (2002). Feature selection algorithms : A survey and
experimental evaluation, in: Proc. of the Int. Conf. on Data Mining, Japan, pp. 306–313
Nelder, J. & Mead, R. (1965). A simplex method for function minimization, Computer Journal,
Vol. 15, pp. 1162–1173
Nock, R. & Nielsen, F. (2004). Statistical region merging, Pattern Analysis and Machine
Intelligence, Vol. 26(11), pp. 1452–1458
Pal, N. R. & Pal, S. K. (1993). A review on image segmentation techniques, Pattern
Recognition, Vol. 26(9), pp. 1277–1294
Peng, J. & Bahnu, B. (1998). Delayed reinforcement learning for adaptive image
segmentation and feature extraction, Systems, Man, and Cybernetics, Vol. 28(3), pp.
482-488
Pignalberi, G.; Cucchiara, R.; Cinque, L. & Levialdi, S. (2003). Tuning range image
segmentation by genetic algorithm, EURASIP Journal on Applied Signal Processing,
Vol. (8), pp. 780–790
Prehn, H. & Sommer, G. (2006). An adaptive classification algorithm using robust
incremental clustering, in: Proc. of the Int. Conf. on Pattern Recognition, pp. 896–899
Priese, V.; Rehrmann, L. & Sturm, P. (2002). Color structure code. URL http://www.uni-
koblenz.de/
Reed, T. R. & du Buf, J. M. H. (1993), A review of recent texture segmentation and feature
extraction techniques, CVGIP: Image Underst., Vol. 57(3), pp. 359–372
Rosenberger, C.; Chabrier, S.; Laurent, H. & Emile, B. (2005). Unsupervised and Supervised
Segmentation evaluation, in book: Advances in Image and Video Segmentation, Yu-Jin
Zhang editor, Tsinghua University, Beijing, China, chapter XVIII
www.intechopen.com
294 Tools in Artificial Intelligence
Schnitman, Y.; Caspi, Y.; Cohen-Or, D. & Lischinski, D. (2006). Inducing semantic
segmentation from an example, in: ACCV, Vol. 3852, Springer-Verlag, pp. 393–384
Skarbek, W. & Koschan, A. (1994). Colour image sementation - a survey, Technical report,
Technical University of Berlin
Wu, T.; Lin, C. & Weng, R. (2004). Probability estimates for multi-class classification by
pairwise coupling, The Journal of Machine Learning Research, Vol. 5, pp. 975–1005
Xia, Y.; Feng, D.; Rongchun, Z. & Petrou, M. (2005). Learning-based algorithm selection for
image segmentation, Pattern Recognition Letters, Vol. 26(8), pp. 1059–1068
Yasnoff, W.; Mui, J. & Bacus, J. (1977). Error measures for scene segmentation, Pattern
Recognition, Vol. 9, pp. 217–231
Zhang, Y. (1996). A survey on evaluation methods for image segmentation, Pattern
Recognition, Vol. 29(8), pp. 1335–1346
Zhang, Y. & Luo, H. (2000). Optimal selection of segmentation algorithms based on
performance evaluation, Optical Engineering, Vol. 39, pp. 1450–1456
Zhang, H.; Fritts, J. E. & Goldman, S. A. (2008). Image segmentation evaluation: A survey of
unsupervised methods, Computer Vision and Image Understanding, Vol. 110(2), pp.
260-280
www.intechopen.com
Tools in Artificial Intelligence
Edited by Paula Fritzsche
ISBN 978-953-7619-03-9
Hard cover, 488 pages
Publisher InTech
Published online 01, August, 2008
Published in print edition August, 2008
This book offers in 27 chapters a collection of all the technical aspects of specifying, developing, and
evaluating the theoretical underpinnings and applied mechanisms of AI tools. Topics covered include neural
networks, fuzzy controls, decision trees, rule-based systems, data mining, genetic algorithm and agent
systems, among many others. The goal of this book is to show some potential applications and give a partial
picture of the current state-of-the-art of AI. Also, it is useful to inspire some future research ideas by identifying
potential research directions. It is dedicated to students, researchers and practitioners in this area or in related
fields.
How to reference
In order to correctly reference this scholarly work, feel free to copy and paste the following:
Vincent Martin and Monique Thonnat (2008). A Cognitive Vision Approach to Image Segmentation, Tools in
Artificial Intelligence, Paula Fritzsche (Ed.), ISBN: 978-953-7619-03-9, InTech, Available from:
http://www.intechopen.com/books/tools_in_artificial_intelligence/a_cognitive_vision_approach_to_image_segm
entation