Home A detailed and comparative work for retinal vessel segmentation based on the most effective heuristic approaches
Article Open Access

A detailed and comparative work for retinal vessel segmentation based on the most effective heuristic approaches

  • Mehmet Bahadır Çetinkaya EMAIL logo and Hakan Duran
Published/Copyright: November 26, 2020

Abstract

Computer based imaging and analysis techniques are frequently used for the diagnosis and treatment of retinal diseases. Although retinal images are of high resolution, the contrast of the retinal blood vessels is usually very close to the background of the retinal image. The detection of the retinal blood vessels with low contrast or with contrast close to the background of the retinal image is too difficult. Therefore, improving algorithms which can successfully distinguish retinal blood vessels from the retinal image has become an important area of research. In this work, clustering based heuristic artificial bee colony, particle swarm optimization, differential evolution, teaching learning based optimization, grey wolf optimization, firefly and harmony search algorithms were applied for accurate segmentation of retinal vessels and their performances were compared in terms of convergence speed, mean squared error, standard deviation, sensitivity, specificity. accuracy and precision. From the simulation results it is seen that the performance of the algorithms in terms of convergence speed and mean squared error is close to each other. It is observed from the statistical analyses that the algorithms show stable behavior and also the vessel and the background pixels of the retinal image can successfully be clustered by the heuristic algorithms.

Introduction

Retinal vessel segmentation is one of the most important areas of retinal image analysis because some attributes of retinal vessels are usually important symptoms of diseases [1]. Due to the disadvantages of manual retinal image analysis automatic analysis of retinal images becomes an important area of research. Automated segmentation of retinal vessels is accepted as the first step of developments in the area of computer-aided diagnosis systems for ophthalmic disorders [2], The abnormalities caused by diseases of obesity [3], hypertension [4], glaucoma [5] and diabetic retinopathy [6], [7], [8] are able to display with higher accuracy by means of the automated segmentation of retinal vessels. It also plays an important role in the areas such as evaluation of retinopathy of prematurity [9], vessel diameter measurement [10]. fovea region detection [11], arteriolar stenosis [12], computer assisted laser surgery [13], treatment of ophthalmologic diseases [14], [15], [16] and optic disk detection [17].

The conventional classical algorithms are well developed for clustering based retinal image analysis. However, there are few works in literature including retinal image analysis by using heuristic algorithms. In this work clustering based heuristic artificial bee colony (ABC), particle swarm optimization (PSO), differential evolution (DE), teaching learning based optimization (TLBO), grey wolf optimization (GWO), firefly (FA) and harmony search (HS) algorithms have been applied to retinal vessel segmentation. The retinal images used in the simulations are taken from the DRIVE and STARE databases. In order to perform a detailed analysis simulations are realized for both normal and abnormal retinal images. The normal and abnormal retinal images taken from DRIVE and STARE databases are given in Figures 1 and 2, respectively.

Figure 1: Retinal images taken from DRIVE database.(a) Normal retinal image; (b) Abnormal retinal image.
Figure 1:

Retinal images taken from DRIVE database.

(a) Normal retinal image; (b) Abnormal retinal image.

Figure 2: Retinal images taken from STARE database.(a) Normal retinal image; (b) Abnormal retinal image.
Figure 2:

Retinal images taken from STARE database.

(a) Normal retinal image; (b) Abnormal retinal image.

In order to improve the performance of retinal vessel segmentation, some pre-processing operations have to be applied before clustering. Then the pre-processed retinal images are evaluated by the algorithms. Pre-processing operations used in this work can be described as the following.

Pre-processing operations

The retinal images analyzed in this work consist of red (R), green (G) and blue (B) color components and when each layer of these images was examined separately, it was seen that the highest clustering performance was obtained in the green layer. The G layer has higher illuminance and it also provides optimal contrast and brightness levels when compared to R and B layers. For these reasons, after this step, the image analysis was continued on the green layer. This pre-processing operation can be called as band selection.

As a result, after switching from a three-dimensional RGB image format to a one-dimensional image by band selection process; i.) the image becomes suitable for two-dimensional bottom-hat filtering, ii.) the pixel differences between the image background and the vessels become a bit more evident.

The contrast difference between the vessels and the image background was not found to be sufficient for a high clustering performance even after pre-processing operation mentioned above. For this reason, the next step is to apply a second pre-processing operation called bottom-hat transformation. This transformation can be described as extraction of the original image from the morphologically closed image. A bottom-hat filter enhances black spots in a white background by subtracting the morphological Close of the image from the original image. Bottom-hat filter increases the contrast between the bright levels in the retinal image and as a result of this, regions with different brightness levels can successfully be distinguished. Eq. (1) represents the bottom-hat transformation,

(1)bottomhat(g)=g(gnB)

where g is the retinal image. B is the structural element to be used (filter structure), and n is the bottom-hat transformation. In this work, a disk with r=8 pixel radius have been used as building element and bottom-hat transformation is applied as to be n=8.

The last pre-processing applied before clustering is the brightness correction. During the photo shoots of biomedical images, low contrast values or high pixel levels may occur due to the device or environment. In brightness correction process, these extreme pixel values creating noise effect on the retinal image are being optimized by using the following Equation for a successful segmentation.

(2)Pout=(Pinc)(ba)(dc)+a

In the equation given above, a and b represent the minimum and maximum pixel values of the image, respectively. Also, c and d are respectively representing the lowest and highest limit values of the pixel range that we want to normalize the image and in which the noise effects are minimized. As a result of brightness correction the contrast between the vessels and the image background becomes optimal for clustering.

Figure 3 (b) and (e) show the green layer images obtained as a result of band selection for the retinal images taken from the DRIVE database given in Figure 1 (a) and (b). The enhanced retinal images obtained after the subsequent two pre-processing operations have been represented in Figure 3 (c) and (f).

Figure 3: The enhanced retinal images obtained after pre-processing operations. (a) and (d) are the normal and abnormal retinal images, respectively, taken from the DRIVE database, (b) and (e) are the green layer images obtained as a result of band selection for the retinal images taken from the DRIVE database given in Figure 1 (a) and (b), (c) and (f) are the enhanced retinal images obtained after the bottom-hat filtering and brightness correction, (g) and (j) are the normal and abnormal retinal images, respectively, taken from the STARE database, (h) and (k) are the green layer images obtained as a result of band selection for the retinal images taken from the STARE database given in Figure 2 (a) and (b), (i) and (l) are the enhanced retinal images obtained after the bottom-hat filtering and brightness correction.
Figure 3: The enhanced retinal images obtained after pre-processing operations. (a) and (d) are the normal and abnormal retinal images, respectively, taken from the DRIVE database, (b) and (e) are the green layer images obtained as a result of band selection for the retinal images taken from the DRIVE database given in Figure 1 (a) and (b), (c) and (f) are the enhanced retinal images obtained after the bottom-hat filtering and brightness correction, (g) and (j) are the normal and abnormal retinal images, respectively, taken from the STARE database, (h) and (k) are the green layer images obtained as a result of band selection for the retinal images taken from the STARE database given in Figure 2 (a) and (b), (i) and (l) are the enhanced retinal images obtained after the bottom-hat filtering and brightness correction.
Figure 3:

The enhanced retinal images obtained after pre-processing operations. (a) and (d) are the normal and abnormal retinal images, respectively, taken from the DRIVE database, (b) and (e) are the green layer images obtained as a result of band selection for the retinal images taken from the DRIVE database given in Figure 1 (a) and (b), (c) and (f) are the enhanced retinal images obtained after the bottom-hat filtering and brightness correction, (g) and (j) are the normal and abnormal retinal images, respectively, taken from the STARE database, (h) and (k) are the green layer images obtained as a result of band selection for the retinal images taken from the STARE database given in Figure 2 (a) and (b), (i) and (l) are the enhanced retinal images obtained after the bottom-hat filtering and brightness correction.

The green layers of the retinal images given in Figure 2 (a) and (b) which are taken from the STARE database have been shown with Figure 3 (h) and (k). The enhanced retinal images obtained after bottom-hat and brightness correction operations have been given in Figure 3 (i) and (l).

Materials and methods

Artificial bee colony algorithm

Swarm intelligence simulates the collective behaviour of animal societies [18] and it provides effective solutions in all areas of engineering. Artificial bee colony algorithm which is proposed by Karaboga in 2005 [19] is one of the most effective and widely used swarm based algorithms.

The pseudo-code of a basic ABC algorithm can be given as the following.

Randomly create an initial population consisting of solutions xi{1,2,,SN}

Calculate the fitness value of each xi solution in the population.

Cycle=1

REPEAT

Produce new solutions vij in the neighbourhood of xijusing vij=xij+ϕij(xijxkj) and calculate the fitness values of these solutions.

Apply the greedy selection betweenxi andviand save the selected new solution to memory.

Calculate the probability values pi for the solutions xi by means of their fitness values using the following Eq. (3).

(3)pi=fitii=1SNfitn

where the pi values are normalized into [0,1]. The fitness value of each solution is calculated by using the following Eq. (4).

(4)fiti=11+fi

Produce the new solutions (new positions) vi from the solutions xi. selected depending on pi. and evaluate them.

Apply the greedy selection between xi and vi then memorize the selected new solution.

Determine the abandoned solution (source), if exists, and replace it with a new randomly produced solution xi using Eq. (5) given below,

(5)xij=xminj+rand[0,1](xmaxjxminj)

So far, memorize the best food source position (solution) achieved

Cycle=Cycle + 1

UNTIL Cycle=Maximum Cycle Number

After an initial population is created randomly the ABC algorithm performs three steps at each cycle of the search. At the first step the fitness (quality) values of the initial solutions are calculated and memorized. Then a new possible solution within the neighborhood of the present solution is determined and its fitness value is calculated. If the fitness value of the new solution is higher, the previous solution is forgotten and the new solution is memorized (greedy selection). At last step new possible solutions are determined within the neighborhood of the solutions evaluated so far and then their fitness values are calculated.

In ABC algorithm, if a solution cannot be improved by a predetermined number of trials, it means that the associated solution has been exhausted and will no more be evaluated. The memorized information about this abandoned solution is replaced with information of a randomly produced new solution. The number of trials for releasing a solution is equal to the value of “limit” which is an important control parameter of ABC algorithm. These three steps are repeated until the termination criteria are satisfied.

Assume that wi is the ith solution to the problem and f(wi) represents the fitness of this solution. Let P(c)={wi(c)|i=1,2,,s} (c: cycle. s: number of possible solutions) represent the population of possible solutions. If the fitness value of a solution is high, the probability of this solution to be transferred to the next generation becomes high. The probability function can be defined as,

(6)pi=f(wi)i=1sf(wk)

where s is the number of solutions in the population. A new possible solution is selected depending on the probabilities calculated and then a neighbour solution around the chosen one is determined. This process is repeated until all possible solutions evaluated. Assume that a possible solution selected is wi. The neighbour solution of wi might be determined as the following,

(7)wi(c+1)=wi(c)+ϕi(wi(c)wk(c))

where ϕi is a randomly generated number in the interval [−1, +1] and k is a randomly produced index different from i. If f(wi(c+1)) is higher than f(wi(c)), then the algorithm memorizes wi(c+1). Namely, wi(c) is replaced with wi(c+1), otherwise wi(c) is kept as it is. If wi cannot be improved through the predetermined number of trials, it is abandoned. The algorithm starts to search for a new possible solution randomly. and after finding the new one, the new solution is accepted to be wi(c+1).

Similar to PSO, DE, TLBO, GWO, FA and HS algorithms, while performing retinal vessel segmentation with basic ABC algorithm, optimal pixel range consisting of optimal lower and upper pixel values is investigated. In basic ABC algorithm, the segmentation process is initialized via a randomly selected xi pixel position. Firstly, the fitness value of each xi pixel position creating the retinal image is calculated and then these position values are updated by using Eq. (5) until the optimum fitness values for lower and upper bounds is reached. Finally, the retinal vessel segmentation is realized in the interval between the optimum xi lower and upper pixel positions obtained so far.

Particle swarm optimization algorithm

Particle swarm optimization is a population based stochastic optimization technique developed by Eberhart and Kennedy in 1995, inspired by social behavior of bird flocking or fish schooling [20]. In PSO, a population of particles (solutions) is made to move across D-dimensional space. The particles move across the problem space by getting information from the current optimum particles. The position of a given particle at any cycle represents a possible solution of the problem. Evaluating the objective function at a given position gives the fitness of the corresponding solution. The particles also have velocities which direct particles searching in the space.

PSO is initialized with a population including randomly produced particles and then searches for optimal solutions by updating positions in each generation. In every cycle, each particle is updated by best solution [p(t)] that it has achieved so far. and the best solution of the population. [g(t)]. When a particle takes part of the population as its topological neighbors, the best value is a local best.

Intuitively, the information about good solutions is distributed through the swarm and thus the particles are directed to move to good areas in the search space. At each time step t. the velocity υ(t) is updated and the particle is moved to a new position. x(t+1). This new position is calculated as the sum of the previous position and the new velocity:

(8)x(t+1)=x(t)+υ(t+1)

The update of the velocity from the previous velocity to the new velocity is determined by the following Eq. (9):

(9)υ(t+1)=ωυ(t)+r1c1(p(t)x(t))+r2c2(g(t)x(t))

where r1 and r2 are uniformly distributed random numbers. The parameter ω is called the inertia weight and controls the magnitude of the old velocity υ(t) in updating velocity. whereas c1 and c2 determine the significance of p(t) and g(t), respectively. Furthermore, at any time step of the algorithm υi is constrained by the υmax parameter. The swarm in PSO is initialized by assigning each particle to a uniformly and randomly chosen position in the search space. Velocities are initialized randomly in the [υmin, υmax] range.

The pseudo code of the procedure:
Initialize
REPEAT
 Calculate fitness values of all particles
 Modify the best particles in the swarm
 Choose the particle with the best fitness value of all the particles
 Calculate the velocities of particles
 Update the particle positions
UNTIL(maximum generations or minimum error criteria is satisfied)

If the sum of accelerations would cause the velocity of that parameter to exceed υmax, the velocity on that dimension is limited to υmax.

While performing retinal vessel segmentation by using basic PSO algorithm, firstly, for each xi pixel position value that is randomly selected and constitutes the initial population, fitness values are calculated separately and stored in the p(t) matrix. Then, the optimum solution in the p(t) matrix obtained at each cycle is transferred to the g(t) matrix in which the best lower and upper pixel values are kept. At each cycle, the fitness value of each solution in the updated p(t) matrix is compared separately with the memorized fitness value kept into the g(t) matrix and the solutions producing the minimum fitness value is selected as the optimal pixel values. Thus, when the PSO algorithm completes the maximum number of cycles, the best lower and upper pixel position values obtained so far are stored into the g(t) matrix. Retinal vessel segmentation is then realized by using these optimal pixel values stored in g(t).

Differential evolution algorithm

Differential evolution algorithm which has been introduced by Storn and Price [21], is a method with crossover and mutation operations that work directly on continuous-valued vectors. An optimization task consisting of D parameters can be represented by a D dimensional vector. Therefore, in DE algorithm a population of NP solution vectors is randomly created at the start. This population is successfully improved by applying mutation, crossover and selection operators.

The main steps of a basic DE algorithm are given below:

Initialization
Evaluation
REPEAT
 Mutation
 Recombination
 Evaluation
 Selection
UNTİL(maximum generations or minimum error criteria is satisfied)

In the mutation operator, for each target vector xi,G, a mutant vector is produced by

(10)vi,G+1=xi,G+K(xr1,Gxi,G)+F(xr2,Gxr3,G)

where i,r,1r2,r3{1,2,...,NP} that randomly chosen and must be different from each other. In Eq. (10). F is the scaling factor [0,2] affecting on difference vector (xr2,Gxr3,G), K is the combination factor. In the crossover operator, the parent and the mutated vectors are evaluated together in order to produce a trial vector of uji,G+1,

(11)uji,G+1={vji,G+1if(rndjCR)orj=rnixji,Gif(rndj>CR)andjrni

where j=1,2, … , D; rj[0,1] random number; CR stands for the crossover constant [0,1] and rni ∈(1,2, … , D) randomly chosen index. Performance of the trial vector and its parent is compared and the better one is selected. All solutions have the same chance of being selected as parents without dependence of their fitness value.

While DE algorithm is applied to retinal vessel segmentation, firstly, an initial population consisting of vectorially defined pixel locations is created. Then, by using the mutation and crossover operations represented in Eqs. (10) and (11), new pixel values are randomly created in the neighbourhood of each pixel value available in the initial population. The fitness values of the previous and the new produced pixel values are compared and the better solutions are stored as the optimal lower and upper pixel values. At each cycle, the optimal pixel values in the memory are updated by mutation and crossover, and as a result of this the most appropriate lower and upper pixel values for retinal vessel segmentation are obtained. Finally, by using the pixel value interval between the lower and upper pixel values obtained optimal retinal vessel segmentation is achieved.

Teaching learning based optimization algorithm

TLBO is an approach based on the learning-teaching interaction between teachers and students in order to solve multi-dimensional, linear and nonlinear problems [22]. The algorithm consists of two main phases so as to be teacher phase and learners phase.

In teacher phase, the teacher transfers knowledge to the students in order to increase the mean knowledge level of the class. Since the teacher is the most experienced and knowledgeable person in a class, he or she represents the best solution of the entire population. Let m be the number of subjects which corresponds to the parameter number to be optimized. Also, if n is the number of learners, population size can be determined as k=1, 2, … , n. At any cycle of i, Mj.i represents the mean result of the learners in a particular subject of j=1, 2, … , m. The average difference between the teacher’s knowledge capacity and the learning capacity of students is given in Eq. (12).

(12)Difference_Meanj,i=ri(Xj,kbest,iTfMj,i)

where, Xj,kbest,i is the best solution obtained for subject j. Also, ri is a randomly produced coefficient in the interval of [0,1]. Finally, the variable Tf is called as learning factor and its value is randomly decided by the algorithm according to Eq. (13).

(13)Tf=round[1+rand(0,1){1,2}]

As seen from Eq. (12), the value of Tf can either be 1 or 2. The available solution is updated according to Difference_Meanj,i as given in Eq. (14),

(14)Xj,k,i=Xj,k,i+Difference_Meanj,k,i

where, Xj,k,i is the updated value of Xj,k,i.

The learner phase aims to increase the knowledge owned by the students by means of interaction with the teacher or between themselves. Let, P and Q be the randomly selected students, such that,

(15)XtotalP,iXtotalQ,i

where XtotalP,i and XtotalQ,i are the updated values of XtotalP,i and XtotalQ,i, at the end of the teacher phase. The candidate solution is obtained as the following.

(16)If, XtotalP,i<XtotalQ,ithen,Xj,P,i=Xj,P,i+ri(Xj,P,iXj,Q,i)
(17)If, XtotalP,i<XtotalQ,ithen,Xj,P,i=Xj,P,i+ri(Xj,Q,iXj,P,i)
The pseudo code of the TLBO algorithm can be given as following.
Randomly create an initial population
Calculate fitness values of all individuals
Teacher phase: Modify the population based on Eq. (14) and discretize it
REPEAT
 Calculate the fitness values of all individuals
 Accept each new individual if it gives a better fitness value
 Learner phase: Modify the population based on Eqs. (15)–(17) and discretize it
 Calculate the fitness value of all individuals
 Accept each new individual if it gives a better fitness value
UNTIL(maximum cycles or minimum error criteria is satisfied)

In retinal vessel segmentation based on TLBO algorithm, each pixel in the retinal image represents a student. The pixel with the highest fitness value among these pixels represents the teacher. At each cycle, the new pixel values representing possible solutions are produced by Eq. (14) and then fitness values of these new solutions are calculated. Among these new fitness values, the best two is selected and stored to the memory as the best lower and upper pixel values by using Eqs. (15)–(17). By repeating these steps at each cycle, the optimal lower and upper pixel values producing the best fitness values are found and the retinal vessel segmentation is applied according to these optimal solutions.

Grey wolf optimization algorithm

GWO is a heuristic optimization algorithm inspired from the collective behaviour of gray wolves in nature and proposed by Mirjalili et al. in 2014 [23], [24], [25]. It simulates the leadership hierarchy consisting of four types of wolves named as alpha (α), beta (β), delta (δ) and omega (w). In the population, α wolves represent the best solution for the problem to be optimized while the β and δ wolves represent the second and the third best solutions, respectively. The rest of the populations are determined as the w wolves. In GWO algorithm, an optimization task consists of three phases to be encircling, hunting and attacking.

Encircling phase includes the optimization of the positions of gray wolf and prey as given below,

(18)D=|CXp(t)X(t)|
(19)X(t+1)=Xp(t)AD

where, t represents the current cycle, X(t) and Xp(t) are the position vectors of the grey wolf and prey, respectively. Also, A and C are the coefficient vectors which can be determined as the following.

(20)A=2ar1a
(21)C=2r2

where, r1 and r2 are randomly produced vectors in the range [0,1] and they provide reach to random positions in the solution space. Also, a are the linearly decreased coefficients from 2 to 0 over the course of cycles.

In hunting phase, the position of the prey corresponds to the global solution. At the beginning of the search it is supposed that the α, β and δ have better knowledge about the position of the prey. The fitness values of each position information are calculated and the solution with the best fitness value is defined as α. The solution with second best fitness value is assigned to β and the solution with third best fitness value is defined as δ. At each cycle of the search, the position of the current optimal solution can be updated according to the Eqs. (22)–(24) in order to reach to the global solution.

(22)D=|C1XX(t)|,Dβ=|C2XβX(t)|,Dδ=|C3XδX(t)|

where, X, Xβ and Xδ represents to the initial positions of the α, β and δ.

(23)X1=|Xa1(D)|,X2=|Xβa2(Dβ)|,X3=|Xδa3(Dδ)|
(24)X(t+1)=X1+X2+X33

As seen from Eq. (24), the position updated would be in a random place within a circle which is defined by the positions of α, β and δ in the search space.

Attacking phase can be simulated by reducing the distance between the gray wolf and the prey. This phase can be modeled by decreasing the value of a in Eq. (20) from 2 to 0 over the course of cycles. As seen from Eq. (20), decreasing a will lead to fluctuations in A. If the values of A are in [−1,1], the new position will be in any position between current position and the position of the prey.

The pseudo code of a basic GWO algorithm can be given as the following.

Randomly create an initial population, Xi(i=1,2,.,n)
Determine the initial values of a, A and C
Calculate the fitness value of each initial position (Xa. Xβ and Xδ)
while(t < Max number of cycles)
     for each search agent
      Update the current optimal position by Eqs. (22)(24).
     end for
     Update a, A and C
     Calculate the fitness value of current positions
     Update (Xa, Xβ and Xδ)
     t=t + 1
end

Return Step 3 for current Xa

In GWO algorithm, an initial population consisting of Xi pixel values each representing position vectors is randomly created. At each cycle of the algorithm the population Xi is evaluated and the solution with the best pixel value range producing the highest fitness value is determined as Xa, the solution with the second best pixel value range is determined as Xβ and finally the solution with the third best pixel value range is determined as Xδ. At each cycle, the values of Xa, Xβ and Xδ are updated by using Eqs. (22)(24). in order to reach lower and upper bounds of the best pixel range producing the optimal segmentation. As a result, retinal vessel segmentation is realized by using Xa which is producing the optimal fitness value.

Firefly algorithm

FA, which was proposed by Xin-She Yang, is a novel swarm intelligence based algorithm inspired by the phenomenon of bioluminescent communication behaviour of fireflies [26].

The following rules create the philosophy of the FA algorithm [26].

  • Fireflies are unisex and can attract any fellow FF.

  • Attractiveness depends on ones brightness.

  • The brightness or light intensity of a firefly is influenced by the landscape of fitness/cost function.

The pseudo code of a basic FA can be given as the following.

BEGIN
  Initialisation max cycle, α, βo, γ
  Generate initial population
  Define the Objective function f(x).
  Determine Intensity (I) at cost (x) of each individual determined by f(xi)
  while (t<Iter max)
        fori=1 to n
           forj=1 to n
               if (Ij>Ii)
               Move firefly i towards j in K dimension
             end if
             Evaluate new solutions and update light intensity
           end forj
        end fori
      Rank the fireflies and find the current best
   end while
   Post process results and visualization
END (if end procedure satisfied)

In retinal vessel segmentation based on FA algorithm, the optimization task begins with creating an initial population consisting of Xi pixel values in an interval limited by lowest and highest pixel values and then the fitness value of each Xi is calculated. Then the fitness value of each Xi is compared to that of the fitness value of new possible solution produced in the neighbourhood of the relevant Xi by the following Eq. (25).

(25)Xi,k=Xi,k+β(Xj,kXi,k)+αi,k

where, β represents the value of attractiveness parameter changing in the interval of [0,1], i represents the index of the selected pixel value, j represents the index of the new pixel produced as a possible solution in the neighbourhood of the ith pixel, k represents the cycle number, Xi,k represents the value of ith pixel for kth cycle, Xj,k represents the value of jth pixel for kth cycle, α represents the value of attractiveness parameter changing in the interval of [0,1] and finally ∈ is a randomly produced number in the interval of [−0.5,0.5].

Among the pixel values updated at each cycle, the pixel values producing the highest fitness value for lower and upper pixel value bounds are selected and retinal vessel segmentation is applied for this interval.

Harmony search algorithm

Harmony search (HS) algorithm was developed by Geem at al. in an analogy with music improvisation process where music players improvise the pitches of their instruments to obtain better harmony [27].

The steps in the procedure of HS algorithm can be given as the following [28].

  • Step 1. Initialize the problem and algorithm parameters.

  • Step 2. Initialize the harmony memory.

  • Step 3. New harmony improvisation.

  • Step 4. Update the harmony memory.

  • Step 5. Check the stopping criterion.

The pseudo code of a basic HS algorithm can be given as the following [28].

Objective function f (Xi), i=1 to N
Define HS parameters: HMS. HMCR. PAR. and BW
Generate initial harmonics (for i=1 to HMS)
Evaluate f(Xi)
while (t<Max number of iterations)
   Create a new harmony: Xinew, i=1 to N
   if (U(0.1)>HMCR),
      Xinew=Xjold, where Xjold is a random value in the interval of {1, … , HMS}
   else if If (U (0.1)>PAR),
      Xinew=Xjold + BW [(2 × U (0.1)) – 1]
   else
      Xinew=XL(i) + U (0.1) × [XU(i) – XL(i)]
   end if
   Evaluation f (Xinew)
   Accept the new harmonics (solutions) if better
end while

Find the current best estimates

While applying HS algorithm to retinal vessel segmentation, firstly, an initial population consisting of HMS position vectors within an interval limited by lowest and highest pixel values is randomly created and then these position vectors representing possible solutions are registered into the harmony memory. After calculating the fitness values for each pixel representing possible solutions, new position vectors representing possible solutions (pixel values) are created by using the equations given in the pseudo code. The fitness value of the older and the new produced position vectors are compared and the pixel value with higher fitness is transferred to the next generation by replacing with the relevant pixel value in the harmony matrix. Otherwise, the old pixel value in the harmony matrix is preserved. Thus, after a predetermined cycle number the algorithm will be reached the most appropriate lower and upper limits of the pixel interval in which the optimum retinal vessel segmentation can be performed.

Retinal vessel segmentation

The performances of the algorithms have been tested on the images taken from two public databases. The first one is the STARE (Structured Analysis of the Retina) database which is created by Hoover et al. [29]. STARE database contains 20 raw retinal images for blood vessel segmentation and 10 of these images have pathologies. All images were captured by a TopCon TRV-50 fundus camera at 35° field of view (FOV) and digitized to 700 × 605 pixels with 8 bits per color channel.

The second database used in this work is the DRIVE database which was established by Alonso-Montes et al. [30]. In DRIVE database, 20 images are employed for training and 20 images for testing. These images are captured by a CanonCR5 3CCD camera with a 45° FOV and size is 700 × 605 pixels per color channel.

The flowchart of the method applied for retinal vessel segmentation is shown in Figure 4.

Figure 4: Flowchart of retinal vessel segmentation.
Figure 4:

Flowchart of retinal vessel segmentation.

At first step, due to its higher contrast the green layer of the RGB retinal image is extracted and in the next steps analysis have been carried out on this layer. Although its higher rate of contrast, it is seen from the analyses that the G layer alone is insufficient for a successful clustering. For this reason, in order to clarify the vessels and the image background and as a result of this to improve the clustering performance, two more pre-processing were applied to the retinal image. The first process is the bottom-hat filtering while the second process is the brightness correction. Retinal images obtained after the pre-processing phase are subjected to segmentation process by the ABC, PSO, DE, TLBO, GWO, FA and HS algorithms. In the segmentation phase. optimal clustering centers have been determined by using ABC, PSO, DE, TLBO, GWO, FA and HS algorithms in order to obtain the highest clustering performance and clustering has been done according to these centers.

In the clustering process cluster centers are randomly determined and all pixels are indexed to the closest cluster center. In order to index each pixel to the nearest cluster center, the Mean-Squared Error (MSE) function is used. Namely, the quality of each solution has been calculated as mean squared error. The MSE function can be expressed as the following,

(26)MSE=1Ni=1N(fiyi)2

As seen from the equation given above, MSE error function calculates the quality of a solution based on the distance between each pixel and its related cluster center. In this equation, N represents the total number of pixels in the retinal image. Cluster centers are determined by algorithm and fi represents the value of the cluster center closest to the pixel i. Also, yi represents the pixel value of the ith pixel. By using MSE error function it is aimed to group the pixels around the appropriate cluster centers with the minimum error.

The control parameter values of the algorithms used in the simulations are given in Table 1.

Table 1:

Control parameter values used in the simulations.

ABC
  • Colony size=10

  • Limit value=12

  • Xmax and Xmin represent the highest and lowest pixel values of the image, respectively.

  • Maximum cycle=100

GWO
  • Search agents size=10

  • The random parameters A and C assist candidate solutions to have hyper-spheres with different random radii.

  • Exploration and exploitation are guaranteed by the adaptive values of a and A.

  • The GWO has only two main parameters to be adjusted (a and C).

  • Xmax and Xmin represent the highest and lowest pixel values of the image, respectively.

  • Maximum cycle=100

PSO
  • Swarm size=10

  • Inertia factor, ω=0.72

  • Cognitive factor. c1=1.49

  • Social factor. c2= 1.49

  • Xmax and Xmin represent the highest and lowest pixel values of the image, respectively.

  • Vmax=25, Vmin=−25

  • Maximum cycle=100

FA
  • Swarm size=10

  • Maximum attractiveness=1

  • Time varying algorithm parameter initial value=0.2

  • Absorption coefficient=1

  • Xmax and Xmin represent the highest and lowest pixel values of the image, respectively.

  • Maximum cycle=100

DE
  • Population size=10

  • Crossover rate=0.8

  • Scaling factor=0.2

  • Xmax and Xmin represent the highest and lowest pixel values of the image, respectively.

  • Maximum cycle=100

HS
  • Harmony memory size=25

  • Number of new harmonies=10

  • Harmony memory consideration rate=0.9

  • Pitch adjustment rate=0.1

  • Xmax and Xmin represent the highest and lowest pixel values of the image, respectively.

  • Maximum cycle=100

TLBO
  • Population size=10

  • This algorithm requires only the common control parameters such as the population size and the number of generations and does not require any algorithm-specific control parameters.

  • Xmax and Xmin represent the highest and lowest pixel values of the image, respectively.

  • Maximum Cycle=100

Results

Figure 5 represents the resulting retinal images obtained after applying segmentation to the images Figure 1 (a) and (b) by using the clustering based ABC, PSO, DE, TLBO, GWO, FA and HS algorithms. From the figures it is seen that some regions that are not vessel but whose pixel values are close to the pixel value of the vessels seems to be segmented as vessel. There are post-processing methods in literature in order to eliminate these regions. But in this work in order to represent the pure performance of the algorithm these post-processing methods have not been applied. In general, as seen from the figures all three algorithms show a similar performance in terms of clustering. Also it is seen that ABC, PSO, DE, TLBO, GWO, FA and HS algorithms are able to cluster close pixel values at high accuracy.

Figure 5: Retinal images obtained after applying segmentation to the images in Figure 1a and 1b by using the clustering based ABC (a and h), PSO (b and i), DE (c and j), TLBO (d and k), GWO (e and l), FA (f and m) and HS (g and n) algorithms.
Figure 5: Retinal images obtained after applying segmentation to the images in Figure 1a and 1b by using the clustering based ABC (a and h), PSO (b and i), DE (c and j), TLBO (d and k), GWO (e and l), FA (f and m) and HS (g and n) algorithms.
Figure 5:

Retinal images obtained after applying segmentation to the images in Figure 1a and 1b by using the clustering based ABC (a and h), PSO (b and i), DE (c and j), TLBO (d and k), GWO (e and l), FA (f and m) and HS (g and n) algorithms.

The clustering performance of the algorithms for Figure 2 (a) and (b) are given in Figure 6. In these images taken from the STARE database, it is seen that the performance of the algorithms is lower than the images taken from the DRIVE database. On the other hand, the results obtained show that the clustering performance of the algorithms is similar but a bit higher when compared with the results obtained in the literature.

Figure 6: Retinal images obtained after applying segmentation to the images Figure 2a and 2b by using the clustering based ABC (a and h), PSO (b and i), DE (c and j), TLBO (d and k), GWO (e and l), FA (f and m) and HS (g and n) algorithms.
Figure 6: Retinal images obtained after applying segmentation to the images Figure 2a and 2b by using the clustering based ABC (a and h), PSO (b and i), DE (c and j), TLBO (d and k), GWO (e and l), FA (f and m) and HS (g and n) algorithms.
Figure 6:

Retinal images obtained after applying segmentation to the images Figure 2a and 2b by using the clustering based ABC (a and h), PSO (b and i), DE (c and j), TLBO (d and k), GWO (e and l), FA (f and m) and HS (g and n) algorithms.

Performance measures

The statistical analysis and performance comparison of the clustering based algorithms are realized based on the sensitivity (Se), specificity (Sp), accuracy (Acc), precision, standard deviation and mean squared error. Firstly, the performance of the algorithms is compared in terms of the Se, Sp and Acc and precision. These measures have been computed individually for each image and on average for the whole test images. The expressions are given below,

(27)Se=TP/(TP+FN)
(28)Sp=TN/(TN+FP)
(29)Acc=(TP+TN)/(TP+FN+TN+FP)
(30)Precision=TP/(TP+FP)

where the true positives (TP) represent the number of pixels that are actually vessels and have been detected as vessels. The false negatives (FN) determine the number of pixels that are actually vessels but which have been detected as background pixels. The true negatives (TN) are the pixels classified as non-vessel and they are really background pixels. The false positives (FP) represent the number of pixels that are actually background pixel and have been detected as vessels. So the sensitivity (Se) is the ratio of correctly classified vessel pixels while specificity (Sp) is the ratio of correctly classified background pixels and accuracy (Acc) is the ratio of correctly classified both the vessels and background pixels. Finally, precision determines the number of accurately true predicted TP pixels.

Performance analysis for DRIVE and STARE databases

Table 2, Table 3, Table 4, Table 5 show the performance of the ABC, PSO, DE, TLBO, GWO, FA and HS algorithms in terms of sensitivity, specificity, accuracy and precision for 20 retinal images taken from both DRIVE and STARE databases. These tables also contain the mean values of the results obtained for 20 retinal images. From the results obtained for Se, it is seen that each algorithm is able to reach high performance in terms of the ratio of correctly classified vessel pixels. The Sp values reached represents that each algorithm can successfully distinguish the vessel pixels and background pixels. Also, when the Acc values examined the results show that the ratio of correctly classified both the vessels and background pixels is very high for each algorithm. Finally, it is seen from the results obtained for precision that heuristic algorithms show similar and good performances in terms of correctly classified vessel pixels.

Table 2:

The performance measures of the different images in the DRIVE database in terms of sensitivity, specificity and accuracy.

ImageSensitivitySpecificityAccuracy
ABCPSODETLBOGWOFAHSABCPSODETLBOGWOFAHSABCPSODETLBOGWOFAHS
10.89130.89130.89130.89130.86060.89130.89130.98840.98840.98840.98840.98470.98840.98840.97910.97910.97910.97910.97240.97910.9791
20.83020.78570.83020.78570.83020.78570.78570.97760.97030.97760.97030.97760.97030.97030.96040.94790.96040.94790.96040.94790.9479
30.94080.87870.87870.87870.53920.87870.87870.99390.98670.98670.98670.9230.98670.98670.98890.9760.9760.9760.8680.9760.976
40.7230.7230.7230.7230.77930.7230.7230.96450.96450.96450.96450.97350.96450.96450.93710.93710.93710.93710.95270.93710.9371
50.95140.95140.95140.95140.95140.95140.95140.99550.99550.99550.99550.99550.99550.99550.99180.99180.99180.99180.99180.99180.9918
60.95720.95720.95720.95720.86050.95720.95720.99540.99540.99540.99540.98360.99540.99540.99170.99170.99170.99170.97070.99170.9917
70.95930.88740.88740.88740.73310.88740.88740.99670.99030.99030.99030.97270.99030.99030.9940.98220.98220.98220.95040.98220.9822
80.55060.55060.55060.55060.86720.55060.55060.94490.94490.94490.94490.98920.94490.94490.90190.90190.90190.90190.980.90190.9019
90.78180.78180.78180.78180.82510.78180.78180.97610.97610.97610.97610.98170.97610.97610.95680.95680.95680.95680.96690.95680.9568
100.89320.89320.7180.89320.89320.89320.89320.99080.99080.97050.99080.99080.99080.99080.98310.98310.94670.98310.98310.98310.9831
110.7870.7870.7870.7870.98790.7870.7870.9760.9760.9760.9760.99890.9760.9760.95690.95690.95690.95690.9980.95690.9569
120.66240.66240.66240.66240.68240.66240.86090.95740.95740.95740.95740.96090.95740.98610.92430.92430.92430.92430.93050.92430.9747
130.74070.79920.79920.79920.64890.79920.79920.96210.97250.97250.97250.94260.97250.97250.93380.95160.95160.95160.90130.95160.9516
140.74390.74390.74390.74390.92920.74390.74390.9730.9730.9730.9730.99390.9730.9730.95110.95110.95110.95110.98870.95110.9511
150.89660.94440.94440.94440.55510.94440.94440.99080.99530.99530.99530.93930.99530.99530.98310.99130.99130.99130.89320.99130.9913
160.77840.83870.83870.83870.70840.83870.83870.97420.98240.98240.98240.96310.98240.98240.95380.96830.96830.96830.93450.96830.9683
170.70170.70170.70170.70170.70170.8740.70170.96720.96720.96720.96720.96720.98860.96720.94090.94090.94090.94090.94090.97910.9409
180.89720.89720.89720.89720.74380.89720.89720.98850.98850.98850.98850.96630.98850.98850.97940.97940.97940.97940.94050.97940.9794
190.81610.81610.81610.81610.50840.81610.81610.97580.97580.95780.97580.90370.97580.97580.95720.95720.95720.95720.8390.95720.9572
200.96460.74770.96460.74770.74640.96460.74770.99630.99670.99630.96670.96650.99630.96670.99330.94120.99330.94120.94080.99330.9412
Mean0.823230.811930.816240.811930.76760.831390.821850.9792550.9793850.9778150.9778850.9687350.980430.979320.962930.960490.96190.960490.945190.965050.96301
Table 3:

The performance measures of the different images in the DRIVE database in terms of precision.

ImagePrecision
ABCPSODETLBOGWOFAHS
10.73540.73540.73540.73540.82570.73540.7354
20.72410.68260.72410.68260.72410.68260.6826
30.68420.75290.75290.75290.86420.75290.7529
40.58380.58380.58380.58380.81770.58380.5838
50.74650.74650.74650.74650.74650.74650.7465
60.72690.72690.72690.72690.66440.72690.7269
70.70660.75390.75390.75390.79880.75390.7539
80.82650.82650.82650.82650.64910.82650.8265
90.74790.74790.74790.74790.6180.74790.6841
100.71210.71210.81780.71210.71210.71210.7121
110.64140.64140.64140.64140.75920.64140.6414
120.82710.82710.82710.82710.50450.82710.774
130.5820.63350.63350.63350.44200.63350.6335
140.82750.82750.82750.82750.73320.82750.8275
150.72830.6830.6830.6830.81590.6830.683
160.66180.70890.70890.70890.84960.70890.7089
170.81250.81250.81250.81250.81250.76130.8125
180.71060.71060.71060.71060.81830.71060.7106
190.68250.68250.68250.68250.88520.68250.6825
200.72130.78270.72130.78270.58560.72130.7827
Mean0.719450.728910.73320.728910.731330.723280.723065
Table 4:

The performance measures of the different images in the STARE database in terms of sensitivity, specificity and accuracy.

ImageSensitivitySpecificityAccuracy
ABCPSODETLBOGWOFAHSABCPSODETLBOGWOFAHSABCPSODETLBOGWOFAHS
10.72050.72050.72050.72050.96110.74980.74980.97020.97020.97020.97020.99680.97420.97420.94610.94610.94610.94610.99410.95330.9533
20.68760.68760.68760.68760.84610.68760.68760.97010.97010.97010.97010.98780.97010.97010.94540.94540.94540.94540.97740.94540.9454
30.64230.64230.64230.64230.97610.64230.64230.95680.95680.95680.95680.9980.95680.95680.92290.92290.92290.92290.99630.92290.9229
40.35130.66340.35130.35130.66340.35130.35130.85080.9540.85080.85080.9540.85080.85080.75750.91910.75750.75750.91910.75750.7575
50.84070.84070.84070.84070.75720.84070.84070.98250.98250.98250.98250.97080.98250.98250.96850.96850.96850.96850.94790.96850.9685
60.82110.82110.87890.82110.71040.82210.82210.9810.9810.98790.9810.96490.9810.9810.96560.96560.97790.96560.93740.96560.9656
70.93240.99450.99450.99450.79990.99450.93240.99370.99950.99950.99950.97850.99950.99370.98840.99910.99910.99910.96120.99910.9884
80.89550.89550.89550.89550.79470.89550.89550.99060.99060.99060.99060.97950.99060.99060.98280.98280.98280.98280.96270.98280.9828
90.99620.92810.92810.92810.92810.92810.92810.99970.99320.99320.99320.99320.99320.99320.99940.98760.98760.98760.98760.98760.9876
100.81430.81430.81430.81430.9250.81430.81430.97960.97960.97960.97960.99270.97960.97960.96330.96330.96330.96330.98660.96330.9633
110.93770.93770.93770.93770.76540.93770.93770.9950.9950.9950.9950.97720.9950.9950.99070.99070.99070.99070.95840.99070.9907
120.81630.81630.78620.81630.81630.81630.81630.97850.97850.97420.97850.97850.97850.97850.96150.96150.95390.96150.96150.96150.9615
130.88130.88130.88130.88130.70340.88130.88130.98830.98830.98830.98830.96430.98830.98830.97870.97870.97870.97870.93620.97870.9787
140.98220.98220.89180.98220.71020.98220.89180.99860.99860.99050.99860.96870.99860.99050.99730.99730.98250.99730.94340.99730.9825
150.79680.79680.79680.79680.93930.79680.79680.97940.97940.97940.97940.99470.97940.97940.96250.96250.96250.96250.99020.96250.9625
160.90950.99050.90950.90950.90950.90950.90950.99130.99130.99130.99130.99130.99130.99130.98410.98410.98410.98410.98410.98410.9841
170.96120.94490.94490.94490.96120.94490.94490.99690.99550.99550.99550.99690.99550.99160.99420.99160.99160.99160.99420.99160.9916
180.75930.75930.75930.75930.75930.95470.75930.97480.97480.97480.97480.97480.99610.97480.95440.95440.95440.95440.95440.99290.9544
190.90550.90950.90950.90950.90550.90550.90550.98960.98960.98960.98960.98960.98960.98960.98130.98130.98130.98130.98130.98130.9813
200.88170.88170.76210.88170.64380.90140.90140.98690.98690.97010.98690.94820.98930.98930.97640.97640.94680.97640.90950.98070.9807
Mean0.826670.845410.816640.8257550.8237950.8378250.820430.9777150.982770.9764950.977610.980020.9789950.977040.961050.9689450.958880.9608650.9641750.9633650.960165
Table 5:

The performance measures of the different images in the STARE database in terms of precision.

ImagePrecision
ABCPSODETLBOGWOFAHS
10.68870.68870.68870.68870.65240.68860.6887
20.76520.76520.76520.76520.70350.76510.7652
30.70870.70870.70870.70870.70870.70870.7087
40.76840.62340.76840.76840.76840.76840.7684
50.78340.78340.78340.78340.71690.78330.7834
60.67970.67970.61770.67970.67970.67960.6797
70.67710.70370.70370.70370.67710.70370.6771
80.7430.7430.7430.7430.78370.7430.743
90.730.70140.70140.70140.730.70140.7014
100.69550.69550.69550.69550.58770.69550.6955
110.69050.69050.69050.69050.66620.69050.695
120.71220.71220.68210.71120.71220.71220.7112
130.67390.67390.67390.67390.67390.67390.6739
140.69270.72930.72930.69270.820.69270.7293
150.73620.73620.73620.73620.68910.73620.7362
160.71340.71340.71340.71340.71340.71340.7134
170.72240.68550.68550.68550.72240.68550.6855
180.64970.64970.64970.64970.64970.63130.6497
190.66410.66410.66410.66410.66410.66410.6641
200.70420.70420.72480.70420.70420.69040.6904
Mean0.709950.7025850.706260.7079550.7011650.7063750.70799

In order to compare the performances of the algorithms, the mean values obtained are combined in Table 6 and Table 7. When the mean Se, Sp and Acc values obtained for the DRIVE database are compared it is seen that all the algorithms produce similar results but FA performs a bit better classification. However, the performance of DE algorithm for DRIVE database in terms of precision seems only a bit better than the other algorithms. In the retinal images taken from the STARE database, the PSO algorithm showed the best results in terms of Se, Sp and Acc. It is also seen that the algorithm with the best precision performance for the STARE database is the HS algorithm. In general, it can be said that the clustering performances of the algorithms are successful and too close to each other.

Table 6:

The statistical performances of the ABC, PSO, DE, TLBO, GWO, FA and HS algorithms for the retinal images taken from DRIVE database.

MethodSensitivitySpecificityAccuracyPrecision
ABC0.8232370.9792550.962930.71945
PSO0.811930.9793850.960490.72891
DE0.816240.9778150.96190.7332
TLBO0.811930.9778850.960490.72891
GWO0.76760.9687350.945190.73133
FA0.831390.980430.965050.72328
HS0.821850.979320.963010.723065
Table 7:

The statistical performances of the ABC, PSO, DE, TLBO, GWO, FA and HS algorithms for the retinal images taken from STARE database.

MethodSensitivitySpecificityAccuracyPrecision
ABC0.826670.9777150.961050.70995
PSO0.845410.982770.9689450.702585
DE0.816640.9764950.958880.70626
TLBO0.8257550.977610.9608650.707955
GWO0.8237950.980020.9641750.701165
FA0.8378250.9789950.9633650.706375
HS0.820430.977040.9601650.70799

A detailed statistical performance comparison has been realized between the heuristic approaches analyzed in this work and the studies published in literature. Table 8 and Table 9 contain the results obtained for DRIVE and STARE databases, respectively. As seen from Table 8, the sensitivity of the heuristic algorithms is close to [33], but higher than the other approaches. On the other hand, for the STARE database, the sensitivity values produced by the heuristic algorithms are similar to [33] and [49], but higher than the other approaches. Namely, the ABC, PSO, DE, TLBO, GWO, FA and HS algorithms considerably improve the sensitivity of the clustering process for both DRIVE and STARE databases. Furthermore, the clustering performance of the heuristic algorithms in terms of the specificity seems similar with the methods published in literature for both DRIVE and STARE databases. Finally, the higher accuracy values produced by the heuristic algorithms represent that the pixel classification performance of the heuristic methods is higher than or similar to the other methods in literature.

Table 8:

Performance comparison of ABC, PSO, DE, TLBO, GWO, FA and HS algorithms and other methods for DRIVE database.

MethodsSensitivitySpecificityAccuracy
You et al. [31]0.74100.97510.9434
Roychowdhury et al. [32]0.7250.9830.952
Wang et al. [33]0.81730.97330.9767
Mendonca et al. [34]0.73440.97640.9452
Azzopardi et al. [35]0.7650.9700.944
Fraz et al. [36]0.74060.98070.9480
Odstrcilik et al. [37]0.7060.96930.934
Marin et al. [38]0.70680.98010.9452
Kaba et al.[39]0.74660.9680.941
Dash et al. [40]0.7190.9760.955
Argüello et al. [41]0.72090.97580.9431
Asad et al. [42]0.73880.92880.9028
Zhao et al. [43]0.7350.9780.947
Zhang et al. [44]0.71200.97240.9382
Imani et al. [45]0.75240.97530.9523
Heuristic algorithmsABC0.8232370.9792550.96293
PSO0.811930.9793850.96049
DE0.816240.9778150.9619
TLBO0.811930.9778850.96049
GWO0.76760.9687350.94519
FA0.8378250.980430.96505
HS0.821850.979320.96301
Table 9:

Performance comparison of ABC, PSO, DE, TLBO, GWO, FA and HS algorithms and other methods for STARE database.

MethodsSensitivitySpecificityAccuracy
You et al. [31]0.7260.9760.950
Roychowdhury et al. [32]0.7720.9730.952
Wang et al. [33]0.81040.97910.9813
Odstrcilik et al. [37]0.78470.95120.9341
Kaba et al. [39]0.76190.9670.9456
Argüello et al. [41]0.73050.96880.9448
Zhao et al. [43]0.7190.9770.951
Zhang et al. [44]0.7180.9750.948
Imani et al. [45]0.75020.97450.959
Mendonca et al. [46]0.7180.9730.946
Xiao et al. [47]0.7150.9740.948
Fraz et al. [48]0.7550.9760.953
Yin et al.[49]0.8540.9420.933
Lázáret al. [50]0.7250.9750.949
Strisciuglio et al. [51]0.8010.9720.954
Heuristic algorithmsABC0.826670.9777150.96105
PSO0.845410.982770.968945
DE0.816640.97644950.95888
TLBO0.8257550.977610.960865
GWO0.8237950.980020.964175
FA0.8378250.9789950.963365
HS0.820430.977040.960165

Figure 7 represents the convergence speeds of the algorithms for the retinal images given in Figure 1 (b) and Figure 2 (a). For a fair and detailed comparison, the convergence speeds have been given for both an abnormal retinal image from DRIVE database and a normal retinal image from STARE database. The evolution of best solutions has been obtained based on MSE. As seen from Figure 7 (a) which shows the convergence speeds obtained for the normal retinal image taken from DRIVE database, each algorithm needs approximately 10 cycles to converge to the optimal solution and the MSE performance of the GWO algorithm seems a bit better than the other algorithms. On the other hand, Figure 7 (b) shows the convergence speeds of the algorithms for the abnormal retinal image taken from STARE database. It is seen that each algorithm reaches to the global solution at 15 cycles. Also, the MSE performance of GWO seems a bit better than the other algorithms. According to the simulation results it is clear that convergence speeds and MSE performances of the algorithms in terms of clustering is close to each other.

Figure 7: Convergence speed of the algorithms.(a) Convergence speeds of the algorithms for the retinal image taken from DRIVE database given in Figure 1b. (b) Convergence speeds of the algorithms for the retinal image taken from STARE database given in Figure 2a.
Figure 7:

Convergence speed of the algorithms.

(a) Convergence speeds of the algorithms for the retinal image taken from DRIVE database given in Figure 1b. (b) Convergence speeds of the algorithms for the retinal image taken from STARE database given in Figure 2a.

Another important performance criterion for heuristic algorithms is the standard deviation which determines the stability of the algorithm. The low standard deviation indicates that the algorithm approximates similar error values at each random run. Table 10 and Table 11 contain the standard deviations obtained after 20 random runs for the retinal images taken from DRIVE and STARE databases. In addition to the standard deviation, the minimum MSE values reached by the algorithms have been shown in the tables. When the minimum MSE values reached by the algorithms are examined it is seen that all the algorithms produce similar MSE values, in other words, their clustering performance are very close to each other. On the other hand, due to their lower standard deviation values for the image taken from DRIVE database PSO and HS algorithms seem a little more stable in terms of clustering when compared to other algorithms. Similarly PSO, TLBO and HS algorithms are said to be a little more stable than the other algorithms for the image taken from STARE database.

Table 10:

Performance comparison of the heuristic algorithms for DRIVE database.

ABCMinimum MSE0.7722
Standard deviation7.18762e-05
PSOMinimum MSE0.752
Standard Deviation2.58125e-08
DEMinimum MSE0.8212
Standard Deviation2.15721e-05
TLBOMinimum MSE0.7689
Standard Deviation5.46974e-05
GWOMinimum MSE0.7253
Standard Deviation5.45881e-06
FAMinimum MSE0.8334
Standard Deviation4.93834e-05
HSMinimum MSE0.8123
Standard Deviation1.94538e-08
Table 11:

Performance comparison of the heuristic algorithms for STARE database.

ABCMinimum MSE0.7578
Standard deviation6.76833e-05
PSOMinimum MSE0.7337
Standard Deviation3.06854e-08
DEMinimum MSE0.7732
Standard Deviation3.41578e-05
TLBOMinimum MSE0.74
Standard Deviation2.9575e-08
GWOMinimum MSE0.719
Standard Deviation9.51344e-06
FAMinimum MSE0.7823
Standard Deviation4.87728e-05
HSMinimum MSE0.764
Standard Deviation5.53854e-07

Discussion

In retinal image analysis it is important to improve algorithms that are flexible and capable of accurate vessel segmentation. The results obtained in this work represent that the clustering based heuristic algorithms provide satisfactory performances in retinal vessel segmentation.

In this work, ABC, PSO, DE, TLBO, GWO, FA and HS algorithms which are most commonly used algorithms in literature for engineering problems are applied for the aim of retinal vessel segmentation. The analyses have been realized for both the normal and abnormal retinal images taken from the DRIVE and STARE databases. From the simulation results it is observed that all the algorithms successfully performed the retinal vessel segmentation. They produced similar MSE values and were able to converge to the global solutions. The standard deviation values reached by the algorithms have been representing that each algorithm was performing a stable behavior. On the other hand, statistical results obtained for sensitivity, specificity, accuracy and precision have been representing that the pixels with value very close to each other can successfully be distinguished with ABC, PSO, DE, TLBO, GWO, FA and HS algorithms.

The conventional gradient based algorithms may usually stuck into the local minimum and also their high dependence to the parameter values is an another disadvantage of these algorithms. Also, the nonflexible structures have been decreasing their optimization performance. Due to these disadvantages of conventional gradient based algorithms, more effective methods that can be used in image analysis have been needed. In this work, it has been shown that heuristic approaches can effectively be used in image analysis.

Conclusion

In this work, novel approaches based on ABC, PSO, DE, TLBO, GWO, FA and HS algorithms for clustering based retinal vessel segmentation is described in the fundus fluorescein angiography retinal images. It is seen from the simulation results that each algorithm converge to the global solutions at similar cycles and the final MSE error values reached by the algorithms are very close to each other. The statistical analyses based on the sensitivity, specificity, accuracy and precision show that ABC, PSO, DE, TLBO, GWO, FA and HS algorithms can successfully be used in analyses of retinal images. On the other hand, since the standard deviation values of the PSO, TLBO and HS algorithms are similar but a bit lower than ABC, DE, GWO and FA algorithms and it can be concluded that PSO, TLBO and HS algorithms are a little bit more stable in terms of clustering based retinal vessel segmentation. As a result, simulation results show that the performance of ABC, PSO, DE, TLBO, GWO, FA and HS algorithms in terms of clustering is too similar and they can successfully be used for the aim of retinal vessel segmentation.


Corresponding author: Mehmet Bahadır Çetinkaya, Department of Mechatronics Engineering, Faculty of Engineering, University of Erciyes, 38039 Melikgazi, Kayseri, Turkey, E-mail:

  1. Research funding: No financial support was received from any person or organization in this work.

  2. Author contributions: All authors have accepted responsibility for the entire content of this manuscript and approved its submission.

  3. Competing interests: The authors do not have financial and personal relationships with other people or organizations that could inappropriately influence their work.

References

1. Fong, DS, Aiello, L, Gardner, TW. Retinopathy in diabetes. Diabetes Care 2004;27:84–7. https://doi.org/10.2337/diacare.27.2007.s84.Search in Google Scholar

2. Fraz, MM, Remagnino, P, Hoppe, A, Uyyanonvara, B, Rudnicka, AR, Owen, CG, et al.. Blood vessel segmentation methodologies in retinal images-a survey. Comput Methods Progr Biomed 2012;108:407–33. https://doi.org/10.1016/j.cmpb.2012.03.009.Search in Google Scholar

3. Wang, JJ, Taylor, B, Wong, TY, Chua, B, Rochtchina, E, Klein, R, et al.. Retinal vessel diameters and obesity: a population-based study in older persons. Obesity 2006;14:206–14. https://doi.org/10.1038/oby.2006.27.Search in Google Scholar

4. Foracchia, M, Grisan, E, Ruggeri, A. Extraction and quantitative description of vessel features in hypertensive retinopathy fundus images. In: 2th Int workshop on comput assisted fundus image anal (CAFIA-2). Copenhagen, Denmark; 2001:5–7 pp.Search in Google Scholar

5. Mitchell, P, Leung, H, Wang, JJ, Rochtchina, E, Lee, AJ, Wong, TY, et al.. Retinal vessel diameter and open-angle glaucoma: the blue mountains eye study. Ophthalmology 2005;112:245–50. https://doi.org/10.1016/j.ophtha.2004.08.015.Search in Google Scholar

6. Goatman, K, Charnley, A, Webster, L, Nussey, S. Assessment of automated disease detection in diabetic retinopathy screening using two-field photography. PloS One 2011;6:275–84. https://doi.org/10.1371/journal.pone.0027524.Search in Google Scholar

7. Asad, AH, Elamry, E, Hassanien, AE, Tolba, MF. New global update mechanism of ant colony system for retinal vessel segmentation. In: 13th IEEE Int conf hybrid intell syst (HIS13) Tunisia; 2013:222–8 pp.10.1109/HIS.2013.6920486Search in Google Scholar

8. Verma, K, Deep, P, Ramakrishnan, AG. Detection and classification of diabetic retinopathy using retinal images flight. In: IEEE annu India conf (INDICON). Hyderabad, India; 2011:1–6 pp.10.1109/INDCON.2011.6139346Search in Google Scholar

9. Heneghan, C, Flynn, J, OKeefe, M, Cahill, M. Characterization of changes in blood vessel width and tortuosity in retinopathy of prematurity using image analysis. Med Image Anal 2002;6:407–29. https://doi.org/10.1016/s1361-8415(02)00058-0.Search in Google Scholar

10. Lowell, J, Hunter, A, Steel, D, Basu, A, Ryder, R, Kennedy, RL. Measurement of retinal vessel widths from fundus images based on 2D modeling. IEEE Trans Med Imag 2004;23:1196–204. https://doi.org/10.1109/tmi.2004.830524.Search in Google Scholar

11. Haddouche, A, Adel, M, Rasigni, M, Conrath, J, Bourennane, S. Detection of the foveal avascular zone on retinal angiograms using Markov random fields. Digit Signal Process 2010;20:149–54. https://doi.org/10.1016/j.dsp.2009.06.005.Search in Google Scholar

12. Grisan, E, Ruggeri, A. A divide etimpera strategy for automatic classification of retinal vessels into arteries and veins. Engineering in medicine and biology society. In: 25th IEEE annu Int conf. Cancun, Mexico; 2003:890–93 pp.Search in Google Scholar

13. Kanski, JJ. Clinical ophthalmology, 6th ed. London: Elsevier Health Sciences; 2007.Search in Google Scholar

14. Archer, DB. Diabetic retinopathy: some cellular. molecular and therapeutic considerations. Eye 1999;13:497–523. https://doi.org/10.1038/eye.1999.130.Search in Google Scholar

15. Antal, B, Hajdu, A. An ensemble-based system for microaneurysm detection and diabetic retinopathy grading. IEEE Trans Biomed Eng 2012;59:1720–6. https://doi.org/10.1109/tbme.2012.2193126.Search in Google Scholar

16. Abramoff, MD, Folk, JC, Han, DP. Automated analysis of retinal images for detection of referable diabetic retinopathy. JAMA Ophthalmol 2013;131:351–7. https://doi.org/10.1001/jamaophthalmol.2013.1743.Search in Google Scholar

17. Foracchia, M, Grisan, E, Ruggeri, A. Detection of optic disc in retinal images by means of a geometrical model of vessel structure. IEEE Trans Med Imag 2004;23:1189–95. https://doi.org/10.1109/tmi.2004.829331.Search in Google Scholar

18. Engelbrecht, AP. Fundamentals of computational swarm intelligence. Chichester, UK: John Wiley & Sons Publication; 2005.Search in Google Scholar

19. Karaboga, D. An idea based on honey bee swarm for numerical optimization. Technical Report-TR06. Kayseri, Turkey: Erciyes University, Engineering Faculty, Computer Engineering Department; 2005.Search in Google Scholar

20. Eberhart, RC, Kennedy, J. A new optimizer using particle swarm theory. In: 6th IEEE Int symp micromachine hum sci. Nagoya, Japan; 1995:39–43 pp.10.1109/MHS.1995.494215Search in Google Scholar

21. Storn, R, Price, K. Differential evolution- a simple and efficient adaptive scheme for global optimization over continuous spaces. Technical Report TR-95-012. Berkeley, California USA; 1995.Search in Google Scholar

22. Rao, RV, Patel, V. An elitist teaching-learning-based optimization algorithm for solving complex constrained optimization problems. Int J Ind Syst Eng Comput 2012;3:535–60. https://doi.org/10.5267/j.ijiec.2012.03.007.Search in Google Scholar

23. Mirjalili, SM, Lewis, A. Grey wolf optimizer. Adv Eng Software 2014;69:46–61. https://doi.org/10.1016/j.advengsoft.2013.12.007.Search in Google Scholar

24. Song, X, Tang, L, Zhao, S, Zhang, X, Li, L, Huang, J, et al.. Grey wolf optimizer for parameter estimation in surface waves. Soil Dynam Earthq Eng 2015;75:147–57. https://doi.org/10.1016/j.soildyn.2015.04.004.Search in Google Scholar

25. Mirjalili, S. How effective is the grey wolf optimizer in training multi-layer perceptrons. Appl Intell 2015;43:150–61. https://doi.org/10.1007/s10489-014-0645-7.Search in Google Scholar

26. Yang, XS. Firefly algorithms for multimodal optimization in stochastic algorithms: foundations and applications. Lect Notes Eng Comput Sci 2009;5792:169–78. https://doi.org/10.1007/978-3-642-04944-6_14.Search in Google Scholar

27. Lee, KS, Geem, ZW. A new meta-heuristic algorithm for continuous engineering optimization: harmony search theory and practice. Comput Methods Appl Mech Eng 2005;194:3902–33. https://doi.org/10.1016/j.cma.2004.09.007.Search in Google Scholar

28. Mahdavi, M, Fesanghary, M, Damangir, E. An improved harmony search algorithm for solving optimization problems. Appl Math Comput 2007;188:1567–79. https://doi.org/10.1016/j.amc.2006.11.033.Search in Google Scholar

29. Hoover, A, Kouznetsova, V, Goldbaum, M. Locating blood vessels in retinal images by piece wise threshold probing of a matched filter response. IEEE Trans Med Imag 2000;19:203–10. https://doi.org/10.1109/42.845178.Search in Google Scholar

30. Alonso-Montes, C, Vilariño, DL, Dudek, P, Penedo, MG. Fast retinal vessel tree extraction: a pixel parallel approach. Int J Circ Theor Appl 2008;36:641–51. https://doi.org/10.1002/cta.512.Search in Google Scholar

31. You, X, Peng, Q, Yaun, Y, Cheng, Y, Lei, J. Segmentation of retinal blood vessels using the radial projection and semi-supervised approach. Pattern Recogn 2011;44:2314–24. https://doi.org/10.1016/j.patcog.2011.01.007.Search in Google Scholar

32. Roychowdhury, S, Koozekanani, DD, Parhi, KK. Blood vessel segmentation of fundus images by major vessel extraction and subimage classification. IEEE J Biomed Health Inf 2015;19:1118–28.10.1109/JBHI.2014.2335617Search in Google Scholar

33. Wang, S, Yin, Y, Cao, G, Wei, B, Zheng, Y, Yang, G. Hierarchical retinal blood vessel segmentation based on feature and ensemble learning. Neurocomputing 2015;149:708–17. https://doi.org/10.1016/j.neucom.2014.07.059.Search in Google Scholar

34. Mendonca, AM, Campilho, A. Segmentation of retinal blood vessels by combining the detection of centerlines and morphological reconstruction. IEEE Trans Med Imag 2006;25:1200–13. https://doi.org/10.1109/tmi.2006.879955.Search in Google Scholar

35. Azzopardi, G, Strisciuglio, N, Vento, M, Petkov, N. Trainable COSFIRE filters for vessel delineation with application to retinal images. Med Image Anal 2015;19:46–57. https://doi.org/10.1016/j.media.2014.08.002.Search in Google Scholar

36. Fraz, MM, Remagnino, P, Hoppe, A, Uyyanonvara, B, Rudnicka, AR, Owen, CG, et al.. An ensemble classification-based approach applied to retinal blood vessel segmentation. IEEE Trans Biomed Eng 2012;59:2538–48. https://doi.org/10.1109/tbme.2012.2205687.Search in Google Scholar

37. Odstrcilik, J, Kolar, R, Budai, A, Hornegger, J, Jan, J, Gazarek, J, et al.. Retinal vessel segmentation by improved matched filtering: evaluation on a new high-resolution fundus image database. IET Image Process 2013;7:373–83. https://doi.org/10.1049/iet-ipr.2012.0455.Search in Google Scholar

38. Marín, D, Aquino, A, Gegúndez-Arias, ME, Bravo, JM. A new supervised method for blood vessel segmentation in retinal images by using gray-level and moment invariants-based features. IEEE Trans Med Imag 2011;30:146–58. https://doi.org/10.1109/tmi.2010.2064333.Search in Google Scholar

39. Kaba, D, Wang, C, Li, Y, Salazar-Gonzalez, A, Liu, X, Serag, A. Retinal blood vessels extraction using probabilistic modelling. Health Inf Sci Syst 2014;27:2.10.1186/2047-2501-2-2Search in Google Scholar PubMed PubMed Central

40. Dash, J, Bhoi, N. A thresholding based technique to extract retinal blood vessels from fundus images. Future Comp Inf J 2017;2:103–9. https://doi.org/10.1016/j.fcij.2017.10.001.Search in Google Scholar

41. Argüello, F, Vilariño, DL. Heras, DB. Nieto, A. GPU-based segmentation of retinal blood vessels. J Real-Time Image Process 2014;14:1–10. https://doi.org/10.1007/s11554-014-0469-z.Search in Google Scholar

42. Asad, AH, Azar, AT, Hassanien, AE. Ant colony-based system for retinal blood vessels segmentation. In: 7th Int conf bio-inspired comput theor and appl (BIC-TA 2012). Gwalior, India; 2012:441–52 pp.10.1007/978-81-322-1038-2_37Search in Google Scholar

43. Zhao, YQ, Wang, XH, Wang, XF, Shih, FY. Retinal vessels segmentation based on level set and region growing. Pattern Recogn 2014;47:2437–46.10.1016/j.patcog.2014.01.006Search in Google Scholar

44. Zhang, B, Zhang, L, Karray, F. Retinal vessel extraction by matched filter with first-order derivative of Gaussian. Comput Biol Med 2010;40:438–45. https://doi.org/10.1016/j.compbiomed.2010.02.008.Search in Google Scholar

45. Imani, E, Javidi, M, Pourreza, HR. Improvement of retinal blood vessel detection using morphological component analysis. Comput Methods Progr Biomed 2015;118:263–79. https://doi.org/10.1016/j.cmpb.2015.01.004.Search in Google Scholar

46. Mendonca, A, Dashtbozorg, B, Campilho, A. Segmentation of the vascular network of the retina. Image Anal Model Ophthalmol 2014:85–109. https://doi.org/10.1201/b16510-6.Search in Google Scholar

47. Xiao, Z, Adel, M, Bourennane, S. Bayesian method with spatial constraint for retinal vessel segmentation. Comput Math Methods Med 2013. https://doi.org/10.1109/icassp.2013.6637899.Search in Google Scholar

48. Fraz, MM, Remagnino, P, Hoppe, A, Uyyanonvara, B, Rudnicka, AR, Owen, CG, et al.. An ensemble classification-based approach applied to retinal blood vessel segmentation. IEEE Trans Biomed Eng 2012;59:2538–48. https://doi.org/10.1109/tbme.2012.2205687.Search in Google Scholar

49. Yin, B, Li, H, Sheng, B, Hou, X, Chen, Y, Wu, W, et al.. Vessel extraction from non-fluorescein fundus images using orientation-aware detector. Med Image Anal 2015;26:232–42. https://doi.org/10.1016/j.media.2015.09.002.Search in Google Scholar

50. Lázár, I, Hajdu, A. Segmentation of retinal vessels by means of directional response vector similarity and region growing. Comput Biol Med 2015;66:209–21. https://doi.org/10.1016/j.compbiomed.2015.09.008.Search in Google Scholar

51. Strisciuglio, N, Azzopardi, G, Vento, M, Petkov, N. Multiscale blood vessel delineation using b-cosfire filters. In: Springer Int conf comput anal images and patterns. Valletta, Malta; 2015:300–12 pp.10.1007/978-3-319-23117-4_26Search in Google Scholar

Received: 2020-04-03
Accepted: 2020-09-28
Published Online: 2020-11-26
Published in Print: 2021-04-27

© 2020 Mehmet Bahadır Çetinkaya and Hakan Duran, published by De Gruyter, Berlin/Boston

This work is licensed under the Creative Commons Attribution 4.0 International License.

Downloaded on 19.9.2025 from https://www.degruyterbrill.com/document/doi/10.1515/bmt-2020-0089/html
Scroll to top button