A New Approach for Biomedical Image Segmentation: Combined Complex-Valued Artificial Neural

A New Approach for Biomedical Image Segmentation: Combined Complex-Valued Artificial Neural



A New Approach for Biomedical Image Segmentation: Combined Complex-Valued Artificial Neural Network Case Study: Lung Segmentation on Chest CT Images

Murat Ceylan,Yüksel Özbay, Erkan Yıldırım

Abstract— The principal goal of the segmentation process is to partition an image into classes or subsets that are homogeneous with respect to one or more characteristics or features. In medical imaging, segmentation is important for feature extraction, image measurements, and image display. This study presents a new version of complex-valued artificial neural networks (CVANN) for the biomedical image segmentation. Proposed new method is called as combined complex-valued artificial neural network (CCVANN) which is a combination of two complex-valued artificial neural networks. To check the validation of proposed method, lung segmentation is realized. For this purpose, we used 32 chest CT images of 6 female and 26 male patients. These images were recorded from Baskent University Radiology Department in Turkey.The accuracy of the CCVANN model is more satisfactory as compared to the single CVANN model.

I.INTRODUCTION

S

egmentation that is one of the most difficult tasks in the image processing subdivides an image into its constituent characteristics or features. Accuracy of segmentation determines the success of image analysis process. In other words, error in segmentation level of complete computerized analysis procedures will impact all higher level activities. In many medical image processing applications, segmentation may be useful to detection of organs such as brain, heart, lungs or liver in computerized tomography (CT) or magnetic resonance (MR) images, while in others into distinguish pathological regions, such as tumor, normal tissue, tissue deformities. A wide variety of segmentation techniques has been proposed in the literature [1-2]. However, there is no one standard segmentation technique that can produce satisfactory results for all imaging applications. Compared with the classical methods in the literature, the ANN approach has the advantage of paralel processing (with appropriate hardware), robustness, noise tolerance, and adaptability. Neural networks provide pixel classification paradigm that can be used for image segmentation [3-4]. Neural network based segmentation approaches may provide good results for medical images with considerable variance in structures of interest [3]. Sha and Sutton [5] proposed a neural network system for segmentation and classificiation of digital brain images. The other application of neural network for automatic segmentation was done by Nattkemper et al. [6]. Papadopoulos et al. [7], was used a hybrid neural network, which consist of two components: a rule-based and a neural network. In [8], ultrasound images were segmented using hybrid neural network. Vilarino et al. [9], were applied cellular neural network to image segmentation based on active contour techniques. Middleton et al. [10], used combination of neural networks and active contour models for segmentation of MR images. In [11], region-based segmentation and neural network edge detection were used. For MR, CT and ultrasound images segmentation, a new method which called as incremental neural network was developed by Dokur [12] and Kurnaz et al. [13 ]. Wismuller et al. [14] and Ong et al. [15] were used self-organized model for the segmentation.

A complex-valued artificial neural network (CVANN) is neural network which consist of complex-valued inputs, weights, thresholds, activation functions and outputs. CVANN have been widening the scope of applications not only in signal processing [16] but also in image processing [17]. For complex signal and image processing problems, many existing neural networks cannot directly be applied. Although for certain applications it is possible to reformulate a complex signal processing problem so that a real-valued network and learning algorithm can be used to solve the problem, it is not always feasible to do so.

In this paper, a novel cascade structure is proposed, called as combined complex-valued artificial neural networks (CCVANN). General scheme of CCVANN is implemented in two levels. In the first level, learning of CVANN is realized using the original data set. After that, prediction of the first level and target of the original data are presented to the second level CVANN as inputs. In this study, segmentation of lung images isrealizedusing CCVANN.

II.MATERIAL and METHODS

A.Image Data

For the development and evaluation of the proposed system we used our image collection (32 chest CT images of 6 female and 26 male patients). These images were recorded from Baskent University Radiology Department in Turkey [17]. This collection includes 10 images with benign nodules and 22 images with malign nodules. Averaged age of patients is 64. Each CT slice used in this study has dimensions of 752 x 752 pixels with grey level.

B.Complex Wavelet Transform (CWT)

Wavelet techniques are successfully applied to various problems in signal and image processing [18-19]. It is perceived that the wavelet transform is an important tool for analysis and processing of signals and images. In spite of its efficient computational algorithm, the wavelet transform suffers from three main disadvantages. These disadvantages are shift sensitivity, poor directionality and absence of phase information. CWT overcomes these disadvantages [20].

Recent research in the development of CWTs can be broadly classified in two groups; RCWT (Redundant CWTs) and NRCWT (Non-redundant CWTs). The RCWT include two almost similar CWT. They are denoted as DT-DWT (Dual-Tree DWT based CWT, with two almost similar versions namely Kingsbury’sand Selesnick’s[20]. In this paper, we used Kingsbury’s CWT [20] for feature extraction of image to be segmented.

C.Combined Complex-Valued Artificial Neural Network (CCVANN)

In this study, a complex back-propagation (CBP) algorithm has been used for pattern recognition. We will first give the theory of the CBP algorithm as applied to a multi layer CVANN. Figure 1 shows a CVANN model. The input signals, weights, thresholds, and output signals are all complex numbers. The activity Yn of neuron n is defined as: (1)

where Wnm is the complex-valued (CV) weight connecting neuron n and m, Xmis the CV input signal from neuron m, and Vn is the CV threshold value of neuron n.To obtain the CV output signal, the activity value Yn is converted into its real and imaginary parts as follows:

(2)

where i denotes . Although various output functions of each neuron can be considered, the output function used in this study is defined by the following equation:

(3)

where fR(u)=1/(1+exp(-u)) and is called the sigmoid function.

Summary of CBP algorithm:

  1. Initialization

Set all the weights and thresholds to small complex random values.

2. Presentation of input and desired (target) outputs

Present the input vector X(1), X(2),….,X(N) and corresponding desired (target) response T(1), T(2),….T(N), one pair at a time, where N is the total number of training patterns.

3. Calculation of actual outputs

Use the formula in (3) to calculate output signals.

4. Adaptation of weights and thresholds

If this condition is satisfied, algorithm is stopped and weights and biases are frozen.

/ (4)

where Tn(p)and On(p)are complex numbers and denote the desired and output value, respectively. The actual output value of the neuron n for the pattern p, i.e the left side of (4) denotes the error between the desired output pattern and the actual output pattern. N denotes the number of neurons in the output layer [21].

Fig.1. CVANN model

The general CCVANN model which is a combination of two CVANN used in this study is illustrated in Fig.2. The CVANNs were used at the first and second levels for the complex-valued pattern recognition.

Fig.2. CCVANN model

III.Measurement for performance evaluation

In this paper, we used an algorithm to evaluate the classification results of CVANN’s outputs for training and test images that include lung region pixels and surrounding pixels of lung-region [17]. Number of correct classified pixels in segmented image (752x752) was calculated according to the following algorithm:

IF

IT - IO = 0

THEN

Number of correct classified pixel (CCP) increase by 1

ELSE

Number of incorrect classified pixel (iCCP) increase by 1

where, IT and IO are target of network and actual output of network, respectively. Finally, we measured the accuracy of proposed methods using following equation:

Accuracy (%) = (CCP / (CCP + iCCP)) x 100 (5)

IV.results and dıscussıon

In this paper, segmentation of lung region was realized using CWT based CCVANN. Complex wavelet transform was used to reduce the size of input matrix of training and test images. In proposed cascade structure, feature vector of original CT images (size of 752x752) was extracted using CWT with 2nd, 3rd and 4th level. Size of feature vectors was obtained as 188x188, 94x94 and 47x47 for 2nd, 3rd and 4th level, respectively. Obtained complex-valued feature vectors were presented as inputs to CCVANN. Figure 3 shows block representation of proposed method.The complex-valued backpropagation algorithm was used for training of the proposed networks.

Fig 3. Block representation of proposed method for lung segmentation.

The training of CWT-CVANN was stopped when the error goal (4) was achieved. In training phase, half of total 32 images were used. After that, the performance of this network was tested by presentingother 16 test images.Averaged accuracy rate of test is calculated as 99.80 %. Fig.4 shows the segmented images using4th level CWT and CCVANN with best accuracy rate.

Fig. 4. Lung images with benign nodules (a) (image no: B10) and malign nodules (b)(image no: M16) and their results, c and d, respectively.

To compare with lung segmentation using “single” CVANN, learning rate, number of hidden nodes and maximum iteration number of first CVANN was chosen as 0.1, 10 and 10, respectively, similarly [17]. These parameters for second CVANN were determined via experimentation. Network structures are given in Table 1.

TABLE I

NETWORK STRUCTURES

Network / Learning Rate / No. of
Hidden Nodes / No. of Iteration
1th CVANN[17] / 0.1 / 10 / 10
2nd CVANN / 1e-6 / 40 / 10

Results of CCVANN and CVANN [17] withthree levels CWT were presented in Table 2 and Table 3. These results are given for best and worst accuracy rates of CVANN. In Table2, B signify an image with benign nodule and M signify an image with malign nodule.According to Table 2 and Table 3, CCVANN can segment lung region better than CVANN for all of the CWT’s level.

V.conclusıons

In this paper, a combined complex-valued artificial neural network model is proposed for biomedical image segmentation. Following conclusions may be drawn based on the results presented;

  1. The results of the CCVANN compared to experimental results are found to be more satisfactory.The proposed CCVANN methods’ results have a lower number of incorrect classification pixels than using a single CVANN model.
  2. Although the performance of the developed CCVANN model is limited to the range of input data used in the training and testing process, the method can easily be applied with additional new set of data.

TABLE II

COMPARISON OF CVANN AND CCVANN

USING NUMBER OF CCP AND iCCP

CWT
Level / Image No. / Number of CCP / Number of iCCP
CVANN / CCVANN / CVANN / CCVANN
2 / B10 (best) / 564279 / 564477 / 1225 / 1027
B1 (worst) / 563483 / 563609 / 2021 / 1895
M16 (best) / 564150 / 564255 / 1354 / 1249
M19 (worst) / 563483 / 563732 / 2021 / 1772
3 / B10 (best) / 564322 / 564380 / 1182 / 1124
B1 (worst) / 562808 / 562942 / 2696 / 2562
M6 (best) / 564226 / 564460 / 1278 / 1044
M10 (worst) / 563548 / 563907 / 1956 / 1597
4 / B10 (best) / 564169 / 564559 / 1335 / 945
B1 (worst) / 561201 / 561480 / 4303 / 4024
M16 (best) / 564079 / 564463 / 1425 / 1041
M4 (worst) / 563349 / 563872 / 2155 / 1632

TABLE III

COMPARISON OF CVANN AND CCVANN

USING ACCURACY RATE

CWT
Level / Image No. / Accuracy (%)
CVANN / CCVANN
2 / B10 (best) / 99.7834 / 99.8184
B1 (worst) / 99.6426 / 99.6649
M16 (best) / 99.7606 / 99.7791
M19 (worst) / 99.6426 / 99.6867
3 / B10 (best) / 99.7910 / 99.8012
B1 (worst) / 99.5233 / 99.5470
M6 (best) / 99.7740 / 99.8154
M10 (worst) / 99.6541 / 99.7176
4 / B10 (best) / 99.7639 / 99.8329
B1 (worst) / 99.2391 / 99.2884
M16 (best) / 99.7480 / 99.8159
M4 (worst) / 99.6189 / 99.7114

Acknowledgment

This work is supported by the Coordinatorship of Selcuk University's Scientific Research Projects.

References

[1]K. S. Fu, J. K. Mui, “A survey on image segmentation”, Pattern Recognition, vol. 13-1, pp. 3-16 , 1981.

[2]R. M. Haralick , L. G. Shapiro: “Survey: image segmentation techniques”, Comp. Vision Graph Image Proc vol. 29, pp. 100-132, 1985.

[3]A. P.Dhawan,Medical Image Analysis, Wiley-Int., USA, 2003.

[4]B. M. Ozkan, R. J.Dawant,, “Neural-network based segmentation of multi-modal medical images: a comparative and prospective study”, IEEE Trans. Med. Imagingvol. 12, 1993.

[5]D. D.Sha, J. P.Sutton,“Towards automated enhancement, segmentation and classification of digital brain images using networks of networks”, Information Sciences,vol. 138, pp. 45-77, 2001.

[6]T. W.Nattkemper, H.Wersing, W.Schubert, H.Ritter,“ A neural network architecture for automatic segmentation of fluorescence micrographs”, Neurocomputingvol. 48, pp. 357-367, 2002.

[7] A.Papadopoulos, D.I.Fotiadis, A.Likas, “An automatic microcalcification detection system based on a hybrid neural network classifier”, Artificial Int. in Medicinevol. 25, pp. 149-167 2002

[8]Z.Dokur, T.Olmez,“Segmentation of ultrasound images by using a hybrid neural network”, Pattern Recognition Lettersvol. 23, pp. 1825-1836, 2002

[9]D. L.Vilarino, D.Cabello, X. M.Pardo, V. M.Brea,“Cellular neural networks and active contours: a tool for image segmentation”, Image and Vision Computingvol. 21, pp. 189-204, 2003.

[10]I.Middleton, R. I.Damper,“Segmentation of magnetic resonance images using a combination of neural networks and active contour models”, Medical Engineering and Physicsvol. 26, pp. 71-86, 2004.

[11]M. I.Rajab, M. S.Woolfson, S. PMorgan,.”Application of region-based segmentation and neural network edge detection to skin lesions”, Comp. Med. Imaging and Graphicsvol. 28, pp. 61-68, 2004

[12]Z.Dokur, “A unified framework for image compression and segmentation by using an incremental neural network”, Expert Systems with Applicationsvol. 34, pp. 611-619, 2008.

[13]M. N.Kurnaz,, Z.Dokur, T.Olmez,“An incremental neural network for tissue segmentation in ultrasound images”, Computer Methods an Programs in Biomedicinevol. 85, pp. 187-195, 2007.

[14]A.Wismuller, F.Vietze, J.Behrens, A.Meyer-Baese, M.Reiser, H.Ritter,“Fully automated biomedical image segmentation by self-organized model adaptation”, Neural Networksvol. 17, pp.1327-1344, 2004.

[15]S. H.Ong, N.C.Yeo, K. H.Lee, Y. V.Venkatesh, D. M.Cao,“Segmentation of color images using a two-stage self-organizing network”, Image and Visual Computing vol. 20, pp. 279-289, 2002.

[16]Y.Özbay, M.Ceylan,“Effects of window types on classification of carotid artery Doppler signals in the early phase of atherosclerosis using complex-valued artificial neural network”, Computers in Biology and Medicine,vol. 37,pp. 287-295, 2007.

[17]M. Ceylan, Y. Özbay, O. N. Uçan, E. Yıldırım, “A novel method for lung segmentation on chest CT images: Complex-valued artificial neural network with complex wavelet transform”,Turk. J. Elec. and Comp. Sciences, vol. 18 (4), 2010.

[18]G.Beylkin, R.Coifman, V.Rokhlin, “Fast wavelet transforms and numerical algorithms”, Communications on Pure and Applied Mathematics, pp. 141-183, 1991.

[19]M. Unser, “Texture classification and segmentation using wavelet frames”, IEEE Trans. Image Processing, vol.4, pp. 1549-1560, 1995.

[20]I.Selesnick, W. R. G.Baraniuk, N. G.Kingsbury, “The dual-tree complex wavelet transform”, IEEE Signal Processing Magazine, vol.22, pp. 123-151, 2005.

[21]T. Nitta (Edt.), Complex-Valued Neural Networks: Utilizing High-Dimensional Parameters, Inf. Science Reference, Penns., USA, 2009.

Manuscript received July 25, 2010. This work was supported by the Coordinatorship Selcuk University’s Scientific Research Projects.

Murat Ceylan is with the Electrical and Electronics Engineering Department, Selcuk University, Konya, 42075 TURKEY (phone: +90-332-2232042; fax: +90-332-2410635; e-mail: ).

Yüksel Özbay is with Electrical and Electronics Engineering Department, Selcuk University, Konya, 42075 TURKEY (e-mail: ).

.Erkan Yıldırım is with Radiology Department, Baskent University, Ankara, TURKEY, (e-mail:).