About this Journal Submit a Manuscript Table of Contents
BioMed Research International
Volume 2013 (2013), Article ID 175271, 25 pages
Research Article

New Morphological Features for Grading Pancreatic Ductal Adenocarcinomas

Department of Computer & Information Engineering, Inha University, 253 Yonghyun-dong, Nam-gu, Incheon 402-751, Republic of Korea

Received 17 January 2013; Revised 7 April 2013; Accepted 24 April 2013

Academic Editor: Xin-yuan Guan

Copyright © 2013 Jae-Won Song and Ju-Hong Lee. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.


Pathological diagnosis is influenced by subjective factors such as the individual experience and knowledge of doctors. Therefore, it may be interpreted in different ways for the same symptoms. The appearance of digital pathology has created good foundation for objective diagnoses based on quantitative feature analysis. Recently, numerous studies are being done to develop automated diagnosis based on the digital pathology. But there are as of yet no general automated methods for pathological diagnosis due to its specific nature. Therefore, specific methods according to a type of disease and a lesion could be designed. This study proposes quantitative features that are designed to diagnose pancreatic ductal adenocarcinomas. In the diagnosis of pancreatic ductal adenocarcinomas, the region of interest is a duct that consists of lumen and epithelium. Therefore, we first segment the lumen and epithelial nuclei from a tissue image. Then, we extract the specific features to diagnose the pancreatic ductal adenocarcinoma from the segmented objects. The experiment evaluated the classification performance of the SVM learned by the proposed features. The results showed an accuracy of 94.38% in the experiment distinguishing between pancreatic ductal adenocarcinomas and normal tissue and a classification accuracy of 77.03% distinguishing between the stages of pancreatic ductal adenocarcinomas.

1. Introduction

Pathological diagnosis is currently performed subjectively via the knowledge and experience of doctors after inspection of tissue slides through light microscopes. This subjective diagnosis has some problems. First, tumor screening at high magnification by light microscope requires a lot of time and effort [1, 2]. Also, the individual competence of doctors has a decisive effect on the final diagnosis. This means that the pathological diagnosis of the same tumor by two different doctors may vary, because it is not based on objective quantitative feature analysis [35].

The advent of digital pathology has led to a new type of pathological diagnosis. Digital pathology changes glass slides into digital images. It provides a convenient screening environment that also makes it possible to objectively diagnose tumors through quantitative feature analysis using a computer [68]. Nevertheless, a pathological diagnosis needs much time and workload because the diagnosis is manually processed. Therefore, there are still problems related to the subjective diagnosis of tumors. To overcome these problems, studies on computer-aided diagnosis (CAD) techniques based on digital pathology are growing. Many studies of CAD-based digital pathology have covered not only detection of tumors but also grading the stages of tumors. In pathology, tumor stage grading is useful to identify the extent of the disease and determine the appropriate treatment for a patient [9]. In fact, it is known that pathologists receive intensive trainings on grading stages of tumors in order to prescribe the correct type of treatment [7]. Currently, many studies of pathological diagnosis with CAD techniques are being carried out on breast and prostate tumors.

First, most CAD studies related to prostate cancer are based on the Gleason grading system [10]. Tabesh et al. [11] proposed an automated system based on machine learning to diagnose prostate cancer and grade the stages (low and high) of cancer with the Gleason grading system. They extracted color, texture, and morphometric features at global and object levels of a given tissue image. Classifying algorithms such as Gaussian, -nearest neighbor (-NN), and Support Vector Machine (SVM) learned the features of cancer diagnosis and Gleason grading. Naik et al. [12] proposed a diagnostic system for distinguishing between intermediate Gleason grades. They identified the candidate gland region using a Bayesian classifier with low-level information and eliminated false positive regions identified as glands using empirical domain information. After that, the morphologic features were extracted from the identified glands, and Gleason grades 3, 4, and benign were classified through the SVM that learned the features. Huang and Lee [13] classified prostate cancer images into 4 grades based on a Gleason grading system. They used Bayesian, -NN, and SVM classifiers for classifying stages of the cancer. And, to teach the classifiers Gleason grading, features are proposed by using differential box-counting and entropy-based dimension estimation techniques. In addition to that, there are many studies for the grading and diagnosis of prostate tumors [1417].

There are also CAD studies related to breast tumors. Anderson et al. [18] worked on a problem for distinguishing ductal hyperplasia (DH), which is benign, and ductal carcinoma in situ (DCIS), which is malignant. In this study, they automatically segmented breast ducts using knowledge-guided machine vision and proposed measuring duct cribriformity and architectural complexity to quantitatively analyze the duct patterns in proliferative lesions to distinguish between DH and DCIS. Bilgin et al. [19] proposed a method to diagnose breast cancer using graph theory techniques. They segmented given tissue images using a -means algorithm and generated different cell graphs using the positional coordinates of cells for each segmented image. An SVM model that can classify given tissue images into benign, invasive, and non-invasive (ductal carcinoma in situ) was learned by quantitative metrics that are computed from the generated cell graphs. Basavanhally et al. [20] proposed a grading system that identifies and grades the extent of lymphocytic infiltration (LI), a known viable prognostic indicator. First, they detected lymphocytes using region growing and Markov random field algorithms. Then, the architectural features were extracted from the detected lymphocytes, and the extent of LI was classified into low, medium, and high grades by an SVM classifier learned by the features. In addition, there are a lot of studies for the diagnosis and grading of breast tumors [2124].

Many pathological CAD studies make an effort to analyze the pathological characteristics of and design the methods for quantitatively measuring each disease, because there are many methods for pathological diagnosis according to the type of disease and lesion. Currently, in addition to the breast and prostate tumors mentioned above, some studies on colonic [25], bladder [26], neuroblastoma [2729], and follicular lymphoma [30, 31] tumors have been performed. Pathological CAD studies are still in the early stages and focus on a few tumors. There are many methods for diagnosing different types of tumors. Therefore, more studies of pathological CADs must be performed.

The aim of this study is to detect pancreatic ductal adenocarcinoma (PDAC) and classify them by stages. To achieve this, we propose new morphological features for diagnosing and grading PDAC. The region that is inspected to diagnose PDAC is a duct that consists of lumen and epithelium. Therefore, this paper segments the given image into lumen, epithelial nuclei, and nonepithelial nuclei and extracts the morphological features for diagnosing PDAC from the segmented objects. After that, the diagnosis and grading stages of PDAC are performed using the SVM model learned by the extracted features. This paper has several sections. Section 2 describes the pathological characteristics and the morphological features needed in diagnosing PDAC. Section 3 shows the configuration of systems used to diagnose PDAC and the segmentation methods of objects. Section 4 discusses the proposed new morphological features for quantitatively measuring the pathological characteristics of PDAC described in Section 2. In Section 5, the performances between SVM classifiers learnt by the proposed and existing classical morphological features are compared with each other to show the suitability of the proposed features to detect and grade the stages of PDAC. Section 6 evaluates and statistically analyzes the results. Finally, Section 7 presents a conclusion of this study.

2. Pathological Characteristics of PDAC

Pancreatic cancer is the second most common gastrointestinal neoplasm that causes death, after colon cancer [33]. And of all pancreatic neoplasms, PDAC accounts for 85–95%. Approximately 80% of all PDAC patients are between 60 and 80 years of age, and cases in people below the age of 40 are rare. The incidence of PDAC is about 50% higher in men than in women. By race, those of African ancestry have the highest rate of PDAC [34]. The best way of treating PDAC is known as curative resection. However, because of the rare possibility of diagnosing PDAC in the early stage, only 5–22% of PDAC patients can take the curative resection at the time the cancer is discovered [35]. Therefore, an accurate determination of the degree of cancer development is crucial factor for the treatment.

PDAC progression is divided by histological and cytological features and mitotic activity into Grade 1, well-differentiated carcinomas; Grade 2, moderately differentiated carcinomas; and Grade 3, poorly differentiated ductal adenocarcinomas [36, 37]. Grade 1 consists of a duct-like structure combined with medium-sized neoplastic glands. Tubular or cribriform patterns are typical. There may also be small irregular papillary projections without a distinct fibrovascular stalk, particularly in large duct-like structures. Mitotic activity is low. The mucin-producing neoplastic cells tend to be columnar, have eosinophils, and occasionally exhibit pale or even clear cytoplasm. Some neoplastic cell nuclei show loss of polarity. Grade 2 is characterized by a mixture of medium-sized duct-like and tubular structures of variable shapes, embedded in desmoplastic stroma. The duct shape is commonly that of incompletely formed glands. Compared with Grade 1, Grade 2 shows a greater variation in nuclear size, chromatin structure, and prominence of the nucleoli. The cytoplasm is usually slightly eosinophilic, but clear cells are occasionally abundant. Mucin production appears to be decreased, and intraductal in situ components are somewhat less frequent than in Grade 1. Grade 3 is infrequent. It is composed of a mixture of densely packed, small, and irregular glands as well as solid tumor cell sheets and nests that entirely replace the acinar tissue. While typical large, duct-like structures and intraductal tumor components are absent, there may be small squamoid features, spindle cells, or anaplastic foci. The neoplastic cells show marked pleomorphism, little or no mucin production, and brisk mitotic activity. Figure 1 shows Normal, Grade 1, Grade 2, and Grade 3 tissue images.

Figure 1: (a) Normal; (b) Grade 1, well differentiated; (c) Grade 2, moderately differentiated; (d) Grade 3, poorly differentiated.

As described above, Grade 3 is not common. Also, Figure 1 shows that Grade 3 is certainly morphologically different from Grades 1 and 2. Therefore, this paper focuses on the detection of PDAC and the differentiation of Grades 1 and 2.

3. System Overview

In this paper, system configuration to diagnose PDAC consists of three phases as follows: segmentation and feature extraction, model learning and validation, and diagnosis. In the first phase, after preprocessing the given tissue image, the image is segmented into three parts. These three parts are the lumen region, epithelial nuclei, and nonepithelial nuclei. Then, according to the characteristics of each part, the features to be used for the classification model are extracted and stored into a feature database. The second phase is the learning and validation of the SVM classification model using the features extracted in the previous step. The final phase carries out PDAC diagnosis for a tissue sample using the generated SVM classification model.

3.1. Segmentation for Major Interest Objects

In this section, we describe the method of segmenting three object types in a tissue image. Two of three object types are the lumen and epithelial nucleus constituting a duct. The last one is nonepithelial nucleus. Figure 2 shows the overall process of identifying three object types from a tissue image.

Figure 2: System overview for diagnosing PDAC.
3.1.1. Lumen Segmentation

In this paper, the lumen of the tissue image is segmented by a seeded region growing (SRG) algorithm [38]. A beginning point should be designated for the use of the SRG. In the previous research [32], segmentation of the lumen region was automated by identifying candidate seed points within the lumen region. The proposed method is as follows. First of all, in order to facilitate the application of SRG, median filtering algorithm and background correction algorithm [39] are applied on a given image, and then maximum Entropy Threshold [40] is applied to produce a binary image . From the produced binary image , Direction Cumulative Map is generated to find seed points. The is generated by cumulating only white pixels of four directions (left, right, up, and down) of the binary image and calculating the sum of the square root of the cumulated values. The will have higher values around the central area of lumen region. Therefore, the local maximum points of would be used as candidate seed points for the SRG algorithm. However, if candidate seed points are generated directly from , it might generate candidate seed points for unnecessarily narrow areas. As a solution for this problem, with a threshold of lower value is, instead, employed to acquire candidate seed points. The Otsu method [41] is used to determine the threshold. With having the acquired candidate seed points set as a beginning point of parameter, the lumen region can be segmented by SRG algorithm. The boundary of segmented lumen region will be denotated as . Figure 3 shows the process of segmenting the lumen boundary that has been explained so far.

Figure 3: (a) The preprocessed binary image for identifying the lumen boundary; (b) and candidate seed points (yellow points) for the image; (c) 3D plot for . It is scaled as a range from 0 to 255; (d) the boundary of lumen segmented by candidate seed (green line).
3.1.2. The Identification of Epithelial Nuclei and Nonepithelial Nuclei

In this phase, we segmented the nuclei of tissue images and separate them into epithelial and nonepithelial nuclei. The process is as follows.

(1) Nuclei Segmentation. This step identifies all nuclei in a tissue image. First, the impurities shown on the tissue image are eliminated by median filter. Then, the color thresholding based on -means [42] removes the parts such as cytoplasm and lumen that are unnecessary to identify nuclei. Next, the holes of the nuclei in the thresholded image are filled with a hole filling algorithm [43]. Finally, the nuclei are separated using a Watershed algorithm [44]. Then, a set of the segmented nuclei are denoted by . Figure 4(a) shows the segmented nuclei.

Figure 4: (a) A set of the segmented nuclei, ; (b) a set of the epithelial nuclei selected from , (marked as red).

(2) Division the Epithelial Nuclei and the Nonepithelial Nuclei. This step divides the segmented nuclei acquired in the previous step into the epithelial nuclei and the nonepithelial nuclei. Epithelial cells surround the lumen. Therefore, epithelial nuclei are identified by selecting the nearest nuclei to the lumen boundary, , from a set of nuclei . A set of epithelial nuclei is denoted by and defined as where is a point in and the Centroid(·) is a function returning the center point of a given object. The Distance(·,·) is a function returning the Euclidean distance between two given points. Segregation of epithelial nuclei from is a procedure in which the nearest nucleus from each point is firstly selected, and the selected nucleus is then included in a set of epithelial nuclei, . Algorithm 1 presents this procedure. Figure 4(b) shows the selected epithelial nuclei (marked as red).

Algorithm 1: Selection_Epithelial_Nuclei .

Nonepithelial nuclei are acquired by eliminating the identified epithelial nuclei from a set of nuclei as follows:

3.2. Notations

Table 1 summarizes the notations used in this paper.

Table 1: List of notations.

4. Proposed Features

A major object examined to diagnose PDAC in a tissue sample is a duct. As described in Section 2, PDAC is classified as Grade 1, 2, or 3 by morphological changes of the duct which is composed of the lumen and the epithelial cells. Therefore, in this section, we propose methods to extract the specific morphological features of the segmented lumen and epithelial nuclei for PDAC diagnosis.

4.1. Lumen Features

In PDAC, a duct seems to have the shape of an incomplete gland with a papillary form and a loss of nucleus polarity. As the processing stage progresses, the shape of the duct becomes more irregular with various atypia. In this subsection, the method representing atypia of a duct and the morphological features for measuring it are discussed.

4.1.1. Representing Atypia of Duct

Generally, a lumen of normal duct seems like a convex hull because atypia rarely appears. Unlike normal formation, as PDAC progresses, atypia of the lumen boundary in the duct becomes more and more irregular. From this standpoint, estimation of ideal lumen boundary of a given duct will be possible and portrayal of atypia of the original lumen will be feasible based on the boundary.

(1)  Ideal Lumen Boundary. In this step, an original lumen boundary and an estimated ideal lumen boundary are represented as and , respectively. and are, respectively, a sequence of points constituting each lumen boundary. The procedure to estimate the ideal lumen boundary, , is as follows. First, the convex hull, , is obtained from the original lumen boundary . Then, because is bigger than , the ideal lumen boundary, , is acquired through downsizing to . The scaling factor, , for scaling to is calculated as follows: where represents an area bordering while does an area bordering . is a function returning area of given region. Therefore, the ideal lumen boundary is a sequence of points that consist of the boundary of the downsized region by the scaling factor about the center of . Figure 5 shows the original lumen boundary, , with a green line and the ideal lumen boundary, , with a red line.

Figure 5: Original lumen boundary (green line) and ideal lumen boundary (red line).

(2)  Atypia-Amplitude Signature. In this stage, 1D signature will be proposed as a means to depict atypia of a lumen employing an original lumen boundary and an ideal lumen boundary. The proposed 1D signature visualizes atypia of a lumen by measuring atypia-amplitude between an original lumen boundary and an ideal lumen boundary. The atypia-amplitude is an orthogonal distance with sign between an original lumen boundary and an ideal lumen boundary. It is measured by an atypia-amplitude function as follows: where is an index variable, indicating an order of points within an ideal lumen boundary and is a function representing sign of a vertical distance between and , which returns +1 or −1, respectively, when a point, , is located either outside or inside the boundary with a point on . Figure 6(a) shows a process of measuring atypia-amplitude at . In this example, has positive value because is located outside the boundary of .

Figure 6: The atypia-amplitude signature with .

The 1D signature using the function, , is named as the atypia-amplitude signature. The atypia-amplitude signature is plotted with points of and . Then -coordinate is the meaning the perimeter from (starting point of ) to , and -coordinate is the implying atypia-amplitude at point of . , the perimeter from (start point of ) to , is a sum of the Euclidean distances of each points within the given breadth. The is as follows: Figure 6 shows how to plot the atypia-amplitude signature by . Table 2 shows the atypia-amplitudes for Figures 1(a), 1(b), and 1(c).

Table 2: The atypia-amplitude signature for Figures 1(a), 1(b), and 1(c).
4.1.2. Features for Measuring Atypia of Duct

This section introduces features that measure atypia of the lumen quantitatively using atypia-amplitude signature and the ideal lumen boundary developed in the previous phase. The proposed features are RMSAA (root-mean-squared atypia-amplitude), TSAV (total sum of atypia volatilities), AtypiaRatio, and #AtypiaRegions (it means the number of atypia regions).

(1) RMSAA (Root-Mean-Squared Atypia-Amplitude). RMSAA is measured by atypia-amplitude signature developed in the previous phase. It is the square root of the mean-squared atypia-amplitude (MSAA). , the vertical distance between and , can be interpreted as the residual that represents the difference between the sample value of and the fitted value of the estimated regression model. Likewise, MSAA corresponds to a mean-squared error (MSE) [45] that measures the average squares error of the regression model. RMSAA is defined as follows: Here, is the number of points in .

(2) TSAV (Total Sum of Atypia Volatilities). Variation of the lumen boundary becomes more irregular since the lumen becomes more complex as PDAC develops. TSAV measures the degree of irregularity of the lumen boundary shown in the progress of PDAC. For the calculation of TSAV, major inflection points of atypia-amplitude signature are identified and the sum of diversion at those points is taken into account. In this paper, we use the Perceptual Important Point (PIP) method [46, 47] to find the major inflection points of the atypia-amplitude signature. The PIP method finds critical points that represent important trends of time series data. In this paper, the conventional PIP algorithm that detects a fixed number of PIPs has been modified to find all critical points in the atypia-amplitude signature. The detail about the modified PIP algorithm is included in Appendices A and B. Figure 7 shows a part of atypia-amplitude signature of Grade 2 in Table 2 and PIPs observed by the modified PIP algorithm.

Figure 7: at a PIP, , is measured as the angle between and .

The TSAV is computed by (7) as the total sum of atypia volatilities (AVs) at PIPs. at a PIP, , is defined by angle between two vectors, and : Here, is the number of PIPs detected from the atypia-amplitude signature.

(3) Atypia Ratio and the Number of Atypia Regions. The shape of duct becomes more complex and papillary becomes more vivid as PDAC develops. The original lumen region of the developed PDAC does not fit into the ideal lumen region, , extending beyond or contracting into the . Thus, in this section, we measure AtypiaRatio and #AtypiaRegions (the number of atypia regions) to assess such characteristics. First, a set of atypia regions, , is composed with regions which are generated by separating results of using . Figure 8 shows the identified atypia regions in Grade 2 tissue image.

Figure 8: (a) Grade 2 tissue image, (b) atypia regions that are generated by region and region of (a).

The obtained atypia regions are used to come up with AtypiaRatio and #AtypiaRegions: where is a function returning the size of a given region. The means the cardinality of the . The AtypiaRatio feature represents the overall degree of distortion within a duct, and the #AtypiaRegions feature quantitatively measures the papillary duct by counting the atypia regions. Then, small atypia regions representative of the papillary are excluded from counting by thresholding. The value for thresholding is set 300 μm2 by consensus of pathologists at the Pathology Department of Yeongnam University.

4.2. Epithelial Cell Feature

Epithelium is another component composing a duct. In most cases, the epithelial cells of PDAC are the mucin-producing neoplastic cells that tend to be columnar, and their nuclei show loss of polarity [34, 48]. In this phase, we introduce the methods of extracting these features from the segmented epithelial nuclei.

(1) Cytoplasm Length. A duct of normal tissue is surrounded by cube-like epithelial cells. In the PDAC, a duct has columnar epithelium with abundant cytoplasm. The nuclei of columnar cells are oval-shaped. The cytoplasm length of columnar epithelium with abundant cytoplasm is longer than cuboidal epithelium. So, measuring the cytoplasm length of epithelial cells represents whether or not the epithelial cell trend is columnar. We proposed the feature, CytoplasmLength measuring the cytoplasm length of the epithelial cells in [32]. The CytoplasmLength is the orthogonal distance between the epithelial nucleus and the original lumen boundary : where is a point in that is orthogonal to Centroid. Figure 9 shows the measured CytoplasmLengths of Normal and Grade 1. In Figure 9, red regions are epithelial nuclei and green line is the identified lumen boundary. The blue lines between nuclei and lumen boundary are the measured CytoplasmLengths.

Figure 9: CytoplasmLengths (blue lines) of epithelial cells for Normal and Grade 1.

(2) The Standard Deviation of CytoplasmLength. This feature measures the loss of nuclear polarity that is one of the features of PDAC. The epithelial nuclei with the loss of nuclear polarity have a large deviation between cytoplasm lengths of them because epithelial nuclei are arranged irregularly along the lumen boundary. In contrast, the CytoplasmLengths of normal epithelium have a small deviation because the epithelial nuclei are arranged along the lumen boundary. Therefore, we measure the loss of nuclear polarity by calculating the standard deviation of the CytopalsmLength of epithelial nuclei: where is the cardinality set and is the average of CytoplasmLength.

5. Experiments

5.1. Image Acquisition and Experimental Environment

We received 21 normal tissue slides and 26 PDAC tissue slides from the Pathology Department of Yeongnam University for our experiments. Those received tissue slides were stained via hematoxylin and eosin. Those tissue slides were scanned into digital slides using the ScanScope CS System [49] at 20x magnification. Each digital image of slides is variable depending on acquired tissue. Table 3 shows the information of digital slide.

Table 3: Information of the digital slide.

In order to assess proposed features, we manually generated images for the experiments from these digital slides to make sure that each includes a duct. Each of the experiment images was formatted into a 24-bit tiff, and their size varied depending on the size of each duct. Important issue in diagnosis is inter- and intraobserver variability leading the diagnosis to be inconsistent, inaccurate, and biased [50]. A similar issue arises from ground truth data of experts for configuration of and performance assessment of a diagnosis system. A feasible way of reducing variability issue of ground truth is to construct ground truth with participation of several experts [51]. This article has three pathologists of the Pathology Department of Yeongnam University participated in an assessment of ground truth. Each image of duct generated from digital slides has been labeled into class with discreet consensus of those three pathologists. Table 4 presents the number of experiment images labeled by the experts for the assessment of ground truth.

Table 4: The obtained experimental images.

We segmented the given tissue images into three parts (lumen, epithelial nuclei, and nonepithelial nuclei) and extracted existing classical morphological features with the proposed features from each part. Table 5 shows the features used in the experiments for diagnosing PDAC. The features extracted from each segment are asterisked. 1~12 rows in Table 5 are the existing classical features [5, 24, 39, 52]. 13~18 rows are the proposed features in this paper and our previous study [32]. Because a number of epithelial and nonepithelial cells were found in a captured tissue image, the features of each object were extracted and then averaged to represent features of the tissue image. The experiment environment for feature extraction was performed on a computer with an AMD Athelon II 3 GHZ CPU and 2 G RAM running Windows7 64 bit. The existing and proposed methods for extracting features were implemented by using ImageJ [53], an image processing package based on the JAVA programming language.

Table 5: Morphological features used in the classification experiment.
5.2. Experiment Design

We compared the performance of the classifiers learned by the classical and proposed features to demonstrate the quality of morphological features that are proposed to diagnose PDAC. In this paper, SVM, a well-founded learning technique based on statistical learning theory [54], was employed as the learning method of the classifier. The SVM shows good generalized performance, because it minimizes the combination of the empirical risk and the VC (Vapnik-Cheronenkis) dimension [55].

The experiment evaluated the classification performance for two cases: classification between Normal and PDAC tissues and classification between Grade 1 and Grade 2 of PDAC. To measure how the proposed features improve the accuracy of classification, the classifiers were learned by feature sets that are configured as existing classical, proposed, and combination features for the three segmented objects (lumen, epithelial nuclei, and nonepithelial nuclei). Table 6 shows the symbol and dimension for the configured feature sets used in classification experiments.

Table 6: The symbol and the dimension of feature sets that are configured by the object features and combined features.

Experiment data were generated according to each feature set in Table 6 for the experiments in two cases (Normal versus PDAC and Grade 1 versus Grade 2). Thus, for the experiments to diagnose PDAC in the first case, 13 data sets for 13 feature sets were generated as follows: D(CLF), D(PLF), , , , , , , , , , , and . D is the data set that is configured by the given feature set as a parameter. It is denoted as , where is the th feature vector corresponding to a given feature set of parameters (symbols of Table 6) and is its class label. The is either −1, which is Normal, or 1, which is PDAC.

Similarly, for the experiments to grade stages of PDAC in the second case, 13 data sets were generated. In these data sets, means Grade 1 whereas means Grade 2. To evaluate the performance of the SVM classifier for each feature set, we configured a training set and test set from the generated data set for the feature set. The ratio of training set to test a set was 60 to 40. In the first experiment (Normal versus PDAC), the experiment data set of PDAC is configured by sampling 80 data from 160 PDAC data of either Grade 1 or Grade 2. Because the number of Normal sample is 80, we limit the number of PDAC for fair evaluation of classifiers. Table 7 shows the number of training and test data sets used in the experiments of two cases.

Table 7: The number of training and testing data sets for learning and evaluating the SVM classifier.

In this paper, SVM classifiers used the soft margin method and RBF kernel [54]. Therefore, model parameter and kernel parameter are required. The optimal classifier parameters in which the classification accuracy for 10 cross-validation [56] in the training set is maximized are selected from parameter pairs of by Grid Search [57].

The number of experiment images used in this study is insufficient. Therefore, the classification accuracy of the generated model might be biased [58]. In statistics, to solve this problem, the bootstrap resampling technique [59] is used. We used the bootstrap resampling technique for the unbiased evaluation of classifiers for each feature set. First, we generated 10 training sets and 10 testing sets from data set corresponding to a given feature set for bootstrap evaluation (refer to Table 6). Therefore, the classification performance for each feature set is measured by averaging evaluation results of individually optimized classifiers for 10 training sets and 10 testing data sets. The performance measures used in the experiment are as follows: true positive (TP), true negative (TN), false positive (FP), false negative (FN), sensitivity (SN), Specificity (SP), positive predictive value (PPV), negative predictive value (NPV), and accuracy (ACC). The descriptions for TP, TN, FP, and FN are explained in each experiment. The rest of performance measures are defined as follows:

5.3. Experimental Results

Case 1 (Normal versus PDAC). Table 8 and Figure 10 show the bootstrap evaluation of classifier learned by each feature set, for distinguishing between the Normal and PDAC. The standard deviations of the evaluation results are displayed in parentheses. First, we will compare of classifiers that are learned by feature set extracted from lumen object. In this comparison, the accuracy of classifier learnt by PLF is 91.56% which is about 18% higher than the classification accuracy with CLF (73.44%). The classification accuracy with ALF feature set configured as combination of CLF and PLF has, contrarily, decreased to 87.35%. The results showed that the PLF feature set is more suitable for diagnosing PDAC than classifiers with CLF. However, no improvement in the performance of the classifier with the ALF that is a combination of PLF and CLF was revealed in the results.
Secondly, we will compare classifiers with feature set extracted from epithelial nuclei objects. PEF and AEF show the same accuracy of 87.50%. However, in case of a classifier with AEF, the standard deviation of accuracy was observed as 1.47 which is far stable than PEF of 3.21. Interestingly, AEF is a combination of the PEF and the CEF, its accuracy is dependent on the proposed PEF. Also, the notable point is that the feature dimension of PEF is only two. These results do not only show that the PEF is very suitable for identifying the PDAC but they also prove their effectiveness in diagnostic cost aspect.
In diagnosis of PDAC, a duct that is composed of lumen and epithelium is an important region. Therefore, we thought that the experiment with the combination of features extracted from lumen and epithelial nuclei is of very meaning. For this, we prepared three combination feature sets which include , , and . With these sets, we performed classification. In these experiments, the classification accuracy is improved to 94.38% when the PDF feature set was used. The classification accuracy was measured about 3~7% higher than PLF and AEF (or PEF) that showed best classification performance in each object. It showed that combination of lumen and epithelial nuclei features helps diagnose PDAC.
Consequently, the experimental results of CTF, PTF, and ATF using combination of the feature sets extracted from three objects in a tissue image depended on experimental results conducted in duct object. Thus, it is implied that there are no improvements as a result of combining all features. Further, the experiment with PTF that consists of PDF and CNF showed 2% lower in its accuracy than PDF alone.
Subsequently, ROC (receiver-operating characteristic) analysis with regard to classifiers learned by each feature set was performed. ROC analysis is being widely used in medical study as a benchmark of accuracy and comparison of diagnosis. ROC analysis examines ROC curves drawn by TP rate (Sensitivity) and FP rate (1-Sensitivity). The examination presents diagnosis accuracy with area under the ROC curve. Swets classified the degree of the accuracy, according to the value of AUC (area under the ROC curve), into noninformative (AUC = 0.5), less accurate (0.5 < AUC ≤ 0.7), moderately accurate (0.7 < AUC ≤ 0.9), highly accurate (0.9 < AUC < 1), and perfect tests (AUC = 1) [60, 61]. In other words, as ROC curve approaches to left hand corner, the accuracy is interpreted as higher. Figure 11 shows the average ROC graph and value of AUC of 13 classifiers learnt by features of each object.
In the ROC analysis, a classifier learned by PDF displays the highest value of AUC with 0.96. It proves the meaningfulness of combination of epithelial nuclei and lumen features likewise the performance evaluation of the classifier. Subsequent to PDF, a classifier with PEF shows a slightly higher AUC value of 0.94 than that of PLF with 0.93. It is interpreted that classifiers learned by features inclusive of the proposed PDF and PEF show fairly accurate diagnosis with AUC value of above 0.9.
Overall, experiments including proposed feature set show better performance than classifiers with classical feature set. As mentioned in Section 2, the experiments showed that the duct is an important region in diagnosing PDAC. PDF that is composed of PLF and PEF has led to improvement of classifier performance. Also, the classifiers with the proposed PLF and PEF extracted from lumen and epithelial nuclei, respectively, show higher performance than classical feature sets, CLF, CEF, and CNF. It implies that simple morphological features such as Area and Perimeter, are inadequate for finding complicated characteristic of PDAC.

Table 8: Evaluation results for distinguishing Normal and PDAC to each feature set.
Figure 10: Comparison of classification accuracy for the Normal and PDAC for each feature set.
Figure 11: Comparison of ROC curves and AUC values for classifiers in Case 1.

Case 2 (Grade 1 versus Grade 2). In this step, we distinguished between two stages, Grade 1 and Grade 2, of PDAC. As with the experiments in Case 1, we generated 10 training and testing sets for Grade 1 and Grade 2, and we evaluated classification accuracy. The results are shown in Table 9 and Figure 12. The classification results of each feature sets for distinguishing between Grade 1 and Grade 2 show lower classification accuracy than the experiments of distinguishing between Normal and PDAC. Firstly, in the experiments with the lumen object, the classification accuracy with PLF (77.03%) was about 19% higher than that with CLF (57.97%). In particular, specificity of PLF (79.69%) is measured 34% higher than CLF (45.94%), showing performance gain by 73%. In the experiment with ALF, combination of CLF and PLF, the classifier accuracy showed poorer classification performance than the classifier learnt by PLF.
Next, the experiments with the feature sets extracted from epithelial nuclei showed lower accuracy in classification than the experiments with epithelial cells in Case 1. As epithelium cells in all stages of Grade 1 and Grade 2 of PDAC showed columnar and loss of polarity characteristics, distinguishing of stages through them is difficult. Nevertheless, accuracy of the classifier with PEF (70.78%) using only two features increased by about 14% compared to CEF (56.41%). In an experiment using AEF, the accuracy was measured lower than PEF, indicating no performance improvement through the combination of CEF and PEF.
As opposed to the previous experiment that distinguished Normal from PDAC using PDF, an experiment of this case with PDF for classifying stages did not lead to improved performance of a classifier and the evaluation results were the same as experiments with PLF. It proves that the performance enhancement of classifier through PDF, a combination of PLE and PEF, is dependent entirely on PLF while PEF not showing any contribution.
The experiments with combination of feature sets (CTF, PTF, and ATF) extracted from three object types of tissue image showed the same results as those with combination of features sets (CDF, PDF, and ADF) extracted from a duct that consists two object types, lumen and epithelial nucleus.
Figure 13 presents ROC graphs and values of AUC averaged by classifiers that distinguishes PDAC stage based on sets of features of each object. In overall, lower performance was detected than Case 1. Classifiers with the classic morphologic features such as CNF, CLF, and CEF provide less accurate diagnosis with AUC values of 0.61, 0.45, and 0.61, respectively. On the contrary, the AUC values of classifiers learned by PLF and PEF are respectively 0.79 and 0.7, showing moderately accurate test results that is one-step higher than AUC values of classifiers learned by existing feature sets such as CEF, CLF, and CNF.
From these experiments of classifying PDAC, classifiers with PLF-contained feature sets show the best performances. As opposed to Case 1 distinguishing Normal from PDAC, no improvements were found with the combination of PEF and PLF. One particular aspect in these experiments is that experiment results of feature sets of lumen object are same as the results of experiments with feature sets of duct object and of tissue object. The feature sets extracted from duct and tissue are composed of the mix of feature sets from lumen object and other objects. It attests that features extracted from lumen are of positive influence to the classifier performance and contain the most information necessary for diagnosing stages of PDAC.
The experiments to classify Grade 1 and Grade 2 in Case 2 showed lower classification performance than the experiments to differentiate between Normal and PDAC in Case 2, both when the proposed features were used and when the existing features were used. This can be explained by the fact that the characteristics of the PDAC commonly appear in Grade 1 and Grade 2 stages. Furthermore, the proposed features lack diagnosing stages of PDAC consisting of similar morphological characteristics since they are designed most of all to distinguish PDAC from Normal tissues. Even if it is so, the proposed feature sets perform better than the classical feature sets.

Table 9: Evaluation results for distinguishing between Grade 1 and Grade 2 to each feature set.
Figure 12: Comparison of classification accuracy for distinguishing between Grade 1 and Grade 2 for each feature set.
Figure 13: Comparison of ROC curves and AUC values for classifiers in Case 2.

6. Discussion

In this section, we statistically analyzed features that were extracted from the three segmented parts (lumen, epithelial nuclei, and nonepithelial nuclei). Firstly, we assumed that if the extracted features are appropriate for diagnosing PDAC, then the value of features will be different among three populations (Normal, Grade 1, and Grade 2). To statistically show whether the extracted features are different among populations, we performed the ANOVA (ANalysis Of VAriance) for each feature. The null hypothesis for testing if the features are different among populations is as follows: where , , and are mean population means of Normal, Grade 1, and Grade 2, respectively. The significant test of ANOVA for the features is tested by -statistic. Tables 10, 11, and 12 show -test results of null hypothesis (12) for features of each of three object types at the 0.01 level of significance. The -test results of features for three object types attest in statistics that there is difference among features in each group (Normal, Grade 1, or Grade 2) in most cases. In -test of features extracted from lumen, the existence of disparities was confirmed between features of all groups (Normal, Grade 1, and Grade 2), except for Roundness and Solidity. In -test for features of epithelial nuclei, all features show statistical difference between groups. As for the test of nonepithelial nuclei, MinorAxis and Skewness were only features not showing statistically difference.

Table 10: The statistics for nonepithelial nuclei features and the results of F-test.
Table 11: The statistics for lumen features and the results of F-test.
Table 12: The statistics for the epithelial nuclei features and the results of F-test.

Next, for the post hoc analysis of features that reject the null hypothesis in -test, we performed the multirange test to find whether there are significant differences between means of any population of two. In this paper, the commonly used Fisher’s LSD (least significant difference) test was employed for the post hoc analysis of -test [62]. In this paper, there are three populations (Normal, Grade 1, and Grade 2) examined. So, LSD-test for total pairs was performed. The null hypothesis for testing each pair is as follows: The LSD test results features for each of three object types are shown in Tables 13, 14, and 15. The bold values in Tables 13, 14, and 15 mean that the features of all the three hypotheses (13) are rejected in LSD test. Features marked with “-” are those not performed for LSD test since hypothesis test in (12) of -test was rejected.

Table 13: LSDtest for F-test of nonepithelial nuclei features.
Table 14: LSD test for F-test of lumen features.
Table 15: LSD-test for F-test of epithelial nuclei features.

Firstly, in LSD test for features extracted from lumen, 12 features except for Circularity and AspectRatio have rejected the three null hypotheses of (13). Of which, 8 features are classical features and 4 features are proposed features. Although there is a number of classical features that have difference between groups, the experiment results using PLF showed more improved performance than experiments using CLF (refer to Tables 8 and 9) in Case 1 and Case 2.

In epithelial nuclei, 5 features including CytoplasmLength and CytoplasmLengthSD rejected the null hypotheses (13). In the case of nonepithelial nuclei, only features of Perimeter, Width, MajorAxis, and Fereter’s Diameter rejected the all null hypothesis of (13). The results of LSD test show that null hypothesis, , of (13) to test significant difference between Grade 1 and Grade 2 were not rejected in many features from two distinct object types, epithelial nucleus and nonepithelial nucleus.

In LSD test of lumen features, features that rejected null hypothesis (12) of -test but not in (13) are only two, Circularity and AspectRatio. However, in LSD tests of epithelial nucleus features, only 6 features of 14 features that rejected null hypothesis (12) of -test reject . In LSD-test of nonepithelial nuclei, only 4 features rejected . Through LSD test, it has been confirmed that the lumen is the most important object in the diagnosis of PDAC and its stages. Furthermore, LSD test results describe the reason why classification performance of the experiments in Case 2 (Grade 1 versus Grade 2) is lower than that of them in Case 1 (Normal versus PDAC).

7. Conclusions

This paper proposed features to diagnose PDAC and to identify the stages of PDAC. PDAC is mainly diagnosed by investigating a duct that consists of lumen and epithelial cells. We segmented a tissue image into three parts: lumen, epithelial nuclei, and nonepithelial nuclei. Then, we proposed methods for extracting new morphological features from the epithelial cells and lumen parts that are segmented. In PDAC, the shape of the duct is more complex than Normal. Thus, this paper proposed the features for measuring atypia of the duct based on this perspective. We transformed the lumen into the atypia-amplitude signature with the atypia-amplitude function to intuitively represent the variation of a duct and proposed RMSAA for measuring the deviation of the aytpia-amplitudes and TSAV for measuring the volatility at PIP points of it. And, using the ideal lumen and original lumen regions, we measured AtypiaRatio, which represents the overall degree of distortion of a duct and #AtypiaRegions that quantify the papillary ducts. Also, we used features such as CytoplasmLength and CytoplasmLengthSD to quantitatively measure the morphological features from segmented epithelial nuclei. The experiments’ results show that the proposed features are suitable to diagnose PDAC and to distinguish between the two stages, Grade 1 and Grade 2, of PDAC.


In this paper, we use PIP method [46, 47] to find critical points of atypia-amplitude signature. However, the existing PIP method may not be able to find points that variation of the lumen boundary becomes more complex because it detects a fixed number of critical points. Therefore, we introduce the modified PIP detection method that finds all critical points of given atypia-amplitude signature. The modified PIP detection method is processed by the following two steps: (1) finding all PIPs of atypia-amplitude signature and (2) postprocessing for eliminating unnecessary PIPs.

A. Finding All PIPs of Atypia-Amplitude Signature

The PIP method finds critical points referred to as PIP (Perceptual Important Points) that represent important trends of time series data. In this paper, the existing PIP detection algorithm that detects a fixed number of PIPs has been modified to find all critical points in the atypia-amplitude signature. The modified PIP detection algorithm detects all PIPs with maximum vertical distances (VDs) [46, 47] above Threshold between adjacent PIPs. PIP_Detection_For_Atypia_Amplitude_Signature (Algorithm 2) shows the modified PIP detection algorithm in this paper.

Algorithm 2: PIP_Detection_For_Atypia_ Amplitude_Signature  (amp_list, T).

As inputs, Algorithm 2 takes a sequence of points (denoted by amp_list) that forms the atypia-amplitude signature and the threshold (denoted by ) to detect PIPs. The algorithm output is a sequence of the detected PIPs (denoted by pip_list). The first step of Algorithm 2 initializes pip_list to the first point of amp_list (Algorithm 2, line 3). After that, all PIPs in a given amp_list are detected by Sub_PIP_Detection algorithm (Algorithm 3).

Algorithm 3: Sub_PIP_Detection  (s, e, amp_list, T, pip_list).

The Sub_PIP_Detection (Algorithm 3) finds a PIP within a given range of amp_list. The first and second inputs of the algorithm are start-index (denoted by s) and end-index (denoted by e) for the range of the amp_list to detect a PIP. Algorithm 3 first finds the location that has the maximum VD value in a given range (Algorithm 3, line 1). Then if the VD value of the found location is greater than a threshold T, the location is used as a pivot. The given range of the algorithm is split into two ranges, and the Sub_PIP_Detection (Algorithm 3) is called with the new two ranges (Algorithm 3 lines 5–7). If the VD value is less than T, the partition of the range is stopped and the point at end-index of the given range for the algorithm is added to pip_list (Algorithm 3, line 9). The Sub_PIP_Detection (Algorithm 3) is recursively called until all PIPs satisfying the threshold condition are detected in amp_list (Algorithm 4).

Algorithm 4: Max_ VD _Idx  (s, e, amp_list).

The distance metric VD used in the PIP detection is the vertical distance between the test point and the line connecting the two adjacent PIPs. That is, VD at a PIP between two adjacent PIPs, and , is as follows: where is the value of a linear function determined by two points, and , when is given. Figure 14 shows the VD between the line connecting the two adjacent PIPs ( and ) and the test point .

Figure 14: VD at between PIPs, and .

B. Postprocessing for Eliminating Unnecessary PIPs

Figure 15(a) shows the PIPs identified by the modified PIP algorithm (Algorithm 2). But some unnecessary PIPs can be found because the modified PIP algorithm identifies all PIPs satisfying the condition that the maximum VD is greater than a threshold. That is, several PIPs can be found where the change of the big trend has not occurred, as shown in Figure 15(a). These PIPs are points that have a trend in the same direction. To remove such unnecessary PIPs, postprocessing for the identified PIPs (pip_list) by the modified PIP algorithm (Algorithm 2) is performed. The postprocessing algorithm for PIPs is as Algorithm 5. The Post_Processing_PIPs (Algorithm 5) assesses whether each PIP of pip_list identified by Algorithm 2 is a maxima or minima when trend is reversed (Algorithm 5, lines 4–10). Then, if a PIP of the pip_list is not a maxima or minima, the PIP is removed in the pip_list (Algorithm 5, lines 11–14). Figure 15(b) shows the result of Algorithm 5 for Figure 15(a).

Algorithm 5: Post_Porocessing_PIPs  (pip_list).

Figure 15: (a) PIPs identified by Algorithm 2; (b) PIPs after applying Algorithm 5 to (a) (red circles represent detected PIPs).


This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Science and Technology (2010-0024216). This work was supported by INHA University Research Grant.


  1. M. N. Gurcan, L. E. Boucheron, A. Can, A. Madabhushi, N. M. Rajpoot, and B. Yener, “Histopathological image analysis: a review,” IEEE Reviews in Biomedical Engineering, vol. 2, pp. 147–171, 2009. View at Publisher · View at Google Scholar
  2. A. E. Tutac, D. Racoceanu, T. Putti, W. Xiong, W. K. Leow, and V. Cretu, “Knowledge-guided semantic indexing of breast cancer histopathology images,” in Proceedings of the International Conference on BioMedical Engineering and Informatics (BMEI '08), vol. 2, pp. 107–112, May 2008. View at Publisher · View at Google Scholar · View at Scopus
  3. S. M. Ismail, A. B. Colclough, J. S. Dinnen et al., “Observer variation in histopathological diagnosis and grading of cervical intraepithelial neoplasia,” British Medical Journal, vol. 298, no. 6675, pp. 707–710, 1989. View at Scopus
  4. A. Andrion, C. Magnani, P. G. Betta et al., “Malignant mesothelioma of the pleura: interobserver variability,” Journal of Clinical Pathology, vol. 48, no. 9, pp. 856–860, 1995. View at Scopus
  5. C. Demir and B. Yener, “Automated cancer diagnosis based on histopathological systematic images: a systematic survey,” Rensselaer Polytechnic Institute, 2005.
  6. S. Al-Janabi, A. Huisman, and P. J. Van Diest, “Digital pathology: current status and future perspectives,” Histopathology, vol. 61, no. 1, pp. 1–9, 2012. View at Publisher · View at Google Scholar · View at Scopus
  7. A. Madabhushi, “Digital pathology image analysis: opportunities and challenges,” Imaging in Medicine, vol. 1, no. 1, pp. 7–10, 2009. View at Publisher · View at Google Scholar
  8. R. S. Weinstein, A. R. Graham, L. C. Richter et al., “Overview of telepathology, virtual microscopy, and whole slide imaging: prospects for the future,” Human Pathology, vol. 40, no. 8, pp. 1057–1069, 2009. View at Publisher · View at Google Scholar · View at Scopus
  9. J. I. Epstein and G. J. Netto, Biopsy Interpretation of the Prostate, Lippincott Williams & Wilkins, Philadelphia, Pa, USA, 4th edition, 2007.
  10. D. F. Gleason and G. T. Mellinger, “Prediction of prognosis for prostatic adenocarcinoma by combined histological grading and clinical staging,” Journal of Urology, vol. 167, no. 2, pp. 953–959, 2002. View at Scopus
  11. A. Tabesh, M. Teverovskiy, H. Y. Pang et al., “Multifeature prostate cancer diagnosis and gleason grading of histological images,” IEEE Transactions on Medical Imaging, vol. 26, no. 10, pp. 1366–1378, 2007. View at Publisher · View at Google Scholar · View at Scopus
  12. S. Naik, S. Doyle, M. Feldman, J. Tomaszewski, and A. Madabhushi, “Gland segmentation and computerized gleason grading of prostate histology by integrating low-, high-level and domain specific information,” in Proceedings of the Microscopic Image Analysis with Applications in Biology (MIAAB '07), 2007.
  13. P.-W. Huang and C. H. Lee, “Automatic classification for pathological prostate images based on fractal analysis,” IEEE Transactions on Medical Imaging, vol. 28, no. 7, pp. 1037–1050, 2009. View at Publisher · View at Google Scholar · View at Scopus
  14. A. W. Wetzel, “Evaluation of prostate tumor grades by content-based image retrieval,” in Proceedings of the 27th AIPR Workshop on Advances in Computer-Assisted Recognition, vol. 3584, pp. 244–252, 1999.
  15. S. Doyle, M. Hwang, K. Shah, A. Madabhushi, M. Feldman, and J. Tomaszeweski, “Automated grading of prostate cancer using architectural and textural image features,” in Proceedings of the 4th IEEE International Symposium on Biomedical Imaging: From Nano to Macro (ISBI '07), pp. 1284–1287, April 2007. View at Publisher · View at Google Scholar · View at Scopus
  16. S.-K. Tai, Y.-C. Wu, C.-Y. Li, Y. J. Jan, and S. C. Lin, “Computer-assisted detection and grading of prostatic cancer in biopsy image,” 2010, http://www.libsearch.com/view/1046167.
  17. Y. Peng, Y. Jiang, S. T. Chuang, and X. J. Yang, “Computer-aided detection of prostate cancer on tissue sections,” Applied Immunohistochemistry & Molecular Morphology, vol. 17, no. 5, pp. 442–450, 2009. View at Publisher · View at Google Scholar
  18. N. H. Anderson, P. W. Hamilton, P. H. Bartels, D. Thompson, R. Montironi, and J. M. Sloan, “Computerized scene segmentation for the discrimination of architectural features in ductal proliferative lesions of the breast,” The Journal of Pathology, vol. 181, no. 4, pp. 374–380, 1997. View at Publisher · View at Google Scholar
  19. C. Bilgin, C. Demir, C. Nagi, and B. Yener, “Cell-graph mining for breast tissue modeling and classification,” in Proceedings of the 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBS '07), pp. 5311–5314, 2007.
  20. A. N. Basavanhally, S. Ganesan, S. Agner et al., “Computerized image-based detection and grading of lymphocytic infiltration in HER2+ breast cancer histopathology,” IEEE Transactions on Biomedical Engineering, vol. 57, no. 3, pp. 642–653, 2010. View at Publisher · View at Google Scholar · View at Scopus
  21. S. Doyle, S. Agner, A. Madabhushi, M. Feldman, and J. Tomaszewski, “Automated grading of breast cancer histopathology using spectral clustering with textural and architectural image features,” in Proceedings of the 5th IEEE International Symposium on Biomedical Imaging: From Nano to Macro (ISBI '08), pp. 496–499, May 2008. View at Publisher · View at Google Scholar · View at Scopus
  22. W. H. Wolberg, W. N. Street, D. M. Heisey, and O. L. Mangasarian, “Computer-derived nuclear features distinguish malignant from benign breast cytology,” Human Pathology, vol. 26, no. 7, pp. 792–796, 1995. View at Scopus
  23. R. Fernandez-Gonzalez, T. Deschamps, A. Idica, R. Malladi, and C. Ortiz de Solorzano, “Automatic segmentation of histological structures in mammary gland tissue sections,” Journal of Biomedical Optics, vol. 9, no. 3, pp. 444–453, 2004. View at Publisher · View at Google Scholar · View at Scopus
  24. W. H. Wolberg, W. N. Street, and O. L. Mangasarian, “Breast cytology diagnosis with digital image analysis,” Analytical and Quantitative Cytology and Histology, vol. 15, no. 6, pp. 396–404, 1993. View at Scopus
  25. A. N. Esgiar, R. N. G. Naguib, B. S. Sharif, M. K. Bennett, and A. Murray, “Fractal analysis in the detection of colonic cancer images,” IEEE Transactions on Information Technology in Biomedicine, vol. 6, no. 1, pp. 54–58, 2002. View at Publisher · View at Google Scholar · View at Scopus
  26. H. K. Choi, T. Jarkrans, E. Bengtsson et al., “Image analysis based grading of bladder carcinoma. Comparison of object, texture and graph based methods and their reproducibility,” Analytical Cellular Pathology, vol. 15, no. 1, pp. 1–18, 1997. View at Scopus
  27. M. N. Gurcan, J. Kong, O. Sertel, B. B. Cambazoglu, J. Saltz, and U. Catalyurek, “Computerized pathological image analysis for neuroblastoma prognosis,” Annual Symposium proceedings, pp. 304–308, 2007. View at Scopus
  28. O. Sertel, J. Kong, H. Shimada, U. V. Catalyurek, J. H. Saltz, and M. N. Gurcan, “Computer-aided prognosis of neuroblastoma on whole-slide images: classification of stromal development,” Pattern Recognition, vol. 42, no. 6, pp. 1093–1103, 2009. View at Publisher · View at Google Scholar · View at Scopus
  29. B. B. Cambazoglu, O. Sertel, J. Kong, J. Saltz, M. N. Gurcan, and U. V. Catalyurek, “Efficient processing of pathological images using the grid: computer-aided prognosis of neuroblastoma,” in Proceedings of the 16th International Symposium on High Performance Distributed Computing (HPDC '07), pp. 35–41, New York, NY, USA, June 2007. View at Publisher · View at Google Scholar · View at Scopus
  30. K. Belkacem-Boussaid, M. Pennell, G. Lozanski, A. Shana'ah, and M. N. Gurcan, “Effect of pathologist agreement on evaluating a computer-aided assisted system: recognizing centroblast cells in follicular lymphoma cases,” in Proceedings of the 7th IEEE International Symposium on Biomedical Imaging: From Nano to Macro (ISBI '10), pp. 1411–1414, April 2010. View at Publisher · View at Google Scholar · View at Scopus
  31. O. Sertel, G. Lozanski, A. Shanáah, and M. N. Gurcan, “Computer-aided detection of centroblasts for follicular lymphoma grading using adaptive likelihood-based cell segmentation,” IEEE Transactions on Biomedical Engineering, vol. 57, no. 10, pp. 2613–2616, 2010. View at Publisher · View at Google Scholar · View at Scopus
  32. J. W. Song, J. H. Lee, T. S. Park, S. J. Chun, and J. H. Choi, “Mucinous cystadenoma classification system using automated epithelial tissue detection,” in Proceedings of the International Conference on Machine Vision, 2010.
  33. G. Klöppel, R. H. Hruban, D. S. Longnecker, G. Adler, S. E. Kern, and T. J. Partanen, “Ductal adenocarcinoma of the pancreas,” in Word Health Organization Classification of Tumours. Pathology and Genetics of Tumours of the Digestive System, International Agency for Research on Cancer, Lyon, France, 3rd edition, 2000.
  34. R. H. Hruban, P. Boffetta, C. Iacobuzio-Donahue et al., “Ductal adenocarcinoma of the pancreas,” in WHO Classification of Tumours of the Digestive System, vol. 3, pp. 217–224, International Agency for Research on Cancer, Lyon, France, 4th edition, 2010.
  35. C. J. Yeo, J. L. Cameron, T. A. Sohn et al., “Six hundred fifty consecutive pancreaticoduodenectomies in the 1990s: pathology, complications, and outcomes,” Annals of Surgery, vol. 226, no. 3, pp. 248–260, 1997. View at Publisher · View at Google Scholar · View at Scopus
  36. G. Klöppel, G. Lingenthal, M. Von Bulow, and H. F. Kern, “Histological and find structural features of pancreatic ductal adenocarcinoma in relation to growth and prognosis: studies in xenografted tumours and clinico-histopathological correlation in a series of 75 cases,” Histopathology, vol. 9, no. 8, pp. 841–856, 1985. View at Scopus
  37. J. Lüttges, S. Schemm, I. Vogel, J. Hedderich, B. Kremer, and G. Klöppel, “The grade of pancreatic ductal carcinoma is an independent prognostic factor and is superior to the immunohistochemical assessment of proliferation,” Journal of Pathology, vol. 191, no. 2, pp. 154–161, 2000. View at Publisher · View at Google Scholar
  38. R. Adams and L. Bischof, “Seeded region growing,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 16, no. 6, pp. 641–647, 1994. View at Publisher · View at Google Scholar · View at Scopus
  39. K. Rodenacker and E. Bengtsson, “A feature set for cytometry on digitized microscopic images,” Analytical Cellular Pathology, vol. 25, no. 1, pp. 1–36, 2003. View at Scopus
  40. P. K. Sahoo, S. Soltani, A. K. C. Wong, and Y. C. Chen, “A survey of thresholding techniques,” Computer Vision, Graphics, and Image Processing, vol. 41, no. 2, pp. 233–260, 1988.
  41. N. Otsu, “A threshold selection method from gray-level histograms,” IEEE Trans Syst Man Cybern, vol. 9, no. 1, pp. 62–66, 1979. View at Scopus
  42. A. Z. Chitade and S. K. Katiyar, “Color based image segmentation using K-means clustering,” International Journal of Engineering Science and Technology, vol. 2, no. 10, pp. 5319–5325, 2010.
  43. R. C. González and R. E. Woods, Digital Image Processing, Prentice Hall, New York, NY, USA, 2008.
  44. C. Pan, C.-X. Zheng, and H.-J. Wang, “Robust color image segmentation based on mean shift and marker-controlled watershed algorithm,” in Proceedings of the International Conference on Machine Learning and Cybernetics, vol. 5, pp. 2752–2756, November 2003. View at Scopus
  45. R. J. Hyndman and A. B. Koehler, “Another look at measures of forecast accuracy,” International Journal of Forecasting, vol. 22, no. 4, pp. 679–688, 2006. View at Publisher · View at Google Scholar · View at Scopus
  46. F.-L. Chung, T. C. Fu, V. Ng, and R. W. P. Luk, “An evolutionary approach to pattern-based time series segmentation,” IEEE Transactions on Evolutionary Computation, vol. 8, no. 5, pp. 471–489, 2004. View at Publisher · View at Google Scholar · View at Scopus
  47. K.-P. Chan and A. W. Fu, “Efficient time series matching by wavelets,” in Proceedings of the 15th International Conference on Data Engineering (ICDE '99), pp. 126–133, 1999.
  48. M. Kosmahl, U. Pauser, M. Anlauf, and G. Klöppel, “Pancreatic ductal adenocarcinomas with cystic features: neither rare nor uniform,” Modern Pathology, vol. 18, no. 9, pp. 1157–1164, 2005. View at Publisher · View at Google Scholar · View at Scopus
  49. “ScanScope CS-Asperio,” 2012, http://www.aperio.com/lifescience/capture/cs.
  50. M. Kallergi, “Evaluation strategies for medical-image analysis and processing methodologies,” in Medical Image Analysis Methods, L. Costaridou, Ed., vol. 18, CRC Press, New York, NY, USA, 2005.
  51. J. K. Udupa, V. R. LeBlanc, H. Schmidt et al., “A methodology for evaluating image segmentation algorithms,” in Proceedings of the Medical Imaging 2002: Image Processing, pp. 266–277, February 2002. View at Publisher · View at Google Scholar · View at Scopus
  52. T. Allen, Particle Size Measurement: Volume 1: Powder Sampling and Particle Size Measurement, Springer, New York, NY, USA, 1996.
  53. W. Rasband, “ImageJ, 2012-1997,” 2012, http://imagej.nih.gov/ij/.
  54. V. N. Vapnik, Statistical Learning Theory, Wiley-Interscience, New York, NY, USA, 1st edition, 1998.
  55. C. J. C. Burges, “A tutorial on support vector machines for pattern recognition,” Data Mining and Knowledge Discovery, vol. 2, no. 2, pp. 121–167, 1998. View at Publisher · View at Google Scholar
  56. T. Hastie, R. Tibshirani, and J. Friedman, The Elements of Statistical Learning: Data Mining, Inference, and Prediction, Second Edition, Springer, New York, NY, USA, 2nd edition, 2009.
  57. C.-C. Chang and C.-J. Lin, “LIBSVM:a library for support vector machines,” ACM Transactions on Intelligent Systems and Technology, vol. 2, no. 3, pp. 1–27, 2011.
  58. D. L. Verbyla and J. A. Litvaitis, “Resampling methods for evaluating classification accuracy of wildlife habitat models,” Environmental Management, vol. 13, no. 6, pp. 783–787, 1989. View at Scopus
  59. H. J. Adér, G. J. Mellenbergh, and D. J. Hand, Advising on Research Methods: a Consultant's Companion, Johannes van Kessel, Huizen, Netherlands, 2008.
  60. J. A. Swets, “Measuring the accuracy of diagnostic systems,” Science, vol. 240, no. 4857, pp. 1285–1293, 1988. View at Publisher · View at Google Scholar
  61. M. Greiner, D. Pfeiffer, and R. D. Smith, “Principles and practical application of the receiver-operating characteristic analysis for diagnostic tests,” Preventive Veterinary Medicine, vol. 45, no. 1-2, pp. 23–41, 2000. View at Publisher · View at Google Scholar
  62. L. J. Williams and H. Abdi, “Fisher's least significance difference (LSD) test,” in Encyclopedia of Research Design, pp. 491–494, SAGE, Thousand Oaks, Calif, USA, 1st edition, 2010.