BioMed Research International

BioMed Research International / 2013 / Article

Research Article | Open Access

Volume 2013 |Article ID 175271 | https://doi.org/10.1155/2013/175271

Jae-Won Song, Ju-Hong Lee, "New Morphological Features for Grading Pancreatic Ductal Adenocarcinomas", BioMed Research International, vol. 2013, Article ID 175271, 25 pages, 2013. https://doi.org/10.1155/2013/175271

New Morphological Features for Grading Pancreatic Ductal Adenocarcinomas

Academic Editor: Xin-yuan Guan
Received17 Jan 2013
Revised07 Apr 2013
Accepted24 Apr 2013
Published25 Jul 2013

Abstract

Pathological diagnosis is influenced by subjective factors such as the individual experience and knowledge of doctors. Therefore, it may be interpreted in different ways for the same symptoms. The appearance of digital pathology has created good foundation for objective diagnoses based on quantitative feature analysis. Recently, numerous studies are being done to develop automated diagnosis based on the digital pathology. But there are as of yet no general automated methods for pathological diagnosis due to its specific nature. Therefore, specific methods according to a type of disease and a lesion could be designed. This study proposes quantitative features that are designed to diagnose pancreatic ductal adenocarcinomas. In the diagnosis of pancreatic ductal adenocarcinomas, the region of interest is a duct that consists of lumen and epithelium. Therefore, we first segment the lumen and epithelial nuclei from a tissue image. Then, we extract the specific features to diagnose the pancreatic ductal adenocarcinoma from the segmented objects. The experiment evaluated the classification performance of the SVM learned by the proposed features. The results showed an accuracy of 94.38% in the experiment distinguishing between pancreatic ductal adenocarcinomas and normal tissue and a classification accuracy of 77.03% distinguishing between the stages of pancreatic ductal adenocarcinomas.

1. Introduction

Pathological diagnosis is currently performed subjectively via the knowledge and experience of doctors after inspection of tissue slides through light microscopes. This subjective diagnosis has some problems. First, tumor screening at high magnification by light microscope requires a lot of time and effort [1, 2]. Also, the individual competence of doctors has a decisive effect on the final diagnosis. This means that the pathological diagnosis of the same tumor by two different doctors may vary, because it is not based on objective quantitative feature analysis [35].

The advent of digital pathology has led to a new type of pathological diagnosis. Digital pathology changes glass slides into digital images. It provides a convenient screening environment that also makes it possible to objectively diagnose tumors through quantitative feature analysis using a computer [68]. Nevertheless, a pathological diagnosis needs much time and workload because the diagnosis is manually processed. Therefore, there are still problems related to the subjective diagnosis of tumors. To overcome these problems, studies on computer-aided diagnosis (CAD) techniques based on digital pathology are growing. Many studies of CAD-based digital pathology have covered not only detection of tumors but also grading the stages of tumors. In pathology, tumor stage grading is useful to identify the extent of the disease and determine the appropriate treatment for a patient [9]. In fact, it is known that pathologists receive intensive trainings on grading stages of tumors in order to prescribe the correct type of treatment [7]. Currently, many studies of pathological diagnosis with CAD techniques are being carried out on breast and prostate tumors.

First, most CAD studies related to prostate cancer are based on the Gleason grading system [10]. Tabesh et al. [11] proposed an automated system based on machine learning to diagnose prostate cancer and grade the stages (low and high) of cancer with the Gleason grading system. They extracted color, texture, and morphometric features at global and object levels of a given tissue image. Classifying algorithms such as Gaussian, -nearest neighbor (-NN), and Support Vector Machine (SVM) learned the features of cancer diagnosis and Gleason grading. Naik et al. [12] proposed a diagnostic system for distinguishing between intermediate Gleason grades. They identified the candidate gland region using a Bayesian classifier with low-level information and eliminated false positive regions identified as glands using empirical domain information. After that, the morphologic features were extracted from the identified glands, and Gleason grades 3, 4, and benign were classified through the SVM that learned the features. Huang and Lee [13] classified prostate cancer images into 4 grades based on a Gleason grading system. They used Bayesian, -NN, and SVM classifiers for classifying stages of the cancer. And, to teach the classifiers Gleason grading, features are proposed by using differential box-counting and entropy-based dimension estimation techniques. In addition to that, there are many studies for the grading and diagnosis of prostate tumors [1417].

There are also CAD studies related to breast tumors. Anderson et al. [18] worked on a problem for distinguishing ductal hyperplasia (DH), which is benign, and ductal carcinoma in situ (DCIS), which is malignant. In this study, they automatically segmented breast ducts using knowledge-guided machine vision and proposed measuring duct cribriformity and architectural complexity to quantitatively analyze the duct patterns in proliferative lesions to distinguish between DH and DCIS. Bilgin et al. [19] proposed a method to diagnose breast cancer using graph theory techniques. They segmented given tissue images using a -means algorithm and generated different cell graphs using the positional coordinates of cells for each segmented image. An SVM model that can classify given tissue images into benign, invasive, and non-invasive (ductal carcinoma in situ) was learned by quantitative metrics that are computed from the generated cell graphs. Basavanhally et al. [20] proposed a grading system that identifies and grades the extent of lymphocytic infiltration (LI), a known viable prognostic indicator. First, they detected lymphocytes using region growing and Markov random field algorithms. Then, the architectural features were extracted from the detected lymphocytes, and the extent of LI was classified into low, medium, and high grades by an SVM classifier learned by the features. In addition, there are a lot of studies for the diagnosis and grading of breast tumors [2124].

Many pathological CAD studies make an effort to analyze the pathological characteristics of and design the methods for quantitatively measuring each disease, because there are many methods for pathological diagnosis according to the type of disease and lesion. Currently, in addition to the breast and prostate tumors mentioned above, some studies on colonic [25], bladder [26], neuroblastoma [2729], and follicular lymphoma [30, 31] tumors have been performed. Pathological CAD studies are still in the early stages and focus on a few tumors. There are many methods for diagnosing different types of tumors. Therefore, more studies of pathological CADs must be performed.

The aim of this study is to detect pancreatic ductal adenocarcinoma (PDAC) and classify them by stages. To achieve this, we propose new morphological features for diagnosing and grading PDAC. The region that is inspected to diagnose PDAC is a duct that consists of lumen and epithelium. Therefore, this paper segments the given image into lumen, epithelial nuclei, and nonepithelial nuclei and extracts the morphological features for diagnosing PDAC from the segmented objects. After that, the diagnosis and grading stages of PDAC are performed using the SVM model learned by the extracted features. This paper has several sections. Section 2 describes the pathological characteristics and the morphological features needed in diagnosing PDAC. Section 3 shows the configuration of systems used to diagnose PDAC and the segmentation methods of objects. Section 4 discusses the proposed new morphological features for quantitatively measuring the pathological characteristics of PDAC described in Section 2. In Section 5, the performances between SVM classifiers learnt by the proposed and existing classical morphological features are compared with each other to show the suitability of the proposed features to detect and grade the stages of PDAC. Section 6 evaluates and statistically analyzes the results. Finally, Section 7 presents a conclusion of this study.

2. Pathological Characteristics of PDAC

Pancreatic cancer is the second most common gastrointestinal neoplasm that causes death, after colon cancer [33]. And of all pancreatic neoplasms, PDAC accounts for 85–95%. Approximately 80% of all PDAC patients are between 60 and 80 years of age, and cases in people below the age of 40 are rare. The incidence of PDAC is about 50% higher in men than in women. By race, those of African ancestry have the highest rate of PDAC [34]. The best way of treating PDAC is known as curative resection. However, because of the rare possibility of diagnosing PDAC in the early stage, only 5–22% of PDAC patients can take the curative resection at the time the cancer is discovered [35]. Therefore, an accurate determination of the degree of cancer development is crucial factor for the treatment.

PDAC progression is divided by histological and cytological features and mitotic activity into Grade 1, well-differentiated carcinomas; Grade 2, moderately differentiated carcinomas; and Grade 3, poorly differentiated ductal adenocarcinomas [36, 37]. Grade 1 consists of a duct-like structure combined with medium-sized neoplastic glands. Tubular or cribriform patterns are typical. There may also be small irregular papillary projections without a distinct fibrovascular stalk, particularly in large duct-like structures. Mitotic activity is low. The mucin-producing neoplastic cells tend to be columnar, have eosinophils, and occasionally exhibit pale or even clear cytoplasm. Some neoplastic cell nuclei show loss of polarity. Grade 2 is characterized by a mixture of medium-sized duct-like and tubular structures of variable shapes, embedded in desmoplastic stroma. The duct shape is commonly that of incompletely formed glands. Compared with Grade 1, Grade 2 shows a greater variation in nuclear size, chromatin structure, and prominence of the nucleoli. The cytoplasm is usually slightly eosinophilic, but clear cells are occasionally abundant. Mucin production appears to be decreased, and intraductal in situ components are somewhat less frequent than in Grade 1. Grade 3 is infrequent. It is composed of a mixture of densely packed, small, and irregular glands as well as solid tumor cell sheets and nests that entirely replace the acinar tissue. While typical large, duct-like structures and intraductal tumor components are absent, there may be small squamoid features, spindle cells, or anaplastic foci. The neoplastic cells show marked pleomorphism, little or no mucin production, and brisk mitotic activity. Figure 1 shows Normal, Grade 1, Grade 2, and Grade 3 tissue images.

As described above, Grade 3 is not common. Also, Figure 1 shows that Grade 3 is certainly morphologically different from Grades 1 and 2. Therefore, this paper focuses on the detection of PDAC and the differentiation of Grades 1 and 2.

3. System Overview

In this paper, system configuration to diagnose PDAC consists of three phases as follows: segmentation and feature extraction, model learning and validation, and diagnosis. In the first phase, after preprocessing the given tissue image, the image is segmented into three parts. These three parts are the lumen region, epithelial nuclei, and nonepithelial nuclei. Then, according to the characteristics of each part, the features to be used for the classification model are extracted and stored into a feature database. The second phase is the learning and validation of the SVM classification model using the features extracted in the previous step. The final phase carries out PDAC diagnosis for a tissue sample using the generated SVM classification model.

3.1. Segmentation for Major Interest Objects

In this section, we describe the method of segmenting three object types in a tissue image. Two of three object types are the lumen and epithelial nucleus constituting a duct. The last one is nonepithelial nucleus. Figure 2 shows the overall process of identifying three object types from a tissue image.

3.1.1. Lumen Segmentation

In this paper, the lumen of the tissue image is segmented by a seeded region growing (SRG) algorithm [38]. A beginning point should be designated for the use of the SRG. In the previous research [32], segmentation of the lumen region was automated by identifying candidate seed points within the lumen region. The proposed method is as follows. First of all, in order to facilitate the application of SRG, median filtering algorithm and background correction algorithm [39] are applied on a given image, and then maximum Entropy Threshold [40] is applied to produce a binary image . From the produced binary image , Direction Cumulative Map is generated to find seed points. The is generated by cumulating only white pixels of four directions (left, right, up, and down) of the binary image and calculating the sum of the square root of the cumulated values. The will have higher values around the central area of lumen region. Therefore, the local maximum points of would be used as candidate seed points for the SRG algorithm. However, if candidate seed points are generated directly from , it might generate candidate seed points for unnecessarily narrow areas. As a solution for this problem, with a threshold of lower value is, instead, employed to acquire candidate seed points. The Otsu method [41] is used to determine the threshold. With having the acquired candidate seed points set as a beginning point of parameter, the lumen region can be segmented by SRG algorithm. The boundary of segmented lumen region will be denotated as . Figure 3 shows the process of segmenting the lumen boundary that has been explained so far.

3.1.2. The Identification of Epithelial Nuclei and Nonepithelial Nuclei

In this phase, we segmented the nuclei of tissue images and separate them into epithelial and nonepithelial nuclei. The process is as follows.

(1) Nuclei Segmentation. This step identifies all nuclei in a tissue image. First, the impurities shown on the tissue image are eliminated by median filter. Then, the color thresholding based on -means [42] removes the parts such as cytoplasm and lumen that are unnecessary to identify nuclei. Next, the holes of the nuclei in the thresholded image are filled with a hole filling algorithm [43]. Finally, the nuclei are separated using a Watershed algorithm [44]. Then, a set of the segmented nuclei are denoted by . Figure 4(a) shows the segmented nuclei.

(2) Division the Epithelial Nuclei and the Nonepithelial Nuclei. This step divides the segmented nuclei acquired in the previous step into the epithelial nuclei and the nonepithelial nuclei. Epithelial cells surround the lumen. Therefore, epithelial nuclei are identified by selecting the nearest nuclei to the lumen boundary, , from a set of nuclei . A set of epithelial nuclei is denoted by and defined as where is a point in and the Centroid(·) is a function returning the center point of a given object. The Distance(·,·) is a function returning the Euclidean distance between two given points. Segregation of epithelial nuclei from is a procedure in which the nearest nucleus from each point is firstly selected, and the selected nucleus is then included in a set of epithelial nuclei, . Algorithm 1 presents this procedure. Figure 4(b) shows the selected epithelial nuclei (marked as red).

Input: A set of nuclei , A original lumen boundary
Output: A set of epithelial nuclei
Method:
(1)  
(2) for all point   in     do
(3) n getNearestNuclus (p, )
(4)  n
(5)   end for
(6)   return  

Nonepithelial nuclei are acquired by eliminating the identified epithelial nuclei from a set of nuclei as follows:

3.2. Notations

Table 1 summarizes the notations used in this paper.


SymbolDescription

p The is a pair of numbers that are and coordinates p = (x, y)
AThe matrix representation for preprocessed binary image
H(A)The direction cumulative histogram for A
(A)The thresholded H(A)
A sequence of points that consist of the original lumen boundary
A sequence of points for the convex hull of
A sequence of points for the ideal lumen boundary that is estimated from
A set of nuclei , ,…,
A set of epithelial nuclei
A set of nonepithelial nuclei
Distance(·,·)The Euclidean distance function
Centroid(·)The function returning center point of given object
A(t)The atypia-amplitude function
L(t)The perimeter from   (start point) to (tth point)
A set of atypia regions    ar1, ar2,…,
The region surrounded by
The region surrounded by
The region surrounded by
Area (·)The function returning area of given region
(·)The data set for given feature set (·) = {(x 1(·), ), …, ( (·), )}, x(·) is feature vector for given feature set and is class label for x(·)

4. Proposed Features

A major object examined to diagnose PDAC in a tissue sample is a duct. As described in Section 2, PDAC is classified as Grade 1, 2, or 3 by morphological changes of the duct which is composed of the lumen and the epithelial cells. Therefore, in this section, we propose methods to extract the specific morphological features of the segmented lumen and epithelial nuclei for PDAC diagnosis.

4.1. Lumen Features

In PDAC, a duct seems to have the shape of an incomplete gland with a papillary form and a loss of nucleus polarity. As the processing stage progresses, the shape of the duct becomes more irregular with various atypia. In this subsection, the method representing atypia of a duct and the morphological features for measuring it are discussed.

4.1.1. Representing Atypia of Duct

Generally, a lumen of normal duct seems like a convex hull because atypia rarely appears. Unlike normal formation, as PDAC progresses, atypia of the lumen boundary in the duct becomes more and more irregular. From this standpoint, estimation of ideal lumen boundary of a given duct will be possible and portrayal of atypia of the original lumen will be feasible based on the boundary.

(1)  Ideal Lumen Boundary. In this step, an original lumen boundary and an estimated ideal lumen boundary are represented as and , respectively. and are, respectively, a sequence of points constituting each lumen boundary. The procedure to estimate the ideal lumen boundary, , is as follows. First, the convex hull, , is obtained from the original lumen boundary . Then, because is bigger than , the ideal lumen boundary, , is acquired through downsizing to . The scaling factor, , for scaling to is calculated as follows: where represents an area bordering while does an area bordering . is a function returning area of given region. Therefore, the ideal lumen boundary is a sequence of points that consist of the boundary of the downsized region by the scaling factor about the center of . Figure 5 shows the original lumen boundary, , with a green line and the ideal lumen boundary, , with a red line.

(2)  Atypia-Amplitude Signature. In this stage, 1D signature will be proposed as a means to depict atypia of a lumen employing an original lumen boundary and an ideal lumen boundary. The proposed 1D signature visualizes atypia of a lumen by measuring atypia-amplitude between an original lumen boundary and an ideal lumen boundary. The atypia-amplitude is an orthogonal distance with sign between an original lumen boundary and an ideal lumen boundary. It is measured by an atypia-amplitude function as follows: where is an index variable, indicating an order of points within an ideal lumen boundary and is a function representing sign of a vertical distance between and , which returns +1 or −1, respectively, when a point, , is located either outside or inside the boundary with a point on . Figure 6(a) shows a process of measuring atypia-amplitude at . In this example, has positive value because is located outside the boundary of .

The 1D signature using the function, , is named as the atypia-amplitude signature. The atypia-amplitude signature is plotted with points of and . Then -coordinate is the meaning the perimeter from (starting point of ) to , and -coordinate is the implying atypia-amplitude at point of . , the perimeter from (start point of ) to , is a sum of the Euclidean distances of each points within the given breadth. The is as follows: Figure 6 shows how to plot the atypia-amplitude signature by . Table 2 shows the atypia-amplitudes for Figures 1(a), 1(b), and 1(c).


175271.tab.002

4.1.2. Features for Measuring Atypia of Duct

This section introduces features that measure atypia of the lumen quantitatively using atypia-amplitude signature and the ideal lumen boundary developed in the previous phase. The proposed features are RMSAA (root-mean-squared atypia-amplitude), TSAV (total sum of atypia volatilities), AtypiaRatio, and #AtypiaRegions (it means the number of atypia regions).

(1) RMSAA (Root-Mean-Squared Atypia-Amplitude). RMSAA is measured by atypia-amplitude signature developed in the previous phase. It is the square root of the mean-squared atypia-amplitude (MSAA). , the vertical distance between and , can be interpreted as the residual that represents the difference between the sample value of and the fitted value of the estimated regression model. Likewise, MSAA corresponds to a mean-squared error (MSE) [45] that measures the average squares error of the regression model. RMSAA is defined as follows: Here, is the number of points in .

(2) TSAV (Total Sum of Atypia Volatilities). Variation of the lumen boundary becomes more irregular since the lumen becomes more complex as PDAC develops. TSAV measures the degree of irregularity of the lumen boundary shown in the progress of PDAC. For the calculation of TSAV, major inflection points of atypia-amplitude signature are identified and the sum of diversion at those points is taken into account. In this paper, we use the Perceptual Important Point (PIP) method [46, 47] to find the major inflection points of the atypia-amplitude signature. The PIP method finds critical points that represent important trends of time series data. In this paper, the conventional PIP algorithm that detects a fixed number of PIPs has been modified to find all critical points in the atypia-amplitude signature. The detail about the modified PIP algorithm is included in Appendices A and B. Figure 7 shows a part of atypia-amplitude signature of Grade 2 in Table 2 and PIPs observed by the modified PIP algorithm.

The TSAV is computed by (7) as the total sum of atypia volatilities (AVs) at PIPs. at a PIP, , is defined by angle between two vectors, and : Here, is the number of PIPs detected from the atypia-amplitude signature.

(3) Atypia Ratio and the Number of Atypia Regions. The shape of duct becomes more complex and papillary becomes more vivid as PDAC develops. The original lumen region of the developed PDAC does not fit into the ideal lumen region, , extending beyond or contracting into the . Thus, in this section, we measure AtypiaRatio and #AtypiaRegions (the number of atypia regions) to assess such characteristics. First, a set of atypia regions, , is composed with regions which are generated by separating results of using . Figure 8 shows the identified atypia regions in Grade 2 tissue image.

The obtained atypia regions are used to come up with AtypiaRatio and #AtypiaRegions: where is a function returning the size of a given region. The means the cardinality of the . The AtypiaRatio feature represents the overall degree of distortion within a duct, and the #AtypiaRegions feature quantitatively measures the papillary duct by counting the atypia regions. Then, small atypia regions representative of the papillary are excluded from counting by thresholding. The value for thresholding is set 300 μm2 by consensus of pathologists at the Pathology Department of Yeongnam University.

4.2. Epithelial Cell Feature

Epithelium is another component composing a duct. In most cases, the epithelial cells of PDAC are the mucin-producing neoplastic cells that tend to be columnar, and their nuclei show loss of polarity [34, 48]. In this phase, we introduce the methods of extracting these features from the segmented epithelial nuclei.

(1) Cytoplasm Length. A duct of normal tissue is surrounded by cube-like epithelial cells. In the PDAC, a duct has columnar epithelium with abundant cytoplasm. The nuclei of columnar cells are oval-shaped. The cytoplasm length of columnar epithelium with abundant cytoplasm is longer than cuboidal epithelium. So, measuring the cytoplasm length of epithelial cells represents whether or not the epithelial cell trend is columnar. We proposed the feature, CytoplasmLength measuring the cytoplasm length of the epithelial cells in [32]. The CytoplasmLength is the orthogonal distance between the epithelial nucleus and the original lumen boundary : where is a point in that is orthogonal to Centroid. Figure 9 shows the measured CytoplasmLengths of Normal and Grade 1. In Figure 9, red regions are epithelial nuclei and green line is the identified lumen boundary. The blue lines between nuclei and lumen boundary are the measured CytoplasmLengths.

(2) The Standard Deviation of CytoplasmLength. This feature measures the loss of nuclear polarity that is one of the features of PDAC. The epithelial nuclei with the loss of nuclear polarity have a large deviation between cytoplasm lengths of them because epithelial nuclei are arranged irregularly along the lumen boundary. In contrast, the CytoplasmLengths of normal epithelium have a small deviation because the epithelial nuclei are arranged along the lumen boundary. Therefore, we measure the loss of nuclear polarity by calculating the standard deviation of the CytopalsmLength of epithelial nuclei: where is the cardinality set and is the average of CytoplasmLength.

5. Experiments

5.1. Image Acquisition and Experimental Environment

We received 21 normal tissue slides and 26 PDAC tissue slides from the Pathology Department of Yeongnam University for our experiments. Those received tissue slides were stained via hematoxylin and eosin. Those tissue slides were scanned into digital slides using the ScanScope CS System [49] at 20x magnification. Each digital image of slides is variable depending on acquired tissue. Table 3 shows the information of digital slide.


Digital slide sizeVariable size
Image resolution0.492 μm/pixel
Image typeSVS/JPEG2000
Image channels3
Image bit depth8 bits
Magnification20x
OrganizationTiled
Tile width240 pixels
Tile height240 pixels

In order to assess proposed features, we manually generated images for the experiments from these digital slides to make sure that each includes a duct. Each of the experiment images was formatted into a 24-bit tiff, and their size varied depending on the size of each duct. Important issue in diagnosis is inter- and intraobserver variability leading the diagnosis to be inconsistent, inaccurate, and biased [50]. A similar issue arises from ground truth data of experts for configuration of and performance assessment of a diagnosis system. A feasible way of reducing variability issue of ground truth is to construct ground truth with participation of several experts [51]. This article has three pathologists of the Pathology Department of Yeongnam University participated in an assessment of ground truth. Each image of duct generated from digital slides has been labeled into class with discreet consensus of those three pathologists. Table 4 presents the number of experiment images labeled by the experts for the assessment of ground truth.


TypeNumber of images

Normal
Grade 1
Grade 2

We segmented the given tissue images into three parts (lumen, epithelial nuclei, and nonepithelial nuclei) and extracted existing classical morphological features with the proposed features from each part. Table 5 shows the features used in the experiments for diagnosing PDAC. The features extracted from each segment are asterisked. 1~12 rows in Table 5 are the existing classical features [5, 24, 39, 52]. 13~18 rows are the proposed features in this paper and our previous study [32]. Because a number of epithelial and nonepithelial cells were found in a captured tissue image, the features of each object were extracted and then averaged to represent features of the tissue image. The experiment environment for feature extraction was performed on a computer with an AMD Athelon II 3 GHZ CPU and 2 G RAM running Windows7 64 bit. The existing and proposed methods for extracting features were implemented by using ImageJ [53], an image processing package based on the JAVA programming language.


No.FeatureDescriptionNEN1LumenEN2

1 AreaArea of selection in square pixels * * *
2 PerimeterThe length of the outside boundary of the selection * * *
3 WidthWidth of the smallest rectangle enclosing the selection * * *
4 HeightHeight of the smallest rectangle enclosing the selection * * *
5MajorAxis Major (primary) axis length of the best fitting ellipse * * *
6MinorAxis Minor (secondary) axis length of the best fitting ellipse * * *
7 Circularity4π × (Area/Perimeter2). A value of 1.0 indicates a perfect circle * * *
8 Feret’s diameterThe longest distance between any two points along the selection boundary * * *
9 AspectRatioMajorAxis/MinorAxis * * *
10 SkewnessThe third order moment about the mean * * *
11 Roundness4 × Area/(π × MajorAxis2). The measure of the sharpness of particle’s edge and corners * * *
12 SolidityArea/ConvexArea * * *
13 RMSAARoot-mean-squared atypia-amplitude *
14 TSAVTotal sum of atypia volatilities for PIPs *
15 AtypiaRatioThe ratio of atypia region *
16 #AtypiaRegionsThe number of atypia regions for identifying papillary *
17CytoplasmLengthThe cytoplasm length of epithelial nucleus [32] *
18CytoplasmLengthSDThe standard deviation of CytoplasmLength *

NEN1: nonepithelial nuclei; EN2: epithelial nuclei.
5.2. Experiment Design

We compared the performance of the classifiers learned by the classical and proposed features to demonstrate the quality of morphological features that are proposed to diagnose PDAC. In this paper, SVM, a well-founded learning technique based on statistical learning theory [54], was employed as the learning method of the classifier. The SVM shows good generalized performance, because it minimizes the combination of the empirical risk and the VC (Vapnik-Cheronenkis) dimension [55].

The experiment evaluated the classification performance for two cases: classification between Normal and PDAC tissues and classification between Grade 1 and Grade 2 of PDAC. To measure how the proposed features improve the accuracy of classification, the classifiers were learned by feature sets that are configured as existing classical, proposed, and combination features for the three segmented objects (lumen, epithelial nuclei, and nonepithelial nuclei). Table 6 shows the symbol and dimension for the configured feature sets used in classification experiments.


ObjectSymbolFeature setDimension

NENCNFClassical nonepithelial nuclei features12

LumenCLFClassical lumen features12
PLFProposed lumen features4
ALFCLF  +  PLF16

ENCEFClassical epithelial nuclei features12
PEFProposed epithelial nuclei features2
AEFCEF  +  PEF14

DuctCDFClassical duct features (CLF  +  CEF)24
PDFProposed duct features (PLF  +  PEF)6
ADFCDF  +  PDF30

TissueCTFClassical features extracted from three objects (CLF  +  CEF  +  CNF)36
PTFProposed features extracted from three objects (PLF  +  PEF  +  CNF)18
ATFAll features (CLF  +  CEF  +  PLF  +  PEF  +  CNF) 42

Experiment data were generated according to each feature set in Table 6 for the experiments in two cases (Normal versus PDAC and Grade 1 versus Grade 2). Thus, for the experiments to diagnose PDAC in the first case, 13 data sets for 13 feature sets were generated as follows: D(CLF), D(PLF), , , , , , , , , , , and . D is the data set that is configured by the given feature set as a parameter. It is denoted as , where is the th feature vector corresponding to a given feature set of parameters (symbols of Table 6) and is its class label. The is either −1, which is Normal, or 1, which is PDAC.

Similarly, for the experiments to grade stages of PDAC in the second case, 13 data sets were generated. In these data sets, means Grade 1 whereas means Grade 2. To evaluate the performance of the SVM classifier for each feature set, we configured a training set and test set from the generated data set for the feature set. The ratio of training set to test a set was 60 to 40. In the first experiment (Normal versus PDAC), the experiment data set of PDAC is configured by sampling 80 data from 160 PDAC data of either Grade 1 or Grade 2. Because the number of Normal sample is 80, we limit the number of PDAC for fair evaluation of classifiers. Table 7 shows the number of training and test data sets used in the experiments of two cases.


ExperimentClassNumber of training dataNumber of testing data

Case 1:
Normal versus PDAC
Normal4832
PDAC4832

Case 2:
Grade 1 versus Grade 2
Grade 14832
Grade 24832

In this paper, SVM classifiers used the soft margin method and RBF kernel [54]. Therefore, model parameter and kernel parameter are required. The optimal classifier parameters in which the classification accuracy for 10 cross-validation [56] in the training set is maximized are selected from parameter pairs of by Grid Search [57].

The number of experiment images used in this study is insufficient. Therefore, the classification accuracy of the generated model might be biased [58]. In statistics, to solve this problem, the bootstrap resampling technique [59] is used. We used the bootstrap resampling technique for the unbiased evaluation of classifiers for each feature set. First, we generated 10 training sets and 10 testing sets from data set corresponding to a given feature set for bootstrap evaluation (refer to Table 6). Therefore, the classification performance for each feature set is measured by averaging evaluation results of individually optimized classifiers for 10 training sets and 10 testing data sets. The performance measures used in the experiment are as follows: true positive (TP), true negative (TN), false positive (FP), false negative (FN), sensitivity (SN), Specificity (SP), positive predictive value (PPV), negative predictive value (NPV), and accuracy (ACC). The descriptions for TP, TN, FP, and FN are explained in each experiment. The rest of performance measures are defined as follows:

5.3. Experimental Results

Case 1 (Normal versus PDAC). Table 8 and Figure 10 show the bootstrap evaluation of classifier learned by each feature set, for distinguishing between the Normal and PDAC. The standard deviations of the evaluation results are displayed in parentheses. First, we will compare of classifiers that are learned by feature set extracted from lumen object. In this comparison, the accuracy of classifier learnt by PLF is 91.56% which is about 18% higher than the classification accuracy with CLF (73.44%). The classification accuracy with ALF feature set configured as combination of CLF and PLF has, contrarily, decreased to 87.35%. The results showed that the PLF feature set is more suitable for diagnosing PDAC than classifiers with CLF. However, no improvement in the performance of the classifier with the ALF that is a combination of PLF and CLF was revealed in the results.
Secondly, we will compare classifiers with feature set extracted from epithelial nuclei objects. PEF and AEF show the same accuracy of 87.50%. However, in case of a classifier with AEF, the standard deviation of accuracy was observed as 1.47 which is far stable than PEF of 3.21. Interestingly, AEF is a combination of the PEF and the CEF, its accuracy is dependent on the proposed PEF. Also, the notable point is that the feature dimension of PEF is only two. These results do not only show that the PEF is very suitable for identifying the PDAC but they also prove their effectiveness in diagnostic cost aspect.
In diagnosis of PDAC, a duct that is composed of lumen and epithelium is an important region. Therefore, we thought that the experiment with the combination of features extracted from lumen and epithelial nuclei is of very meaning. For this, we prepared three combination feature sets which include , , and . With these sets, we performed classification. In these experiments, the classification accuracy is improved to 94.38% when the PDF feature set was used. The classification accuracy was measured about 3~7% higher than PLF and AEF (or PEF) that showed best classification performance in each object. It showed that combination of lumen and epithelial nuclei features helps diagnose PDAC.
Consequently, the experimental results of CTF, PTF, and ATF using combination of the feature sets extracted from three objects in a tissue image depended on experimental results conducted in duct object. Thus, it is implied that there are no improvements as a result of combining all features. Further, the experiment with PTF that consists of PDF and CNF showed 2% lower in its accuracy than PDF alone.
Subsequently, ROC (receiver-operating characteristic) analysis with regard to classifiers learned by each feature set was performed. ROC analysis is being widely used in medical study as a benchmark of accuracy and comparison of diagnosis. ROC analysis examines ROC curves drawn by TP rate (Sensitivity) and FP rate (1-Sensitivity). The examination presents diagnosis accuracy with area under the ROC curve. Swets classified the degree of the accuracy, according to the value of AUC (area under the ROC curve), into noninformative (AUC = 0.5), less accurate (0.5 < AUC ≤ 0.7), moderately accurate (0.7 < AUC ≤ 0.9), highly accurate (0.9 < AUC < 1), and perfect tests (AUC = 1) [60, 61]. In other words, as ROC curve approaches to left hand corner, the accuracy is interpreted as higher. Figure 11 shows the average ROC graph and value of AUC of 13 classifiers learnt by features of each object.
In the ROC analysis, a classifier learned by PDF displays the highest value of AUC with 0.96. It proves the meaningfulness of combination of epithelial nuclei and lumen features likewise the performance evaluation of the classifier. Subsequent to PDF, a classifier with PEF shows a slightly higher AUC value of 0.94 than that of PLF with 0.93. It is interpreted that classifiers learned by features inclusive of the proposed PDF and PEF show fairly accurate diagnosis with AUC value of above 0.9.
Overall, experiments including proposed feature set show better performance than classifiers with classical feature set. As mentioned in Section 2, the experiments showed that the duct is an important region in diagnosing PDAC. PDF that is composed of PLF and PEF has led to improvement of classifier performance. Also, the classifiers with the proposed PLF and PEF extracted from lumen and epithelial nuclei, respectively, show higher performance than classical feature sets, CLF, CEF, and CNF. It implies that simple morphological features such as Area and Perimeter, are inadequate for finding complicated characteristic of PDAC.


ObjectFeature setTNFPFNTPSN (%)SP (%)PPV (%)NPV (%)ACC (%)

NENCNF
(1.58)(2.11)(1.58)(2.11)(4.94)(6.60)(5.17)(4.22)(3.95)

CLF
LumenPLF29.302.702.7029.3091.5791.5791.9991.8891.56
(1.77)(2.00)(1.77)(2.00)(5.52)(6.26)(5.45)(4.70)(3.14)
ALF

CEF
ENPEF27.503.504.5028.5085.9489.0789.5286.6787.50
(1.51)(2.95)(1.51)(2.95)(4.72)(9.23)(6.80)(3.25)(3.21)
AEF

CDF
DuctPDF29.801.402.2030.6093.1395.6395.7893.5094.38
(1.40)(1.51)(1.40)(1.51)(4.37)(4.71)(4.34)(3.80)(2.35)
ADF

CTF
TissuePTF29.802.502.2029.5093.1392.1992.7493.2692.66
(1.32)(2.22)(1.32)(2.22)(4.11)(6.95)(6.15)(3.76)(3.21)
ATF

TP (true positive): the number of PDACs that are correctly classified as PDACs.
FP (false positive): the number of Normals that are incorrectly classified as PDACs.
FN (false negative): the number of PDACs that are incorrectly classified as Normals.
TN (true negative): the number of Normals that are correctly classified as Normals.

Case 2 (Grade 1 versus Grade 2). In this step, we distinguished between two stages, Grade 1 and Grade 2, of PDAC. As with the experiments in Case 1, we generated 10 training and testing sets for Grade 1 and Grade 2, and we evaluated classification accuracy. The results are shown in Table 9 and Figure 12. The classification results of each feature sets for distinguishing between Grade 1 and Grade 2 show lower classification accuracy than the experiments of distinguishing between Normal and PDAC. Firstly, in the experiments with the lumen object, the classification accuracy with PLF (77.03%) was about 19% higher than that with CLF (57.97%). In particular, specificity of PLF (79.69%) is measured 34% higher than CLF (45.94%), showing performance gain by 73%. In the experiment with ALF, combination of CLF and PLF, the classifier accuracy showed poorer classification performance than the classifier learnt by PLF.
Next, the experiments with the feature sets extracted from epithelial nuclei showed lower accuracy in classification than the experiments with epithelial cells in Case 1. As epithelium cells in all stages of Grade 1 and Grade 2 of PDAC showed columnar and loss of polarity characteristics, distinguishing of stages through them is difficult. Nevertheless, accuracy of the classifier with PEF (70.78%) using only two features increased by about 14% compared to CEF (56.41%). In an experiment using AEF, the accuracy was measured lower than PEF, indicating no performance improvement through the combination of CEF and PEF.
As opposed to the previous experiment that distinguished Normal from PDAC using PDF, an experiment of this case with PDF for classifying stages did not lead to improved performance of a classifier and the evaluation results were the same as experiments with PLF. It proves that the performance enhancement of classifier through PDF, a combination of PLE and PEF, is dependent entirely on PLF while PEF not showing any contribution.
The experiments with combination of feature sets (CTF, PTF, and ATF) extracted from three object types of tissue image showed the same results as those with combination of features sets (CDF, PDF, and ADF) extracted from a duct that consists two object types, lumen and epithelial nucleus.
Figure 13 presents ROC graphs and values of AUC averaged by classifiers that distinguishes PDAC stage based on sets of features of each object. In overall, lower performance was detected than Case 1. Classifiers with the classic morphologic features such as CNF, CLF, and CEF provide less accurate diagnosis with AUC values of 0.61, 0.45, and 0.61, respectively. On the contrary, the AUC values of classifiers learned by PLF and PEF are respectively 0.79 and 0.7, showing moderately accurate test results that is one-step higher than AUC values of classifiers learned by existing feature sets such as CEF, CLF, and CNF.
From these experiments of classifying PDAC, classifiers with PLF-contained feature sets show the best performances. As opposed to Case 1 distinguishing Normal from PDAC, no improvements were found with the combination of PEF and PLF. One particular aspect in these experiments is that experiment results of feature sets of lumen object are same as the results of experiments with feature sets of duct object and of tissue object. The feature sets extracted from duct and tissue are composed of the mix of feature sets from lumen object and other objects. It attests that features extracted from lumen are of positive influence to the classifier performance and contain the most information necessary for diagnosing stages of PDAC.
The experiments to classify Grade 1 and Grade 2 in Case 2 showed lower classification performance than the experiments to differentiate between Normal and PDAC in Case 2, both when the proposed features were used and when the existing features were used. This can be explained by the fact that the characteristics of the PDAC commonly appear in Grade 1 and Grade 2 stages. Furthermore, the proposed features lack diagnosing stages of PDAC consisting of similar morphological characteristics since they are designed most of all to distinguish PDAC from Normal tissues. Even if it is so, the proposed feature sets perform better than the classical feature sets.


ObjectFeature setTNFPFNTPSN (%)SP (%)PPV (%)NPV (%)ACC (%)

NENCNF

CLF
LumenPLF
ALF

CEF
ENPEF
AEF

CDF
DuctPDF
ADF