Abstract

Traditionally, for diagnosing patellar dislocation, clinicians make manual geometric measurements on computerized tomography (CT) images taken in the knee area, which is often complex and error-prone. Therefore, we develop a prototype CAD system for automatic measurement and diagnosis. We firstly segment the patella and the femur regions on the CT images and then measure two geometric quantities, patellar tilt angle (PTA), and patellar lateral shift (PLS) automatically on the segmentation results, which are finally used to assist in diagnoses. The proposed quantities are proved valid and the proposed algorithms are proved effective by experiments.

1. Introduction

Patellar dislocation occurs when the patella slips out from the patellar surface of the femur. It is a common knee injury that may happen when people, especially teenagers and athletes, do vigorous physical exercises, e.g., playing basketball and football. To help the diagnosis, computerized tomography (CT) images are often taken at the knee area. On the knee CT images, clinicians usually make manual measurements and make diagnosis according to the measured results. The manual measurement is often complex, tedious, and error-prone. Therefore, a fully automatic approach by computers is highly wanted.

Computed tomography has been widely used to diagnose knee joint pathologies. Correspondingly, knee CT images have been automatically or semiautomatically processed and analyzed (e.g., [15]) for computer-aided diagnosis. Subburaj et al. [1] proposed a computer graphics-based method to automatically localize and label anatomical landmarks on the 3D bone model reconstructed from knee CT images of a patient. Krcah et al. [2] proposed to segment the femur in 3D CT volumes based on graph cuts and a bone boundary enhancement filter. Jang et al. [3] compared and validated various segmentation algorithms to segment the knee CT images and construct a corresponding 3D model. Wu et al. [4] proposed to segment multiple bones around the knee joint with severe pathologies to help patient-specific orthopedic knee surgery planning. Mezlini et al. [5] proposed to measure the knee joint space based on semiautomatic CT image segmentation for the monitoring of osteoarthritis progression. However, to the best of our knowledge, no efforts have been published specifically for automatic measurement on knee CT images for the purpose of patellar dislocation diagnosis.

The major contributions of our work reside in the following aspects. Firstly, we propose two quantities, patellar tilt angle (PTA) and patellar lateral shift (PLS), to measure on the knee CT images. Secondly, in order to make the automatic measurement, we propose computing algorithms to segment the patella and femur regions in the CT images and measure the proposed quantities on the segmented regions. Finally, we make experiments to verify the validity and effectiveness of the measured results for the computer-aided diagnosis (CAD). Note that a preliminary version of our work has been published in reference [6]. Extending the preliminary work [6], we utilize the correlation between adjacent CT images by bone region prediction for bone region segmentation and make more complete experimental validation in terms of accuracy of measurement and applicability for CAD in this work.

2. Scheme Overview

The proposed scheme takes a specific portion of knee CT images as input and conducts a complete and automatic process of bone regions segmentation and geometric measurement.

2.1. Input Images

The source CT images for a patient are acquired by scanning the middle part of his or her leg. While being scanned, the patient may move his or her leg naturally through a range of knee angles, resulting in multiple sequences of CT images sampled at a preset temporal frequency. For each image sequence acquired at a time instance, we only use a portion that corresponds to cross sections through the femur and the patella. As an example, the anatomical structure of the middle part of a leg is shown in Figure 1(a) with the femur, patella, and tibia labelled. As shown in Figure 1(a), the portion of CT images that we use corresponds to the cross sections between the two planes as marked with blue parallelograms. Examples of the CT images in this portion are shown in Figure 1(b), and the ideal segmentation result for a CT image is shown in Figure 1(c) where the femur and the patella regions are neatly segmented and the narrow gap between them corresponds to the sutura.

We presume that the input CT images are ordered such that images of higher scanning positions on the leg go earlier in the sequence.

2.2. Working Process

We use Figure 2 to illustrate the automatic segmentation and measurement process. For an input CT image sequence as shown in Figure 2(a), we first segment the femur and the patella regions on each image to get the result as shown in Figure 2(b). Based on the profiles of the segmented regions, we use least squares fitting to find the central planes of the femur and the patella, respectively, as shown in Figure 2(c). Finally, we quantify the geometric relationship between the two central planes by PTA and PLS, which provide the basis for patellar dislocation diagnosis.

3. Segmentation of Femur and Patella Regions

Segmenting two solid bone regions corresponding to the femur and the patella, respectively, in each knee CT image is a key step in our scheme. It is a challenging task due to the following characteristics of knee CT images (as illustrated by Figure 3): (1) a single CT image usually contains responses of bones and other tissues (e.g., soft tissues) simultaneously and is contaminated with noises; (2) the patella and the femur regions may be very close to each another (e.g., only a couple of pixels apart or even locally fused) in many CT images; and (3) different parts (e.g., cortical bone tissue, spongy bone tissue, bone marrow, and bone cavity) inside a bone usually have different radiological densities, leading to highly variant gray levels of pixels in one bone region.

The generic problem of image segmentation has been researched for decades. For a survey of the early-days algorithms, we refer to references [7, 8]. Later on, with the development of medical imaging technology, intensive and specific efforts have been made to segment various types of medical images. The existent medical image segmentation algorithms can be classified as threshold-based methods [2], region-based methods [911], edge-based methods [12], active-contour-model-based methods [1322], hybrid methods [2326], and others [2732].

Among the various methods, the active-contour-model-based ones appear more advantageous to us. Relatively speaking, they handle structures with high topological complexity well and achieve subpixel accuracy and robustness against noise. In addition, they incorporate easily with other segmentation techniques and facilitate intuitive interaction [33, 34]. In particular, we choose the Chan–Vese (C-V) region-based active contour models [17] for our knee CT image segmentation, as it is in general less sensitive to initialization and noise than many other methods [1319] of its category. Further, according to our experiments (see Section 5), it yields better segmentation results than the other selected active-contour-model-based methods [18, 21, 22], when used with our proposed framework.

The existent image or medical image segmentation algorithms usually assume that the pixels inside a meaningful region have highly uniform levels of intensity. In addition, the contrast between meaningful and nonmeaningful regions and the noise level in an image also influence the segmentation results. These algorithms cannot be directly applied for our purpose due to the highly challenging characteristics of the knee CT images, as described earlier in the text. Therefore, we propose to improve the quality of the knee CT images first, by increasing the uniformity of pixel intensities, enhancing the contrast, and suppressing the noises, before making the final segmentation. Specifically, we process the CT images in an input sequence one by one in spatial order. For each CT image, we enhance its contrast to increase (decrease) the gray levels of the bone tissue (soft tissue and noise) pixels, predict the bone and sutura regions in it and modify its pixel values accordingly, utilizing the segmentation result of the previously processed CT image, if any, and employ the C-V region-based active contour method to make the final segmentation on the modified image. Details of these steps are given in the following sections.

3.1. Contrast Enhancement

The common global contrast enhancement method based on histogram manipulation does not work well for our case. The reason is that the pixels’ gray levels concentrate around very low and very high values (see Figure 4), leaving little room for contrast enhancement. Instead, we propose a contrast enhancement method based on local characteristics around each pixel. Observing that higher (lower) gray levels correspond to bone tissues (soft tissues and probably noise), we increase (decrease) the gray level of a pixel with brighter (darker) neighborhood.

Specifically, we perform a nonlinear scaling of each pixel’s gray level according to its neighboring pixels’ gray levels [6]. For a pixel, , we denote its gray level as and the gray levels of all the other pixels in ’s neighborhood as (). Assuming that the maximum gray level is 255, we update to according to

We find by experiments that the above process, when iterated for two or three passes, yields good results.

We show the effect of contrast enhancement on an example CT image in Figure 5, where the original image and the contrast-enhanced image are shown in Figures 5(a) and 5(b), respectively, and the enlarged views of the corresponding sutura areas are shown in Figures 5(c) and 5(d), respectively. Comparing Figures 5(a) and 5(b), we see that the bone pixels are emphasized while the soft tissue and noise pixels are suppressed in general. Comparing Figures 5(c) and 5(d), however, we find that the intensity of some pixels in the sutura region is unwantedly increased at the same time, reducing the gap between two bone regions and adding to the difficulty of bone regions segmentation. This issue is addressed by the proposed bone regions prediction technique, as described in the following section.

3.2. Prediction of Bone Regions

Narrow and vague gap between bone regions and inhomogeneous pixel intensity within bone regions are limiting factors for accurate bone region segmentation. In order to address these issues, we propose a bone region prediction process that further improves the CT image quality to facilitate accurate segmentation, as detailed below.

On any input CT image sequence used in our experiments, we observe two facts: firstly, the femur and the patella regions are relatively small and wide apart and contain highly homogeneous pixel intensity in the initial CT images of the sequence; secondly, the shape and the position of a bone’s profile vary only slightly between two adjacent CT images in the sequence. The former implies that we may apply a prevalent image segmentation algorithm on the first CT image (after contrast enhancement) in the sequence to obtain a good result, while the latter implies that a good segmentation result on a CT image may be utilized to predict the bone and sutura regions in the next image to be segmented.

Assume that we are currently processing the -th () original CT image, , in the sequence. After the contrast enhancement, we obtain . If , we simply use as the modified image, , which is to be segmented. Otherwise, we already have the segmentation result for the n-th image, which is a binary image, , with “255”-pixels for the bone regions and “0”-pixels for the background. Using , we improve the quality of by a proposed process of bone regions prediction to obtain , as detailed below.

Firstly, based on , we predict in the sutura region, , and the local bone region, , around the sutura, enabling us to treat these local regions with special care in the following steps. Specifically, in , we morphologically dilate the femur region, , and the patella region, , to and , respectively, by a disk with a radius of pixels. Empirically, we take . Locations in which correspond to nonbone pixels (“0”-pixels) in form the predicted sutura region, , and gives the predicted local bone region around the sutura in . An example of the local regions prediction is shown in Figure 6(a) with and colored in yellow and blue, respectively.

Secondly, we selectively revert pixels in which fall inside Q by , . This is based on the observation that the contrast enhancement tends to narrow the gap between the two bone regions (see Figures 5(c) and 5(d)), adding to the difficulty of bone regions segmentation.

Thirdly, in order to increase the bone regions’ density homogeneity, we combine with to obtain according towhere α, β, and are parameters to control the degree of fusion. We empirically use , , and . In extreme cases, if and if .

Fourthly, we reduce the intensities of the predicted local region, Q, around the sutura in , making a clearer separation of the two bone regions. This is achieved by selectively updating pixels in according towhere is a threshold set to the mean gray level of all our test CT images, and we empirically use and . By equation (3), we weaken pixels in Q whose original intensities are below a threshold. If a pixel’s original intensity is above the threshold, however, it is probably a bone pixel and we leave it untouched. Note that we weaken pixels in two subregions, and , since the predicted sutura region, , is usually not completely precise and we choose to weaken selected pixels in a wider local area, i.e., .

Lastly, based on , we predict in a thin layer, B, of pixels between the two bone regions when they get close to each other and set these pixels to “0” for further separation of the bone regions. Specifically, in , we morphologically dilate and by a disk with a radius of pixels to and , respectively, and obtain . Empirically, we take . An example is shown in Figure 6(b) with B colored in red. Depending on the shapes of and the distance between the two bones regions, there may be none, one, or multiple connected components in B.

3.3. C-V Region-Based Active Contour Segmentation

After () is modified to , we employ the C-V models to segment . The C-V model was originally proposed by Chan and Vese [17] and is based on the following energy model:where , , µ, and are constants, and represent the regions inside and outside contour C, respectively, and and correspond to the average pixel intensity in and , respectively. The solution of optimal contour C is reached by minimizing the energy function , resulting in an optimal segmentation of the image I.

As an example, we show an original CT image in Figure 7(a), the image modified by the contrast enhancement in Figure 7(b), and the image modified further by the bone regions prediction in Figure 7(c). We observe in Figure 7(c) that the bone region prediction leads to improved intensity homogeneity of the bone regions, suppressed soft tissue intensities, and well cleared sutura between the bone regions. Applying C-V models on the three images, we obtain the corresponding segmentation results as shown in Figures 7(d), 7(e), and 7(f), respectively. Comparing these three figures, we observe that the proposed contrast enhancement and bone regions prediction techniques lead to significantly improved segmentation results.

4. Automatic Measurement

In a segmented CT image, we expect to have two major regions with right shapes, corresponding to the femur and the patella, respectively. In rare cases, it may happen that more or less than two regions are segmented on a CT image or/and the shapes of segmented bone regions change tremendously between CT images, mostly due to low CT image quality. These cases can be easily detected based on the number of and the geometric properties (e.g., position and area) of the segmented regions. We simply discard these outlier cases and do not use them for measurement.

The CT images are acquired on parallel cross sections of the knee region, as shown in Figure 1. As such, we locate a few key points on each CT image on the boundaries of the femur region and the patella region, respectively, and then compute the central planes for the femur and the patella bones by optimally fitting those key points on the CT images.

4.1. Selection of Key Points

For the femur region in each CT image, we select three points as the key points: the two central valley points along the boundary and the middle of the leftmost and the rightmost points. For the patella region in each CT image, we select three points as the key points: the two central peak points along the boundary and the middle of the leftmost and the rightmost points. These key points can be easily identified through boundary tracking and inflection point detection. This key point selection scheme is illustrated in Figure 8(a).

4.2. Plane Fitting

The central plane of the femur (patella) bone is determined by optimally fitting a plane to the key points of the femur regions (patella regions) on the stack of CT images. In general, denoting the points as () and the plane equation as , the plane that optimally fits those points can be obtained bywhich can be solved with the least squares method.

4.3. PTA and PLS Measurement

We measure the patella tilt angle, θ, between the femur and the patella’s central planes, as illustrated in Figure 8(b). It is measured by the angle between the normals of the two bone’s central planes.

Further, we measure the patella lateral shift, D, between a pair of parallel approximate central planes of the femur and the patella, as illustrated in Figure 8(c). For this purpose, we fit a pair of parallel planes to the femur regions’ and the patella regions’ key points, respectively, and measure the distance, D, between the planes. Assuming that the equations of the two parallel planes are and , given the femur regions’ key points as () and the patella regions’ key points as (), the parallel plane fitting is done byusing the least squares method.

5. Results and Discussion

In the experiments, we conduct automatic segmentation and measurement on our dataset of knee CT images using the proposed scheme and validate the results of both the segmentation and the measurement.

5.1. Dataset

Our dataset is composed of fifteen patients’ knee CT images that were acquired using the Toshiba Aquilion ONE CT scanner in the affiliated hospital of Shandong University of TCM. Among the fifteen patients, ten are female and five are male. While being scanned, each patient was asked to move her/his legs freely from 0° to about 90°, and 22 CT image sequences were sampled at 22 time instances, one at each, during the scanning process. Each CT image sequence includes 320 images, 70 of which corresponding to the upper part of the leg (ref. Figure 1(a)) are used as the input to our system. The CT scanner is set up such that the thickness of each slice and the interval between two adjacent slices are both 0.5 mm, the default window width is 30 HU, the window level is 320 HU, and every CT image has a resolution of .

5.2. Validation of Bone Region Segmentation

In this section, we validate the bone region segmentation results both visually and quantitatively. In order to validate our choice of the C-V models [17], we compare with the following benchmark methods for image segmentation: the bias-corrected fuzzy c-means method (BCFCM) proposed by Mohamed et al. [35], the updated region-based active contour method using region-scalable fitting (RSF) energy function proposed by Li et al. [18], the level set method with bias field (LSEBFE) proposed by Li et al. [21], and the active contours driven by local image fitting energy (LIF) proposed by Zhang et al. [22]. For each image segmentation method, we run it both without and with our proposed framework, meaning that we run it both directly on the original CT images and on the CT images after modification with the approach proposed in Sections 3.1 and 3.2.

BCFCM modifies the objective function of the standard fuzzy c-means (FCM) algorithm to compensate for intensity inhomogeneities and allows the labeling of a pixel (voxel) to be influenced by the labels in its immediate neighborhood, which leads to better segmentation results than the standard FCM. RSF is a modified region-based active model using local intensity information at a controllable scale, which can preserve local details better and have higher robustness to intensity inhomogeneity. Note that BCFCM and RSF have been widely used in medical image segmentation. LSEBFE is a region-based level set method with bias field. It derives a local intensity clustering property of the image intensities and defines a local clustering criterion function, which are integrated with respect to the neighborhood center to give a global criterion of image segmentation. This criterion defines an energy in terms of the level set functions and a bias field that accounts for the intensity inhomogeneity of the image. It is more robust to initialization, faster, and more accurate than the well-known piecewise smooth model. LIF is a region-based active contour model that embeds the image local information. It uses Gaussian filtering for variational level set to regularize the level set function. It can not only ensure the smoothness of the level set function but also eliminate the requirement of reinitialization. Both LSEBFE and LIF are proposed to segment images with intensity inhomogeneities.

5.2.1. Visual Validation

In this section, we present the segmentation results of C-V, BCFCM, RSF, LSEBFE, and LIF on two representative challenging CT images, as shown in Figure 9. In the first image (in Figure 9(a)), the two bone regions are very close to each other while in the second image (in Figure 9(m)), there is more significant noise and weaker bone boundary response. Besides, both images have a high level of intensity inhomogeneity. In Figure 9, the first column shows the original CT images and the ground truth of their segmentations provided by experienced clinicians (i.e., Jiushan Yang, Shaoshan Wang, and Ruiqi Zou), and the following columns show the segmentation results by the five image segmentation methods, respectively. Further, the segmentation results in the first and the third rows are obtained with our framework (i.e., CT image modification followed by image segmentation) while those in the second and the fourth rows are obtained without our framework (i.e., they are obtained by direct image segmentation).

Comparing the segmentation results with and without our framework in Figure 9, we observe that for any of the image segmentation methods, our proposed framework promotes the performance by a large margin, leading to more neatly and accurately segmented femur and patella regions. This also demonstrates the robustness of the proposed framework to soft tissues responses, intensity inhomogeneities, and noises in the CT images. Comparing the segmentation results of all the five image segmentation methods with our framework, we observe that the C-V method is the most advantageous in terms of accuracy and smoothness of segmented bone boundaries, confirming our choice of the C-V method in the proposed scheme.

5.2.2. Quantitative Validation

For the quantitative validation, we randomly choose the automatic segmentation results on 30 CT images from each leg of each patient’s dataset and also manually mark the bone regions segmentation on each chosen CT image which is used as the ground-truth reference. Similar to Yao et al. [36], we use three metrics, i.e., overlap rate (), false-positive rate (), and Dice similarity coefficient (), to quantitatively validate the segmentation accuracy. On a CT image, if we denote the automatic and the ground-truth segmentations of a bone region as and , respectively, these metrics are defined as , , and . In addition, we measure the separation rate, , of the patella and the femur regions in the segmentation result. If they are completely separated, we set ; otherwise, .

In Table 1, we show the mean and standard deviation statistics of , , and and the mean statistics of for all the five image segmentation methods (C-V, BCFCM, RSF, LSEBFE, and LIF) with and without our framework (i.e., CT image modification followed by binary image segmentation) on all the test CT images. Here, , , and metrics are computed by treating the patella and the femur regions as one united bone region in each CT image.

From Table 1, we observe that (1) for any of the five methods, its mean , mean , and mean values are all increased and its mean value is decreased when our framework is used, showing the effectiveness of our proposed CT image modification technique; (2) for any of the five methods, its mean value with our framework reaches 100%, showing the effectiveness of our proposed bone regions prediction technique in separating the two bone regions; and (3) when used with our framework, the C-V method yields the largest mean , a value of 100%, the second largest mean , and the third smallest mean value and appears superior to the other methods considering all metrics overall.

In Table 2, we show the overlap rate and false-positive rate statistics for the femur and the patella regions on both legs of all the patients. For each bone region on each leg of each patient, we compute the two rates on all the 30 chosen CT images, average the rates over the 30 samples, and list the average in Table 2 where () and () mean the overlap rate and the false-positive rate of the femur (patella) region, respectively. Further, we compute the mean and the standard deviation on each of the , , , and statistics of each leg and place them at the bottom two rows in Table 2.

From Table 2, we observe that (1) for either leg and either bone region, the mean overlap rate is close to 95% and the mean false positive rate is close to 2%, showing the high accuracy of our bone regions segmentation scheme; (2) the standard deviations of the various rate statistics are all below or slightly above 3%, showing the stability and robustness of our bone regions segmentation scheme; and (3) rates of the same type on both legs are quite comparable, again confirming the stability and robustness of our bone regions segmentation scheme.

We show the statistics for the femur and the patella regions on both legs of all the patients in Table 3. For each bone region on each leg of each patient, we compute the on all the 30 chosen CT images, average them over the 30 samples, and list the average in Table 3.

From Table 3, we see that all the statistics are above or slightly below 0.96 and the standard deviations of the Dice coefficient are close to 0.02, further confirming the high accuracy, stability, and robustness of our bone region segmentation scheme.

5.3. Validation of PTA and PLS Measurement
5.3.1. Measured Angles and Distances

In this validation, we pick up the CT image sequences taken at four random time instances, , , , and , for four randomly picked patients’ left or right legs. For the CT image sequence at each time instance, we use our system to automatically measure the angle (i.e., the PTA), θ, and the distance (i.e., the PLS), D, between the two bones’ central planes. Note that when , we do not measure D. For the purpose of comparison, we ask several radiologists to measure the same parameters on the CT images manually. We use the unit of degree for angle measurement and the unit of millimeter for distance measurement. Note that for the automatic measurement, we have converted the unit of pixel to the unit of millimeter, knowing that one pixel corresponds to 0.95 millimeters in the photographing. The corresponding statistics is given in Table 4, from which we see that there is very little difference between the automatically and the manually measured angle numbers. Similarly, the automatically and the manually measured distances also closely match each other.

5.3.2. Diagnosis by Measured Results

We further test the accuracy and reliability of using the automatically measured results for diagnosis. According to orthopedists, the angle, θ, between the femur and the patella bones’ central planes provides the most important basis for patellar dislocation diagnosis. Thus, patellar dislocation may be straightforwardly diagnosed by comparing the measured angle against threshold values. Specifically, as an initial test, we set our system to automatically diagnose normal if , patellar subluxation if , and patellar dislocation if .

For this test, we have the dataset of 30 legs from 15 patients. For each leg, we use all the 22 CT image sequences and the average results of the 22 image sequences to make the diagnosis. Among the 30 samples, 11 samples (36.7%) are diagnosed as normal, 16 samples (53.3%) as patellar subluxation, and 3 samples (10%) as patellar dislocation, based on the automatically measured angles and the above-described diagnosis rule. On the same set of data, our orthopedists made diagnosis as well through manual measurement and clinic analysis and diagnosed 10 samples (33.3%) as normal, 17 samples (56.7%) as patellar subluxation, and 3 samples (10%) as patellar dislocation. Comparing the automatic and the manual diagnosis results, we find that the error rates of the automatic diagnosis on normal, patellar subluxation, and patellar dislocation are 9.1%, 7.1%, and 0%, respectively. Further, we visualize the distribution of the automatically measured angles with respect to the orthopedists’ manual diagnosis results in Figure 10. We see from Figure 10 that all the cases with are diagnosed by the orthopedists as patellar dislocation, and the majority of the cases with and are diagnosed by the orthopedists as normal and patellar subluxation, respectively. There is fuzziness only for a small portion of cases with θ closely around .

As a refined test, we further investigate the effectiveness of using distance as an auxiliary means for the patellar dislocation diagnosis. We only focus on the samples with , as there is fuzziness for samples with θ around in our initial test. For these samples, the distances between the two bones’ central planes are automatically measured and their distribution with respect to the orthopedists’ diagnosis results are plotted in Figure 11. From Figure 11, we see that a distance threshold of will accurately separate the cases of normal and patellar subluxation, thus eliminating the errors of diagnosis in our initial test where only angles are used.

6. Conclusions

In this work, we have developed a system for automatic segmentation and measurement on knee CT images. Firstly, on each CT image in an input sequence, we segment the femur and the bone regions; thereafter, we identify key points on the bone regions’ boundaries and conduct optimal fitting to obtain the central planes of the two bones; finally, angles and distances between the central planes are measured which can be used to assist doctors in patellar dislocation diagnosis.

Of the whole process, the biggest challenges exist with the bone region segmentation, due to the confusion from soft tissue responses and noises, inhomogeneity of bone region intensities, and close or even fused bone regions in the sutura area. To overcome these challenges, novel and effective methods are proposed to improve the quality of input CT images by enhancing the contrast of each CT image and predicting the bone regions in a CT image utilizing the coherence between adjacent CT images. The improved CT images are finally segmented using a region-based active contour method. The accuracy and robustness of the automatic segmentation and measurement results are validated in our experiments.

In the future, we will extend our system to measure more parameters as needed for the diagnosis. Furthermore, we will investigate reconstructing a 3D volume of the bones from the CT images and conduct measurements on this 3D volume with increased capability and flexibility.

Abbreviations

CT:Computerized tomography
CAD:Computer-aided diagnosis
PTA:Patellar tilt angle
PLS:Patellar lateral shift
C-V:Chan–Vese
BCFCM:Bias-corrected fuzzy c-means method
RSF:Region-scalable fitting
FCM:Fuzzy c-means
LSEBFE:Level set method with bias field
LIF:Local image fitting
OLP:Overlap rate
FPR:False-positive rate
Dice:Dice similarity coefficient
SR:Separation rate.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Disclosure

A preliminary version of this work has been presented in 2013 IEEE International Conference on Image Processing (https://ieeexplore.ieee.org/document/6738233).

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The authors thank Xian Wu for his help in rendering the images for Figure 2. This work was supported by the National Natural Science Foundation of China (grant no. 61872398).