Table of Contents Author Guidelines Submit a Manuscript
The Scientific World Journal
Volume 2014, Article ID 376569, 13 pages
http://dx.doi.org/10.1155/2014/376569
Research Article

Development of Vision Based Multiview Gait Recognition System with MMUGait Database

Faculty of Computing and Informatics, Multimedia University, 63100 Cyberjaya, Malaysia

Received 29 August 2013; Accepted 23 January 2014; Published 27 March 2014

Academic Editors: J. Hu and L. Xiao

Copyright © 2014 Hu Ng et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Abstract

This paper describes the acquisition setup and development of a new gait database, MMUGait. This database consists of 82 subjects walking under normal condition and 19 subjects walking with 11 covariate factors, which were captured under two views. This paper also proposes a multiview model-based gait recognition system with joint detection approach that performs well under different walking trajectories and covariate factors, which include self-occluded or external occluded silhouettes. In the proposed system, the process begins by enhancing the human silhouette to remove the artifacts. Next, the width and height of the body are obtained. Subsequently, the joint angular trajectories are determined once the body joints are automatically detected. Lastly, crotch height and step-size of the walking subject are determined. The extracted features are smoothened by Gaussian filter to eliminate the effect of outliers. The extracted features are normalized with linear scaling, which is followed by feature selection prior to the classification process. The classification experiments carried out on MMUGait database were benchmarked against the SOTON Small DB from University of Southampton. Results showed correct classification rate above 90% for all the databases. The proposed approach is found to outperform other approaches on SOTON Small DB in most cases.

1. Introduction

Biometrics is a way to identify individuals through their physical and behavioral characteristics such as fingerprint, gait, face, iris, and spoken speech. These characteristics are known as biometric identifiers that are distinctive and will be attached to a person permanently. Gait is a biometric modality that gained public recognition and is well accepted as security assessment. This is mainly because the user does not require any contact or intervention with the capturing device. Furthermore, it is still capable of identifying people at a distance even if other biometrics identifiers are intentionally obscured (hand gloves to cover finger print and mask to cover face).

The development of gait database has started since 1998 with the University of California San Diego (UCSD) gait database [1] which consists of six subjects with a total of 42 outdoor walking sequences. In 2002, the University of Southampton released the SOTON gait database [2] that contains the Large DB with 115 subjects and the Small DB with 11 subjects involving 15 covariate factors. The HumanID Gait Challenge [3] was released in 2005. It comprised 122 subjects with five major covariate factors (surface, shoes, carrying, camera angles, and time). In 2005, CASIA database B [4] was published with three covariate factors (bag, normal, and coat) with 11 different view angles. In 2012, the Treadmill Dataset B from OU-ISIR Gait Database [5] provided 64 subjects with 32 combinations of clothing variations while user was walking on treadmill. However, Lee and Hidler [6] demonstrated that there are differences in optical flow between subjects walking on treadmill and solid ground. As such, this database may not be suitable for gait recognition evaluation as it does not reflect real world scenarios.

While many gait databases are available, we argue that all of them have neglected one important challenge: none of the male subjects are wearing long fabrics covering the legs. Most of the databases contain male subjects that are only wearing trousers. The closest breakthrough is the SOTON database, where there are two female subjects (out of 115 subjects) wearing long blouse and Indian traditional garment (“salwar kameez”). On the other hand, OU-ISIR Treadmill Database B contains short skirt as a covariate factor for the participants. However, the short skirt is only partially covering the legs with the knees still clearly visible. In reality, males do wear long fabrics such as “sarong” and “kain samping” for ethnic Malays in South East Asia, long “dhoti” in Southern Asia, and “kilt” in Scotland. Due to this reason, it motivates us to build database with subjects wearing “sarong” and “kain samping”.

Our database consists of 82 subjects on normal condition for reliable performance evaluation on the large population and 19 subjects with 11 covariate factors as emphasis for the evaluation on the exploratory factor analysis of gait recognition. It contains side-view and oblique-view videos, the extracted silhouette frames, subject still face images, and ancillary data like subject information, cameras setup, and floor measurements.

We also propose a multiview gait recognition system, tested on this database, and benchmarked with the SOTON Small DB. This system comprises database construction, viewpoint normalization, gait feature extraction, feature processing, and classification. In the next Section, we present existing research works on gait recognition.

2. Related Work

Various research works have been conducted in gait recognition system. This section reviews the related literature on gait features extraction approaches and multi-view normalization approaches.

2.1. Gait Features Extraction Review

Generally, the approaches on gait features extraction are divided into two major groupings, which are model-based approaches and model-free approaches. Model-based approaches [7, 8] normally imitate the human body structure as two dimensional boxes. Then, it integrates with the knowledge of the body shape and match as model components. The capability to acquire the dynamic gait features from model parameters is the main advantage of these approaches. It is competent to liberate the background noise and is also able to avoid the bad effects from the changes of the camera view-points or subject’s apparel. Conversely, it generates complex model during the extraction process, which required high computational power, massive storage space, and huge experimental cost.

On the other hand, model-free approaches [912] simplify the entire human body to a concise representation using silhouette or skeleton. The significant advantages of these approaches are that they are related straightforward and only require minimum memory usage with low computational cost. Nevertheless, their performance is heavily defected by the background noise and the variations of the subject’s apparel.

In this paper, we employed model-based approach for the extraction of gait features. The approach is based on human body proportion [13], which is similar to the work of Goffredo et al. [10] and Yoo and Nixon [12]. Compared to their work, our approach is related more straightforward as it does not involve the measurement of gait cycle for kinematic parameter calculation. Moreover, manually labelling of the model template is not required, as performed by Goffredo et al.

2.2. Multiview Normalization Review

There are three major types of approaches that solved the issues in the multiview normalization, namely, view invariant gait feature, view synthesis, and view transformation.

In the first approach [1417], they are able to extract gait features that are invariant to changes in walking trajectory and camera view-point. However, these approaches can only be applied to a few limited viewing angles and their feature extraction processes can disrupted due to occlusion.

In the second approach [18, 19], they reconstruct gait motion by 3D information from calibrated multiple view-point cameras. They manage to generate precise synthetic images, but it requires heavy computational resource and complicated technical setup due to camera calibration.

In the third approach [20, 21], they reconstruct the gait features into the same view by mapping the relation between the gait features with the subject. However, these approaches propagate numerous noises during the reconstruction process, which degrades the recognition performance.

We employ perspective correction for view invariant gait feature extraction, which is comparable to Jean et al. [15]. However, we do not extract the spatiotemporal trajectories of body parts for gait modeling. We managed to mitigate the problems with missing heads or feet in the silhouettes as faced by Jean et al. In addition, our approach does not require the detection of half gait cycle as proposed by in Jean et al. We also do not attempt to detect each of the lower limbs. Thus, it can handle occluded silhouette either from self-occlusion or those occluded by apparels, which are normally disastrous for other view invariant gait feature approaches.

Our approach is also free from view synthesizing and camera calibration processes and is related more straightforward and faster than the view synthesis and view transformation approaches.

3. Methodology

According to Murray [22], it is not practical to measure pelvic and thorax rotations. Furthermore, they were found to be inconsistent after repeated tests. Thus, we only consider gait features from the lower limbs. Nonetheless, our system does not attempt to detect each of the lower limbs. Therefore, it can handle occluded silhouette which may be either due to self-occlusion or external occlusion, such as occlusion by subject’s apparel (long blouses or baggy trousers) or load carrying. These conditions are normally disastrous for model-free approaches. It is also straightforward, faster, and simpler than other model-free approaches.

In view of the fact that the static body parameters and the dynamics of human walking stance are the main structures in gait configuration [22], we present a model-based approach to extract subject’s height, width, step-size, and crotch height as the static features and compute joint angular trajectories as the dynamic features. Figure 1 illustrates the flow of the processes.

376569.fig.001
Figure 1: Proposed processes flowchart.
3.1. Acquisition of the MMUGait Database

The MMUGait database is part of the MMU GASPFA database [23]. It was captured and recorded in the Set and Background Studio located in the Faculty of Creative Multimedia, Multimedia University. The acquisition was done over a period of four days in December 2011 with involving 82 participants. Ethical approval was obtained from the participants by signing a standard university approval consent form prior to volunteering. The recording of MMUGait database was done in an indoor environment with green backdrop and white solid surface. Figure 2(a) shows the acquisition environment.

fig2
Figure 2: (a) The acquisition environment. (b) The recording layout.

Two SONY HDR-XR160E full HD video camera recorders (camcorders) were used during the filming. The recorded videos are in MPEG Transport Stream (MTS) format with resolution of 1920 (Height) × 1080 (Width) pixels. The video stream was captured using progressive scan with a frame rate of 50 frames per seconds (fps). The camcorders captured the walking sequence from two different view angles, which were side-view (frontal parallel) and 60° oblique-view.

The subjects walked back and forth on a track continuously and captured in both directions. Both ends of the walking track were the turning point that subjects must turn around and repeat the walking sequence in another direction. Figure 2(b) shows the recording layout.

There were 82 participants that took part in the recording of the MMUGait database. They were from different nationalities, genders, and age groups. There were 67 subjects from Malaysia, while the remaining 15 subjects came from six other countries (the Middle East, Central Asia, and Africa). For the gender class, 15 subjects were female, while the remaining 67 were male. 71 subjects were aged between 15 and 25 years, while the rest fell between the ages of 26–40 years.

There are two categories of database collected in this phase. The first set (MMUGait Large DB) contains 82 subjects that walked for at least 20 sequences in both directions with personal clothing (own shoes and own cloth). On the other hand, the second set (MMUGait Covariate DB) contains 19 subjects that walked in both directions wearing different types of clothes, shoes, and carrying various types of bags, with varying walking speed. This set includes 11 covariant factors, which are “sarong”, “kain samping”, personal clothing, hand bag (held in hand), barrel bag (slung over shoulder), barrel bag (carried by hand), rucksack, walking slowly, walking quickly, walking in flip flops, and bare feet. In total, there were approximately 110 sequences per subject.

To the best of our knowledge, this is the first paper that introduces biometric gait recognition with Malays’ traditional costumes, which are common attire for ethnic Malays in South East Asia, especially during Friday prayers and religious festivals. In this case, we recorded subjects with “kain samping” and “sarong” as covariate factors as changes in apparel. We believe that it can act as a benchmark database for performance evaluation by other gait recognition systems.

3.2. Data Processing

The original video format filmed was in MPEG transport stream format (MTS) with 50 frames per second (FPS). For further processing, all the recorded videos were converted to Audio Video Interleave (AVI) format with resolution of 1920 (High) * 1080 (Width) pixels. After that, the videos were extracted into individual frames in Joint Photographic Experts Group (JPEG) format.

3.3. Silhouette Generation

To extract the human silhouette from each extracted frame, background subtraction technique was employed. We have improved the conventional background subtraction technique by summing up the gray level results from (a) background image subtracted by foreground image and (b) foreground image subtracted by background image. This approach is able to amplify the difference between the foreground and the background. As a result, the foreground object is more distinctive.

Next, the pixels’ intensity was adjusted to increase the contrast of the foreground object. The process was carried out by obtaining optimal thresholding via Otsu’s method [24]. Once the threshold was found, it was used to rescale the gray level values to new values such that values between the lowest input value and the threshold were scaled to values between 0 and 1.

By applying the Otsu’s method [24] again, the gray scale image was then converted to binary image using the new threshold. In this case, if a pixel value was below the threshold, it was set to zero; otherwise, it was set to one in the binary image.

After that, the morphological operations with a 7 × 7 diamond shape structuring element are applied to enhance the generated foreground object. Morphological opening was applied to separate the shadow into isolated regions, while morphological closing is used to close the small gaps in the foreground object.

In some cases, noise still exists as pseudoobjects in the image. Connected component labeling is applied to label all regions within the image. Subsequently, those regions with area smaller than 1500 pixels are removed. Figure 3 shows the resulting images at different stages of silhouette generation.

fig3
Figure 3: (a) Foreground image. (b) Color image of foreground subtracts background. (c) Color image of background subtracts foreground. (d) Gray level image of (b). (e) Gray level of (c). (f) Addition of (d) and (e). (g) Foreground object after intensity adjustment. (h) Binary image of foreground object. (i) Foreground object after morphological opening. (j) Extracted silhouette.

The pseudocodes of the processes involved are described below.

Definition of Notations: running pixel;: domain of image;: background image;: foreground image;: foreground object with pixels of higher intensity than background image;: foreground object with pixels of lower intensity than background image;: foreground object in gray scale;: thresholds generated from Otsu’s method;: foreground object after contrast enhancement;: binary image of foreground object;: binary image after morphological opening with structuring element ;: binary image after morphological closing with structuring element ;: labeled image after connected component labeling;: labeled image after blobs removal;trun(): truncate negative pixel values to zero;rgb2gray(): convert RGB values to gray level values;Ostu(): Ostu’s thresholding technique;scale(): rescale the gray level values from 0 and to 0 and 1, respectively;concom(): connected component labeling;blobremove(): remove blobs with size smaller than ;

Pseudocode

Δ1() = trun Δ2() = trun Δ() = rgb2gray (Δ1()) + rgb2gray (Δ2())

.

3.4. View-Point Normalization

To normalize the oblique walking sequence into the side-view plane, the perspective correction technique is employed. First, all silhouettes in a walking sequence are superimposed into a single image, as shown in Figure 4(a).

fig4
Figure 4: (a) Superimposed silhouettes from one walking sequence. (b) Superimposed silhouettes after perspective correction.

Next, lines X and Y are drawn horizontally based on the highest and lowest point among the silhouettes. As the normal gait cycle is periodic, a sinusoidal line is formed when the highest points of all silhouettes in a walking sequence are connected. Line Z is then drawn by connecting the first peak and the last peak of the sinusoidal line.

The perspective correction technique consists of two stages: vertical and horizontal adjustments. For vertical adjustment, each silhouette is vertically stretched from line Z towards line X. In addition, each silhouette is also vertically stretched from the bottom towards line Y. Figure 4(b) shows superimposed silhouettes after perspective correction.

3.5. Feature Extraction

Several processes were involved in the extraction of the gait features from the human silhouette images. The details of these processes are discussed in the following subsections.

3.5.1. Enhancement of Silhouettes

Occasionally, the foreground object segmentation does not work well due to noise. As a result, the boundaries of the segmented human silhouette are coarse with jagged edges. Sometime, spurious regions may also present near the border of the human silhouette or even in the background. If horizontal lines are drawn across the human silhouette for automatic body joint identification, the jagged edges and spurious regions will hinder the process as they may be erroneously considered as parts of the human silhouette.

To mitigate these problems, the width of the regions crossed by the horizontal lines should be determined. To determine the width of these regions, the distance between the rising edge and falling edge of a region is determined. If the width of a region is less than a threshold value, that region is considered as invalid and will be excluded during the body joint identification process. In the proposed technique, the threshold value has been empirically determined as five percent of the human silhouette height. Figure 5 shows the improvement on joint detection with jagged edge exclusion. In Figure 5(a), there is a jagged edge at the left side of the back knee. When the proposed body joint identification technique draws a horizontal line across the human silhouette at knee height, the horizontal line crosses the jagged edge as well as both knees. The horizontal line is then scanned from left to right and the middle points of the first two regions that it crosses are identified as the center of the knees. In this example, the jagged edge has been erroneously identified as the center of a knee while the actual center of the knee at the front leg has been completely ignored even though the jagged edge is only a few pixels wide. By excluding the regions narrower than the threshold value, the jagged edge has been excluded from the body joint identification process and the centers of both knees have been correctly identified as shown in Figure 5(b).

fig5
Figure 5: (a) Results without jagged edge exclusion. (b) Results with jagged edge exclusion.
3.5.2. Bounding Box Measurement

After that, the height () and width () of the body are obtained from the bounding box of the improved silhouette. Figure 6 shows the two extracted gait features.

376569.fig.006
Figure 6: The height and width of a human silhouette with the image profile of horizontal line drawn through the hip.
3.5.3. Hip Joint Estimation

By referring to a priori knowledge of the body proportions [13], the vertical position of hip, knee, and ankle are estimated as 0.48, 0.285, and 0.039 with referring to the body height . Once these are known, the horizontal position of these body joints can be determined. Figure 6 shows a horizontal dashed line that goes through the hip with the resulting image profile shown below the silhouette. The horizontal center position of the hip can be found by determining the midpoint between the positive and negative edge by applying the equation below: where is the horizontal position of the positive edge, is the horizontal position of the negative edge, and is the horizontal center position of the hip.

3.5.4. Knee and Ankle Joints Estimation for Nonoccluded Silhouettes

To locate the center horizontal positions of the knees, a horizontal line is outlined at knee height all the way through the silhouette. The silhouette without self or external occlusion, four edges can be found on the image profile along this horizontal line, as indicated by two dots beside each knee in Figure 7(a). The horizontal center knee positions are discovered by computing the midpoint flanked by two adjacent edges on each knee by the equations below: where and are the horizontal center positions of the front and back knee for normal silhouette, and are the horizontal positions of the positive edge on the front and back knee, and and are the horizontal positions of the negative edge on the front and back knee.

fig7
Figure 7: (a) Knee positions on normal silhouette. (b) Ankle positions on normal silhouette. (c) Knee positions on self-occluded silhouette. (d) Ankle positions on self-occluded silhouette.

To locate the horizontal center position of the ankles, the same method is applied. If a horizontal line is outlined at ankle height, four edges are found on the image profile along the horizontal line on a normal silhouette, as illustrated in Figure 7(b). As a result, the horizontal center ankle positions will be detected by employing the subsequent equations where and are the horizontal center positions of the front and back ankle for normal silhouette, and are the positive edge on the front and back ankle, and and are the negative edge on the front and back ankle.

3.5.5. Knee and Ankle Joints Estimation for Occluded Silhouettes

For occluded silhouettes, the horizontal center knee positions are located by determining the midpoint between each edge with referring to the horizontal center position of the hip (as only two edges can be found on the image profile), as shown in Figure 7(c). Consider where and are the horizontal center positions of the front and back knee for occluded silhouette, is the horizontal center position of the hip, is the horizontal position of the positive edge, and is the horizontal position of the negative edge on the corresponding image profile.

Since there are only two edges on the image profile, as highlighted in Figure 7(d), the horizontal center ankle positions are found by determining the midpoint flanked by both edges with employing the equations as where and are the horizontal center positions of the front and back ankle for occluded silhouette, and are the horizontal positions of the positive and negative edge on the image profile, and 0.25 and 0.75 are chosen to compute the first quarter and third quarter points between these edges. is not used in the formulation as it does not reflect the middle point between and .

3.5.6. Joint Angular Trajectory, Step-Size, and Crotch Height Calculation

Figure 8(a) illustrates in what manner the joint angular trajectory is determined from two joints. The joint angular trajectory () can be determined by using the following equation: where , , and are the -coordinates of joint , , and , respectively, and , , and are the -coordinates of joint , , and , respectively.

fig8
Figure 8: (a) Computation of joint angular trajectory. (b) Nine extracted gait features.

In our system, five joint angular trajectories have been extracted, as there are the five main joints on the limbs. These angular trajectories are hip angular trajectory (), front knee angular trajectory (), back knee angular trajectory (), front ankle angular trajectory (), and back ankle angular trajectory ().

The Euclidean distance between the ankles is used to represent the subject’s step-size (). Then, the Euclidean distance between the ground and the subject’s crotch is being calculated as crotch height (CH). If the crotch height is found lower than the height of knee, we will assumed that it is equal to zero, as the crotch is considered occluded. Figure 8(b) shows a sample of a human silhouette with the extracted nine gait features.

To construct the feature vector, maximum hip angular trajectory () was determined during a walking sequence. When was identified, the corresponding , , , , , , , and CH were also determined. To better describe the human gait, 24 features were used to construct the feature vector as shown: where , , , , , , , , and are the average of the local maxima detected for width, height, crotch height, hip angular trajectory, front knee angular trajectory, back knee angular trajectory, front ankle angular trajectory, back ankle angular trajectory, and step-size, respectively; , , , , , and are the ratio of , , , CH, , and to , respectively.

3.6. Features of Smoothing and Normalization

As the presence of outliers in the extracted features would hinder the classification process, Gaussian filter with sigma values () equal to 2.5 is applied to remove them. In order to normalize the extracted features from various dimensions to be independent and standardized, linear scaling technique [25] has been applied to normalize each feature component to the range between 0 and 1.

3.7. Features Selection

The performance of a recognition system is determined by the effectiveness of the selected features, which can maximize interclass variance. In other contexts, the redundant and inappropriate features which degrade the classification rate would be found and removed.

In the proposed system, Ranker algorithm proposed by Hall et al. [26] is used to rank features by their individual evaluations, which helps to identify those extracted features that contribute positively in the recognition process. Based on the scores obtained, all twenty-four features have exhibited positive contribution. Thus, all of them are used in our system. Figure 9 shows examples of successful joint detection from self-occluded silhouettes and silhouettes with external occlusion.

fig9
Figure 9: Occluded human silhouettes with joints detection. (a) Wearing “kain samping”. (b) Wearing “sarong”. (c) Barrel bag slung over shoulder. (d) Walking in bare feet. (e) Carrying rucksack. (f) Carrying barrel bag by hand.
3.8. Classification Techniques

To study the performance of our gait recognition system, four classification techniques were applied to find the best correct classification rate and to verify the consistency of the results. In our work, multiclass Support Vector Machine (SVM), Back-propagation artificial neural network (BPANN), Fuzzy k-nearest neighbor (Fuzzy k-NN) with Euclidean distance metrics, and Linear Discriminant Analysis (LDA) classifiers are employed.

For SVM, experiments were carried out to examine the effects on kernel functions—Linear (Ln), Polynomial (Poly), and Radial Basis Function (RBF). The kernel’s parameters such as (degree), (gamma), and (coefficient) and regularization parameter were trained to find the best correct classification rate. For Fuzzy k-NN, numerous numbers of neighbors have been tested. For BPANN, various numbers of hidden layers have been trained to find the best classification rate.

As cross-validation procedure is essential to evaluate the accuracy of the classification performance. We employed tenfold cross-validation for this project, where the walking sequences from the gait databases were randomly divided into ten disjoint subsets, nine subsets used for analysis training and one subset used for validation. The cross-validation process was iterated for 10 turns with features vectors of each disjointed subset channeled into classifiers as the validation test. Then, the single mean correct classification rate can be obtained by averaging the cross-validation results.

The experiments were carried out on four classification techniques with various optimization parameters that were obtained during the training. Three quality measures were used in the experiment: correct classification rate (CCR), true positive rate (TPR), and false positive rate (FPR).

4. Experimental Results and Discussion

In order to evaluate the performance of the proposed gait recognition system on the new database, numerous experiments have been conducted. This section presents and discusses the results of these experiments which were aimed to assess the recognition rate of the proposed system with respect to view normalization, large population, and covariate factors.

Three databases were employed for performance evaluation; MMUGait Large DB, MMUGait Covariate DB and one database from SOTON Small DB for comparison. In the evaluation of each database, the analysis is performed on walking sequences captured from side-view (Side), normalized oblique-view (NorOb) and a combination of both views (Com). All walking sequences from three databases were used during the training and testing stages.

For group covariate factor analysis, the walking sequences were categorized into five groups: Group 1 (G1)—different speeds; Group 2 (G2)—variety of shoes; Group 3 (G3)—various objects carrying; Group 4 (G4)—various types of apparel; Group 5 (G5)—personal clothing without carrying any object.

4.1. Experimental Results of MMUGait Large DB

The performance is evaluated on 80 subjects from MMUGait Large DB. The side-view consists of 2961 walking sequences, the normalized oblique-view consists of 2843 walking sequences and the combination of both views will give a total of 5804 walking sequences. The overall CCR results are summarized in Figure 10. The best overall performance came from SVM RBF with CCRs of 95.4%, 91.6% and 93.3% for side-view, normalized oblique-view and combination of both views, respectively.

376569.fig.0010
Figure 10: CCRs of MMUGait Large DB.
4.2. Experimental Results of MMUGait Covariate DB

The performance is evaluated on 19 subjects from MMUGait Covariate DB. The side-view consists of 3780 walking sequences, normalized oblique-view consists of 3713 walking sequences, and the combination of both views will give a total of 7493 walking sequences. The overall CCR results are summarized in Figure 11. In overall, the best performance came from SVM RBF with CCRs of 96.0%, 93.6%, and 94.2% for side-view, normalized oblique-view and, combination of both views, respectively.

376569.fig.0011
Figure 11: CCRs of MMUGait Covariate DB.

Spurred by the encouraging results, subsequent experiments using SVM RBF were conducted to evaluate the system on various covariate factors. The overall CCRs for group and individual covariate factor are summarized in Figures 12 and 13, respectively.

376569.fig.0012
Figure 12: CCRs of group covariate factors for MMUGait Covariate DB.
376569.fig.0013
Figure 13: CCRs of individual covariate factor for MMUGait Covariate DB.
4.3. Experimental Results of SOTON Small DB

The performance is evaluated on the complete 11 subjects from SOTON Small DB [2]. The side-view consists of 3178 walking sequences, normalized oblique-view consists of 3036 walking sequences, and the combination of both views will give a total of 6214 walking sequences. The overall CCR results are summarized in Figure 14. The best overall performance came from SVM RBF with CCRs of 96.0%, 92.5%, and 92.9% for side-view, normalized oblique-view, and combination of both views, respectively.

376569.fig.0014
Figure 14: CCRs of SOTON Small DB.

Motivated by the encouraging results, subsequent experiments using SVM RBF were conducted to evaluate the performance on multiple covariate factors. The overall CCR results for group and individual covariate factor are summarized in Figures 15 and 16, respectively.

376569.fig.0015
Figure 15: CCRs of group covariate factors for SOTON Small DB.
376569.fig.0016
Figure 16: CCRs of individual covariate factor for SOTON Small DB.
4.4. Results Discussion

In general, our gait recognition system managed to obtain high CCRs in all the experiments. All the highest CCRs from each experiment are above 90%. The TPRs obtained are identical to the CCRs. The system achieved low FPRs, which are in the range of 0.1% to 2.6%.

From Figures 12 and 15, it can be observed that our system is robust to covariant factors as it has resulted in high CCRs. For that reason, we found that Group G1 generated high CCR as the duration of the walking cycle was not included as a feature. Similarly, the CCRs generated from group G2 and G3 were high as well. This showed that shoe and bags did not affect the extracted feature. In contrast, group G4 generated lower CCRs; this was because of the crotch height that was unidentified due to the occlusion by apparel. Group G5 presented the best results as there was no covariate factor involved.

From Figures 13 and 16, it can be observed that our system is able to provide high CCRs even when the subjects were wearing long fabrics or carrying objects. However, the clothing factor (“sarong”) in MMUGait Covariate DB and the objects carrying factor (carrying handbag) in SOTON Small DB have resulted the lowest CCRs among other covariates factors. This is mainly because it was not possible to measure the crotch height of the subject due to the occlusion by clothing and bag.

In general, SVM outperforms all the other three classification techniques. In particularly, the nonlinear SVM with RBF or Poly kernel outperforms SVM with Ln kernel. As our generated gait feature vectors are not linear, the kernel trick in the non-linear SVM permits the algorithm to adapt the maximum-margin hyperplane in a transformed feature space [27]. As expected, the RBF kernel produces better results as it normally reigned over Poly kernel [28]. Overall, nonlinear SVM outperforms BPANN, as the BPANN suffers from the existence of multiple local minimal solutions. However, the solution of SVM is unique and global [29].

LDA performed poorer than the other three classification techniques. This may be due to the reason that the hyperplane computed by LDA is only optimal when the covariance matrices for all of the classes are identical [30], whereas the performance of Fuzzy k-NN was stable and it was the second best among the four techniques.

To the best of our knowledge, no other researcher has published results on oblique-view SOTON Small DB. We can only compare finding on side-view SOTON Small DB. Table 1 shows the comparison of our best experiment results (SVM with RBF) with other gait recognition systems in detail.

tab1
Table 1: Comparison with other systems employing SOTON Small DB.

From Table 1, the highest CCR (95.97%) obtained outperforms the results obtained by Bouchrika and Nixon [7] and Pratheepan et al. [11]. The poorer result by Bouchrika and Nixon is due to the requirement to manually label model template to describe joints’ motion. Conversely, our results are better than Pratheepan et al., as we do not incorporate the selection or estimation of gait cycle. Furthermore, we are the only group that have tested the complete database with 11 subjects, 15 covariate factors, and 3178 walking sequences compared with Bouchrika and Nixon (10 subjects, 11 covariate factors, and 440 walking sequences), Pratheepan et al. (10 subjects, 4 covariate factors, and 180 walking sequences), and Bashir et al. (11 subjects, unknown number of covariate factors, and 373 walking sequences).

Even though Bashir et al. [9] produced better results, it is inappropriate to compare our results with them as they have only tested about 10% of the full walking sequences and the total numbers of the covariate factors being employed is unclear.

5. Conclusion

We presented an automated multiview gait recognition system by employing hybrid approach (model-free and model based). It was tested on a new gait database, MMUGait which consists of new variations in clothing. In the proposed approach, the joint angular trajectories can be computed from the detected body joints even on self-occluded or external occluded silhouettes. Thus, it has shown to be more effective than other approaches in related research.

In addition, the high CCRs and TPRs and low FPRs also show that it is robust and can achieve good performance either in gait databases with various covariate factors or large population of subjects and multiple view angles. For upcoming research, more gait databases will be tested by the proposed approach to evaluate its performance.

The authors are planning to allow public access to the MMUGait database in the near future. We believe that the walking sequences with special apparels will be invaluable for performance evaluation of other gait recognition systems.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

The authors would like to thank all the anonymous participants, voluntarily helpers, and the university in this research work. The authors would also like to thank Professor Mark Nixon, University of Southampton, United Kingdom, for providing the databases in this research work.

References

  1. J. Little and J. Boyd, “Recognizing people by their gait: the shape of motion,” Videre: Journal of Computer Vision Research, vol. 1, no. 2, pp. 1–32, 1998. View at Google Scholar
  2. J. Shutler, M. Grant, M. S. Nixon, and J. N. Carter, “On a large sequence-based human gait database,” in Proceedings of the 4th International Conference on Recent Advances in Soft Computing, pp. 66–71, Nottingham, UK, 2002.
  3. S. Sarkar, P. J. Phillips, Z. Liu, I. R. Vega, P. Grother, and K. W. Bowyer, “The humanID gait challenge problem: data sets, performance, and analysis,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 27, no. 2, pp. 162–177, 2005. View at Publisher · View at Google Scholar · View at Scopus
  4. S. Yu, D. Tan, and T. Tan, “A framework for evaluating the effect of view angle, clothing and carrying condition on gait recognition,” in Proceedings of the18th International Conference on Pattern Recognition (ICPR '06), pp. 441–444, Hong Kong, August 2006. View at Publisher · View at Google Scholar · View at Scopus
  5. Y. Makihara, M. Hidetoshi, T. Akira et al., “The OU-ISIR gait database comprising the treadmill dataset,” IPSJ Transactions on Computer Vision and Applications, vol. 4, pp. 53–62, 2012. View at Publisher · View at Google Scholar
  6. S. J. Lee and J. Hidler, “Biometrics of overground vs. treadmill walking in healthy individuals,” Journal of Applied Physiology, vol. 104, no. 3, pp. 747–755, 2008. View at Google Scholar
  7. I. Bouchrika and M. S. Nixon, “Exploratory factor analysis of gait recognition,” in Proceedings of the 8th IEEE International Conference on Automatic Face and Gesture Recognition, pp. 1–6, September 2008. View at Publisher · View at Google Scholar · View at Scopus
  8. D. Cunado, M. S. Nixon, and J. N. Carter, “Automatic extraction and description of human gait models for recognition purposes,” Computer Vision and Image Understanding, vol. 90, no. 1, pp. 1–41, 2003. View at Publisher · View at Google Scholar · View at Scopus
  9. K. Bashir, T. Xiang, S. Gong, and Q. Mary, “Gait representation using flow fields,” in British Machine Vision Conference, pp. 113.1–113.11, BMVA Press, Dundee, UK, 2009. View at Publisher · View at Google Scholar
  10. M. Goffredo, I. Bouchrika, J. N. Carter, and M. S. Nixon, “Self-calibrating view-invariant gait biometrics,” IEEE Transactions on Systems, Man, and Cybernetics B: Cybernetics, vol. 40, no. 4, pp. 997–1008, 2010. View at Publisher · View at Google Scholar · View at Scopus
  11. Y. Pratheepan, J. V. Condell, and G. Prasad, “Individual identification using Gait sequences under different covariate factors,” in Computer Vision Systems, Fritz, M. Schiele B, and J. H. Piater, Eds., vol. 5815 of Lecture Notes in Computer Science, pp. 84–93, Springer, Berlin, Germany, 2009. View at Publisher · View at Google Scholar
  12. J. H. Yoo and M. S. Nixon, “Automated markerless analysis of human gait motion for recognition and classification,” ETRI Journal, vol. 33, no. 2, pp. 259–266, 2011. View at Publisher · View at Google Scholar · View at Scopus
  13. W. T. Dempster and G. Gaughran, “Properties of body segments based on size and weight,” The American Journal of Anatomy, vol. 120, pp. 33–54, 1967. View at Publisher · View at Google Scholar
  14. I. Bouchrika, M. Goffredo, J. Carter, and M. S. Nixon, “Covariate analysis for view-point independent Gait recognition,” in Advances in Biometrics, M. Tistarelli and M. S. Nixon, Eds., vol. 5558 of Lecture Notes in Computer Science, pp. 990–999, Springer, Berlin, Germany, 2009. View at Publisher · View at Google Scholar
  15. F. Jean, A. B. Albu, and R. Bergevin, “Towards view-invariant gait modeling: computing view-normalized body part trajectories,” Pattern Recognition, vol. 42, no. 11, pp. 2936–2949, 2009. View at Publisher · View at Google Scholar · View at Scopus
  16. A. Kale, A. K. R. Chowdhury, and R. Chellappa, “Towards a view invariant Gait recognition algorithm,” in Proceedings of the IEEE Conference on Advanced Video and Signal Based Surveillance, pp. 143–150, Miami, Fla, USA, 2003. View at Publisher · View at Google Scholar
  17. C. S. Lee and A. Elgammal, “Towards scalable view-invariant Gait recognition, multilinear analysis for Gait,” in Audio-and Video-Based Biometric Person Authentication, T. Kanade, A. Jain, and N. K. Ratha, Eds., vol. 3546 of Lecture Notes in Computer Science, pp. 395–405, Springer, Berlin, Germany, 2005. View at Publisher · View at Google Scholar
  18. R. Bodor, A. Drenner, D. Fehr, O. Masoud, and N. Papanikolopoulos, “View-independent human motion classification using image-based reconstruction,” Image and Vision Computing, vol. 27, no. 8, pp. 1194–1206, 2009. View at Publisher · View at Google Scholar · View at Scopus
  19. G. Shakhnarovich, L. Lee, and T. Darrell, “Integrated face and gait recognition from multiple views,” in Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, pp. I439–I446, Hawaii, Hawaii, USA, December 2001. View at Publisher · View at Google Scholar · View at Scopus
  20. W. Kusakunniran, Q. Wu, H. Li, and J. Zhang, “Multiple views Gait recognition using view transformation model based on optimized Gait energy image,” in Proceedings of the IEEE 12th International Conference on Computer Vision Workshops, pp. 1058–1064, Kyoto, Japan, October 2009. View at Publisher · View at Google Scholar · View at Scopus
  21. Y. Makihara, R. Sagawa, Y. Mukaigawa, T. Echigo, and Y. Yagi, “Gait recognition using a view transformation model in the frequency domain,” in Computer Vision, A. Leonardis, H. Bischof, and A. Pinz, Eds., vol. 3953 of Lecture Notes in Computer Science, pp. 151–163, Springer, Berlin, Germany, 2006. View at Publisher · View at Google Scholar
  22. M. P. Murray, “Gait as a total pattern of movement,” The American Journal of Physical Medicine, vol. 46, no. 1, pp. 290–333, 1967. View at Google Scholar · View at Scopus
  23. C. C. Ho, H. Ng, W. H. Tan et al., “MMU GASPFA: a COTS multimodal biometric database,” Pattern Recognition Letters, vol. 34, no. 15, pp. 2043–2050, 2013. View at Publisher · View at Google Scholar
  24. N. Otsu, “A threshold selection method from gray-level histograms,” Automatica, vol. 9, no. 1, pp. 62–66, 1979. View at Publisher · View at Google Scholar
  25. S. Aksoy and R. M. Haralick, “Feature normalization and likelihood-based similarity measures for image retrieval,” Pattern Recognition Letters, vol. 22, no. 5, pp. 563–582, 2001. View at Publisher · View at Google Scholar · View at Scopus
  26. M. Hall, E. Frank, G. Holmes, B. Pfahringer, P. Reutemann, and I. Witten, “The WEKA data mining software: an update,” ACM SIGKDD Explorations Newsletter, vol. 11, no. 1, pp. 10–18, 2009. View at Publisher · View at Google Scholar
  27. J. Suykens, V. Gestel, J. D. Brabanter, B. D. Moor, and J. Vandewalle, Least Squares Support Vector Machines, World Scientific, Singapore, 2002.
  28. H. Byun and S. W. Lee, “A survey on pattern recognition applications of support vector machines,” International Journal of Pattern Recognition and Artificial Intelligence, vol. 17, no. 3, pp. 459–486, 2003. View at Publisher · View at Google Scholar · View at Scopus
  29. B. Schölkopf and C. Burges, Advances in Kernel Methods: Support Vector Learning, MIT Press, Boston, Mass, USA, 1999.
  30. I. Gokcen and J. Peng, “Comparing linear discriminant analysis and support vector machines,” in Advances in Information Systems, T. Yakhno, Ed., vol. 2457 of Lecture Notes in Computer Science, pp. 104–113, Springer, Berlin, Germany, 2002. View at Publisher · View at Google Scholar