Abstract

Object detection in real images is a challenging problem in computer vision. Despite several advancements in detection and recognition techniques, robust and accurate localization of interesting objects in images from real-life scenarios remains unsolved because of the difficulties posed by intraclass and interclass variations, occlusion, lightning, and scale changes at different levels. In this work, we present an object detection framework by learning-based fusion of handcrafted features with deep features. Deep features characterize different regions of interest in a testing image with a rich set of statistical features. Our hypothesis is to reinforce these features with handcrafted features by learning the optimal fusion during network training. Our detection framework is based on the recent version of YOLO object detection architecture. Experimental evaluation on PASCAL-VOC and MS-COCO datasets achieved the detection rate increase of 11.4% and 1.9% on the mAP scale in comparison with the YOLO version-3 detector (Redmon and Farhadi 2018). An important step in the proposed learning-based feature fusion strategy is to correctly identify the layer feeding in new features. The present work shows a qualitative approach to identify the best layer for fusion and design steps for feeding in the additional feature sets in convolutional network-based detectors.

1. Introduction

Object detection in natural scenes is an important problem that drives many real-life applications. In the last few decades, a significant amount of research effort has gone into the understanding of object detection problem with general as well as domain-specific challenges [15]. As a result, several novel and innovative object detection methods have been developed. Despite continuous efforts from researchers from diverse backgrounds, the present state of the art is far from satisfactory as seen in recent results on standard datasets PASCAL-VOC [6] and MS-COCO [7].

Most recent works on object detection have invariably applied the convolutional neural networks (CNNs) based detector model. These models are learned end-to-end addressing feature extraction, parameter learning, and postprocessing in one- or at most two-stage training process. These methods have significantly improved the state of the art in object detection. The representational capability of features extracted by a CNN depends on the complexity of the detection problem in testing time and variability captured within the training dataset. In real-world applications, a target object’s appearance undergoes significant variations due to view angles, lighting, background clutter, and occlusions. Handcrafted features that are designed with domain understanding have often shown to be more distinctive and reliable in many situations. An intelligent fusion of both the modalities of features is expected to achieve better detection performance. In this work, we propose feature enhanced CNN based object detection framework by learning-based fusion of handcrafted features with deep features in the embedding space. We use fundamental color channels: RGB, HSV, and LBP in combination with gradient and orientation histograms to enhance deep features for overcoming the prediction errors due to appearance variations. To effectively combine handcrafted features with deep learning-based features, we also investigate the problem of identifying the optimum layer to inject the handcrafted features for feature fusion. Our detection framework is based on state-of-the-art YOLO [8] architecture where we inject the handcrafted features at appropriate layer(s) to regularize network weights during the training procedure. We hypothesise that fusing handcrafted features during network learning guides the CNN to extract a more accurate and robust feature set by exploiting upon the complementary information available in the handcrafted feature set. The major contributions in this paper are summarized as follows:(1)We propose a novel object detection approach based on CNN learning by fusing simple feature descriptors like color channels and gradient histograms [9] by learning-based fusion to train a more robust and accurate detection model.(2)We use the latest YOLO objection detection architecture as our base. We describe the proposed learning-based feature fusion strategy that uses a qualitative approach to select the best layer for feature injection. In this work, the presented work presents a novel strategy to fuse features in CNN based object detectors.(3)We demonstrate the comparative performance of the proposed feature fusion approach on PASCAL-VOC and MS-COCO datasets which shows improvement in the original YOLO object detection rate by a significant measure.

The paper is organized as follows: Section 2 presents the relevant previous research works. We briefly discuss the latest YOLO architecture and Integral channel features in Section 4. Section 5 discussed the details and implementation issues of the proposed object detection method. Experimental evaluation of the proposed object detection framework on different datasets is presented in Section 6. We conclude the paper in Section 7 presenting the perspective of our work and future research directions.

Our work in this paper and the proposed approach therein are related to the CNN based object detection using vision. There has been a significant amount of work in this research area. A thorough review of the literature on object detection and CNN based classification and regression is beyond the scope of this paper. The following discussion reviews the prominent and critical works in the context of the proposed approach for object detection.

Object detection in natural scenes can be pursued using different approaches: model-based, learning-based, and auxiliary [10]. Model-based methods depend on heuristic determined models using color, shape, and intensity attributes. The object characterization in such methods is also referred to as handcrafted features where they are based on the experts’ engineering features that best describe representations in images. Learning-based methods are where features from objects are automatically learned from classifiers and are later used for detection. Last, the auxiliary approach is where the location of objects is used as prior information for visualizing them. The auxiliary approach is expensive to deploy as the infrastructure of objects needs to be replaced for them to be effective. Prior knowledge of the object’s location reduces the computational search cost and improves accuracy. Furthermore, it helps in eliminating a large number of false positives.

The last few decades of advancements in the state of the art in object detection solutions have seen novel design of handcrafted features such as histogram of gradients (HOG) [11], scale-invariant feature transform (SIFT) [12], and pyramid HOG [13]. These methods used simple discriminative classifiers by scanning the image space, or by feature matching. Jones et al. [14] used integral images for feature extraction for face detection which can be quickly computed. These works generate fixed or variable size object descriptors as a set of local feature vectors. In [15], the authors proposed local-contour based features with two-staged, partially supervised learning for object detection. Among the earliest works on learning-based object detection, the authors in reference [16] proposed learning a sparse part-based object representation. In [17], the authors introduced multiple-component learning-based parts based object detection by automatic learning and combination of individual classifiers. The authors in reference [9] subsequently combined HOG with heterogeneous color channels for pedestrian detection. These features widely known as integral channel features (ICF) again utilize integral images for fast feature computation which is an important requirement for image scan based object detection.

In [18], the authors proposed to learn structure between different object parts using SVM formulation for modeling of object structures at multiple scales using mixtures of deformable part models. An extensive review of pedestrian detection using handcrafted features has been presented by [1]. In [19], the visual attention based modeling is used for salient object detection by using a bootstrap learning model. In [20], the authors proposed regionlets—the integration of different types of features that were computed locally. These features were used to model an object class by a cascaded boosting classifier.

With the recent rise in CNN based learning methods, several CNN based object detection models have been proposed. Girshick et al. [21] proposed region-based convolutional neural networks (R–CNN) as object detection models. The R–CNN model trained independent components for generating region proposals as bounding boxes by using selective search-based image scanning and for classifying the region proposals to one of the object categories. The model was further improvised as Fast R–CNN [22]; nevertheless these models were slow and were difficult to optimize. Faster R–CNN [3] alleviated the difficulties with its previous versions by replacing region proposal network (RPN) with selective search which is trained as a single neural network with fast R–CNN sharing the convolutional feature set. The RPN component of the Faster R–CNN guides the unified network to look into different regions of interest. Faster R–CNN hitherto is considered as the most robust and accurate object detector; these models still lack real-time performance because the first proposals are generated and subsequently proposal labeled in the known categories.

Dai et al. [23] extended the shared, fully convolutional network architecture originally proposed for image segmentation ([24]) to two-stage detection strategy having region proposal and region classification. The end-to-end learning of the network constructs a set of position-sensitive score maps to deal with the translation variance of target objects. The scores are generated by a bank of convolutional layers which encodes the relative spatial position information incorporating translation variance in the learning.

Some recent works on object detection also followed a regression-based approach which includes Single Shot Multibox Detector (SSD) [25], Deconvolutional Single Shot Detector (DSSD) [26], and You Only Look Once (YOLO) [8]. SSD combines the Multibox approach for bounding box regression by using a set of default boxes for predicting the shape offset and category confidence at each location in the image. The authors used VGG-16 as the base network architecture where the network combines predictions from different feature maps with different resolutions. DSSD further improvised SSD by shifting to the “encoder-decoder” type of networks to incorporate context information in the learning by replacing the VGG-16 architecture with residual-101.

YOLO [27], unlike its predecessors, trained single regressor network for direct prediction of object bounding boxes with category confidences. This approach of complete regression without any classification step yielded superior real-time performance in terms of the detection speed. Also, YOLO bettered detection accuracy in comparison with all other detection systems except Faster-RCNN. With the incorporation of anchor boxes for guiding the bounding box prediction in a newly designed network, YOLOv2 bettered all other existing state of the art in PASCAL VOC-2017 challenge. The latest YOLO (here onwards we refer to this version as YOLOv3) claims to improve its performance with the incorporation of much deeper network architecture combining predictions at multiple scales. The YOLO algorithms see the entire image all at once through the forward pass of the network which gives more accurate and comprehensive information. This helps the detector in avoiding false positives than the classification based detection systems which focus only on the region proposals. Despite remarkable progress in object detection in natural images due to deep learning-based strategies, the performance of existing models requires improvement from the accuracy and real-time performance standpoint. The latest results on Pascal VOC-challenge [6] and MS-COCO [7] datasets establish that more effort is required to solve the detection problem.

In addition, several CNN based detection models have been designed for specific target objects [2830]. These methods have raised the benchmark in target objection detection; nevertheless, the problem is far from being solved [5,3133]. An important distinction in specific object detection is the availability of auxiliary information sources or user feedback which has helped researchers to solve the problem up to the acceptable limit.

In this work, we propose a feature enriched object detector based on the latest YOLO architecture. Our choice of YOLO architecture is based on its comparative detection accuracy and superior detection rate. It is expected that CNN based detection model should be able to learn the object-specific salient features required for detection; nevertheless, it cannot be guaranteed in single objective-based learning formulation. This direction of research exploring learning-based enrichment of deep features by including additional modalities has been not been explored for object detection. In this context, the recent work on foreground extraction for video analysis [34], a deep neural network based framework, exploits on the multistage fusion of the combination of residual features from different convolutional layers. On the other hand, our approach is to feed in the handcrafted features in the network to guide the feature learning process leading towards more accurate detection. For learning-based feature fusion, we use ICF descriptor which has shown remarkable object detection performance as discussed above.

3. Revisiting the Object Detection Problem

Object detection consists of localization of region capturing the object and assigning their labels by processing localized region. The localization problem can be formulated as (i) regression problem that predicts the region of interests or (ii) binary classification problem focusing on detection of foreground region having object segments. The outcome of any of these methods would predict object boundary as set of four coordinates (, , , and ) as shown in Figure 1.

Label identification of the object region is a -class classification problem where denotes the number of classes, for example, {tree, person, bird, dog, bicycle} or {cat, car, tree}. In both the stages of object detection problem, challenges arise because of the underlying variations in lighting conditions, scaling, occlusion, partial view, and orientation. Figure 2 shows some such challenging situations from the COCO dataset.

4. Preliminaries

Before presenting the object detection methodology using learning-based feature fusion, we briefly discuss the integral channel features and YOLOv3’s architecture. We also make some modifications in the YOLOv3 architecture as proposed in [35] which are also described in this section.

4.1. You Only Look Once Object Detector-YOLOv3

The class of YOLO algorithms [8,27,36] look at the entire image when detecting and recognizing objects and extract deep information about classes and their appearance, unlike other approaches such as sliding window-based methods or R–CNN based algorithms. These algorithms treat the detection of objects as a single regression problem giving faster response with a reduction in the design complexity of the detector. Though the significant speed achievement, the algorithms lag in terms of accuracy especially with small objects.

The latest algorithm in the YOLO family, that is, YOLOv3, has proven its performance compared to other state-of-the-art detectors. The YOLOv3 architecture has 107 layers in total distributed as {convolutional = 75; route = 4; residual = 23; upsample = 2; detection = 3}. The architecture uses a new model for feature extraction referred to as Darknet-53. The new model is significantly larger than models used in the earlier version; nevertheless, it has proven to be more efficient than other state of the arts. The model uses 53 convolution layers which takes an input image of size 416  416. Figure 3 shows the architecture of the YOLOv3 object detector. The Darknet-53 model is pretrained on ImageNet [37]. For the detection task, the network is modified by removing its last layer and then stacking up additional layers resulting in the final network architecture. The first 75 layers in the network represent 52 convolutional layers of the Darknet-53 model pretrained on ImageNet. The remaining 32 layers are added to qualify YOLOv3 for object detection on different datasets with further training. Besides, YOLOv3 applies new residual layers similar to skip connections that combine feature maps from two layers using elementwise addition resulting in finer-grained information.

The YOLOv3 replaces the softmax based activation used in older versions with independent logistic classifiers. Features are extracted using a similar concept to feature pyramid networks. Also, binary cross-entropy loss is now used for class predictions, which is beneficial when faced with images having overlapping labels. The K-means is used for anchor box generation; however, 9 bounding boxes are now used rather than 5. The number of bounding boxes is distributed equally across the three detection scales. In the present design, a route layer is also used which outputs a layer’s feature maps.

4.1.1. Prediction

YOLOv3 processes images by dividing them in grid, and if the center of the object falls in a grid cell, then that cell is responsible for detecting the object. The network predicts bounding boxes at three different scales. The first detection scale is used for detecting large-sized objects. The second detection scale is used for medium-sized objects and the last for small objects. The three detection layers are displayed in red in Figure 3. Each cell predicts bounding boxes, and each prediction has 5 predictions: , and confidence score representing the measure of prediction having an object. The and parameters are the center of the box for the grid cell, while and are the width and height of the predicted box for the entire image. The confidence score is the Intersection over Union (IoU) between the predicted box and the ground truth. The output prediction is tensor where 5 is the number of predictions per bounding box (, and the confidence value) and is the total number of object categories. Figure 4 shows the prediction algorithm of YOLOv3 for cells, where each cell has bounding boxes.

4.1.2. Loss Function

The YOLOv3 includes a loss function (1) that instructs the network to correctly predict bounding boxes and accurately classify the detected objects with a provision to penalize false positives:

The symbols are explained in Table 1. The symbols under hat represent corresponding prediction values. The loss function in the equation has three error components: localization, confidence, and classification as observed in equation (1). Different loss components are combined by sum-squared algorithm as it is easier for optimization. The localization loss is responsible for minimizing the error between the “responsible” bounding box and the ground truth object, if an object is detected in a grid cell.

4.2. Integral Channel Features (ICF)

In this work, we fuse different channels of information proposed in the ICF [9] in the YOLOv3 architecture. ICF descriptor is a series of image channels computed from the input image using linear and nonlinear transformations. A channel refers to a representation of the input image. Next, first-order features are extracted by computing the sum of rectangular regions. The evaluated channels were color channels (RGB, Gray, HSV, and LUV), gradient magnitude, and gradients histogram. The authors in [9] claimed that the most informative channel among all in independent evaluation for pedestrian detection was HOG. Further, a combination of LUV, gradient, and HOG gave the best detection rate. Our proposed work reevaluates the combination of channels because of varying challenges in the present application scenario. In our work, the window size parameter for HOG computation depends on the 2D dimensions of the deep features where they are to be fused with (discussed in Section 4.1). As an example, if the 2D dimensions of a layer where features are to be fused are 13  13 and the input image is of size 416  416, then the window size will be 32  32 so that it results in 1313 HOGs. For each window, we compute the HOG using six bins. Other HOG parameters including cell and block size used for normalization purposes are also set the same as the window size. The block stride parameter is also set equal to the window size having no overlap between two windows. As shown in Figure 5, the ICF for a given image is the linear concatenation of individual color channels after normalization, the corresponding HOG feature. The preprocessing in feature computation includes the resizing of the input image to align it with the network layer that receives this input. For experiments discussed in this paper, we follow the same steps and parameters for HOG computation as discussed in [9].

5. Learning-Based Feature Fusion in YOLOv3

Before we establish the proposed feature fusion, we set out the required steps which will be described subsequently:(1)Identify the candidate convolution layers for feature injection(2)Evaluate the layers for feature injection using complete set of ICF channels(3)Evaluate different combinations of ICF channels using the best network layer obtained in the previous step(4)Train and test the detector injecting the best ICF channel combination at the best layer position

Our proposal for feature fusion starts with identifying a location and space for handcrafted features within the YOLOv3 architecture. For injecting an additional set of features in this architecture, we first need to identify the convolution layer to feed additional information. We adopt a qualitative approach to decide on the layer for feature fusion by validating feature fusion on different layers. The depth space of a convolutional layer represents the number of feature maps that hold the deep feature—hereafter referred to as filters. We need to identify the specific number of filters within the specific layer which will store the additional features. First, we propose to double the original number of filters in the specific layer as shown in Figure 6.

Depending on the number of channels in the ICF set, the same number of filters needs to be reserved for storing handcrafted feature values. The ICF descriptor is injected into the network at the beginning of the training procedure. In the layer selected for ICF injection, the additional filters (introduced by doubling the layer depth) are copied with ICF descriptor from the rear end as shown in Figure 6. The remaining filters are set to zero in the beginning. As the training progresses, the layer weights, that is, filter values update following the learning algorithm. The output generated by this layer depends on the values of all filters. The proposed method for feature fusion diverges from the simple approach of the stacking of handcrafted features with original filters in the selected layer. The extra filters in addition to the filters used for storing ICF descriptor values help regularize the layer weights. We maintain the strategy of doubling the filters for feature fusion regardless of the specific convolutional layer under validation. An example showing the procedure of doubling the number of filters at a specific layer to issue space for ICF fusion is illustrated in Figure 6. In this example, fusion was selected to occur at layer of YOLOv3’s network. The top side of the figure shows a section from the original YOLOv3’s network and the bottom side is where the number of filters for the layer is doubled. The number of filters was changed from 1024 to 2048. The light orange is the new filter depth (2048) and consists of the original filter depth (1024) concatenated with the additional filters of the same depth, represented in green. Additionally, doubling the number of filters in the fusion layer poses fewer design challenges than with filters with direct stacking of handcrafted features.

5.1. Design Issues with Injection of ICF Descriptor

To fuse handcrafted features with deep features, their 2D dimensions have to match the width and height dimensions of the layer. Considering the example in Figure 6 at layer, if the input image size is 416416 then the filter dimensions will be 13  13. Hence, all selected channels for fusion should be of size 13  13 to fit into the additional filters. In the testing phase of YOLOv3, the size of the input image is fixed at 416  416, so the dimensions of the ICF descriptor at the layer will always be 13  13. On the other hand, in the training phase of YOLOv3, the size of input image changes based on a random parameter in every 10 iterations where the size of the image is defined as a factor of 32, starting from 320  320 up to 608  608. If the input size is 320  320 or 608  608, then at the layer the filter size becomes 10  10 or 19  19, respectively. The filter size compared with the input image size is downsized by a factor of 32. After making sure that the handcrafted features are of the right size, their values can replace the additional filters in the chosen layer.

The flowchart of the proposed object detector is shown in Figure 7. For training purposes, the desired ICF channels of input images at different scales are calculated offline. These images at different scales are used to incorporate variability in the training process. In the testing phase, ICF channels are calculated only once for each image at a fixed scale. The right side of the figure shows that as our proposed detector is being trained/tested fusion takes place at the layer named YOLOv3 convolutional layer.” The doubled number of filters and the number of ICF channels are represented as and , respectively. The figure shows how the ICF overlaps with a portion of the deep features residing in some of the additional filters.

6. Experimental Evaluation

A preliminary evaluation of the proposed framework is performed on the PASCAL-VOC dataset (VOC2007 detection task). The original YOLOv3 detector achieved the mAP of 67.5% on the testing set with input image dimension as 416  416. The YOLOv3 implementation available by [35] is used for benchmarking our experimental results. As mentioned in Section 5, we first identify the convolutional layer for injecting the handcrafted features, and the required space (i.e., the number of filters). First, we figure out the best layer location to fuse ICF channels. Subsequently, by another set of experiments, we determine the best combination of ICF channels. We separate a validation set comprising one-tenth of the training examples by random selection. The validation set is used for testing the convolutional layers and combination of ICF channels for optimum selection. For determining the best layer for fusion, we use all channels of information in ICF as used in the original work by Dollar et al. [9]. This is done by doubling the number of filters at the layer under evaluation, fusing 17 channels of ICF set, and checking the resulting mAP. We evaluated the two convolutional layers preceding a detection layer in YOLOv3 architecture. For each experiment, the network is trained for 25,000 iterations. The remaining training parameters are set as the original YOLOv3 detector. We performed the first evaluation on layer which is the last convolutional layer before the detection layer.

Doubling the number of filters for layer and injecting the full set of ICF channels resulted in the mAP of 69.8%. In comparison with the benchmark performance based on YOLOv3 performance, we achieved an increase of 1.7% on mAP scale. The next experiment on layer achieved the mAP of 71.5%, which is more effective than the previous experiment. Table 2 lists the mAP scores for experiments conducted for determining the best layer position for feature fusion. As observed, the best mAP score of 71.5% is achieved at layer.

After determining the most suitable layer for ICF fusion, we conduct experiments to discover the combination of ICF channels for fusion that is most suitable for the object detection task in the VOC dataset. Again, in this set of experiments, we fuse ICF channels by doubling the number of filters at the layer. In this experiment, we run the training process for 50,000 iterations. Fusing all ICF channels, color, gradient magnitude, and HOG, achieved the mAP of 77.7%. The entire ICF descriptor summed up to 17 channels, considering 6 channels for HOG. Removing the RGB channel from the 17 ICF channels increased the mAP to 78.2% showing a minuscule increase of 0.5% from the preceding experiment. Considering original input images are in RGB color values, the increase is not significant. We also explored other channel combinations; the corresponding mAP scores are presented in Table 3.

As seen, the best ICF channel combination for VOC2007 detection task consists of Gray, HSV, LUV, gradient magnitudes and HOG achieving maximum mAP score of 79.1%. We retrain the YOLOv3 detector on complete training set fusing the finalized channels of ICF descriptor. The network achieved the mAP of 78.7% on the testing set which is 11.2% more than the detection rate achieved by original YOLOv3 detector.

6.1. Evaluation on MS-COCO Dataset

MS-COCO dataset consists of 82,783 training, 40,504 validation, and 40,775 testing images belonging to 80 categories. In the literature, the YOLOv3 [8] reported the mAP of 55.3% on the COCO dataset for an input image size of 416  416. All experiments reported in this work were executed on a NVIDIA GeForce GTX 1080 GPU desktop which is unsubstantial in comparison with the infrastructure used in the original paper [8]. Therefore, we adopt an alternate strategy for benchmarking where the YOLO detector (original and with proposed fusion strategy) is retrained on the MS-COCO dataset and we compare the detection performance at 50000 iterations. The original YOLOv3 evaluation on COCO testing dataset achieved the mAP of 36.5% for 50,000 iterations.

After determining the best location for ICF fusion using the mAP values achieved on the validation set as shown in Table 4, we performed experiments to determine the best combination of ICF channel fusion that best suits the COCO dataset (Table 5). As shown in Table 4, doubling the layer resulted in the highest mAP score (37.3%). Therefore for all subsequent experiments related to the COCO dataset were conducted with the doubled number of filters at the layer. All experiments were evaluated at 50,000 iterations. Fusing all ICF channels (color, gradient magnitude, and HOG) resulted in the mAP of 37.7%. Removing RGB from the 17 ICF channels led to the mAP of 37.3%. Other channel combinations were tried out but did not give good results such as fusing only HOG features (6 channels), LUV + HOG (9 channels), gray + LUV + gradient + HOG (11 channels), and gray + HSV + LUV + gradient (8 channels). The best combination that gave the highest mAP score consisted of fusion of the all ICF channels.

Furthermore, the evaluation on testing set with fusion of the complete set of ICF channels at the layer achieved 37.8% of the mAP value. With COCO dataset, we also experimented with the effect of normalization on the ICF channels. To measure the effect of normalization on the performance of the network, we tried out two different normalization techniques. All the results discussed so far were achieved without any normalization. We first experimented with max normalization technique dividing each channel by its maximum range so that all values fall in the range of 0 to 1. The fusion of max normalized ICF channels in YOLOv3 achieved the mAP of 38.4% on the testing set which is 0.6% higher than the best result achieved with unnormalized features and 1.9% with the original YOLOv3 performance. This confirms with the training principle of deep networks which suggests the use of input normalization as one of the tools for performance improvement. In the second normalization technique, we used Z-score based normalization which converts data in all channels to have mean 0 and standard deviation 1. However, the Z-score normalization did not perform as good as the previous normalization technique.

7. Conclusion and Future Work

In this work, we present an approach to fuse handcrafted features in the convolutional neural network based object detector. The fusion of handcrafted features with learning-based features has proved to be effective in many earlier works. In this work, we demonstrate the methodological steps to fuse handcrafted features in the latest version of the YOLO detector. Our experiments with a combination of simple integral channel features fusion in YOLO have yielded substantial improvement in the detection rate on PASCAL-VOC and MS-COCO datasets. In conventional machine learning, early, late, and learning-based fusion have been primary strategies for feature fusion. However, deep learning networks are designed for specific input sizes which pose a challenge for algorithm designer to feed in extra information in the network unless the network is redesigned. The work presented a novel approach based on methodological steps to address both the problems. In future work, we plan to explore the approach for sequence-based learning for object detection and tracking.

Data Availability

This work utilizes standard datasets which are open access.

Conflicts of Interest

The authors declare that they have no conflicts of interest.