Abstract

At present, research on hesitant fuzzy operations and measures is based on equal length processing, and an equal length processing method will inevitably destroy the original data structure and change the data information. This is an urgent problem to be solved in the development of hesitant fuzzy sets. Aiming at solving this problem, this paper firstly defines a hesitant fuzzy entropy function as the measure of the degree of uncertainty of hesitant fuzzy information and then proposes the concept of hesitant fuzzy information feature vector. The hesitant fuzzy distance measure and similarity measure are studied based on the information feature vector. Finally, the hesitant fuzzy network clustering method based on similarity measure is given, and the effectiveness of our algorithm through a numerical example is illustrated.

1. Introduction

Torra and Narukawa [1, 2] extended fuzzy sets [3] to hesitant fuzzy sets (HFSs) because they found that, under a group setting, it is difficult to determine the membership of an element to a set due to doubts between a few different values. For example, two DMs discuss the membership degree of into . One wants to assign 0.4 and the other 0.6, and they cannot persuade with each other; thus the membership degrees of into can be represented by {0.4, 0.6}. This is obviously different from fuzzy number 0.4 (or 0.6) and the intuitionistic fuzzy number (0.4, 0.6). Therefore, hesitant fuzzy sets can better simulate the hesitant preferences of decision-makers. Since it was put forward, the hesitant fuzzy set has received extensive attention from scholars at home and abroad. The main research work is concentrated in the following aspects: (1) research on various measures in the hesitant fuzzy environment [410]; (2) research on the integration operator of hesitant fuzzy information [1116]; and (3) the expansion of hesitant fuzzy set theory [1722].

It should be pointed out that the present researches on the operation, sorting, and various measures of hesitant fuzzy sets require that the hesitant fuzzy elements have the same length. In practical application, the length of hesitant fuzzy element is different. The method proposed in [2] is adding some elements to a shorter hesitant fuzzy element, making it equal to another hesitant fuzzy element, or repeating their elements in order to obtain two series with the same length [23]. Obviously, these methods will destroy the original data structure and change the data information. How to overcome the shortcomings has become an urgent problem to be solved in the development process of hesitant fuzzy sets.

Clustering is a basic technique, which is often utilized in a primary step of analyzing unlabeled data with the goal of summarizing structural information [24]. In practical applications, the clustering data are mostly uncertain or fuzzy. To solve the problem of data clustering in different fuzzy environments, fuzzy clustering algorithms [25], intuitionistic fuzzy clustering algorithms [26], and 2-type fuzzy clustering algorithms [27] have been proposed. However, in the group of decision-making environment, the decision information is more suitable to express hesitant fuzzy sets, and the algorithm mentioned above is not suitable for handling the clustering problem of this type of information. If the fuzzy logic is used to handle it, generally take the average value of preference information that are provided by experts or can take the minimum range containing all of the preference information, that is, convert the hesitant fuzzy information into interval value information for processing. This method of data processing is bound to change the original preference information that provided by experts; as a result, the research of clustering problem under the hesitant fuzzy information has a certain scientific significance. One of the advantages of applying the hesitant fuzzy set is that clustering hesitant and vague information permits us to find patterns among hesitant fuzzy data. At present, the clustering researches under the hesitant fuzzy environment are still at the its initial stage, and Chen et al. [28] used the correlation coefficient of hesitant fuzzy set to construct hesitant fuzzy relationship matrix and then conducted hesitant fuzzy clustering analysis based on the relation of equivalence. In order to obtain an equivalence relation matrix, a fuzzy relation matrix needs to be iterated continuously, which not only loses information but also has a large amount of calculation [29]. Due to the existence of uncertainty for the similarity measure of samples, leading to the clustering, results were not precise enough and the divided categories were inconsistent with the fact. In [4], the hesitant fuzzy similarity measure formula based on distance was proposed. The measurement is inconsistent with the facts sometimes, and the resolution is not high enough; in the literature [29], a hesitant fuzzy clustering method based on agglomerative hierarchical clustering [30] was proposed. This method needs to use a hesitant fuzzy average operator to calculate the clustering center repeatedly, and the calculation amount is large; in the literature [31], a hesitant fuzzy clustering algorithm based on minimal spanning tree was proposed. The distance of hesitant fuzzy set used in this method is put forward based on the literature [4], which also has the shortcoming of low resolution and sometimes inconsistent with the fact; in the literature [32], from the point of view of information theory, hesitant fuzzy relative entropy and symmetric interactive entropy are proposed, a new kind of hesitant fuzzy similarity degree is proposed, which is combined with the idea of TOPSIS, and a hesitant fuzzy clustering method is proposed based on the traditional netting clustering method. The premise of all the above methods in the measurement and operation is that the data are equal in length, which is not satisfied by the hesitant fuzzy set. Therefore, it is necessary to add artificial elements for equal length processing, and the processed data will inevitably change the original data information and affect the clustering results.

Based on the above analysis, this paper firstly proposes the concept of hesitant fuzzy entropy function and hesitant fuzzy information feature vector, aiming at solving the problem of processing data of hesitant fuzzy set, sorting, and various measures in the study of different lengths. Furthermore, the hesitant fuzzy uncertainty measure, distance measure, and similarity measure are studied. Finally, based on the similarity measure and the traditional network clustering method, the network clustering method for hesitant fuzzy information is given. And then we illustrate its effectiveness via numerical examples.

2. Preliminary

Definition 2.1 [1, 2]. Let be a fixed set; a hesitant fuzzy set (HFS) on is represented by a function that when applied to , it returns a subset of [0, 1], which can be expressed by a mathematical symbol:where is a set of some values in , denoting the possible membership degrees of the element to the set . is called the hesitant fuzzy number or hesitant fuzzy element. If it does not cause confusion, it can be abbreviated as . The hesitant fuzzy number can be expressed in more detail as . Among which, denotes the number of elements in a hesitant fuzzy number . Obviously, when , the hesitant fuzzy set degenerates into the traditional fuzzy set.

Definition 2.2 [1, 2]. Set as a given nonempty set, then is the complement of the hesitant fuzzy set , among whichDistance measure and similarity measure are important research contents in fuzzy set theory and have a wide application background. In the literature [4], the axiomatic definitions of distance and similarity measure of hesitant fuzzy sets are given.

Definition 2.3 [4]. Sets A, B be the two hesitant fuzzy sets defined on , and then the distance measure between A and B satisfies the following conditions:(1)(2), if and only if (3)

Definition 2.4 [4]. Sets A, B be the two hesitant fuzzy sets defined on , and then the similarity measure between A and B satisfies the following conditions:(1)(2), if and only if (3)Definition 2.3 is proposed to facilitate the use of distance measures to define similarity measures. In practice, the distance can only be satisfied with .

3. A New Kind of Hesitant Fuzzy Entropy

Entropy is the measurement of the degree of uncertainty of information, and it has always been an important research object in uncertainty decision analysis. A new hesitant fuzzy entropy measure function is proposed by analyzing the shortcomings in the current research results on hesitant fuzzy entropy.

Definition 3.1. Assign the hesitant fuzzy element , where is the number of element in the hesitant fuzzy element, and recordwhere represents the fuzzy degree of the hesitant fuzzy element and represents the hesitant degree of the hesitant fuzzy element . Then the real valued function on the hesitant fuzzy element can be expressed by a binary function , if the meets the following conditions:(1) if and only if and (2) if and only if or (3) and , and (4)Then, can be called as a hesitant fuzzy entropy function.

3.1. Interpretation and Analysis

(1), . Then indicates that is a clear set, then the entropy is 0.(2)When , since , , then it can get , and it is concluded that the domain of the entropy function is because the entropy function is concave increase with respect to and , and the maximum value of is 1 when or is obtained; that is, when or , the uncertainty reaches the maximum. is completely contradictory information, and is completely fuzzy information; in both cases, the uncertainty is maximized and in line with intuitive judgment.(3)It ensures that the entropy function is concavely increased with respect to fuzziness and hesitation degree, conforms to human cognitive characteristics, and improves discrimination.(4)Fuzziness and hesitancy have the same effect on entropy.

Based on the above analysis, function obviously satisfies the above conditions in Definition 3.1, so it can be regarded as an entropy function. For example, if , then ; if , then ; and if , , then , , where . The above judgment results are consistent with the intuition.

Property 3.1. Set hesitant fuzzy element , when ; the hesitant fuzzy element degenerates into a fuzzy number, and the entropy of fuzzy value is .

Proof. (1), that is, or , where is a clear set.(2)According to condition (2) or because when , , so , that is, .(3)According to condition (3), it is known that increases monotonously with respect to , so when is closer to 0.5, the larger the is, the larger the entropy of the fuzzy value is.The property 3.1 indicates that the fuzzy entropy is a special case of the hesitant fuzzy entropy function, and the hesitant fuzzy entropy function can also be applied to the fuzzy set.
In order to illustrate the advantage of the entropy function proposed in this paper in measuring uncertainty, the following is compared with the existing entropy formula: at present, the common formulas of hesitating fuzzy entropy include the entropy formula proposed by Xu and Xia and the entropy formula proposed by Farhadinia, in whichwhere indicates the number of elements in a hesitant fuzzy number and indicates the element of the largest ith in the hesitant fuzzy number where is strictly monotonically decreasing function, which may get , , , ; (, where indicates the number of elements contained in the fuzzy number ).
Set hesitant fuzzy number , , , , , , , and . The entropy formula proposed by Xu and Xia and the entropy formula proposed by Farhadinia are compared with the entropy function proposed in this paper. The results are shown in Table 1.
Because the entropy formula proposed by Farhadinia only considers fuzziness and neglects the influence of hesitancy, the result is quite different from that of the method proposed in this paper and the method proposed by Xu. It is not difficult to find from the above table that the method proposed in this paper is obviously higher in the discrimination than that proposed by Xu, and the comparison result is close to it, and the individual results are inconsistent. For example, ; however, according to the method presented in this paper, the result is ; this is because the starting point is inconsistent and the hesitant fuzzy entropy proposed by Xu requires that the number of elements contained in the two pairs be equal and that the elements should be artificially added when the number of elements is different. Therefore, the proposed method is bound to deviate from intuitive judgment for the comparison of entropy of two hesitant fuzzy numbers with a different number of elements contained. The entropy measure function proposed in this paper not only considers the influence of fuzziness on the entropy value but also considers the effect of hesitation degree on the entropy value, which can more reasonably depict the uncertainty degree of the hesitation fuzzy number, so the result is more consistent with our intuition.

4. Hesitant Fuzzy Distance Measure and Similarity Measure

For a hesitant fuzzy element , the most important information it contains is the size of value and the degree of uncertainty, which is also a common concern in practical applications. Based on this, we introduce the definition of feature vector of hesitant fuzzy information.

Definition 4.1. Set hesitant fuzzy element ; two-dimensional vector is called the information feature vector of hesitant fuzzy element , which is marked aswhere represents the size of a hesitant fuzzy element and is the entropy of hesitant fuzzy element , representing its degree of uncertainty, calculated by Definition 3.1.
The number of elements contained in different hesitant fuzzy elements may be different. In order to facilitate sorting and measurement, it is usually necessary to add elements artificially, which will inevitably destroy the original structure of the data and change the data information. The hesitant fuzzy element is proposed by the information feature vector to solve this kind of problem. The following formulas of the measure and similarity measure of hesitation fuzzy distance based on the feature vector of hesitant fuzzy information are given.
The feature vector of hesitant fuzzy information is to describe the information feature of the hesitant fuzzy element from two different factors, so the dimensions of different components are different; at the same time, there is obviously a correlation between the two components. Therefore, it is not appropriate to choose the traditional distance formula to measure the difference between the two hesitant fuzzy elements. This paper defines a new measure of distance and similarity from the angle of information theory.

Definition 4.2. Set as a nonempty domain; , , is the two hesitant fuzzy sets defined on , and its information feature vectors are separately , . For the convenience of writing, note , then callas the distance measure of . Among which,where is the symbolic variable, .
The distance measure based on the information feature vector is based on the relative entropy idea, and it is easy to verify that it satisfies the following properties.

Property 4.1. Set as a nonempty domain; , , is the two hesitant fuzzy sets defined on , and its distance measure is , then(1)(2), if and only if , (3)

Proof. Make; according to, is the concave function in a defined domain, that is, , among which, and . If and only if , the equal sign is established. Suppose , bring in and getIf and only if when , ; at this moment, . Then, ; if and only if when , .
In the same way, we can get ; if and only if when , . In summary, (1) and (2) can be established.
According to the expression itself, it can be judged that (3) is clearly established.

Note. Property (2), , , is not equivalent to ; for example, take and , it is obvious that . The information feature vector is represented as and , respectively, according to property (2), and then . At this moment, the result is consistent with human intuition because completely ambiguous information and completely contradictory information can convey the same amount of information. This is also the main difference between the distance measure proposed in this paper and other hesitant fuzzy distance measures.
Inspired by TOPSIS, the hesitant fuzzy similarity measure formula based on hesitant fuzzy distance measure is given below.

Definition 4.3. Let as a given nonempty domain; is the two hesitant fuzzy sets defined on , and modified to as the similarity measures of .

Property 4.2. Set as a nonempty domain; , , is the two hesitant fuzzy sets defined on , and the similarity measure of is :(1)(2)(3), if and only if when , (4), if and only if when , (5)If and only if when , Property 4.2 can be determined by the formula itself. The proof process is omitted.
In practical application, different elements in set have different status and should be given different weights. A similarity measure formula considering weights is given below:among which . is the weight of element and satisfies , . Obviously, when , and .

5. Network Clustering Method Based on Hesitant Fuzzy Similarity Measure

The network clustering [33] method is a common method in data clustering analysis, and it is also the best choice to extend the clustering method to the fuzzy environment. The specific process is as follows: the similarity coefficient matrix is constructed by the data similarity measure, and then the cutting level is selected as truncated matrix of , and replace the principal diagonal element with the scheme symbol. In the lower left of the principal diagonal, the symbol “∗” is used instead of “1” to remove the “0” element. The position of the symbol “∗” is called the node. The so called network is to cross the nodes as the latitude and longitude lines and tie the scheme corresponding to the latitude and longitude lines at the nodes to achieve classification. The main advantage is that the clustering results can be obtained quickly and effectively by using the similarity coefficient matrix directly on the table. The method of clustering analysis in the hesitant fuzzy environment is given below. The calculation process is as follows:(1)Let be the set of object to be classified, be the decision factor set, and be the decision factor weight vector. The decision expert group measures the classified objects according to the decision factors and obtains the hesitant fuzzy decision matrix .(2)According to formula (6), the hesitant fuzzy value is expressed by the information feature vector, and then the decision matrix is transformed into the information feature vector matrix.(3)Calculate the hesitant fuzzy similarity coefficient matrix by using formula (11).(4)Remove elements above the principal diagonal and replace the principal diagonal element with the scheme symbol.(5)Select cutting level as the truncated matrix of , in the lower left of the principal diagonal, the symbol “∗” is replaced by “1,” and the “0” element is removed. The position of the symbol “∗” is called a node, the node is the latitude and longitude lines, and the node is over the node. The schemes corresponding to the latitude and longitude lines are bundled into one category.

6. Illustrative Example

In order to facilitate comparative analysis, this paper uses an example from the literature [32]. Through four factors (price , function , after-sales service , and quality ), 7 cell phones are classified. Assume the factor weight vector is . The decision group gives the evaluation value of mobile phone under the decision factor , which is represented by the hesitant fuzzy set , , among which indicates the degree to which the mobile satisfies the decision factor . Then the decision information can be represented by the decision matrix (Table 2). According to the network clustering method, cluster analysis is performed on 7 mobile phones as shown in Table 2.

Step 1. See Table 2.

Step 2. According to the formula (6), the hesitation fuzzy value in the hesitating fuzzy decision matrix is transformed into the information feature vector matrix (Table 3).
For example, the data (0.5, 0.68) in the first column of the first row in Table 3 are the information feature vector corresponding to the data {0.4, 0.6} in the first column of the first row in Table 2. They are calculated according to formula (6), where , .

Step 3. Calculate hesitant fuzzy similarity coefficient matrix by using formula (4):The first row and second column data 0.6183 are the similarity measurement between the date in the first row and the date in the second row in Table 3 and is calculated according to formula (11):

Step 4. Remove elements above the principal diagonal, and replace the principal diagonal element with the scheme symbol, that is:

Step 5. Select the cutting level as thetruncation matrix of and then classify through the network:(1)When , they are divided into 7 categories: , , , , , , (2)When , they are divided into 6 categories: , , , , , (3)When , they are divided into 5 categories: , , , , (4)When , they are divided into 4 categories: , , , (5)When , they are divided into 3 categories: , , (6)When , they are divided into 2 categories: , (7)When , they are divided into 1 category: Next, the clustering results of this paper are compared with those of the literature [28], literature [29], literature [31], and literature [32], and the results are analyzed. The clustering results obtained by other methods are shown in Table 4.
Different literatures choose different measures to measure the degree of closeness between samples; among them, the literature [28] is based on the correlation coefficient; the literature [29] and literature [31] are based on the distance measure; and the literature [32] and this paper are based on the similarity measure. In order to compare the sensitivity of various measures, it is necessary to analyze the variance of the measurement data. The larger the variance, the higher the sensitivity of the corresponding measure. The results are shown in Table 5. Furthermore, in order to compare the effectiveness of various methods, the D-B index [34] of the clustering results is calculated separately. The distance measure selected for calculating the D-B index of the literature [28] is . The results are shown in Table 5.
It can be found from Table 5 that (1) the similarity measure proposed in this paper has higher sensitivity than other measures and the clustering result has better robustness and (2) tThe D-B index of this paper is smaller, indicating that the clustering results are better.
Through comparison, it can be found that (1) the results obtained by using the method proposed in this paper are consistent with those obtained in the literature [32], which to some extent reflects the effectiveness of the method proposed in this paper; (2) the results of the classification in the literature [28] are not precise and accurate, as it is intuitively possible to judge from the data in Table 2, so it is more appropriate to classify them as a group; (3) the methods of the literature [29] and literature [31] are put forward based on the distance formula in the literature [4], but the resolution of the distance formula is not high, and the results are sometimes inconsistent with the facts [32], which will inevitably affect the classification results; (4) the data must be processed by equal length in the literature [28], literature [29], literature [31], and literature [32], which will inevitably affect the clustering results. The reason why the results in this paper are consistent with the literature [32] is that the difference in the number of hesitant fuzzy numbers in Table 2 is small, and the size of each element in the same hesitant fuzzy number is not much different, and if not so, the results must be different from those obtained by the present method.

7. Conclusion

In this paper, the hesitant fuzzy information feature vector is used as the entry point, which provides a new idea for solving various hesitant fuzzy measures. Then the hesitant fuzzy uncertainty measure, distance measure, and similarity measure are studied. Finally, a clustering method for fuzzy information is proposed. Through analyzing the results of the example, it has been proved that this method is faster and more effective in practical applications. The main contributions of this paper are (1) it effectively avoids the problem of processing data with equal length in the research of the measure of hesitant fuzzy set and (2) combined with the similarity measure proposed by TOPSIS idea, the resolution between schemes can be improved. Subsequent research on hesitant fuzzy set theory and application based on hesitant fuzzy information feature vectors will be a meaningful topic.

Data Availability

Previously reported data were used to support this study and are available at DOI: 10.1080/00207721.2013.797073. These prior studies (and datasets) are cited at relevant places within the text as references [29].

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (no. 61350003).