Abstract

This paper describes an enhancement of fuzzy lattice reasoning (FLR) classifier for pattern classification based on a positive valuation function. Fuzzy lattice reasoning (FLR) was described lately as a lattice data domain extension of fuzzy ARTMAP neural classifier based on a lattice inclusion measure function. In this work, we improve the performance of FLR classifier by defining a new nonlinear positive valuation function. As a consequence, the modified algorithm achieves better classification results. The effectiveness of the modified FLR is demonstrated by examples on several well-known pattern recognition benchmarks.

1. Introduction

Much attention has been paid lately to applications of lattice theory [1] in different fields including neural networks [2]. Artificial neural networks whose computation is based on lattice algebra have become known as morphological neural networks [3, 4]. Lattices are popular in mathematical morphology including image processing applications [5, 6]. Moreover, algebraic lattices have been used for modeling associative memories [7]. In [8], the problem of capacity storage limitation in associative memories [9, 10] has been eliminated by proposing one-way and bidirectional lattice associative memories. Furthermore, lattices are used implicitly in some neural networks such as fuzzy ART and min-max [11, 12] as explained in [2, 13]. A practical advantage of lattice theory is the ability to model both uncertain information and disparate types of lattice-ordered data [14]. The term of a fuzzy lattice was proposed by Nanda in 1989 on the basis of the concept of a fuzzy partial-order relation [15]. Several authors have employed the notion β€œfuzzy lattice” in mathematics emphasizing algebraic properties of lattice ideals [16, 17]. Furthermore, the notion of fuzzy concept lattice has been studied in [18–20]. Sussner and Esmi [21] have introduced the morphological perceptron with a fusion of fuzzy lattices for competitive learning. Fuzzy lattices have also been used in clustering and classification algorithms. More specifically, independently from the development of morphological neural networks, Petridis and Kaburlasos [13] have found inspiration in lattice theory and versions of the ART model and have devised another successful approach to lattice-based computational intelligence. Hence, they proposed a fundamentally new and inherently hierarchical approach in neuron computing named fuzzy lattice neurocomputing (FLN) [14]. Moreover, fuzzy lattice reasoning (FLR) classifier was announced for inducing descriptive, decision-making knowledge (rules) in a mathematical data domain including space 𝑅𝑁, and it has been successfully applied to a variety of problems such as ambient ozone estimation [22] as well as air quality assessment [23]. Decision making in FLR is based on an inclusion measure function; moreover, the definition of an inclusion measure is based on a positive valuation function.

The original FLR model employs a linear positive valuation function to define an inclusion measure. Liu et al. [24] proposed a nonlinear valuation function (arctan) for computing the inclusion measure function and successfully applied it to several data set benchmarks.

In this work, we apply FLR algorithm to solve pattern classification problems without feature extraction and improve its performance based on a new nonlinear positive valuation function. As a consequence, the modified algorithm achieves better classification results. The effectiveness of the modified FLR is demonstrated by examples on several well-known benchmarks.

The layout of this paper is as follows. In Section 2, the mathematical background of fuzzy lattices is reviewed. Section 3 explains modified fuzzy lattice reasoning classifier model. Section 4 provides experimental results that demonstrate the performance of modified FLR. Finally, Section 5 summarizes the results of this work.

2. Mathematical Background

A lattice (𝐿,≀) is a partially ordered set (or, simply, poset) such that any two of its elements π‘Ž,π‘βˆˆπΏ have a greatest lower bound π‘Žβˆ§π‘=inf{π‘Ž,𝑏} and a least upper bound π‘Žβˆ¨π‘=sup{π‘Ž,𝑏}. The lattice operations ∧ and ∨ are also called meet and join, respectively. A lattice (𝐿,≀) is called complete when each of its subsets has a least upper bound and a greatest lower bound in 𝐿 [1]. A nonvoid complete lattice has a least element and a greatest element denoted by 𝑂 and 𝐼, respectively. The inverse β‰₯ of an order relation ≀ is itself an order relation. The order β‰₯ is called the dual order of ≀ symbolically β‰€πœ•. A lattice (𝐿,≀) can be Cartesian product of 𝑁 constituent lattices 𝐿1,…,𝐿𝑁, that is, 𝐿=𝐿1×⋯×𝐿𝑁. The lattice operations meet and join of product lattice are defined as follows: ξ€·π‘Žπ‘Žβˆ§π‘=1,…,π‘Žπ‘ξ€Έβˆ§ξ€·π‘1,…,𝑏𝑁=ξ€·π‘Ž1βˆ§π‘1,…,π‘Žπ‘βˆ§π‘π‘ξ€Έ,ξ€·π‘Žπ‘Žβˆ¨π‘=1,…,π‘Žπ‘ξ€Έβˆ¨ξ€·π‘1,…,𝑏𝑁=ξ€·π‘Ž1βˆ¨π‘1,…,π‘Žπ‘βˆ¨π‘π‘ξ€Έ.(1)

A valuation on a crisp lattice 𝐿 is a real-valued function π‘£βˆΆπΏβ†’π‘… which satisfies 𝑣(π‘Ž)+𝑣(𝑏)=𝑣(π‘Žβˆ¨π‘)βˆ’π‘£(π‘Žβˆ§π‘),π‘Ž,π‘βˆˆπΏ. A valuation is called monotone if and only if π‘Žβ‰€π‘ in 𝐿 implies 𝑣(π‘Ž)≀𝑣(𝑏) and positive if and only if π‘Ž<𝑏 implies 𝑣(π‘Ž)<𝑣(𝑏). We remark that the goal of positive valuation function 𝑣 is to deal with lattice elements. Choosing a suitable valuation function is problem dependent. Decision making by the FLR is based on an inclusion measure function; therefore,a proper positive valuation function might improve performance.

Definition 1. An inclusion measure 𝜎 with least element 𝑂 and great element 𝐼 in a complete lattice 𝐿 is a mapping πœŽβˆΆπΏΓ—πΏβ†’[0,1] such that it satisfies the following conditions [2]:(i)𝜎(π‘Ž,𝑂)=0,π‘Žβ‰ 0, (ii)𝜎(π‘Ž,π‘Ž)=1, for all π‘ŽβˆˆπΏ,(iii)π‘Žβ‰€π‘β‡’πœŽ(𝑐,π‘Ž)β‰€πœŽ(𝑐,𝑏), for all π‘Ž,𝑏,π‘βˆˆπΏ (consistency property).

It reveals that an inclusion measure indicates the degree to which one fuzzy set is contained in another one.

Theorem 2. A positive valuation function π‘£βˆΆπΏβ†’π‘… in a lattice (𝐿,≀) with 𝑣(𝑂)=0 is a sufficient condition for two inclusion measures [2]: π‘˜(π‘₯,𝑒)=𝑣(𝑒)𝑣(π‘₯βˆ¨π‘’),(π‘₯βˆ¨π‘’β‰ π‘‚),𝑠(π‘₯,𝑒)=𝑣(π‘₯βˆ§π‘’)𝑣(π‘₯),(π‘₯≠𝑂).(2)

In our experiments, the data have been normalized in lattice 𝐿=[0,1]𝑁, that is, the unit 𝑁-dimensional hypercube, where 𝑁 is the dimension of the input data. Furthermore, we propose the following nonlinear positive valuation function: ξ€·[]βˆ€π‘₯∈0,1𝑁,≀,𝑣(π‘₯)=𝑁𝑖=1π‘₯𝑖π‘₯ln𝑖[+𝛾,π›Ύβˆˆ1,+∞),(3) where 𝛾 is called location parameter. Without loss of generality, let π‘Ž=(π‘Ž1,…,π‘Žπ‘)β‰₯𝑏=(𝑏1,…,𝑏𝑁), that is, π‘Žπ‘–β‰₯𝑏𝑖,𝑖=1,…,𝑁: =𝑣(π‘Žβˆ¨π‘)+𝑣(π‘Žβˆ§π‘)𝑁𝑖=1π‘£π‘–ξ€·π‘Žπ‘–βˆ¨π‘π‘–ξ€Έ+𝑁𝑖=1π‘£π‘–ξ€·π‘Žπ‘–βˆ§π‘π‘–ξ€Έ=𝑁𝑖=1ξ€·π‘Žπ‘–βˆ¨π‘π‘–ξ€Έξ€·π‘Žlnπ‘–βˆ¨π‘π‘–ξ€Έ++𝛾𝑁𝑖=1ξ€·π‘Žπ‘–βˆ§π‘π‘–ξ€Έξ€·π‘Žlnπ‘–βˆ§π‘π‘–ξ€Έ=+𝛾𝑁𝑖=1π‘Žπ‘–ξ€·π‘Žln𝑖++𝛾𝑁𝑖=1𝑏𝑖𝑏ln𝑖=+𝛾𝑁𝑖=1π‘£π‘–ξ€·π‘Žπ‘–ξ€Έ+𝑁𝑖=1𝑣𝑖𝑏𝑖=𝑣(π‘Ž)+𝑣(𝑏).(4) Furthermore, the proposed valuation function is a strictly increasing function, thus for any π‘Ž,π‘βˆˆ([0,1]𝑁,≀),π‘Ž<𝑏⇒𝑣(π‘Ž)<𝑣(𝑏). Finally, the aforementioned function (3) maps the least element of lattice ([0,1]𝑁,≀) to zero. On the other hand, βˆ‘π‘£(𝑂)=𝑁𝑖=10Γ—ln(0+𝛾)=0.

The aforementioned valuation function operates in a more flexible manner compared with other valuations proposed in the literature. First, the performance of FLR can be optimized by selecting different values of the location parameter. Second, if the first variable π‘₯ is assumed to be constant, then it will be converted to 𝑣(π‘₯)=ln(π‘₯+𝛾), in the space 𝑅, and in some special applications, it might be a proper valuation function. One may say that it does not satisfy the condition 𝑣(𝑂)=0, in this case, it can be defined as follows [14]: π‘£ξ…ž(π‘₯)=𝑣(π‘₯)βˆ’π‘£(𝑂),βˆ€π‘₯∈𝐿.(5)

Figure 1(a) plots the positive valuation function 𝑣(π‘₯)=ln(π‘₯+𝛾), whereas Figure 1(b) plots 𝑣(π‘₯)=π‘₯ln(π‘₯+𝛾) for 𝛾=1.

A lattice (𝐿,≀) is totally ordered if and only if for any π‘Ž,π‘βˆˆπΏ, either π‘Žβ‰₯𝑏 or π‘Ž<𝑏. The lattice ([0,1]𝑁,≀) under inequality relation is not a totally ordered lattice.

A fuzzy lattice is a pair (𝐿,πœ‡) where (𝐿,≀) is a crisp lattice and (𝐿×𝐿,πœ‡) is a fuzzy set with membership function πœ‡βˆΆπΏΓ—πΏβ†’[0,1] such that πœ‡(π‘Ž,𝑏)=1β‡”π‘Žβ‰€π‘. Note that given a lattice 𝐿, for which a positive valuation function π‘£βˆΆπΏβ†’π‘… can be defined with 𝑣(𝑂)=0, then both (𝐿,π‘˜) and (𝐿,𝑠) are fuzzy lattices [2].

Consider the set 𝑅 of real numbers. It turns out that (βˆ’π‘…=𝑅βˆͺ{βˆ’βˆž,+∞},≀) under the inequality relation ≀ between π‘Ž,π‘βˆˆπ‘… is a complete lattice with the least element βˆ’βˆž and the greatest element +∞ [25].

For lattice (𝐿,≀), we define the set of (closed) intervals as 𝜏(𝐿)={[π‘Ž,𝑏]βˆ£π‘Ž,π‘βˆˆπΏandπ‘Žβ‰€π‘}. We remark that (𝜏(𝐿),≀) is a lattice with the ordering relation, lattice join and meet defined as follows [25]: []≀[][]∨[]=[],[]∧[]=[].π‘Ž,𝑏𝑐,π‘‘β‰‘π‘β‰€π‘Ž,𝑏≀𝑑,π‘Ž,𝑏𝑐,π‘‘π‘Žβˆ§π‘,π‘βˆ¨π‘‘π‘Ž,𝑏𝑐,π‘‘π‘Žβˆ¨π‘,π‘βˆ§π‘‘(6)

Including a least (the empty) interval, denoted by [𝐼,𝑂], to (𝜏(𝐿),≀) leads to a complete lattice (πœπ‘‚(𝐿),≀)=(𝜏(𝐿)βˆͺ{[𝐼,𝑂]},≀). Note that in case of 𝐿=βˆ’π‘…, the lattice (πœπ‘‚(𝐿),≀) is equal to conventional intervals (sets) in βˆ’π‘…. Our particular interest here is in the complete lattice (πœπ‘‚([0,1]),≀) with greatest element [0,1] and least element [1,0].

An isomorphic function πœ‘ from poset 𝑃 to poset 𝑄 is a map if both β€œπ‘₯β‰€π‘¦π‘–π‘›π‘ƒβ‡”πœ‘(π‘₯)β‰€πœ‘(𝑦)𝑖𝑛𝑄” and β€œπœ‘ is onto Q.” Based on the positive valuation function 𝑣 of lattice (𝐿,≀) and an isomorphic function πœƒβˆΆ(𝐿,β‰€πœ•)β†’(𝐿,≀), a valuation function π‘£πœπ‘‚ in (πœπ‘‚(βˆ’π‘…),≀) is defined as π‘£πœπ‘‚([]π‘Ž,𝑏)=𝑣(πœƒ(π‘Ž))+𝑣(𝑏).(7)

As a consequence, the degree of inclusion of an interval in another one in lattice (πœπ‘‚(𝐿),≀) is computed as follows [25]: π‘˜πœπ‘‚([],[]π‘£π‘Ž,𝑏𝑐,𝑑)=πœπ‘‚([])𝑐,π‘‘π‘£πœπ‘‚([]∨[]),π‘ π‘Ž,𝑏𝑐,π‘‘πœπ‘‚([],[]π‘£π‘Ž,𝑏𝑐,𝑑)=πœπ‘‚([]∧[])π‘Ž,𝑏𝑐,π‘‘π‘£πœπ‘‚([]).π‘Ž,𝑏(8)

For two 𝑁-dimensional hypercubes 𝐴=[π‘Ž1,𝑏1]Γ—β‹―Γ—[π‘Žπ‘,𝑏𝑁] and 𝐡=[𝑐1,𝑑1]Γ—β‹―Γ—[𝑐𝑁,𝑑𝑁], the following inclusion measure between two intervals 𝐴 and 𝐡 is defined: π‘˜πœπ‘‚π‘£(𝐴,𝐡)=πœπ‘‚(𝐡)π‘£πœπ‘‚=βˆ‘(𝐴∨𝐡)𝑁𝑖=1ξ€Ίξ€·π‘π‘£πœƒπ‘–ξ€Έξ€·π‘‘+π‘£π‘–ξ€Έξ€»βˆ‘π‘π‘–=1ξ€Ίξ€·π‘Žπ‘£πœƒπ‘–βˆ§π‘π‘–ξ€Έξ€·π‘+π‘£π‘–βˆ¨π‘‘π‘–.ξ€Έξ€»(9)

3. FLR Model

This section presents a classifier for extracting rules from the input data based on fuzzy lattices. One of FLR important properties is the ability of dealing with disparate type of data, including real vectors, fuzzy sets, symbols, graphs, images, waves, and even any combination of the aforementioned data and this shows the ability of FLR in combining different types of data. Furthermore, FLR can handle both complete and noncomplete lattices, and it can cope with both points and intervals. Moreover, stable learning is carried out both incrementally and fast in a single pass through the training data. In some applications, we might face with β€œmissing” or β€œdo not care” data. In this case, FLR can manage β€œmissing” and β€œdo not care” data by replacing them with least element 𝑂 and great element 𝐼, respectively. For example, if the constituent lattice is ([0,1],≀), then we can replace β€œmissing” and β€œdo not care” data by intervals of 𝑂=[1,0] and 𝐼=[0.1], respectively [13, 14, 22].

It should be mentioned that an input datum to the FLR classifier (model) is represented as (π‘Žπ‘–,𝐢𝐾) where 𝐢𝐾 is the class label of the datum π‘Žπ‘–, and it can be interpreted as a rule β€œif π‘Žπ‘–β€‰β€‰then πΆπ‘˜.” We remark that a single real number π‘Žβˆˆπ‘… corresponds to the trivial interval [π‘Ž,π‘Ž]. Learning and generalization in FLR is based on the computation of hyperboxes in space 𝑅𝑁, that is, a rule induced by FLR corresponds to an N-dimensional hypercube.

Suppose a knowledge base KB={(π‘Ž1,𝐢1),…,(π‘Žπ‘,𝐢𝑐)} is given. KB can be empty at first. Decision making in FLR is based on an inclusion measure. During learning phase, when an input datum (π‘Ž0,𝐢0) is presented to the network, the degree of inclusion between input and each stored rules in KB will be calculated as π‘˜(π‘Ž0,π‘Ž1),…,π‘˜(π‘Ž0,π‘Žπ‘), respectively. The fuzzy lattice reasoning classifier will choose the rule with argmaxπ‘–βˆˆ{1,…,𝑐}{π‘˜(π‘Ž0,π‘Žπ‘–)} as the winner. If the winner rule π‘Žπ½ and input datum π‘Ž0 have the same class label and the size of (π‘Ž0βˆ¨π‘Žπ½), denoted by 𝑍, is less than a user defined threshold, then the winner rule will be updated. Note that the size of an interval [π‘Ž,𝑏] is computed as 𝑍[π‘Žπ‘]=𝑣(𝑏)βˆ’π‘£(π‘Ž). Otherwise, this process is repeated; if no more rules are left, then the input datum (π‘Ž0,𝐢0) will be a new member of KB. Algorithm for training is described in Algorithm 1.

S0. The first input ( π‘Ž 0 , 𝐢 0 ) is memorized. A t an instant, there are 𝑐 Known
Classes 𝐢 1 , … , 𝐢 𝑐 memorized in the memory, initially 𝑐 = 0 .
S1.   Present the next input ( π‘Ž 𝑖 , 𝐢 π‘˜ ) , 𝑖 = 1 , . . . , π‘š to the initially β€œset” family of rules.
S2.  If no rules are β€œset” then
Store input ( π‘Ž 𝑖 , 𝐢 𝐾 ) ,
𝑐 = 𝑐 + 1 ,
     Go to S1.
   Else
Compute π‘˜ ( π‘Ž 0 , π‘Ž 𝑖 ) , 𝑖 = 1 , … , 𝑐 of the β€œset” rules.
S3.    Competition among the β€œset” rules:
     Winner is rule ( π‘Ž 𝐽 , 𝐢 𝐽 ) such that 𝐽 = a r g m a x { π‘˜ ( π‘Ž 0 , π‘Ž 𝑖 ) } , 𝑖 = 1 , … , 𝑐 .
S4.    The Assimilation Condition:
   Both 𝑍 ( π‘Ž 𝑖 ∨ π‘Ž 𝐽 ) ≀ 𝜌 and 𝐢 𝑖 = 𝐢 𝐽 .
S5.     If the Assimilation Condition is satisfied then
   Replace π‘Ž 𝐽 by π‘Ž 0 ∨ π‘Ž 𝐽 .
      Else
    β€œreset” the winner ( π‘Ž 𝐽 , 𝐢 𝐽 ) , Go to S2.

Note that 𝜌 is the threshold size which is used to specify the maximum size of a hyperbox to be learned. The decision boundaries which can be formed by FLR endowed with logarithmic valuation function are illustrated in the following example.

Example 3. The Simpson benchmark is a two-dimensional data set consisting of 24 points which is used for testing the performance of a clustering algorithm [26]. This is a perceptual grouping problem in vision, which deals with the detection of the right partition of an image into subsets [27]. We have divided the data into three classes.

Figure 2 shows the decision surfaces of the FLR endowed with proposed logarithmic valuation function without any misclassified data. Here, due to lack of space, we have classified the data four times for location parameter (𝛾) equal to 𝛾=1 and with four different vigilance parameter 𝜌 values. Note that the size of a hyperbox is tuned by the vigilance parameter 𝜌; more specifically, larger values of 𝜌 result in more hyperboxes.

As it was said in the previous section, one of the FLR properties is the ability of knowledge representation. Indeed, FLR is capable of extracting implicit features beyond the data and represents them as rules. Each rule is represented as if (π‘Ž1π΄π‘π·β‹―π΄π‘π·π‘Žπ‘€)  then 𝐢𝑖 where π‘Žπ‘—,𝑗=1,…,𝑀 are attributes, each one corresponding to an interval, and 𝐢𝑖,𝑖=1,…,𝑐 are class labels. Table 1 shows three induced rules corresponding to Figure 2(d).

4. Experimental Results and Discussions

4.1. Benchmark Dataset Description

In this section, we evaluate the classification performance of the optimized FLR in a series of experiments on six well-known benchmarks.

4.1.1. Object Recognition

We evaluate the classification performance of the FLR model using images of the Columbia Image database [28]. Columbia object image library (COIL-100) is a database of color images of 100 objects. We selected the 10 objects from the dataset shown in Figure 3. The objects were placed on a motorized turntable against a black background. The turntable was rotated through 360Β° to vary object pose with respect to a fixed color camera. Images of the objects were taken at pose intervals of 5Β°. This corresponds to 72 poses per object. The images were size normalized. There are 720 128Γ—128-dimensional instances divided into 10 separate classes; 72 for each class; only six randomly selected instances per each object were sufficient for whole training set, and the remaining patterns were used for testing set. The aim is the correct classification of the testing data to their corresponding classes.

4.1.2. Image Segmentation

The image segmentation data set was donated by the Vision Group, University of Massachusetts, and is included in the Machine Learning Repository of the University of the California, Irvine [29]. The Image Segmentation data set consisted of data relating numerous analyses of the colors in subdivided images to the type of surface in the image. Each image was divided up into small subsections, each of which comprised one point of data. Each data point was composed of 18 different attributes, including one that determined what the image was of: brick face, foliage, grass, sky, window, concrete, and dirt. This data set consists of 210 samples for training and 2100 samples for testing. The goal is to distinguish between seven different classes.

4.1.3. Pen-Based Recognition of Handwritten Digits

The Pen-based recognition of handwritten digits dataset was taken from the UCI repository of machine learning databases [29]. It was created by collecting 250 digit samples from 44 writers. The data have 16 continuous attributes distributed in 10 separated classes. A training set is given explicitly. For a faster simulation, we have resized the training set by selecting randomly six instances per each class. Distribution of digits between 0 and 9 in the dataset is shown in Figure 4.

4.1.4. Letter Recognition

The letter recognition benchmark was employed from the UCI repository of machine learning databases [29]. The data set consists of 20,000 unique letter images generated by randomly distorting pixel images of the 26 uppercase letters from 20 different commercial fonts. The parent fonts represented a full range of character types including script, italic, serif, and gothic. The features of each of the 20,000 characters were summarized in terms of 16 primitive numerical attributes. A training set is not given explicitly. We have divided all the data into a training set consisting of 10 percent of patterns which have been selected randomly and a testing set consisting of the remaining patterns. Examples of the character images are presented in Figure 5.

4.1.5. Semion Hand Recognition

The semion hand recognition benchmark was taken from the UCI repository of machine learning databases [29]. This dataset consists of 1593 handwritten digits from around 80 persons was scanned and stretched in a rectangular box 16Γ—16 in a gray scale of 256 values. Then each pixel of each image was scaled into a boolean value using a fixed threshold. Each person wrote on a paper all the digits from 0 to 9, twice. The commitment was to write the digit the first time in the normal way and the second time in a fast way. We have used 10 percent of data for whole training set.

4.1.6. Optical Recognition of Handwritten Digits

The optical recognition of handwritten digits benchmark was employed from the UCI repository of machine learning databases [29]. In this data set, 32Γ—32 bitmaps of handwritten digits from a total of 43 people are divided into nonoverlapping blocks of 4Γ—4, and the numbers of on pixels are counted in each block. This generates an input matrix of 8Γ—8 where each element is an integer in the range of 0 to 16. This reduces dimensionality and gives invariance to small distortions. Training and testing sets are given explicitly including 3823 and 1797 64-dimensional samples. For a faster simulation, the way we have used the data set was to employ 10 percent of the training set for actual training.

Table 2 shows briefly the characteristic of the selected benchmark data sets.

4.2. Experiments and Results

In order to provide a meaningful comparison, all the algorithms have been implemented in the same environment using the C++ object-oriented programming language, the same partitioning of data sets for training and testing, the same order of input patterns, and a full range of parameters, and we have employed the isomorphic function πœƒ(π‘₯)=1βˆ’π‘₯. Furthermore, all the 𝑁-dimensional data have been normalized into space [0,1]𝑁 by the function π‘₯norm=(π‘₯βˆ’π‘₯min)/(π‘₯maxβˆ’π‘₯min), where π‘₯min and π‘₯max stand for the least and the greatest attribute values, respectively, in a data dimension. In this work, the FLR algorithm endowed with linear valuation function π‘₯, nonlinear valuation arctan, and nonlinear logarithmic valuation function is denoted, respectively, by FLRπ‘₯, FLRπ‘Ž, and FLR𝑙. To compare the learning capability, Table 3 shows the comparison of the experimental results of FLR𝑙 with the ones produced by the FLRπ‘₯ and FLRπ‘Ž, the SOM [30], fuzzy ART [11], and GRNN [31].

In all our experiments in order to achieve the best performance, we have considered GRNN for different values of variance parameter between 0 and 0.5 in steps of 0.001. For fuzzy ART, we have set the choice parameter to 0.01, and the values of vigilance and learning parameters have been adopted between 0 and 1 in steps of 0.01. Computational experiments for the SOM algorithm have been carried out using 𝑀×𝑀(𝑀=1,…,10) grids of units and number of 100 epochs. Since the results produced by SOM depend on the initialization of the weights, we have chosen the weights that yielded best results on the testing set for 10 random initializations. The only parameter for FLR algorithm that should be tuned is the threshold size parameter 𝜌. In our simulations, the size of 𝜌 was set from 0.01 up to 𝑁, where 𝑁 is the dimension of the input data, in steps of 0.01 except for object recognition data set, we have set 𝜌 in steps of 50 due to high-dimensional input data. Just for FLR𝑙, the location parameter 𝛾 should be tuned too. We have set 𝛾 between 1 and 50 in steps of 1.

Table 3 cites the classification accuracy and ranking of different methods for each benchmark. In other words, each table cell, which belongs to a specific learning algorithm and the data set, contains the percentages of correct classification of that model over the corresponding data set. The number in brackets in each table cell shows the ranking of each method after running on a specific data set. The best results have been shown in bold face.

As can be seen in Table 3, FLR𝑙 has obtained acceptable results in comparison with other methods, and in five cases, it gets the first rank. In Table 4, the average classification accuracy on all data sets for each of the different learning algorithms along with the new relative obtained ranks has been shown. In other words, first the average of each column of the previous table has been calculated, then the corresponding ranking is shown within brackets. As it can be seen from among all the methods, FLR𝑙 has achieved the best ranking among all other methods.

In Table 5, this comparison has been made according to the sum of the ranks available in Table 3 per each column. Although this quantity is of lower precision degree for reporting results in some cases, it is common in nonparametric statistics. As can be seen in this table, FLR𝑙, FLRπ‘₯, and FLRπ‘Ž get first, second, and third rankings, respectively. It should be pointed out that although there is no universal learning algorithm that can get the best results on the all benchmarks, the results obtained by FLR𝑙 confirm that our proposed model is an efficient classifier compared with established classifiers from the literature.

5. Conclusion

In this work, we introduced an improvement of fuzzy lattice reasoning (FLR) classifier using a new nonlinear positive valuation function. We have investigated the performance of new FLR model in several well-known classification problems. Experimental results demonstrated that our proposed methods outperformed established classification models in terms of classification accuracy on the testing data.