Abstract

The gradient learning model has been raising great attention in view of its promising perspectives for applications in statistics, data dimensionality reducing, and other specific fields. In this paper, we raise a new gradient learning model for ontology similarity measuring and ontology mapping in multidividing setting. The sample error in this setting is given by virtue of the hypothesis space and the trick of ontology dividing operator. Finally, two experiments presented on plant and humanoid robotics field verify the efficiency of the new computation model for ontology similarity measure and ontology mapping applications in multidividing setting.

1. Introduction and Motivations

The term “ontology” is originally from the field of philosophy and it is used to describe the nature connection of things and the inherent hidden connections of their components. In information and computer science, ontology is a model for knowledge storing and representation and has been widely applied in knowledge management, machine learning, information systems, image retrieval, information retrieval search extension, collaboration, and intelligent information integration. In the past decade, as an effective concept semantic model and a powerful analysis tool, ontology has been widely applied in pharmacology science, biology science, medical science, geographic information system, and social sciences (e.g., see Hu et al., [1], Lambrix and Edberg [2], Mork and Bernstein [3], Fonseca et al., [4], and Bouzeghoub and Elbyed [5]).

The structure of ontology can be expressed as a simple graph. Each concept, object, or element in ontology corresponds to a vertex and each (directed or undirected) edge on an ontology graph represents a relationship (or potential link) between two concepts (objects or elements). Let be an ontology and a simple graph corresponding to . The nature of ontology engineer application can be attributed to get the similarity calculating function which is to compute the similarities between ontology vertices. These similarities represent the intrinsic link between vertices in ontology graph. The goal of ontology mapping is to get the ontology similarity measuring function by measuring the similarity between vertices from different ontologies, such mapping is a bridge between different ontologies, and get a potential association between the objects or elements from different ontologies. Specifically, the ontology similarity function is a semipositive score function which maps each pair of vertices to a nonnegative real number.

Example 1. Ontology technologies are widely used in humanoid robotics in recent years. Different bionic robot has a different structure. Each bionic robot or each component of a bionic robot can be represented as an ontology. Each vertex in ontology stands for a part or a construction, edge between vertices represents a direct physical link between these constructs, or these parts have intrinsic link with its function. Thus, the similarity calculation between vertices in the same ontology allows us to find the degree of association and the potential link between different constructs in bionic robots. Similarity calculation between two different ontologies (i.e., ontology mapping building) allows us to understand the potential association for different components or parts in two biomimetic robots.

Example 2. In information retrieval, ontology concepts are often used in query expansion. The user queries the information related concept . If we manually set the parameters , the ontology algorithm will find that all concepts meet . Then the information related concepts will be returned to the user as the query expansion for concept .

Very recently, ontology technologies are employed in a variety of applications. Ma et al. [6] presented a graph derivation representation based technology for stable semantic measurement. Li et al. [7] raised an ontology representation method for online shopping customers knowledge in enterprise information. Santodomingo et al. [8] proposed an innovative ontology matching system that finds complex correspondences by processing expert knowledge from external domain ontologies and in terms of using novel matching tricks. Pizzuti et al. [9] described the main features of the food ontology and some examples of application for traceability purposes. Lasierra et al. [10] argued that ontologies can be used in designing an architecture for monitoring patients at home.

Traditional methods for ontology similarity computation are heuristic and based on pairwise similarity calculation. With high computational complexity and low intuitive, this model requires large parameters selection. One example of traditional ontology similarity computation method is where and are two vertices corresponding to two concepts; and ; Simname, Siminstance, Simattribute, and Simstructure are functions of name similarity, instance similarity, attribute similarity, and structure similarity, respectively. These similarity functions are determined by experts directly in terms of their experience. Hence, this model has the following deficiencies: (i)many parameters rely heavily on the experts;(ii)high computational complexity and thus being inapplicable to ontology with large number of vertices;(iii)pairwise similarities fall reflect the ontology structure intuitively.Thus, a more advanced way to deal with the ontology similarity computation is using ontology learning algorithm which gets an ontology function . By virtue of the ontology function, the ontology graph is mapped into a line which consists of real numbers. The similarity between two concepts then can be measured by comparing the difference between their corresponding real numbers.

The essence of this algorithm is dimensionality reduction. In order to associate the ontology function with ontology application, for vertex , we use a vector to express all its information (including its name, instance, attribute and structure, and semantic information of the concept which is corresponding to the vertex and that is contained in name and attribute components of its vector). In order to facilitate the representation, we slightly confuse the notations and use to denote both the ontology vertex and its corresponding vector. The vector is mapped to a real number by ontology function , and the ontology function is a dimensionality reduction operator which maps multidimensional vectors into one-dimensional vectors.

There are several effective methods for getting efficient ontology similarity measure or ontology mapping algorithm in terms of ontology function. Wang et al. [11] considered the ontology similarity calculation in terms of ranking learning technology. Huang et al. [12] raised the fast ontology algorithm in order to cut the time complexity for ontology application. Gao and Liang [13] presented an ontology optimizing model such that the ontology function is determined by virtue of NDCG measure, and it is successfully applied in physics education. Since the large part of ontology structure is the tree, Lan et al. [14] explored the learning theory approach for ontology similarity calculating and ontology mapping in specific setting when the structure of ontology graph has no cycle. In the multidividing ontology setting, all vertices in ontology graph or multiontology graph are divided into parts corresponding to the classes of rates. The rate values of all classes are determined by experts. In this way, a vertex in a rate has larger score than any vertex in rate (if ) under the multidividing ontology function . Finally, the similarity between two ontology vertices corresponding to two concepts (or elements) is judged by the difference of two real numbers which they correspond to. Hence, the multidividing ontology setting is suitable to get a score ontology function for an ontology application if the ontology is drawn into a noncycle structure. Gao and Xu [15] studied the uniform stability of multidividing ontology algorithm and obtained the generalization bounds for stable multidividing ontology algorithms.

In the above described ontology learning algorithms, their optimal ontology function calculation model or its solution strategy is done by gradient calculation. Specifically, the ontology gradient learning algorithm obtains the ontology function vector which maps each vertex into a real number (the value corresponds to vertex ). In this sense, it is good or bad policy gradient calculation algorithm that will determine the merits of the ontology algorithm. In this paper, we raise an ontology gradient learning algorithm for ontology similarity measuring and ontology mapping in multidividing setting. The organization of the rest paper is as follows: the notations and ontology gradient computing model are directly presented in Section 2; the detailed description of new ontology algorithms is shown in Section 3; in Section 4, we obtain some theoretical results concerning the sample error and convergence rate; in Section 5, two simulation experiments on plant science and humanoid robotics are designed to test the efficiency of our gradient computation based ontology algorithm, and the data results reveal that our algorithm has high precision ratio for plant and humanoid robotics applications.

2. The Gradient Computation Model for Ontology in Multidividing Setting

In order to combine the machine learning technology and ontology frame, the relevant information for each vertex in ontology graph is represented as an -dimensional vector. Hence the vertex set is a subset of (vertex space or input space for ontology). Assume that is compact. In the supervised learning, let be the label set for . Denote as a probability measure on . Let and be the marginal distribution on and conditional distribution at , respectively. The ontology function associated with is described as .

For each vertex , denote . Then, the gradient of the ontology function is the vector of ontology functions

Let be a random sample independently drawn according to in standard ontology setting. The purpose of standard ontology gradient learning is to learn from the sample set . From the perspective of statistical learning theory, the gradient learning algorithm is based on the Taylor expansion if two vertices have large common information (i.e., ). We expect that and if , . The demand is met by virtue of setting weights Using unknown ontology function vector to replace , then the standard least-square ontology learning algorithm is denoted as where and are two positive constants to control the smoothness of ontology function. Here is a positive semidefinite, continuous, and symmetric kernel (i.e., Mercer kernel) and is the reproducing kernel Hilbert space (for short, RKHS) associated with the Mercer kernel . The notation presented in (4) is the -fold hypothesis space of composing of vectors of ontology functions with norm .

By the representation theory in statistical learning theory, the ontology algorithm (4) can be implemented in terms of solving a linear system for the coefficients of , where for is the ontology function in and . Let be the rank of the matrix ; hence the coefficient matrix for the linear system has size . Therefore, this size will become huge if the size of sample set is large itself. The standard approximation ontology algorithm allows us to solve linear systems with coefficient matrices of smaller sizes.

The gradient learning model for ontology algorithm in standard setting is determined as follows: where the sample set , , , is the sequence of step sizes and is the sequence of balance parameters.

For multidividing ontology setting, the vertex in ontology sample set can be divided into rates. Let with for . Denote , and is the label of for and . Hence, (4) becomes We obtain the following gradient computation model for ontology application in multidividing setting which corresponds to (5): Here in (6) and (7), .

We emphasize that our algorithm in multidividing setting is different from that of Wu et al. [16]. First, the label for ontology vertex is used to present its class information in [16], that is, , while in our setting, . Second, the computation model in [16] relies heavily on the convexity loss function , while our algorithm depends on the weight function .

3. Description of Ontology Algorithms via Gradient Learning

The above raised gradient learning ontology algorithm can be used in ontology concepts similarity measurement and ontology mapping. The basic idea is the following: via the ontology gradient computation model, the ontology graph is mapped into a real line consisting of real numbers. The similarity between two concepts then can be measured by comparing the difference between their corresponding real numbers.

Algorithm 3 (gradient calculating based ontology similarity measure algorithm). For and is an optimal ontology function determined by gradient calculating, we use one of the following methods to obtain the similar vertices and return the outcome to the users.
Method 1. Choose a parameter and return set .
Method 2. Choose an integer and return the closest concepts on the value list in .
Clearly, method 1 looks like fairer, but method 2 can control the number of vertices that return to the users.

Algorithm 4 (gradient calculating based ontology mapping algorithm). Let be ontology graphs corresponding to ontologies . For () and being an optimal ontology function determined by gradient calculating, we use one of the following methods to obtain the similar vertices and return the outcome to the users.
Method 1. Choose a parameter and return set .
Method 2. Choose an integer and return the closest concepts on the list in .
Also, method 1 looks like fairer and method 2 can control the number of vertices that return to the users.

4. Theoretical Analysis

In this section, we give certain theoretical analysis for our proposed multidividing ontology algorithm. Let and . We divide this section into two parts: first, some useful lemmas are prepared; then, main results in our paper concerning approximation conclusions are presented. Our error analysis depends on integral operators and gradient learning, and more references on these tricks can be referred to in Mukherjee and Wu [18], Mukherjee et al. [19], Yao et al. [20], and Rosasco et al. [21].

Set In what follows, , Our tricks of proofs in this paper follow from [22, 23].

4.1. Preliminary Results

Let sequence be the noise-free limit of the sequence (7) which is determined by and Our error analysis for proving main result (Theorems 12 and 13 in the next subsection) consists of two parts: sample error and approximation error.

The main task in this subsection is to estimate the sample error in terms of McDiarmid-Bernstein-type probability inequality and the multidividing sampling operator. For each , the multidividing sampling operator associated with a discrete subset of is defined by The adjoint of the multidividing ontology sampling operator, , is given by where Let us express (7) by virtue of the multidividing ontology sampling operator. Note that For each pair of with , we single out one summation from (7) as We infer that Denote Hence, we have Thus, it confirms the following representation for the sequence . For simplicity, let in the following contents.

Lemma 5. Set If is defined by (7), we deduce

We should discuss the convergence of the multidividing ontology operator to the integral operator determined by where .

Lemma 6. Let be multidividing sample set independently drawn according to a probability distribution on . Denote as a Hilbert space and suppose that is measurable. If there is nonnegative such that for each and almost every , then for every , where For any , with confidence , one gets

By regarding and as elements in and , the space of bounded linear multidividing ontology operators on , Lemma 6 cannot be directly employed because is not a Hilbert space, but a Banach space only. Therefore, we consider a subspace of , which is the space of Hilbert-Schmidt operators on with inner product . As is a subspace of , their norm relations are presented as In addition, is a Hilbert space and contains multidividing ontology operators and . By applying Lemma 6 to this Hilbert space, we obtain the following lemma.

Lemma 7. Let be multidividing sample set independently drawn from . With confidence , one obtains

Proof. Let . Consider the multidividing ontology function with values in defined by For , we confirm that Recall that reproducing property of the RKHS says that It implies that the rank of operator determined by is 1, and also in . Furthermore, . Let be the operator on which maps to . Then the above fact reveals that . Hence for any , we infer that Using the fact that and , we deduce that Since the stated result is held by combining Lemma 6 with and using the bound .

In order to find the difference between and , the convergence of to the ontology function defined by (55) is studied.

Lemma 8. Let be a multidividing ontology sample independently drawn from . With confidence , one has

Proof. By applying Lemma 6 to the Hilbert space and the ontology function given by we yield . Hence, for almost every , we get Lemma 6 implies that for any , with confidence , we obtain Finally, conclusion follows from the fact that .

Obviously, for , the sequence has a similar expression as (20).

Lemma 9. Let be an ontology operator on and suppose that . For the ontology operator determined by (22) and by (10), one obtains

The sample error is stated in the following conclusion.

Theorem 10. Let be obtained by (5) and by (10). Suppose that and for all . Then for any , with confidence , one infers that

Proof. Let Let with measure at least such that (36) establishes for any . Thus, from the positivity of the multidividing ontology operator (for each pair of ) on and the assumption , we have that for any ,
In terms of and , we get By virtue of the assumptions on , , we infer that which implies that for any .
Now, we consider the estimate of . Let with measure at least such that (27) is established for any . In view of (26), for each we yield
Using the fact that , we obtain that for any , By changing the order of summation, we determine that According to (45), we can verify that is bounded by In view of the above fact and (46), we obtain that for any , However, the measure of the subset of is at least . The desired conclusion follows after substituting for .

The following result is Theorem 4 in Dong and Zhou [23]; it also holds in multidividing setting and we skip the detailed proof.

Theorem 11. Let be determined by (53). Then, we deduce that

4.2. Main Results

The first main result in our paper implies that is a good approximation of a noise-free limit for the ontology function (6) as a solution of (8) which we refer as multidividing ontology function .

Theorem 12. Let , , and and satisfy and . For any , take Define by (7) and by (8). If is almost established, then for any , with confidence , one has where constant independent of or and is the multidividing ontology function determined by The proof of Theorem 12 follows from Theorems 10 and 11 and an exact expression for the constant relying on and can be easily determined.

The second main result in our paper follows from Theorem 10 and the technologies raised in [23].

Theorem 13. Assume that for certain , and for any , the marginal distribution satisfies and the density of exists and for any, any satisfies Suppose that the kernel and . Let and . Take , , and and suppose that ; then for any , with confidence , one infers that where and constant is independent of or .

Proof. Obviously, under the assumptions , (56) and (57), we get Furthermore, by virtue of Proposition 15 in Mukherjee and Zhou [22], we have where constant relies on and . Theorem 10 and these estimates reveal that with confidence , we yield The learning rate (58) is determined according to the selection of the parameters.

5. Experiments

To show the effectiveness of our new ontology algorithms, two experiments concerning ontology measure and ontology mapping are designed below.

5.1. Ontology Similarity Measure Experiment on Plant Data

In the first experiment, we use plant “PO” ontology which was constructed in the website http://www.plantontology.org/. The structure of is presented in Figure 1. (precision ratio; see Craswell and Hawking [24]) is used to measure the quality of the experiment data. Here, we take , , , and .

We first give the closest concepts for every vertex on the ontology graph by experts in plant field, and then we obtain the first concepts for every vertex on ontology graph by Algorithm 3 and compute the precision ratio. Specifically, for vertex and given integer . Let be the set of vertices determined by experts and it contains vertices having the most similarity of . Let Then the precision ratio for vertex is denoted by The average precision ratio for ontology graph is then stated as

At the same time, we apply ontology methods in [1113] to the “PO” ontology. Calculating the average precision ratio by these three algorithms and comparing the results to Algorithm 3 rose in our paper, part of the data is referred to in Table 1.

When = 3, 5, or 10, the precision ratio by virtue of our gradient computation based algorithm is higher than the precision ratio determined by algorithms proposed in [1113]. In particular, when increases, such precision ratios are increasing apparently. Therefore, the gradient learning based ontology Algorithm 3 described in our paper is superior to the method proposed by [1113].

5.2. Ontology Mapping Experiment on Humanoid Robotics Data

For the second experiment, we use “humanoid robotics” ontologies and . The structure of and is shown in Figures 2 and 3, respectively. The ontology presents the leg joint structure of bionic walking device for six-legged robot, while the ontology presents the exoskeleton frame of a robot with wearable and power-assisted lower extremities. In this experiment, we take , , , and .

The goal of this experiment is to give ontology mapping between and . We also use precision ratio to measure the quality of experiment. Again, we apply ontology algorithms in [12, 13, 17] on “humanoid robotics” ontology and compare the precision ratio which is gotten from three methods. Some results referred to in Table 2.

Taking = 1, 3, or 5, the precision ratio in terms of our gradient computation based ontology mapping algorithm is higher than the precision ratio determined by algorithms proposed in [12, 13, 17]. Particularly, as increases, the precision ratios in view of our algorithm are increasing apparently. Therefore, the gradient learning based ontology Algorithm 4 described in our paper is superior to the method proposed by [12, 13, 17].

6. Conclusions

As a data structural representation and storage model, ontology has been widely used in various fields and proved to have a high efficiency. The core of ontology algorithm is to get the similarity measure between vertices on ontology graph. One learning trick is mapping each vertex to a real number, and the similarity is judged by the difference between the real number which the vertices correspond to. In this paper, we raise a gradient learning model for ontology application in multidividing setting. The sample error and approximation properties are given in our paper. These results support the gradient computation based ontology algorithm from the theoretical point of view. The new technology contributes to the state of the art for applications and the result achieved in our paper illustrates the promising application prospects for multidividing ontology algorithm.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

Acknowledgments

This work was supported in part by the Key Laboratory of Educational Informatization for Nationalities, Ministry of Education, the National Natural Science Foundation of China (60903131), the College Natural Science Foundation of Jiangsu Province in China (10KJD520002), and the Ph.D. initial funding of the first author. The authors are grateful to the anonymous referee for careful checking of the details and for helpful comments that improved this paper.