Research Article | Open Access
Yanping Chen, Mingdao Zhao, Hong Xia, Xiaodong Jin, Zhongmin Wang, Zhong Yu, "A Method for Extracting High-Quality Core Data from Edge Computing Nodes", Mathematical Problems in Engineering, vol. 2019, Article ID 3834846, 10 pages, 2019. https://doi.org/10.1155/2019/3834846
A Method for Extracting High-Quality Core Data from Edge Computing Nodes
Intelligent factory has the characteristics of wide data sources, high data dimensions, and strong data relevance. Intelligent factories need to make different decisions for different needs, so they need to efficiently analyze these data and explore the inherent laws contained in them. At the same time, the increasing amount of data brings various burdens to the network infrastructure between users and smart devices. For the above needs, this paper proposes a tension-based heterogeneous data fusion model in the edge computing layer, which represents the multisource heterogeneous data in the industrial scene as a tensor model, and uses the incremental decomposition algorithm to extract high-quality core data. The model reduces the data flow between the data center and the central cloud while retaining the core data set. Experiments show that the approximate tensor reconstructed from the tensor with 15% core data can guarantee 90% accuracy.
Driven by emerging technologies such as big data, cloud computing, and the Internet of things, intelligent manufacturing has become the theme of today’s manufacturing industry, in which intelligent factories and intelligent production are the core contents of intelligent manufacturing. Intelligent factories relying on big data and Internet of things technology still face severe technical challenges in the face of massive industrial big data.
Industrial big data is the general term of industrial data generated in the whole life cycle of industrial production. Industrial big data has 4V characteristics of generalized big data: volume, variety, value, and velocity. In addition, it has new characteristics different from other industries :(1)Data sources are extensive, and the proportion of semistructured and unstructured data increases.(2)There is high correlation between data.(3)The characteristics of time and space should be considered in data analysis.(4)The data are specific to industrial scenes and highly professional.
Big data mining aims to acquire knowledge from a large amount of complex data and create new value . In the process of industrial production, data comes from sensor, intelligent equipment, workstation, production process data of production control system, operation monitoring data, and log record data of various production workshops distributed in different geographical locations. These data are partly structured data. At the same time, there are a lot of unstructured and semistructured data, such as text, log file and sound, image, and video. In order to transform data into knowledge, data warehouse, online analytical processing, and data mining technologies are needed. The traditional method of data storage and management is oriented toward relational structured data, where it is difficult to meet such a large and diverse demand for analysis of unstructured data.
In a traditional industrial cloud architecture, all data from physical devices is transferred to the cloud for storage and advanced analysis. Since cloud platforms have higher computing power than devices on the edge of the network, transferring computation-intensive tasks to core cloud computing platforms is an effective method for data processing. However, the transmission of large amounts of data over the network will impose a huge burden on the network equipment. In addition, it does not meet the requirements of location awareness, low latency, and mobility support. Therefore, we need to face the following two challenges:(1)Difficulties in information integration : There are various sources of industrial big data, different data structures, different attributes and standards, production cycle data, relational data from enterprises, unstructured or semistructured data such as video surveillance data, XML log, etc. Each data system is independent of the others, at the same time, due to different data sources, production equipment, and soft. Various factors, such as the diversity of component providers, lead to different data formats, which makes it difficult to achieve information integration. Based on the above challenges, a unified specification of data format is needed.(2)Challenges posed by system complexity : In the face of huge scale, complicated structure, multisource heterogeneous data, industrial big data of value, and the processing of faces including computing complexity, real-time demand is high; the correlation between strong challenges, these problems on the operation of the large cloud data processing system transmission efficiency, and performance of the system put forward the stringent requirements, which need high intelligent analysis of data mining methods.
From the above analysis, it can be seen that the heterogeneous data fusion method is the premise and basis for the efficient analysis of data required by the intelligent factory. However, the current data fusion method is unable to quantify the unstructured, semistructured, and structured data in the industry in a unified way, and it is also difficult to analyze them as a whole. In addition, the increasing data volume, data transmission delay, and data transmission loss also put forward higher requirements for the underlying network transmission of intelligent factories.
2. Related Work
Traditional data analysis deals with problems in a single data domain, while industrial big data comes from a wide variety of sources, including different data domains, different data formats, different characteristics, distributions, scales, and densities. Some applications require low latency and real-time services. The traditional reliance on cloud computing system framework has become the bottleneck of today’s industrial big data platform. In recent years, many scholars at home and abroad have carried out a large number of studies based on the above characteristics.
For the effective representation of large-scale massive data, data description methods based on ontology  in complex unstructured data have been adopted in many researches. Through ontology description, classes, attributes, and relationships can be used to describe things, thus establishing a general reference model. Then, based on ontology, Huang Yinghui et al. proposed semantic Internet of things , but it was difficult to express high-dimensional data. When the semantic deviation between heterogeneous domain data is large, the accuracy of shared feature is difficult to guarantee. In other fields, some scholars use third-order tensors to represent heterogeneous data. Tensor is a big data analysis tool, which can maintain the internal structure of complex type data and is an extension of matrix. It has prominent advantages in the representation and processing of complex, high-order, and multidimensional data. Researcher Liao  used the third-order tensor to solve the recommendation problem of users’ interest points in terms of topics and time. Yilun Wang  used the third-order tensor to predict the driving time of a driver on a certain trajectory through the historical data of the driver’s car. The existing tensor models are all three-dimensional tensor models for specific application scenarios, which are relatively simple low-order mathematical models and cannot achieve unified representation of multisource heterogeneous cross-domain data in big data. However, in the industrial scenario, there is no model to uniformly represent the multisource heterogeneous data in this scenario.
How to effectively extract high-quality core data from large-scale low-quality data sets is a severe challenge for big data analysis. Tensor decomposition is a method of processing large-scale data, which can effectively extract the core data from the data set. Traditional decomposition methods include CP decomposition  and Tucker decomposition . However, CP decomposition algorithm is not stable, and solving the optimal rank of CP decomposition of high-order tensor is an NP hard problem . Although Tucker’s decomposition is a stable decomposition algorithm, the core tensor obtained by Tucker’s decomposition does not reduce the order of the original tensor. It is still a high-order tensor, and the analysis and calculation of the core tensor still bear the dimensional disaster problem. With the emergence of tensor chain decomposition , more and more applications have been proposed for the operation and processing of high-order tensor data based on tensor chain decomposition. Gorodetsky  proposed a method based on tensor chain decomposition to solve the optimal solution of the control synthesis problem efficiently. Chen  constructed the classifier model based on the tensor chain model. Phien  combines data recovery algorithm with tensor chain data, and the application of tense-based algorithm reflects the efficiency advantage of tensor chain decomposition in processing high-order tensors. In short, current algorithms all process data in the form of batch processing, but when the data is presented in the form of streaming data, there is no way to directly conduct incremental operation on tensors, since repeated decomposition of tensor data will consume a lot of computing resources to conduct incremental operation.
Incremental nature is one of the intrinsic characteristics of big data. Real life widely distributed sensing and monitoring equipment and real-time Internet social media have constituted the dynamic growth of big data online scenario. Based on the data analysis of large data, we not only obtain historical data, but also make a dynamic discovery of new data. Incremental nature mainly embodies three aspects: the first is the incremental data samples; the second is the increment of description information of sample features; and the third is the change of category increment and data distribution. Xu and researchers studied  how sampling methods impact on incremental support vector machine (SVM) algorithm and put forward a kind of incremental support vector machine (SVM) based on Markov algorithm of resampling (MR-ISVM), realizing that ISVM learning efficiency improved significantly. Gu , based on cost sensitive hinge loss support vector machine (CSHL-SVM) incremental learning algorithm, constructs the data block. In  Ahmad et al. used concept drift method for streaming data of unsupervised learning, effectively improving the precision of online anomaly detection. The static data analysis method of traditional DSS cannot make the right decisions at the time of onset of concept drift problems; for example, Dong , to study the concept drift problem of data driven decision support system, proposed a detection method based on the concept of data distribution drift, for better, more detailed data flow distribution, making the knowledge of DSS able to adjust the decisions at the right time to adapt to changing circumstances. Lobo  used kernel density estimation method to generate evolutionary diversification and fast adaptation of learning strategies after concept shift in online learning. According to the above analysis, when the data is presented in the form of streaming data, the existing algorithms are all based on learning algorithm. For specific scenarios, the disadvantage is that the accuracy of features is difficult to guarantee. The method proposed in this paper presents heterogeneous data in the form of tensors. There is no existing method that can directly conduct incremental operation on tensor data, since repeated decomposition of tensor data will consume a lot of computing resources to conduct incremental operation.
As a result, this paper proposes a calculation level edge oriented core data extraction method, the data acquisition layer, the heterogeneous data from different data sources through the data acquisition equipment, then, in the edge layer, the acquisition of data preprocessing, with each type of data represented by different order number, dimension of tensor, and then incremental data dimension reduction cleaning process for core tensor data.
With the efforts of many scholars, multisource heterogeneous data fusion and core data extraction methods have achieved certain results, but these methods cannot be directly applied to industrial big data. Therefore, at present, for multisource heterogeneous data fusion methods, domain-specific representation is relatively simple, mainly including feature standardization, common extraction, feature selection of weight, and other methods. These methods will lose or change some features when processing big data. How to fuse multisource heterogeneous data in the industrial big data environment while ensuring the completeness of data and the simplicity of the fusion model is the difficulty to be solved. During the analysis and processing of large-scale data in big data, due to the huge data size, the calculation result will increase too much, which has an impact on the performance of big data processing system. The traditional calculation method is considered to solve the problem by optimizing the calculation process and rescheduling the task. Communication, transmission delay, and other requirements are not considered. However, transmission delay and other factors must be considered in the industrial big data environment. This is also a difficult point.
Therefore, this paper proposes a tensor-based method for edge computing-oriented industrial heterogeneous data fusion and an extraction method for high-quality core data. The main contents are as follows:(i)Data fusion model: different types of data in industrial scenes are quantitatively encoded and low-order tensors are constructed, and then the fusion is extended into high-order tensors to achieve heterogeneous data fusion.(ii)Incremental SVD algorithm based on Lanczos: incremental algorithm is designed to update data.
3. United Data Representation Method
3.1. Big Data Fusion Processing Framework
This paper proposes a tensor-based big data fusion and processing framework (TBFPF), as shown in Figure 1.(1)Big data collection: Unstructured data, semistructured data, and structured data from different fields are acquired by the acquisition device, the data stream is formed and submitted to the edge computing layer for tensorization representation, and the format of source data is not changed during the submission process.(2)Big data tensorization representation: The data of different structure types are first transformed into low-order subtensors in the edge computing level and then merged using tensor extension operators to arrange different data features into different orders of tensor space, and finally we establish high-order big data unified representation model.(3)Big data dimensionality reduction: In the edge computing level, the incremental Lanczos-based high-order singular value decomposition method is used to reduce the dimensionality of low-quality raw data, remove inconsistent, inaccurate, redundant, collinear data, and obtain high-quality core data. From a big data computing perspective, core data is smaller, more accurate, and easier to handle.(4)Big data analysis: This module includes various big data methods, such as multimodal prediction, association analysis, clustering, and classification, to mine the inherent laws implied in large-scale heterogeneous data and user behavior patterns. The analysis results are then presented to the upper application.(5)Big data application: The top layer of the framework is big data applications, including smart city, smart factory, smart healthcare, smart grid, and other types of big data application systems. Through in-depth analysis of big data and analysis of internal laws, the big data application layer will provide users with active, convenient, and intelligent services.
This paper mainly studies the unified representation of data at the edge computing level and the method of extracting core tensor data.
3.2. Data Tensor Representation in the Edge Computing
3.2.1. Unstructured Data Tensor Representation
In subtensorization representation method for unstructured video data (Figure 2), the main features of video data are time frame, the width and height of the frame picture, and color. Therefore, a fourth-order tensor can be used to represent RMVB format video data, and the elements value in the tensor is the encoded value of the video. The formula is as follows:
where , , , and represent video frame, width of image per frame, length of image per frame, and color gamut of image, respectively. For example, an RMVB format video clip with a 400-frame resolution of 768 × 276 pixels can be expressed as a fourth-order tensor by the following formula.
3.2.2. Semistructured Data Tensor Representation
In subtensorization representation method for semistructured JSON data (Figure 3), the database of semistructured data is a collection of nodes. Each node is a leaf node or an internal node. Each semistructured data set has a hierarchical structure, which can be decomposed into a tree structure. This paper represents a JSON document as a third-order subtensor, expressed as the following formula:
where denotes the row of the identification matrix, denotes the column of the identification matrix, and denotes the encoding of the element.
3.2.3. Structured Data Tensor Representation
Structured data is data that is logically expressed and implemented by a two-dimensional table structure, mainly managed and stored through a relational database. In a simple type of database table, a field is often represented by a number or a characters, so that it can be represented as a matrix. For complex types of fields, you can add new tensors order to represent them. The structured data shown in Table 1 can be expressed as the following tensor.
3.2.4. Subtensor Fusion Method
Defining the tensor fusion extension operator: if the tensor and the tensor represent two third-order tensors, the tensor fusion extension operator is defined as follows.
When two tensors have orders of the same attribute, they can be merged by tensor extension, and the order of those different attributes is preserved.
3.2.5. Unified Tensor Fusion Model
To reduce data redundancy and duplication, the subtensor is converted into a uniform tensor using a uniform data tensor function as shown below.
The independent variables in the formula represent structured data, semistructured data, and unstructured data, respectively. These different types of data represent different subtensors and then are fused by tensor extension operators and uniformly expressed as high-order tensors corresponding to unified variable data structures in computer systems.
4. Incremental Dimension Reduction Method
To solve the dimensional reduction problem of large-scale heterogeneous data, two problems need to be faced:(1)Tensor model dimension is too high: In the tensor operation, the expansion of the data dimension is exponential and increasing in terms of the complexity of the tensor operation.(2)Tensor decomposition repeated computation: In the process of calculating the core data of tensor, after the tensor is expanded according to module, the matrix of the tensor is decomposed by high-order singular value, and then the core tensor is obtained. However, in the process of obtaining the core tensor, a large number of intermediate calculation results will be generated between the original data and the newly added data, and these intermediate results will be repeated calculation, thus affecting the calculation efficiency and accuracy of the results.
Therefore, we propose a Lanczos-based incremental high-order singular value decomposition algorithm (IncLHOSVD), and our algorithm was achieved based on two equivalence theorems proposed by Kuang . First, we give the process of the Lanczos-based high-order singular value decomposition algorithm:
In line 1 of Algorithm 1, we first perform modulo expansion on tensor , and the algorithm in lines 2-10 performs bi-Lanczos operation on each modulus expansion matrix to change the matrix into a dual-diagonal matrix. Line 11 of the algorithm initializes each diagonal matrix, line 12 of the algorithm changes the double diagonal matrix into a diagonal matrix, and line 16 of the algorithm obtains the core tensor .
|Output: core tensor , Orthogonal matrix ;|
|The tensor is subjected to a mode expansion operation to obtain an expansion matrix ;|
|Given unit initial vector , let , ;|
|compute the of Wilkinson Displacement|
|Performing a Givens transformation on|
|obtain Orthogonal matrix|
In Algorithm 2 the first line expands the dimension of the original tensor and the new tensor to the same dimension. In the second line of the algorithm, the tensor modulus expansion is added to obtain the modulus expansion matrix . The fourth line of the algorithm maps each module expansion matrix to the left singular matrix of each module expansion matrix to obtain the new mapping matrix, decomposes the mapping matrix, and obtains the left singular matrix for subsequent calculation. The seventh line of the algorithm obtains the updated core tensor .
|Input: Raw tensor, apply the LHOSVD method to to get , ,, and incremental tensor ;|
|Output: Updated core tensor , orthogonal matrix ;|
|Expand the original tensor and the new tensor to the same dimension|
|Perform mode expansion on the new incremental tensor to obtain a mode expansion matrix|
|Map matrix to the left singular matrix of the original tensor expansion matrix|
|(c) Calculate the unit orthogonal matrix of to get|
|(d) Combine the matrix with the matrix to find the singular value decomposition result of , ,|
|Let , , update Combine the original tensor with the new incremental tensor|
|into a tensor using equation (5)|
|return core tensorreturn and orthogonal matrix|
Figure 5 uses a third-order tensor to describe the process of augmented tensor decomposition, where is a primitive tensor, and three truncated unit orthogonal matrices are obtained after decomposition. The new tensor is , and three expansion matrices are obtained after expansion. The three matrices are used to update the obtained unit orthogonal matrix, and the updated core tensor and truncated orthogonal matrices are obtained. The original tensor and the new tensor are expanded, and then the original tensor is decomposed by the HOSVD, and the obtained decomposition result can be reused in the quantitative dimension reduction. The module expansion of the new tensor F is carried out, and the three newly obtained module expansion matrices are projected onto the left singular matrix of the original matrix module expansion matrix. Finally, the updated three left singular matrices and the updated singular value matrices are obtained.
In this section, the norm of Frobenius  is adopted to measure the reconstruction error, and the dimensionality reduction ratio and reconstruction error rate proposed in the paper  are used as experimental comparison conditions.
We take the data in the intelligent factory as an example and evaluate the proposed tensor model (Figure 4) from the approximation rate, the dimensionality reduction ratio, and the reconstruction error rate. The test data in the experiment includes factory equipment, sensors, structured form data generated by user interaction, semistructured XML data, and unstructured video data. The data used in the experiment comes from the factory data supported by the project. The experiment compares the traditional HOSVD with the IncLHOSVD in terms of compression and dimensionality reduction.
The computer used in this section is configured with an Intel Core (TM) i5 CPU, including four 2.8ghz cores, 8GB of memory, and Matlab 2017. In the experiment, one minute was taken as a time statistical scale and 60 minutes were counted. A total of 11,542 pieces of structured form data, semistructured JSON data, and unstructured video data were extracted from the workshop of the intelligent factory. 1000 pieces of data were extracted to construct the unified tensor.
5.1. Algorithm Dimension Reduction Effect
In this section, the unified tensor is decomposed with different dimensionality reduction ratios, and the traditional HOSVD, SVD, and ttr1SVD  and the IncLHOSVD are used to reduce the dimensionality. We use the uniform tensor to reduce the dimensionality with different dimensionality reduction rate, the core tensor obtained after dimensionality reduction is reconstitution into approximate tensor, and the approximate ratio of the approximate tensor and the data contained in the original tensor is compared.
Figure 6 is the approximate ratio obtained after the dimensionality reduction experiment on the original data. Figure 7 is the dimensionality reduction ratio. By comparing the two figures, it can be seen that the IncLHOSVD reduction ratio decreased from 86% to 15%, and the approximate ratio decreased from 99% to 90% (Figure 8). It is clear that nearly 90% of the original data remains after reducing the dimension to 15%. However, the traditional high-order singular value decomposition, singular value decomposition, and ttr1SVD decomposition keep the dimensionality reduction rate close to that of incremental high-order singular value decomposition, and the core data retention result is lower than that of incremental decomposition. The results show that, compared with the existing methods, the proposed method has a larger dimensionality reduction ratio and more core data retention.
5.2. Algorithm Reconstruction Error
We extracted five seconds of video and three users’ semistructured data documents to construct a fifth-order tensor. The rate of reduction of the fifth-order tensor is 42%, 18%, and 5%, respectively. The reconstructed tensor after the dimension reduction is reduced by 4%, 7%, and 24%, respectively.
From Figure 9, we can see that there is a 4% reconstruction error for the 42% reduction rate and a reconstruction error of 6% for the 18% reduction rate. During this period, the reconstruction error is maintained at a low level. When the dimension rate reaches 24%, the reconstruction error rate rises to 24%. Generally speaking, the reconstruction error increases with the reduction of dimension reduction rate. Therefore, in the process of dimension reduction, we should not blindly pursue low dimension, but balance the dimension reduction ratio with the reorganization error rate.
As can be seen from Figure 10, the incremental higher-order SVD algorithm has a higher dimensionality reduction rate in the lower reconstruction error, which is better than the traditional higher-order SVD, SVD, and ttr1SVD decomposition algorithms.
5.3. Algorithm Performance Comparison
In the dimension reduction process, the traditional high-order singular value decomposition method decomposes a tensor repeatedly and then integrates the generated core tensor with the orthogonal matrix to generate a new tensor. The incremental high-order singular value decomposition proposed in this paper is updated orthogonal matrix dynamically generating new tensors. This experiment compares the size of the tensor and the time used by the traditional HOSVCD, SVD, and ttr1SVD algorithms and the incremental high-order singular value decomposition (IncLHOSVD) algorithm. During the experiment, the tensor size and the calculated time normalized standard value were compared.
The experimental results are shown in Figure 11. Starting from the tensor size normalization value of 0.25, the upward trend of the incremental algorithm is more gradual than the traditional algorithm. When the tensor size is higher than the tensor normalization value of 0.75, the time required for the traditional algorithm operation is much longer than the incremental decomposition algorithm, with SVD algorithm and ttr1SVD computations being too much out of memory, and the IncLHOSVD can process more data under the same conditions.
This paper provides a solution to the problem of high-quality core data extraction in edge computing nodes. For the wide range of data sources and characteristics of heterogeneous features in industrial scenarios, a representation model TBFPF is proposed to uniformly represent multisource heterogeneous data as tensor data. The dimension of the tensor data is too high and changes with time; the decomposition method of the tensor data dimension reduction IncLHOSVD method is proposed to extract the core data. Finally, the algorithm model is evaluated by the factory data. The proposed algorithm and IncLHOSVD model are effective according to theoretical analysis and experimental results.
The data used to support the findings of this study are available from the corresponding author upon request.
Conflicts of Interest
The authors declare that they have no conflicts of interest.
This work is partly supported by Science and Technology Project in Shaanxi Province of China (Program No. 2019ZDLGY07-08); the International Science and Technology Cooperation Program of the Science and Technology Department of Shaanxi Province, China (Grant No. 2018KW-049); the Communication Soft Science Program of Ministry of Industry and Information Technology, China (Grant No. 2019-R-29); and the Innovation Fund of Xi’an University of Posts and Telecommunications (Grant No. CXJJLY2018048).
- W. Jianmin, “Survey on industrial big data,” Big Data Research, vol. 3, no. 6, pp. 3–14, 2017.
- T. C. Jepsen, “Just what Is an ontology, anyway?” IT Professional, vol. 11, no. 5, pp. 22–27, 2009.
- H. Wentao and S. Cheng, “The development and challenges of industrial big data analysis technology,” Information and Control, vol. 47, no. 4, pp. 398–410, 2018.
- P. Patel, M. I. Ali, and A. Sheth, “From Raw data to smart manufacturing: AI and semantic web of things for industry 4.0,” IEEE Intelligent Systems, vol. 33, no. 4, pp. 79–86, 2018.
- G. Liao, S. Jiang, Z. Zhou, and C. Wan, “Dual fine-granularity poi recommendation on location-based social networks,” Computer Research and Development, vol. 54, no. 11, pp. 2600–2610, 2017.
- Y. Wang, Y. Zheng, and Y. Xue, “Travel time estimation of a path using sparse trajectories,” in Proceedings of the 20th ACM SIGKDD International Conference, pp. 25–34, ACM, August 2014.
- M. Boizard, R. Boyer, G. Favier, J. E. Cohen, and P. Comon, “Performance estimation for tensor CP decomposition with structured factors,” in Proceedings of the 40th IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP 2015, pp. 3482–3486, Australia, April 2014.
- A. Karami, M. Yazdi, and G. Mercier, “Compression of hyperspectral images using discerete wavelet transform and tucker decomposition,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 5, no. 2, pp. 444–450, 2012.
- D. S. Hochba, “Approximation Algorithms for NP-Hard Problems,” ACM SIGACT News, vol. 28, no. 2, pp. 40–52, 1997.
- S. Holtz, T. Rohwedder, and R. Schneider, “The alternating linear scheme for tensor optimization in the tensor train format,” SIAM Journal on Scientific Computing, vol. 34, no. 2, pp. A683–A713, 2012.
- A. Gorodetsky, S. Karaman, and Y. Marzouk, “High-dimensional stochastic optimal control using continuous tensor decompositions,” International Journal of Robotics Research, vol. 37, no. 2-3, pp. 340–377, 2018.
- Z. Chen, K. Batselier, J. A. Suykens, and N. Wong, “Parallelized tensor train learning of polynomial classifiers,” IEEE Transactions on Neural Networks and Learning Systems, vol. 29, no. 10, pp. 4621–4632, 2018.
- H. N. Phien, H. D. Tuan, J. A. Bengua, and M. N. Do, Efficient Tensor Completion: Low-Rank Tensor Train, Numerical Analysis, 2016.
- J. Xu, C. Xu, B. Zou, Y. Y. Tang, J. Peng, and X. You, “New incremental learning algorithm with support vector machines,” IEEE Transactions on Systems, Man, and Cybernetics: Systems, pp. 1–12, 2018.
- B. Gu, X. Quan, Y. Gu, V. S. Sheng, and G. Zheng, “Chunk incremental learning for cost-sensitive hinge loss support vector machine,” Pattern Recognition, vol. 83, pp. 196–208, 2018.
- S. Ahmad, A. Lavin, S. Purdy, and Z. Agha, “Unsupervised real-time anomaly detection for streaming data,” Neurocomputing, vol. 262, pp. 134–147, 2017.
- F. Dong, G. Zhang, J. Lu, and K. Li, “Fuzzy competence model drift detection for data-driven decision support systems,” Knowledge-Based Systems, vol. 143, pp. 284–294, 2018.
- J. L. Lobo, J. Del Ser, M. N. Bilbao, C. Perfecto, and S. Salcedo-Sanz, “DRED: An evolutionary diversity generation method for concept drift adaptation in online learning environments,” Applied Soft Computing, vol. 68, pp. 693–709, 2018.
- L. Kuang, F. Hao, L. T. Yang, M. Lin, C. Luo, and G. Min, “A tensor-based approach for big data representation and dimensionality reduction,” IEEE Transactions on Emerging Topics in Computing, vol. 2, no. 3, pp. 280–291, 2014.
- C. Meyer, Matrix Analysis and Applied Linear Algebra, Society for Industrial and Applied Mathematics, Philadelphia, Pa, USA, 2000.
- K. Batselier, H. Liu, and N. Wong, “A constructive algorithm for decomposing a tensor into a finite sum of orthonormal rank-1 terms,” SIAM Journal on Matrix Analysis and Applications, vol. 36, no. 3, pp. 1315–1337, 2015.
Copyright © 2019 Yanping Chen et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.