Abstract

This paper looks at techniques to simplify data analysis of large multivariate military sensor systems. The approach is illustrated using representative raw data from a video-scene analyzer. First, develop fuzzy neural net relations using Matlab. This represents the best fidelity fit to the data and will be used as reference for comparison. The data is then converted to Boolean, and using Boolean Decision Diagrams (BDD) techniques, to find similar relations between input vectors and output parameter. It will be shown that such Boolean techniques offer dramatic improvement in system analysis time, and with minor loss of fidelity. To further this study, Boolean Neural Net techniques (BNN) were employed to bridge the Fuzzy Neural Network (FNN) to BDD representations of the data. Neural network approaches give an estimation method for the complexity of Boolean Decision Diagrams, and this can be used to predict the complexity of digital circuits. The neural network model can be used for complexity estimation over a set of BDDs derived from Boolean logic expressions. Experimental results show good correlation with theoretical results and give insights to the complexity. The BNN representations can be useful as a means to FPGA implementation of the system relationships and can be used in embedded processor based multi-variate situations.

1. Introduction

Modern radar systems trace their roots to efforts in several countries to refine existing concepts and theories just prior to WWII. This was the culmination of theory and systems dating back to the 1860s. Simplistically, a transmitted electromagnetic signal reflects from an object back to an antenna where the characteristics of the signal determine properties such as location, speed and direction of the target. We take this simple problem, a show how variable complexity grows, and how the corresponding computational problem also grows.

Consider a simple Time-of-Flight range detector. For a stationary object, the range is simply determined by where = Range; = Speed of propagation in medium ( in air); = Time to receive the reflected signal.

For a moving target, radar systems use the Doppler Effect where the relative motion between source and observer produces a frequency shift where = observed frequency; = transmit frequency; = velocity of propagation in medium; = velocity of target relative to medium; = velocity of antenna relative to medium.

Both equations assume colocated transmit and receive antennas. Azimuth and elevation information comes with knowledge of the antenna beam position. So, what started as a simple calculation with 2 variables, becomes a problem with 4 variables. And, as we add moving sensors, propagation media dielectric variations, frequency jitter, multipath echoes, and so forth, the number of variables grows, and the computation problem quickly overwhelms the processor. Modern day military sensors, with video and other multispectra imagers, capture such large amounts of data that embedded processors are rapidly overwhelmed. The result then is that the embedded processor cannot complete its computations within the prescribed cycle time, and does not return a result or, starts to discard data, resulting in a false answer. Given that outcome, many systems find an “approximate” answer adequate for many mission objectives. In this paper, we employ data from such a video processor and show how Binary Decision Diagrams (BDD) fidelity compares with Fuzzy Neural Network (FNN) result quality and, how the Neural Network Model (NNM) can be used as an indicator of the Field Programmable Gate Array (FPGA) complexity.

The paper describes an environment for the rapid analysis, synthesis and optimization of embedded systems. The simplification strategy involves structure reduction and is carried out through an iterative algorithm aiming at selecting a minimal number of rules for the problem at hand. The selection algorithm allows manipulation of the neuro-fuzzy model or binary decision diagram to minimize complexity and to preserve a good level of accuracy. Since the implementation of these systems is rather complicated, we propose a methodology which automates the entire design flow. Flexibility is achieved by allowing manual intervention which is realized via a modular implementation of algorithms which are being provided.

2. Logic Analysis

An extension of the Time-of-Flight system is a typical track-while-scan (TWS) radar system, which gets a discrete (temporally separated) sequence of detections (“hits”) on a given target. Each detection is of a finite, typically quite short, duration. From each detection a given collection of attributes about that hit is extracted. The attributes typically include detection time, position, and reflected intensity, as well as various statistical moments and other mathematical features of the detection. The goal of these radar detections is typically to identify (detect and classify) the target of interest and its trajectory. By successfully tracking the target in motion, knowledge of the target can usually be substantially improved, since repeated detections of the moving target provide a greater volume of information and allow aggregate, track-based, statistics to also be used.

If we expand the problem to a multidimensional target identification problem, let us consider an example of a cluttered video image, and we need to find and classify the target in the scene. The image attributes include distance (to target it is called distance), aspect ratio of target (called aspect), vertical pixel height of target (called vert), area (in pixels it is called area) covered by target, target luminosity (called target lum), dark area in the image (called dark area lum), the surrounding luminosity (called surround lum), and edge pixels (called edg pts). The object is to classify the target (type) and assess the time taken to complete this chore. Reproduced from [13] is a table, Table 1, showing rows of such data. The data is just for illustration purpose only and is used in calculations to illustrate the comparison between full neuro-fuzzy analysis versus what target classification confidence we can achieve using Boolean or fuzzy techniques. The details of neuro-fuzzy techniques are not reproduced here [4], and, as background, the ANFIS (ANFIS = adaptive neurofuzzy inference system. As deployed in MatLab the Sugeno fuzzy model is where all output membership functions are singleton spikes and the implication and aggregation methods are fixed. The implication method is simply multiplication, and the aggregation operator just includes all of the singletons. The Sugeno method is ideal for acting as an interpolative supervisor of multiple linear controllers that apply different operating conditions of a dynamic nonlinear system. A Sugeno system is also suited for modeling nonlinear systems by interpolating multiple linear models) toolkit in MatLab and VeriMax analysis using MiniTab are employed. The architecture and learning procedure underlying ANFIS (adaptive-network-based fuzzy inference system) is a fuzzy inference system implemented in the framework of adaptive networks. By using a hybrid learning procedure, ANFIS can construct an input-output mapping based on fuzzy if-then rules and stipulated input-output data pairs.

And, as is shown [5, 6], we see that the Boolean and fuzzy approaches increase the linguistic rules as the number of input vectors increases. A neuro-fuzzy approach allows a more compact and computationally efficient representation and lends itself to adaptive schemes. And, where there is no intuitive knowledge of input vector behaviors or relationships to the output, these adaptive techniques in turn help create the entire fuzzy network for us.

Looking at Table 1, for this example, we see that there are 9 target types, and that there is little intuitive correlation between input vectors and target type or time taken to classify the target. Heuristic methods and expert knowledge (described in [7]), exploit a-priori knowledge to make inferences and help cluster the vectors and reduce system size by eliminating some inputs.

Using the Verimax analysis in MiniTab, we calculate, Table 2 [3], the correlation between vectors and see that vert (number of vertical pixels) and area (area covered by target in pixels) information is strongly correlated (95%).

There are other inferences that can also be made. But, none that allows a simple determination of logical connections. Employing the Neuro-fuzzy ANFIS function from the MatLab toolkit, we see in Table 3, that it is possible to get upto 87% confidence in target classification, relying on the full data set.

Where the terms as used by MiniTab are [8] as follows.CORR: Correlation between the original output and the estimated output from the fuzzy neural system using the data from each method.TRMS: total Root mean square for the distance between the original output and the estimated output using the same testing data through the fuzzy neural system where is the estimated value and is the original output value.STD: Standard deviation for the distances between the original output and the estimated output using the same testing data through the fuzzy neural system.MAD: Mean of the absolute distances between the original output and the estimated output using the same testing data through the fuzzy neural system.EWI: The index value from the summation of the values with multiplying the statistical estimation value by its equally weighted potential value for each field.ERR: The error rate is where is the number of testing data, is the estimated output, and is the actual output.

This forms our baseline. We now compare this analysis to what we can get if the data in Table 1 is converted to Boolean. This Boolean data is presented in Table 4.

Verimax analysis when we employ Boolean relationships to determine the correlation functions shows correlation between the original output and the estimated output to be 82%. This compares well with the FNN analysis and is a reasonable approximation of the original data. It is this acceptable degradation that provides an opportunity to use Binary Decision Diagrams and Boolean techniques, to reduce the number of variables and to simplify the computational problem. The general correlation similarities between vectors are approximately preserved. And, again, vert and area vectors have high correlation. Thus, using the premise that an approximate answer is better than no-answer. The simple Boolean quantization of the raw data, provides a reasonable estimate as compared to the full data set, fuzzy inferences.

3. Boolean Decision Diagram Techniques

The analysis techniques used above show that reasonable computation reduction can be achieved by data quantization implying that using full-precision data as baseline, we still achieve approximately 82% fidelity in the correlation data using Boolean quantization of the data. This opens the door to using Boolean Decision tools for rapid, reasonable fidelity and analysis of the system.

Let us again start with the data in Table 4—which is the Boolean representation of the raw data. We find that we have 8 input variables and 1 output vector called search time, that is, the time required to correctly classify the target type. As shown in [2, 7], the correlation matrix for the Boolean data also shows that, because it has such high correlation between adjacent vectors, we can simply eliminate the vert or area vectors and, using a-priori knowledge for this specific problem, that target lum and edge pts vectors also can be combined in an OR function to some extent. Finally, that distance and target lum also can be OR combined to further reduce the matrix. We can then develop a 4-variable matrix to develop a set of linguistic rules.

Table 5 shows such a reduced vector representation of the original raw data. And, this is specific for this problem, other problems, and, using specialized knowledge [7] we can use the analysis techniques shown to reduce the input vector dimension.

Using Boolean Decision Diagram techniques, this is represented as

From this, we can create the K-map (Figure 1) of the above linguistic rules.

And, we can re-write as:

The complexity of circuit and systems design increases rapidly. Therefore, in seeking efficient algorithms and data structures, we choose, binary decision diagrams (BDDs). These have been used in a wide variety of applications and were intensively studied. And, the Boolean Decision Diagram as Figure 2.

Having this Boolean Decision Diagram, we can apply the analysis tools [911] readily available for such analysis. This shows how a complex video scene, once it is translated into desired attributes, can then be analyzed using BDD techniques to yield fast approximate solutions to the data classification problem. Research exists [12] on how to generate FPGA using BDDs, and in this paper we acknowledge that numerous researchers have used BDD synthesis for FPGA performance optimization. Additionally, recent publications [1316] show that such order-reduction models are created and changed throughout the development process process and [13, 14] argue that if one looks at the idea of model checking as a constraint checking problem, then the BDD-based symbolic model checking indeed is capable of verifying systems with a large number of states. This is in agreement with our main thesis as well, that is the analysis of abstract models using such order-reduction analysis will have great impact in reducing computational load on embedded processors. The authors of [15, 16] note that it is indeed possible to reduce (replace) the initial set of input variables with their linear combinations and that such path reduction (order reduction) can be compared using spectral techniques, or weighted autocorrelation techniques, to compare and approximate accuracy degradations.

4. Proposed Algorithm

The proposed algorithm to implement the above technique is illustrated in this section. The input space is reduced and the number of rules is decreased, and the simulation results illustrate the approach is practicable, simple, and effective and the performance index is not significantly degraded. The video scenes, in this illustrative example, are first converted recognition attributes as required by the classification algorithm. And, using FNN and ANFIS techniques, a baseline answer fidelity index is generated. Then, the data streams are converted to Boolean and checked to see that resultant fidelity degradation is acceptable. This assessment can be rule-set or can rely on expert a-priori knowledge to judge when the quantization is acceptable. Then, using Factor and Cluster techniques, the problem size is reduced first by combining or eliminating the input vectors, and then, by selecting orthogonal data rows, such that a full set of outcomes is represented. Finally, use BDD techniques, to draw K-map diagrams and to develop a Boolean equation to solve the problem.

The synthesis, optimization, and implementation of embedded systems is rather complicated, and we propose a methodology which automates the entire design flow. Flexibility is achieved by allowing manual intervention which is realized via a modular implementation of algorithms which are being provided. This is represented in the following flow (see Figure 3).

Assume there are by matrix with -1 input variables and one output.(1)Using the raw data, and FNN, ANFIS function from the MatLab toolkit, calculate the correlation matrix, , between variables of data set. This correlation is the benchmark correlation between variables and can be used later to decide the validation of reducing procedure.(2)Review input vector correlation data, to determine which variables can be reduced (or combined).(3)Use iterative convergence to reduce/combine input vectors. Each time verifying that the resultant data is reasonably correlated to the original data. (4)Convert remaining data stream to Boolean. Typically this binary representation is of a reduced set of input variables.(5)Then, using Factor and Cluster techniques, the problem size is reduced first by combining or eliminating the input vectors, and then, by selecting orthogonal data rows, such that a full set of outcomes is represented. This can be performed, based on heuristic, a-priori, or correlation analysis.(6)Repeat MiniTab/VeriMax analysis to validate that the resultant correlation value is within acceptable range as compared to the full-fidelity FNN correlation value. This step can be automated via a rule set, or can include judgment, depending on confidence in the factoring algorithms.(7)Next, use BDD techniques, to draw K-map diagrams and to develop a Boolean equation to solve the problem.(8)Create the BDD to represent the Boolean relationship.(9)Verify the diagram with representative cases.(10)Use FPGA design techniques, to convert the BDD to code.(11)The system is now ready to take a raw stream of input data and yield a high efficiency solution to the target classification problem.

5. Analysis and Conclusion

Binary Decision Diagrams are generated by applying Shannon expansion [16] repeatedly to a logic function to yield a multivalued decision diagram. This is usually a (very) large matrix, generally unsolvable by embedded processors having finite memory and limited cycle-time before a decision must be rendered. Cluster decomposition (i.e., circuit partitioning) is also possible, but, generally this leads to several smaller matrices that must be solved simultaneously and also requires a fuzzy-logic arbitration algorithm to decide on the outcome. Further, when such subcircuits have fanouts, [16], the number of BDD nodes may increase exponentially thereby defeating the intent of Order Reduction algorithm. The study in [15] also confirms that the number of paths (order) is a function of the weighted autocorrelation coefficients and that it is possible to minimize the number of paths by constructing an ordered set of basis vectors of high weighted autocorrelation value. This is the key to use auto-correlation analysis to find relations between input vectors and output parameter, also to convert dynamic data to Boolean, that is, a 1-bit approach, to create both a reduced-order circuit, and to simplify analysis. This technique of both order reduction (linearly combining input vectors) and input data-resolution reduction (converting from full-dynamic range data to Boolean data) still yields adequate quality results.

In summary, a high number of input parameters can result in a large number of rules that are computationally difficult to handle. Hence, it is imperative to develop the techniques, which can reduce the input parameters such that the original system and reduced systems have approximately the same behavior. In this paper, we have taken multi-variate streams of as-available data from a video scene analyzer and have created a technique for order reduction, and of digitization, such that the problem can reasonably be reduced to a Boolean problem, and, that this Boolean problem can then be solved using well-known techniques for Boolean Decision Diagrams. In the context of the example chosen for this paper, data streams from a video recorder are selected and analyzed for target classification efficiency. It is shown that an original 8 input-parameter data set with an 87% confidence full fuzzy neural net analysis, the reduced order problem was converted to a 4-variable problem, and the technique was able to generate a reasonably accurate 82% accurate solution. And, that such a reduced-order problem can then be converted to a Boolean relationship. Existing research [912] shows how to convert this relationship to a FPGA design, such that, for this example, such video data can be quickly analyzed and a reasonably accurate answer generated. So while it is possible to have larger and larger processors and more and more complex code, often in many military applications, an approximate answer is adequate. And, that answer can be calculated using order-reduction techniques outlined herein, and then completely solved using BDD techniques.

From this study, we may decide the best heuristic and fuzzy techniques and also when the usage of fuzzy or heuristic method is better than the other. These are of significant advantage in satellite sensors where power and compute capacities are limited, and where such approximations can then cue more detailed analysis if warranted.