Abstract

A new framework intended for representing and segmenting multidimensional datasets resulting in low spatial complexity requirements and with appropriate access to their contained information is described. Two steps are going to be taken in account. The first step is to specify ()D hypervoxelizations, , as Orthogonal Polytopes whose th dimension corresponds to color intensity. Then, the D representation is concisely expressed via the Extreme Vertices Model in the -Dimensional Space (D-EVM). Some examples are presented, which, under our methodology, have storing requirements minor than those demanded by their original hypervoxelizations. In the second step, 1-Dimensional Kohonen Networks (1D-KNs) are applied in order to segment datasets taking in account their geometrical and topological properties providing a non-supervised way to compact even more the proposed -Dimensional representations. The application of our framework shares compression ratios, for our set of study cases, in the range 5.6496 to 32.4311. Summarizing, the contribution combines the power of the D-EVM and 1D-KNs by producing very concise datasets’ representations. We argue that the new representations also provide appropriate segmentations by introducing some error functions such that our 1D-KNs classifications are compared against classifications based only in color intensities. Along the work, main properties and algorithms behind the D-EVM are introduced for the purpose of interrogating the final representations in such a way that it efficiently obtains useful geometrical and topological information.

1. Introduction

The representation of a polytope through a scheme of Hyperspatial Occupancy Enumeration is essentially a list of identical hyperspatial cells occupied by the polytope. Specific types of cells, called hypervoxels [1] are hyperboxes of a fixed size that lie in a fixed grid in the -Dimensional space, . By instantiation, it is well known that a 2D hypervoxel is a pixel while a 3D hypervoxel is a voxel; the term rexel is suggested for referencing a 4D hypervoxel [1]. The collection of hyperboxes can be codified as an D array . The array represents the color associated with each hypervoxel. If , the empty hypervoxel represents an unoccupied region from the D space. If , where is in a given color scale (black & white, grayscale, RGB, etc.), then the occupied hypervoxel represents a region from the D space with intensity . In fact, the set of occupied cells defines an orthogonal polytope whose vertices coincide with some of the occupied cells’ vertices. By using the representation through an array, the spatial complexity of a -Dimensional hypervoxelization is at least where , , is the length of the grid along the -axis. For example, a 3D grid with requires to store 1 billion (1 × 109) voxels. Moreover, according to the used color scale each voxel will have a storing requirement: if the color space is RGB then each voxel will require three bytes (four, if the alpha channel is considered) for codifying its corresponding intensity.

It is well known that some devices represent natively datasets through hypervoxelizations. However, sometimes their storing requirements make the manipulation and the extraction of information a difficult task. In this sense, several efforts have been made in order to reduce the spatial complexity of datasets always taking in account the information they contain, due to its importance and relevance, should be compromised as minimum as possible. For example, an algorithm for compression of datasets by means of quadtrees, in order to encode slices of data, is presented in [2]. Such encodings are used for discovering similarities between consecutive slices. In [3], 3D medical datasets are compressed via a method sustained in the use of octrees. Both works share us evidence of the spatial conciseness provided by considering the use of Solid Representation Schemes. This work is devoted to present an alternative representation for datasets whose color scales are not binary (in [4] is presented a methodology designed specifically for 3D datasets with black & white color scale). The main idea is to specify D datasets, , as D polytopes where the th dimension corresponds to color. Then, the D representation can be concisely expressed and manipulated through a polytopes’ representation scheme: the Extreme Vertices Model in the -Dimensional Space (D-EVM). However, we take one additional step: a 1-Dimensional Kohonen Network (1D-KN) will be applied in order to segment datasets taking in account their geometry and topology providing by instance a nonsupervised intelligent way to compact even more the proposed D representations. In this last sense, neural networks-based segmentation is a topic well developed in the literature. In [5] is presented the moving average self-organizing map (MA-SOM) which is applied for segmenting medical images obtained by means of X-ray computer tomography, breast ultrasound imaging, and magnetic resonance imaging. In [6], Kohonen networks are used together with a threshold technique in order to segment satellite images. There is a review of segmenting methods based on clustering and neural networks approaches presented in [7]. In such work some experiments are presented based on the use of Multilayer Perceptrons trained with the well-known Backpropagation Algorithm. The paper by Jiang et al. [8] presents a survey of application of Neural Networks for medical support in processing images in tasks such as Image Segmentation, Edge Detection, and Computer Aided Diagnosis by means of application of the Multilayer Perceptrons, Hopfield Networks, and Self-Organizing Maps, among others. These works exemplify how the Artificial Neural Networks approach has gained confidence among Image Processing researchers due to valuable properties such as robustness to noise, plasticity, parallelism, and computational efficiency.

This work is organized as follows: Section 2 will describe the fundamentals behind the D-EVM. The methodology for the conversion of a D hypervoxelization dataset to a D-EVM is described in Section 3. Also, there will be presented some examples of datasets expressed under the D-EVM. By comparing the storing requirements between the original and the D-EVM representations it is established the first level of conciseness of our proposal. Section 4 describes the fundamentals behind the 1D-KNs. In Section 5 it is described the way a 1D-KN assists us in the automatic nonsupervised segmentation of a dataset. In Section 6 how the classification provided by our 1D-KNs gives shape to the second level of conciseness by describing some examples and their final spatial requirements is presented. Section 7 provides arguments to sustain the segmentations that have been produced have good quality and they are even better than those segmentations produced by simple and direct classification based only in color intensities. For this purpose, some error functions are introduced. Finally, Section 8 presents our conclusions and lines for future research.

2. The Extreme Vertices Model in the -Dimensional Space (D-EVM)

This section is a summary of results originally presented in [9, 10]. For the sake of brevity, only the required concepts for the purpose of this work are presented and some propositions are only enunciated. More details and formal proofs can be found in the two abovementioned references. Furthermore, in the following Sections we will make brief mentions of more concepts, applications, and algorithms, under the context of the D-EVM, and the references where they are presented and described with enough detail.

2.1. Preliminary Background: the -Dimensional Orthogonal Pseudopolytopes (D-OPPs)

Definition 1. A singular -Dimensional hyperbox in the -Dimensional Euclidean Space is the continuous function

Definition 2. For all the two singular D hyperboxes and are defined as follows. If then

Definition 3. In a general singular D hyperbox the cell is defined as .

Definition 4. The orientation of a cell is given by .

Definition 5. An D oriented cell is given by the scalar-function product .

Definition 6. A formal linear combination of singular general D hyperboxes, , for a closed set is called a -chain.

Definition 7 (see [11]). Given a singular D hyperbox the -chain, called the boundary of , is given by

Definition 8 (see [11]). Given a singular general D hyperbox the -chain, called boundary of , is defined by

Definition 9 (see [11]). The boundary of an -chain , where each is a singular general D hyperbox, is given by

Definition 10. A collection , , of general singular D hyperboxes is a combination of D hyperboxes if and only if
In the above definition the first part of the conjunction establishes that the intersection between all the D general singular hyperboxes is the origin, while the second part establishes that there are not overlapping D hyperboxes.

Definition 11. An -Dimensional Orthogonal Pseudopolytope , or just an D-OPP , will be an -chain composed by D hyperboxes arranged in such way that by selecting a vertex, in any of these hyperboxes, we have that such vertex describes a combination of D hyperboxes composed up to hyperboxes.

2.2. The D-EVM: Foundations

Definition 12 (see [10]). Let be a combination of hyperboxes in the -Dimensional Euclidean Space. An Odd Edge will be an edge with an odd number of incident hyperboxes of .

Definition 13 (see [10]). A brink, or extended edge, is the maximal uninterrupted segment, built out of a sequence of collinear and contiguous odd edges of an D-OPP.

Definition 14 (see [10]). The Extreme Vertices of an D-OPP are the ending vertices of all the brinks in .

Definition 15 (see [10]). Let be an D-OPP. The Extreme Vertices Model of , denoted by , is defined as the model as only stores to all the extreme vertices of .

2.3. Sections and Slices of D-OPPs

Definition 16. Let be an D-OPP. A D couplet of , , is the maximal set of D cells of that lies in a D space, such that a D cell belongs to a D extended hypervolume if and only if belongs to an D cell present in .

Definition 17. The Projection Operator for D cells, points, and set of points is, respectively, defined as follows.(i)Let be an D cell embedded in the D space. will denote the projection of the cell onto an D space embedded in D space whose supporting hyperplane is perpendicular to -axis: .(ii)Let a point in the D space. The projection of in the D space, denoted by , is given by .(iii)Let be a set of points in the D space. The projection of the points in , denoted by , is defined as the set of points in the D space such that .Where is the coordinate corresponding to -axis to be suppressed.

Definition 18. Consider an D-OPP .(i)Let be the number of distinct coordinates present in the vertices of along -axis, .(ii)Let , or just if no reference to is needed, be the th D couplet of which is perpendicular to -axis, .

Definition 19. A Section is the D-OPP, , resulting from the intersection between an D-OPP and a D hyperplane perpendicular to the coordinate -axis, , which not coincide with any D-couplet of . A section will be called external or internal section of if it is empty or not, respectively. will refer to the th section of between and , .

Theorem 20 (see [10]). The projection of the set of D-couplets, , of an D-OPP , can be obtained by computing the regularized XOR between the projections of its previous and next sections; that is,

Theorem 21 (see [10]). The projection of any section, , of an D-OPP , can be obtained by computing the regularized XOR between the projection of its previous section, , and the projection of its previous couplet .

2.4. The Regularized XOR Operation on the D-EVM

Theorem 22 (see [9]). Let and be two D-OPPs having and as their respective EVMs in D space; then

This result allows expressing a formula for computing D-OPPs’ sections from couplets and vice-versa, by means of their corresponding Extreme Vertices Models. These formulae are obtained by combining Theorem 22 with Theorem 20, and Theorem 22 with Theorem 21, respectively.

Corollary 23 (see [9]).

Corollary 24 (see [9]).

Finally, the following corollary can be stated, which correspond to a specific situation of the XOR operand. It allows computing the union of two D-OPPs when that specific situation is met.

Corollary 25 (see [9]). Let and be two disjoint or quasidisjoint D-OPPs having and as their respective Extreme Vertices Models; then

2.5. Basic Algorithms for the D-EVM

According to previous sections the primitive operations shown in Algorithm 1 can be defined based in the functions originally presented in [9]. Function MergeXor, according to Theorem 22, performs an XOR between two D-EVMs; that is, it keeps all vertices belonging to either or and discards any vertex that belongs to both and . Since the model is sorted, this function consists of a simple merging-like algorithm, and therefore, it runs on linear time [9, 10].

Output:   An empty nD-EVM.
Procedure   InitEVM( )
{Returns the empty set.                      }
Input:   An nD-EVMp
Output:   A Boolean.
Procedure   EndEVM(EVM  p)
{Returns true if the end of  p  along -axis has been reached. }
Input:   An nD-EVM  p
Output:   An (n1)D-EVM embedded in (n1)D space.
Procedure   ReadHvl(EVM  p)
{Extracts next (n1)D couplet perpendicular to -axis from  p.  }
Input:   Two nD-EVMs  p  and  q.
Output:   An nD-EVM
Procedure   MergeXor(EVM  p, EVM  q)
{Applies the Exclusive OR operation to the vertices of  p  and  q
and returns the resulting set.                   }
Input:   An nD-EVM  p
Output:   A Boolean.
Procedure   IsEmpty(EVM  p)
{Returns true if  p  is an empty set.                 }

From the above primitive operations, Algorithms 2 and 3 are derived. Algorithm 4 computes the sequence of sections of an D-OPP from its D-EVM using the previous functions. It sequentially reads the projections of the D couplets hvl of polytope . Then it computes the sequence of sections using function GetSection. Each pair of sections and (the previous and next sections about the current hvl) is processed by a generic processing procedure (called Process), which performs the desired actions upon and .

Input:   An (n1)D-EVM corresponding to section S.
   An (n1)D-EVM corresponding to couplet hvl.
Output:   An (n1)D-EVM.
Procedure   GetSection(EVM S, EVM hvl)
     // Returns the projection of the next section of an nD-OPP whose previous section is S.
     return   MergeXor(S, hvl)
end-of-procedure

Input:   An (n1)D-EVM corresponding to section .
   An (n1)D-EVM corresponding to section .
Output:   An (n1)D-EVM.
Procedure   GetHvl(EVM , EVM )
     // Returns the projection of the couplet between consecutive sections and .
     return   MergeXor(, )
end-of-procedure

Input:   An  nD-EVM  p.
Procedure   EVM_to_SectionSequence(EVM  p)
    EVM  hvl   // Current couplet.
    EVM  ,    // Previous and next sections about hvl.
    hvl  = InitEVM( )
      = InitEVM( )
      = InitEVM( )
    hvl  = ReadHvl(p)
    while   (Not(EndEVM(p)))
  = GetSection(,  hvl)
Process(,  )
  =  
hvl  = ReadHvl(p)// Read next couplet.
    end-of-while
end-of-procedure

3. Representing Hypervoxelization Datasets through D Orthogonal Polytopes

The conversion of a D hypervoxelization to a D polytope, and therefore to a D-EVM, is in fact a straight procedure. The methodology to be described in this Section was originally presented, for the 4D case, in [12]. As commented in Section 1, the proposal is oriented to color spaces with three or more values. If it is the case that datasets have binary intensity values, then refer to [4], where a more appropriate representation is described. It is assumed the intensities of each hypervoxel are expressed as a single number.

A D hyperprism is a polytope generated by the parallel motion of an D polytope; it is bounded by the D polytope in its initial and final positions and by several D hyperprisms [13, 14]. In this sense, each hypervoxel in a D dataset is extruded towards the th Dimension; that is, it is converted in a D hyperprism whose bases are precisely the original hypervoxel and its height is given by its corresponding intensity-plus-one value in the dataset. The vertices’ coordinates in the hyperprism’s bases correspond to the original hypervoxel’s coordinates. The inferior base’s points have their coordinate equal to zero, while in the remaining vertices the coordinate is equal to the intensity-plus-one value (see Figure 1 for an example under the instance ). Let us call to the set composed by the D hyperprisms (the extruded hypervoxels) of the extruded D dataset.

Let be a D hyperprism in and the number of prisms in that set: is in fact equal to the number of hypervoxels in the original dataset. Since all hyperprisms in are quasidisjoint D-OPPs, the extreme vertices, of the whole D extruded dataset, can be easily obtained by computing the regularized union of all the hyperprisms in . Hence, Corollary 25 is applied in the following way: where is the D-OPP that represents the union of all the hyperprisms in . By this way, it is obtained a representation for a D Dataset through a D-OPP and the D-EVM.

Now we describe some first results related to the conversion from voxelizations (3D hypervoxelizations) to our proposed representation. Such datasets were taken from The Volume Library [15], and the University of Iowa’s Department of Radiology [16]. Datasets’ visualizations shown in Table 1 were obtained via a volume rendering software available at [17]. Table 1 also shows the space complexity measures obtained when the considered voxelizations were converted to 4D-OPPs and the D-EVM. Now, let be a dataset expressed under a D hypervoxelization with size () and with as its corresponding D-EVM. Consider the ratio [12]:

For example, see model Aneurism (Table 1). Its source voxelization has size (256 × 256 × 256) which implies that it is required to store 16,777,216 voxels. The 4D-EVM associated to Aneurism has 1,317,008 extreme vertices. Hence, the proposed ratio gives the value 12.7388 which tell us the number of stored voxels that belong to the original representation of the object is precisely 12.7388 times greater than the number of obtained extreme vertices. Table 2 shows the ratio Number-of-voxels/Number-of-Extreme-Vertices for the models described in Table 1. The value shared by the ratio depends on the topology and geometry of the objects being modeled, but it shows how some conciseness is gained, related to storing requirements, when they are represented through the D-EVM.

Tables 3, 4, and 5 show some 3D couplets obtained from datasets Aneurism, CT-Monkey, and Statue Leg, respectively. 3D-EVM visualizations from Tables 3 to 6 were achieved by means of visualization software developed in [1820]. Table 6 shows some 2D sections corresponding to one of the 3D couplets from the model CT-Monkey. There are shown, by direct application of Algorithm 4, 22 of these 2D sections which allow understand the internal organization of the selected 3D couplet. It is possible to visualize some internal bone structures associated to the monkey’s cranium.

In [21] is presented an D-EVM-based algorithm for computing Connected Components Labeling (CCL) over an D-OPP. It is well known CCL is one of the most important tools in image processing and computer vision. The main idea is to assign labels to the pixels in an image such that all pixels that belong to a same Connected Component (CC) have the same label [22, 23]. The direct application of our CCL-EVM-Based algorithm over the 4D-EVMs associated to our three datasets will report only one component. This is because of the way the original 4D-OPPs were built (see (13)). Now, let CCL_EVM3() be the output shared by our mentioned CCL-EVM-Based algorithm when is sent as input a 3D-OPP , expressed under the D-EVM. In other words, CCL_EVM3() gives the number of 3D CCs present in . For specific details about our CCL-EVM-Based algorithm refer to [21]. In Tables 3 to 5 we also make mention of the number of CCs present in the shown 3D couplets. These data is going to be part of an analysis to be performed in Section 6.

The importance behind a dataset is the information can be obtained from it. If the datasets are represented through the D-EVM then the extraction of its couplets perpendicular to -axis will provide a classification of the elements in the original model according to their color intensities. Or in others words, we obtain a dataset’s segmentation based exclusively on color intensities. Given a dataset based on a scale of intensities, then it is possible to obtain at most, from its corresponding D-OPP,   D couplets perpendicular to the axis associated to color. The “extra” couplet, in fact the first couplet, is the result of the union of all the inferior bases of the D hyperprisms in (the set composed by the dataset’s extruded hypervoxels): the points in such bases have value zero for the coordinate associated to color. Under this reasoning, Tables 3, 4, and 5 present part of the color-intensities-based segmentation achieved in our considered datasets.

It is possible that color intensities correspond to physical properties. For example, in a 3D medical dataset intensities could refer to certain tissues. Hence, by extracting 3D couplets perpendicular to color axis we could obtain, for a given couplet, only those parts of the 3D dataset with the same type of tissue. The type of material is, in this case, assumed to be defined by the color intensity of the voxels in the original dataset. The projection of each couplet, perpendicular to the axis associated to color, in a D-EVM, is in fact a D-EVM which in time corresponds to a binary dataset because it only contains material of the same type. Hence, it is possible to apply Algorithm 4 to these D couplets in order to obtain their associated D sections. These sections will share a description of the interior of the homogeneous object with the objective of performing the appropriate analyses according to the application, as seen previously in the example described in Table 6.

This Section has focused about the representation of hypervoxelizations through D-OPPs which in time are expressed in the D-EVM. Moreover, we have seen how the basic procedures under the Extreme Vertices Model share the extraction of useful information related to the original datasets: the classification of their elements according to intensities via D couplets perpendicular to color axis. Likewise, it is possible to compute other geometrical and topological interrogations over an D-EVM. By this way it could be obtained much more information and properties about the datasets. Furthermore, there are well specified procedures under the D-EVM which allow performing Regularized Boolean Operations, Polytopes Splitting, Discrete Compactness Computation, Morphological Operations, Connected Components Labeling, Boundary Extraction, among others. In [9, 10, 12, 21, 24, 25] there are described with enough detail algorithms based in the D-EVM which are useful and efficient for performing these interrogations and/or manipulations.

According to the above discussion, if datasets are represented through the D-EVM then the extraction of its couplets will provide a classification of the elements in the original model according to their color intensities. The projection of each couplet, perpendicular to the axis associated to color is in fact a D-EVM which in time corresponds to a binary dataset because it only contains material of the same type, in this case, the same intensity. We are now in a critical point where the next logical step is given by the consideration of an additional approach in such way that the points in a dataset could be characterized by taking in account not only their color intensity but their class. Due to the presence of scanning noise and artifacts, a classification based only in intensities is sometimes not enough because points that should belong to the same class could be characterized as distinct points and therefore assigned to distinct classes, hence, our D representation will integrate each point in distinct D couplets. Let an intelligent approach, such as an Artificial Neural Network, be responsible for automatically identifying the classes of points present in a multidimensional dataset by taking in account their geometry, topology, neighborhood, and so forth. Once the classification is achieved, then the corresponding conversion to our proposed -Dimensional representation could take place. It should be possible to assign to each class a coordinate on the th dimension in such way D couplets perpendicular to the th axis now correspond to sets of points, in the original dataset, that belong to the same class. It is also hypothesized that by considering more appropriate points classification, a much better conciseness should be obtained. This is the line or reasoning which is going to be explored in the following Sections.

4. A Survey on 1-Dimensional Kohonen Networks (1D-KNs)

A Kohonen Network with inputs and neurons may be used to classify points embedded in an -Dimensional space into categories [26, 27]. Input points have the form (). Each neuron , , has associated an -Dimensional Weights Vector which has a representation of its corresponding class . All these vectors have the form , .

A set of training points are presented to the network times. According to [28], all values of Weight Vectors should be randomly initialized. In the th presentation, , the neuron whose Weights Vector , , is the most similar to the input point is chosen as Winner Neuron. In the model proposed by Kohonen, such selection is based on the Squared Euclidean Distance. The selected neuron is that with the minimal distance between its Weights Vector and the input point :

Once the th Winner Neuron, , in the th presentation, has been identified each one of the network’s weights vectors is updated according to: where the term is the learning coefficient and is a neighborhood function that denotes the distance between the Winner Neuron and the neuron . For neurons close enough to the winner neuron, should be a value near to 1. On the other hand, is close to zero for those neurons characterized as distant to the Winner Neuron.

When the presentations have been achieved, the values of the Weights Vectors correspond to coordinates of the “gravity centers” of the clusters of the categories.

5. Multidimensional Hypervoxelizations’ Segmentation Achieved by 1D-KNs

Nonsupervised segmentation of datasets has great potential in the practice, but this task is complex for several reasons, being one of them, for example, the datasets’ acquisition process which could introduce noise and artifacts which are difficult to correct. As commented previously, the specific problem now to be under attack is the automatic nonsupervised segmentation based on the assistance of a 1D-KN. It is expected there could be identified, during its training processes, the proper representations for a previously established number of classes. Hence, a dataset can be segmented in such way each type of region is appropriately identified. Many methods for description, object recognition or indexing are sustained on a preprocessing based on automatic segmentation [22, 29, 30]. This Section describes our methodology which is inspired in some facts established originally in [31].

An aspect to be considered respect to the training sets to be used is the fact a first approach could suggest that the color intensity of each hypervoxel in a dataset can be seen as an input vector (formerly an input scalar). However, as discussed in several works, such as [32, 33], a network will be biased towards a characterization based only in intensities. It is clear each hypervoxel has an intensity which, it is understood, captures, or is associated, to a particular property (a type of tissue, material, etc.); however, it is important to consider the hypervoxels around it. The neighborhood surrounding a given hypervoxel complements the information about the properties to be identified. Let be a hypervoxel in an D dataset, . Given it is possible to build a subdataset by taking those hypervoxels inside a hypercubical neighborhood of radius and center at . Hypervoxel and its neighboring hypervoxels will be called a -Dimensional mask. For example, given a 3D dataset, if we refer to 3D masks of radius then we have subdatasets of size (5 × 5 × 5) whose central voxel, with coordinates (, , ), is the one of interest and the remaining 124 voxels compose its cubical neighborhood.

Now, the network’s training set is then composed by all the D masks of radius that can be generated in a given D dataset. As commented in Section 4 a 1D-KN expects as input a vector, or point, embedded in the -Dimensional Space. A D mask is formerly seen as a D array, but it is clear by stacking its columns on top of one another a vector is obtained. This straightforward procedure linearizes a D mask making it a suitable input for the network. The resulting vectors, composed by components, preserve the properties associated to the original D masks. Each component has the color intensity of the corresponding hypervoxel in its associated D mask. By this way, we have specified the elements that conform our proposed approach for automatic segmentation of D hypervoxelizations. We proceed now to describe some experiments and the obtained results.

Our study cases were based in the set of 3D datasets presented in Section 3 plus two additional datasets described in Table 7. Both datasets were taken from The Volume Library [15] and the visualizations shown in Table 7 were achieved via a volume rendering software available at [17].

The implemented 1D-KNs have the following topologies and training conditions:(i)3D masks radius: ,(ii)inputs: = 5 × 5 × 5 = 125,(iii)neurons (classes): ,(iv)presentations: .

According to (16) we must define a neighborhood function. In this work we will use the following rule: That is, once the Winner Neuron has been identified, only its weights are updated as established in (16). As commented above, training sets’ cardinalities clearly depend on the value for radius , but also relies on the dataset’s size. Because we are considering five testing voxelizations, each one to be managed by its corresponding 1D-KN, then we have to think about five different training sets. In Table 8 we can see the corresponding cardinalities of the training sets associated to each considered voxelization.

Table 9 shows the obtained distribution of the training 3D masks in each one of the available 40 classes. Firstly, according to this Table, the networks trained with 3D masks associated to datasets Aneurism, VL-CT-Knee, and VL-Baby, distributed all their 3D masks along their 40 classes. In the case for networks trained with 3D masks associated to datasets Statue Leg and CT Monkey only 22 and 39 classes are, respectively, used. A second observation arises: in all the proposed networks there is a class which has associated more than 45% of the elements in its corresponding training set. Seeing Table 9 it is possible to identify such classes as follows.(i)Dataset Aneurism: class 4, 98.47%.(ii)Dataset Statue Leg: class 24, 93.72%.(iii)Dataset CT Monkey: class 27, 81.69%.(iv)Dataset VL-CT-Knee: class 27, 48.17%.(v)Dataset VL-Baby: class 9, 67.16%.As we will see with more detail in Section 6, we can mention in advance these abovementioned classes grouped 3D masks corresponding to empty regions in the original datasets. Members in these classes are precisely part of the point of discussion relative to the fact hypervoxelizations’ spatial complexities make sometimes difficult their manipulation and storage: under hypervoxelization representation we are also dealing with the processing and “cargo space” of empty regions. However, we will see in next Section by considering the D-EVM model these regions can be taken in account in the representation without impacting spatial conciseness.

6. The Framework D-EVM/Kohonen

The previously defined 1D-KNs have provided us with segmentations associated with datasets by establishing a number of classes and a size for the D masks that describe the used training sets. As commented previously, each one of these D masks describes in fact a subdataset by considering a hypervoxel of interest and its neighborhood of radius . Then, the masks are sent as input to a 1D-KN and through the corresponding training process there are determined their associated classes. More specifically, once the training has finished, given a D mask as input it is obtained as output the number of the corresponding class which is assigned only to the input mask’s central hypervoxel. Our hypervoxelizations were originally expressed with color intensities, however, we have now through a 1D-KN these intensities have been mapped onto a set of elements: the number of classes established for training the 1D-KN. Therefore, the color intensity value for a hypervoxel can now be replaced by its corresponding number of associated class given as output by the network. The hypervoxelization’s original size is not affected except on its boundaries because masks of radius cannot be completed. Therefore, to consider a hypervoxelization again as the appropriate representation for storing the segmentation achieved by means of a 1D-KN does not provides a reduction in terms of spatial complexity.

Now we reconsider our process for expressing D hypervoxelizations using the D-EVM which was described in Section 3. We recall that given a hypervoxel with color intensity it was obtained a D hyperprism whose inferior base contains vertices with the same coordinates just as the vertices in the original D hypervoxel but incorporating a th coordinate with zero value. On the other side, vertices in the upper base also corresponded with vertices in the original hypervoxel, but in this case, the th coordinate has value . Our proposed Framework D-EVM/Kohonen takes place when a dataset’s segmentation achieved by a 1D-KN is represented by using the D-EVM, considering the conversion methodology abovementioned but replacing color intensity values by the class numbers generated as output by the corresponding 1D-KN. From a “-Dimensional point of view”, the -axis now is referring to the number of class which is associated a hypervoxel respect to the segmented dataset.

Summarizing, the following is the workflow for our Framework D-EVM/Kohonen.

Input. A D hypervoxelization under color intensities, , .

Step 1. Given value for radius of D masks it is built a training set.

Step 2. Given values and for number of presentations and number of classes, respectively, it is trained a 1D-KN using as training set all the masks generated in the previous step.

Step 3. The training set is used again as input for the 1D-KN with adjusted weights. Each D mask is sent as input to the 1D-KN and it is obtained the corresponding output: the number of its associated class. This number is assigned only to the input mask’s central hypervoxel (The remaining hypervoxels are classified in time when they take the role of central hypervoxels in their corresponding masks). Coordinates, respect to the original hypervoxelization, of the central hypervoxel in the input D mask, are used together with its class number for building a D hyperprism.

Step 4. All generated hyperprisms are integrated in such way it is obtained the corresponding whole D-OPP which in turn is expressed in the D-EVM.

Output. The final D-EVM.

Just in the way we proceed in Section 3, Tables 10 to 14 show some 3D couplets, perpendicular to -axis, obtained from our considered 5 datasets, Aneurism, CT-Monkey, Statue Leg, VL-Baby, and VL-CT-Knee, respectively, after the application of Framework D-EVM/Kohonen. The 3D-EVM visualizations from these Tables were again achieved by means of visualization software developed in [1820]. We remember the projection of 3D couplets under our 4D-EVM representation collect those voxels that were grouped inside the same class by the corresponding 1D-KN. We can see, for example, in Table 13, how it is possible to appreciate couplets 12, 19, and 28 were bone tissue is only present. On the other hand, 3D couplets 5, 17, 20, and 26 describe the grouping of skin tissue (and other support medical accessories). A very similar situation can be appreciated respect to voxelization VL-CT-Knee (Table 14) where all of the shown 3D couplets, except 12, 21, and 28, correspond to different parts of bone tissue.

In Section 5 we made mention of the fact that all considered 1D-KNs were grouping more than 45% of the elements in their training sets in just one class. This can be visually appreciated in 3D couplets 5, 27, 19, 10, and 28 from Tables 10, 11, 12, 13, and 14, respectively. All of these 3D couplets present a block-like structure whose minimal bounding box has the same dimensions than the original voxelizations minus the volume removed by those 3D masks of radius 2 that could not be completed at the datasets’ boundaries. These are the regions that correspond to empty space in the original voxelizations. In Section 5 we gave an argument regarding the observation that these 3D couplets are considered in our 4D-EVM representation. We also mentioned this empty space is precisely one of the disadvantages behind a voxelization because it only contributes to spatial complexity. Our argument serves for the purpose of touching a crucial point: the spatial complexity of the representations achieved by the Framework D-EVM/Kohonen. Table 15 shows the cardinalities for the 4D-EVMs obtained for expressing our 5 voxelizations. We also compute (14) in order to observe the number of times our 4D models are exceeded by its original voxelizations. We can see for example, dataset CT Monkey originally required to be stored in more than 4 million voxels. In Section 3 we saw this dataset had a direct 4D-EVM representation with 3,573,172 extreme vertices: an slight improvement over the original voxelization. By taking in account the proposed framework we have were used 789,320 extreme vertices, or in other words, the original voxelization is 5.1477 times greater, in terms of spatial storage, than the final 4D-EVM’s cardinality. A more impacting example comes from dataset Aneurism: the original voxelization is formed by 16,777,216 voxels; the direct 4D-EVM representation had 1,317,008 extreme vertices; and finally, the Framework D-EVM/Kohonen shared a 4D-EVM with 517,318 extreme vertices: giving the ratio 32.4311. We should take in account our representations are also considering the storing of the regions corresponding to empty space. This give us an advantage by storing in the same representation both the occupied and the empty space but without affecting the memory needed to store the models. These results share us evidence of the conciseness’ power achieved when the grouping properties of 1D-KNs are combined with the storing properties of a model like the D-EVM: spatial complexity is reduced and a more appropriate segmentation is achieved, which impacts how the elements in the datasets are classified and by consequence accessed. For example, Table 16 shows some 2D sections corresponding to one of the 3D couplets from the model CT-Monkey after the application of Framework D-EVM/Kohonen. There are shown, via direct application of Algorithm 4, 17 of these 2D sections which allow understand the internal organization of the selected 3D couplet. It is possible to visualize some internal bone structures associated to the monkey’s cranium.

7. Quality of Segmentations Achieved by the D-EVM/Kohonen Framework

In some sense, we have sustained, in implicit way, datasets’ representations generated by the Framework D-EVM/Kohonen provide appropriate segmentations that are in fact better than those segmentations produced by simple and direct classification based only in color intensities. Part of the asseveration is sustained in the sense our 1D-KNs classifications are taking in account geometry and topology surrounding each hypervoxel in the dataset: the concept of multidimensional mask as training element. This Section provides more formal arguments to sustain the segmentations that have been produced have good quality and they are even better than those produced by classifications based only in color intensities. For this purpose, some previously defined concepts are revisited and error functions are introduced.

Original 4D-EVM representations for datasets Aneurism, CT-Monkey, and Statue Leg were described in Section 3 (Tables 3, 4, and 5). The respective visual comparison with Tables 10, 11, and 12 clearly show us how our proposed framework has provided more cohesive representations for these voxelizations. We sustain our claim by considering the number of produced 3D CCs. Let function CCL3 be defined as follows: where denotes the number of 3D couplets perpendicular to -axis for 4D-OPP expressed under the EVM, and function CCL_EVM3, originally defined in Section 3, shares the number of 3D CCs for a 3D-OPP, a 3D couplet in the current instance. Then, it is clear CCL3 shares us the total number of 3D CCs present in all couplets, perpendicular to -axis, associated to the input 4D-EVM . We recall CCs computation can be achieved by our previously mentioned CCL-EVM-Based algorithm whose specific details are presented in [21]. We compute CCL3 for representations of datasets Aneurism, Statue Leg, and CT-Monkey. Table 17 presents the corresponding values for both direct 4D-EVM representation (Section 3) and under Framework D-EVM/Kohonen. The same source data were used for building both representations, but as seen in Table 17, our framework produces much more cohesive and compact representations. A counterargument to this claim could arise by establishing the gained compaction and cohesion is product of the fact our 1D-KNs “forces” the data to be classified under 40 classes while the segmentation based on intensity has “more space” because 256 classes are available. We attack this situation by proposing some error functions. Let vector be a linearized D mask in a training set TS, such that it was characterized as part of class once a 1D-KN was trained precisely with TS. Index only refers to an arbitrary position assigned to the mask belonging class . Now consider to Weights Vector associated to class . can be also seen as a linearized D mask and for instance it can be also compared with those D masks located in the corresponding class [34]. We define then to the “Error in Class respect to ”, denoted by , as [34] It is clear by computing we are obtaining the differences existing between each D mask that belongs to class with Weights Vector , which in fact is its representative. We can think if has a low value then is a good representative for all the D masks belonging to . This reasoning can be extended for considering all the representatives of all classes for a given 1D-KN. Therefore, we have a value for measuring the classification quality shared by a network. We say the “Global Error for Training Set TS respect to its Representatives”, denoted by , can be defined as [34] Prior to make a comparison between the two presented 4D representations by using the above error functions we face the situation related to the fact that segmentation based on intensities only considers the color intensity of the hypervoxel of interest and nothing more; while our Framework D-EVM/Kohonen requires, as specified in Section 5, the creation of a training set whose elements are D masks which contain the hypervoxels of interest and also their neighborhoods of radius . This situation recall us we are considering different aspects for the considered segmentations, however, the fundament behind the use of D masks is to take in account geometrical and topological properties such that a 1D-KN use them in order to achieve and to enhance the classification. On the other hand, because of simplicity in implementing color-intensity based segmentation the notion of a representative for a class is absent: the hypervoxels included in a class share the same intensity. In order to apply our proposed error functions to classification achieved only by color intensity we build a representative for elements in a same class. Consider a hypervoxel with color intensity , then, it is located in class . Now, we will consider the neighborhood of radius for the given hypervoxel respect to the original hypervoxelization. Then, we are defining a D mask whose center is . Given value for , the corresponding D mask also have elements (Equation (17)). The mask can be also linearized in such way we obtain the vector representation . Index only refers to an arbitrary position assigned to the mask belonging class . One of the elements in the vector is precisely color intensity, , of central hypervoxel while the rest are the color intensities of the hypervoxels inside its neighborhood. Given the above elements we build a representative of class , but under color-intensity based segmentation (CIBS), as follows: That is, given the members of class , under color-intensity based segmentation, its representative has as components the average color intensity for each one of the positions in each one of the D masks defined for the class’ members. On one side, we have now representatives in order to apply them over our proposed error functions. On the other side, we proceeded in the way described because if it is false our asseveration related to the fact geometry and topology described by a neighborhood enhances classification, in particular with our 1D-KNs, then the masks and representatives defined by use of (22) when applied over our error functions will report results as good, and even better, than those reported by the use of 1D-KNs. Table 18 shows the global error respect to its representatives, obtained for each one of the training sets generated with datasets Aneurism, Statue Leg and CT-Monkey. The results presented are encouraging: global errors under direct 4D-EVM representation are 1.5877, 2.2248, and 1.6362 times greater than those obtained by Framework D-EVM/Kohonen for our datasets Aneurism, Statue Leg and CT-Monkey, respectively. We then recur again to Table 17 where CCL3s under direct 4D-EVM representation are 6.0964, 45.031, and 18.4845 times greater than those generated by Framework D-EVM/Kohonen for the same datasets, respectively. This experimental analysis has share us elements to sustain cohesiveness produced by our proposal is appropriate because presented measures tell us the voxels were better positioned by the 1D-KNs producing better segmentation based in information supported by D neighborhoods.

8. Conclusions and Future Work

In this work we have developed the Framework D-EVM/Kohonen whose purposes are the concise representation of multidimensional datasets, and a modeling of the data they contain in such way the information contained can be accessed in efficient and useful way. These purposes have been achieved by the consideration of the Extreme Vertices Model in the -Dimensional Space as the representation scheme that provides conciseness, as seen in Sections 3 and 6, but providing an appropriate organization of datasets’ elements by means of taking in account an extra geometrical dimension whose objective is allow grouping only those elements characterized as members of the same class. Precisely, the classification process, is achieved by the use of 1-Dimensional Kohonen Networks whose task is the automatic nonsupervised characterization of the elements in a dataset. As seen in Section 5, we have defined training sets taking in account the geometry and topology surrounding each hypervoxel. The experimental results we have obtained show us the use of 1D-KNs trained according to our proposal produce classes whose members form connected components with more cohesion and more compact than those generated by segmentation achieved only by color intensity. The expressed asseverations, discussed in Section 7, have been sustained by the use of function errors that have shown, for our study cases, that classification achieved by means of 1D-KNs is a good option due to the way members in each class relate to their corresponding representatives. Moreover, we have seen these 1D-KN-assisted classifications impact even more the conciseness of the D-EVMs obtained producing compression ratios, for our study cases, between 5.1477 and 32.4311. Summarizing, we have presented evidence related to how our established framework effectively combine the power of the D-EVM and the 1D-KNs in order to produce very concise datasets’ representations but also providing an appropriate segmentation.

In terms of lines of future research we make mention we will direct our efforts into the manipulation of datasets where hypervoxels have associated, besides spatial information, data that requires to be expressed with more than a scalar value. In fact, an immediate line of attack could be the manipulation of datasets in RGB colorspace where precisely the red, green and blue components are separated and forming a vector. We hypothesize the Framework D-EVM/Kohonen could be able to manipulate and represent this kind of situation in such way not only one additional dimension is required, maybe more extra dimensions are going to be taken in account.

We also expect to aboard in the future the study of the utility of the classifications produced by 1D-KNs in terms of application in specific areas and by taking in account evaluation by experts. More specifically, we are going to attack the particular topic of medical datasets’ segmentations. The idea is to represent and manipulate these datasets using our Framework D-EVM/Kohonen but we will analyze particularly the automatic tissue classification task. In a first stage, classification is going to be automatically achieved by our 1D-KNs and it will be evaluated against segmentations produced by physicians. In a second stage there will be studied evaluations’ results and procedures, if it would be necessary, for enhancing them. In this sense we can mention some experiments have been developed, in [31, 34, 35], for automatically classifying brain tissue in Computed Tomography Brain Slices. We can mention the obtained results are encouraging. Such as in our current line of research, we aim to propose concise and useful representations for these medical datasets but without compromising the information they contain because of its obvious importance and relevance. Finally, in a more independent-of-the-application point of view, we are also going to study algorithmic approaches in order to automatically avoid the phenomenon of under and over segmentation and the way these methodologies can be incorporated in the Framework D-EVM/Kohonen.

Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.