- About this Journal ·
- Abstracting and Indexing ·
- Advance Access ·
- Aims and Scope ·
- Annual Issues ·
- Article Processing Charges ·
- Articles in Press ·
- Author Guidelines ·
- Bibliographic Information ·
- Citations to this Journal ·
- Contact Information ·
- Editorial Board ·
- Editorial Workflow ·
- Free eTOC Alerts ·
- Publication Ethics ·
- Reviewers Acknowledgment ·
- Submit a Manuscript ·
- Subscription Information ·
- Table of Contents

Mathematical Problems in Engineering

Volume 2012 (2012), Article ID 712974, 18 pages

http://dx.doi.org/10.1155/2012/712974

## Prediction of Optimal Design and Deflection of Space Structures Using Neural Networks

^{1}The Iranian Academic Center for Education, Culture and Research, Kerman 7616914111, Iran^{2}School of Civil Engineering, Universiti Sains Malaysia, Nibong Tebal, Penang, Malaysia^{3}City University College, Petaling Jaya, Selangor, Malaysia

Received 8 October 2012; Accepted 18 November 2012

Academic Editor: Siamak Talatahari

Copyright © 2012 Reza Kamyab Moghadas et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

#### Abstract

The main aim of the present work is to determine the optimal design and maximum deflection of double layer grids spending low computational cost using neural networks. The design variables of the optimization problem are cross-sectional area of the elements as well as the length of the span and height of the structures. In this paper, a number of double layer grids with various random values of length and height are selected and optimized by simultaneous perturbation stochastic approximation algorithm. Then, radial basis function (RBF) and generalized regression (GR) neural networks are trained to predict the optimal design and maximum deflection of the structures. The numerical results demonstrate the efficiency of the proposed methodology.

#### 1. Introduction

The history of the applications of Artificial Intelligence to civil and structural engineering is simultaneously brief and long. It is brief if compared to the history of civil and structural engineering, whose definition as a discipline can be fixed a very long time ago. It makes sense to consider civil and structural engineering as the most ancient applicative discipline, being founded in preclassical world by Egyptians and Babylonians. It is long, instead, if compared to the history of Artificial Intelligence, whose name first appeared in science at the end of the sixties of the twentieth century. The earliest applications to civil and structural engineering are very likely [1], where authors review tools and techniques for knowledge-based expert system for engineering design. An even earlier paper whose scope was indeed wider, but introduced some fundamental themes, is [2]. We can definitely settle a start date in 1986 when the first International Symposium on this theme took place [3]. In statistical terms, since we can fix Artificial Intelligence engineering applications start date in 1950 when the first attempt to provide a true intelligent program was carried out [4], this means that, nowadays, we are experiencing 18 years of history of the applications of Artificial Intelligence techniques to civil and structural engineering.

As in this study our main aim is to employ neural networks to predict the optimal design and maximum deflection of the double layer grids, the next paragraph is devoted to review the literature about the optimal design of space structures by soft computing techniques.

Erbatur et al*.* [5] reported the development of a computer-based systematic approach for discrete optimal design of planar and space structures composed of one-dimensional elements. Rajasekaran [6] created the input for large space structures using Formian. In this paper, a new optimization technique called cellular automata (CA) has been combined with genetic algorithm (GA) to develop different search and optimization known as cellular genetic algorithm (CGA), which considers the areas of the space structures as discrete variables. Krishnamoorthy et al. [7] proposed GA with objective-oriented framework which was used in space truss optimization. Tashakori and Adeli [8] adopted the patented robust neural dynamics model for optimum design of space trusses made of commercially available cold-formed shapes in accordance with AISC specification. E. Salajegheh and J. Salajegheh [9] achieved the optimal design of space structures while the design variables are continuous and discrete. To reduce the computational work involved in the optimization process they employed a semiquadratic function; also they use a hybrid form of the approximation. Kaveh and Dehkordi [10] trained neural networks for the analysis, design, and prediction of the displacements of domes using the backpropagation and radial basis Functions neural networks. The performance of these networks is compared when applied to domes. Kaveh et al. [11] combined the energy method and the force method in the context of transmission tower optimization in order to form a holistic design and optimization approach, eliminating the need for time-intensive matrix inversion. The addition of a neural network as an analysis tool reduces the overall computational load. Kaveh and Servati [12] trained neural networks for design of square diagonal-on-diagonal double layer grids. They employed backpropagation algorithm for training the networks for evaluation of the maximum deflection, weight, and design of the double layer grids. Salajegheh and Gholizadeh [13] employed a modified genetic algorithm (GA) and radial basis function (RBF) neural networks to optimize space structures. Kaveh et al. [14] employed ant colony optimization (ACO) algorithm for optimal design of space structures with fixed geometry. Gholizadeh et al. [15] employed a combination of GA and wavelet radial basis function (WRBF) neural networks to find the optimal weight of structures subject to multiple natural frequency constraints.

Much more other applications of neural networks in the field of civil engineering can be found in the literature [16–20].

In this investigation, an innovative methodology is proposed to predict the optimal design and maximum deflection of the square-on-square double layer grids. This methodology consists of three stages. In the first stage, a number of the double layer grids with random spans and heights are generated. In the second stage the generated double layer grids are optimized by an optimization algorithm. Although, in the recent years many new structural optimization algorithms have been proposed by the researchers [21–25], in this paper, simultaneous perturbation stochastic approximation (SPSA) [26] algorithm is used due to its computational merits. Also, the maximum deflections of the optimal structures are saved. In the third stage, radial basis function (RBF) [27] and generalized regression (GR) [27] neural networks are trained to predict the optimal design and maximum deflection of the double layer grids. To design neural networks MATLAB [28] is employed.

#### 2. Formulation of Optimization Problem

In optimal design problem of space trusses the aim is to minimize the weight of the truss under constraints on stresses and displacements. This optimization problem can be expressed as follows:
where , and are cross-sectional area of members belonging to group *n*, weight density, and length of *m*th element in this group, respectively;* ng* and *nm* are the total number of groups in the structure and the number of members in group *n*, respectively; *ne* and *nj* are the total number of the elements and nodes in truss, respectively; and are stress in the th element and displacement of the *j*th node, respectively. Also, and are allowable stress in the th member and allowable deflection of the *j*th node, respectively.

In this study, besides cross-sectional areas () the geometry dependent parameters of the double layer grid, *L* and *h*, are also variables. In other words, the aim is to find optimal cross-sectional areas for each set of and *h*. Thus, (2.1) can be reexpressed as follows:

It is obvious that the computational burden of the above optimization problem is very high due to the fact that *L* and *h *are variables. Employing the neural network technique can substantially reduce the computational costs.

As the SPSA requires less number of function evaluations (structural analyses) than the other type of gradient-based methods, it is selected as the optimizer in this study. The basic concepts of the SPSA are explained in the next section.

#### 3. SPSA Optimization Algorithm

SPSA has recently attracted considerable international attention in areas such as statistical parameter estimation, feedback control, simulation-based optimization, signal and image processing, and experimental design. The essential feature of SPSA is the underlying gradient approximation that requires only two measurements of the objective function regardless of the dimension of the optimization problem. This feature allows for a significant reduction in computational effort of optimization, especially in problems with a large number of variables to be optimized. The basic unconstrained SPSA optimization algorithm is in the general recursive stochastic approximation (SA) form [26]:
where represents the estimate of *X* at *k*th iteration, represent a scalar gain coefficient, and represent an approximate gradient at . Under appropriate condition, (3.1) will converge to optimum design in some stochastic sense. The essential part of (3.1) is the gradient approximation that is obtained using the simultaneous perturbation (SP) method. Let *w*(*·*) denote a measurement of objective function at a design level represented by the dot and let be some positive number. The SP approximation has all elements of randomly perturbed together to obtain two measurements of *w*(*·*), but each component is formed from a ratio involving the individual components in the perturbation vector and the difference in the two corresponding measurement. For two sided simultaneous perturbation, we have
where the distribution of the user-specified dimensional random perturbation vector satisfies condition discussed in [26].

It is observed that each iteration of SPSA needs only two objective function measurements independent of because the numerator is the same in all components. This circumstance provides the potential for SPSA to achieve a large savings in the total number of measurements required to estimate when is large.

##### 3.1. Implementation of SPSA

The following step-by-step summary shows how SPSA iteratively produces a sequence of estimates [26].

*Step 1 (initialization and coefficient selection). *Set counter index . Pick initial guess and nonnegative coefficients *a*, *c*, *A*, * α*, and

*in the SPSA gain sequences and . The choice of gain sequences is critical to the performance of SPSA. Spall provides some guidance on picking these coefficients in a practically manner.*

*γ**Step 2 (generation of the simultaneous perturbation vector). *Generate by Monte Carlo an -dimensional random perturbation vector , where each of the components of is independently generated from a zero mean probability distribution satisfying some conditions. A simple choice for each component of is to use a Bernoulli 1 distribution with probability of for each 1 outcome. Note that uniform and normal random variables are not allowed for the elements of by the SPSA regularity conditions.

*Step 3 (objective function evaluations). *Obtain two measurements of the objective function *w*(*·*) based on simultaneous perturbation around the current and with the and from Steps 1 and 2.

*Step 4 (gradient approximation). *Generate the simultaneous perturbation approximation to the unknown gradient:
where is the th component of vector.

*Step 5 (updating X estimate). *Use the standard SA to update to new value.

*Step 6 (iteration or termination). *Return to Step 2 with replacing *k*. Terminate the algorithm if there is little change in several successive iterates or the maximum allowable number of iterations has been reached. Figure 1 shows the flowchart of the SPSA.

In the present work, we suppose that the length and height of the double layer grids are varied in specific ranges. Our aim is to optimize all of the possible structures defined in the ranges. Therefore it can be observed that the additional difficulty is the huge computational burden of the optimization process. In order to mitigate the difficulty, RBF and GR neural networks are employed to predict the optimal design of the double layer grids with various length and height.

#### 4. Neural Networks

In the recent years, neural networks are considered as more appropriate techniques for simplification of complex and time consuming problems. The interest shown to neural networks is mainly due to their ability to process and map external data and information based on past experiences. Neural networks are not programmed to solve specific problems. Indeed, neural networks never use rules or physic equations related to the specific problem in which they are employed. Neural networks use the knowledge gained from past experiences to adapt themselves to solve the new problems.

##### 4.1. Radial Basis Function

The use of RBF in the design of neural networks was first introduced by Wasserman in 1993 [27]. The RBF network basically involves three entirely different layers: an input layer, a hidden layer of high enough dimension, and an output layer. The transformation from the hidden unit to the output space is *linear*. Each output node is the weighted sums of the outputs of the hidden layer. However, the transformation from the input layer to the hidden layer is *nonlinear*. Each neuron or node in the hidden layer forming a linear combination of the basis (or kernel) functions which produces a localized response with respect to the input signals. This is to say that RBF produce a significant nonzero response only when the input falls within a small localized region of the input space. The most common basis of the RBF is a Gaussian kernel function of the form:
where is the output of the *l*th node in hidden layer; is the input pattern; is the weight vector for the *l*th node in hidden layer, that is, the center of the Gaussian for node *l*; is the normalization parameter (the measure of spread) for the *l*th node; and *q *is the number of nodes in the hidden layer. The outputs are in the range from zero to one so that the closer the input is to the center of the Gaussian, the larger the response of the node is. The name RBF comes from the fact that these Gaussian kernels are radially symmetric; that is, each node produces an identical output for inputs that lie a fixed radial distance from the center of the kernel. The network outputs are given by
where is the output of the th node, is the weight vector for this node, and *M* is the number of nodes in the output layer.

There are two common ways to calculate the measure of spread .

(1)Find the measure of spread from the set of all training patterns grouped with each cluster center ; that is, set them equal to the average distance between the cluster centers and the training patterns: where is the number of patterns that belong to the*l*th cluster and

*k*is the index number of a pattern that belongs to the

*l*th cluster.(2)Find the measure of spread from among the centers (

*p*-nearest neighbor heuristic):

##### 4.2. Generalized Regression

Generalized regression network (GR) subsumes the basis function methods. This network does not require iterative training. The structure of GR is designated such that transpose of input matrix and transpose of desired output (target) matrix are chosen as first layer and second layer weight matrices, respectively. GR algorithm is based on nonlinear regression theory, a well established statistical technique for function estimation. Except the approach of adjusting of second layer weights, the other aspects of GR are identical to RBF neural networks.

#### 5. Proposed Methodology

##### 5.1. Double Layer Grid Model

In this section dimensions of considered double layer grid structure and its corresponding model are described. The model considered here is a double layer grid with bar elements connected by pin joints. The length of the spans, *L*, is varied between 25 and 75 m with step of 5 m. The height is varied between 0.035 and 0.095 *L* with steps of 0.2 m. The smallest and biggest structures in this interval are shown in Figure 2. The sum of dead and live loads equal to 250 kg/m^{2} is applied to the nodes of the top layer.

In order to satisfy practical demands, in the optimization of large-scaled structure such as space structures, the structural elements should be divided into some groups. In this study the elements are put into 18 different groups. For this purpose a step-by-step summary defined bellow is employed.

*Step 1. *A similar cross sectional area is initially assigned to all elements of the structure.

*Step 2. *The structure is analyzed through FE and axial stresses of all members are obtained.

*Step 3. *All tension members of the structure are put into 6 groups according to their stress states as follows:

*Step 4. *All compressive members of top and bottom layer elements of structure are put into 6 deferent groups according to their stress values as follows:

*Step 5. *All compressive members of middle layer elements of structure are also put into 6 deferent groups based on their stresses as follows:

Preparing a neural network is achieved in three stages: data generating, training, and testing. In the first stage, a number of input and output pairs are provided and divided into training and testing sets. In the second stage, the training set is used and the modifiable parameters of the neural network are adjusted. In the last stage the performance generality of the trained neural network is examined through the testing set.

In order to provide the required data (data generation), a number of double layer grids according to their *L* and *h* are randomly selected. All of the selected structures are optimized using SPSA. Optimal designs of the selected structures and their corresponding maximum deflections are saved. This process is shown in Figure 3.

In order to train neural networks, the generated data should be separated to training data and testing data as follows.

Training data for optimal design predictor networks:
Training data for maximum deflection predictor networks:
Testing data for optimal design predictor networks:
Testing data for maximum deflection predictor networks:

##### 5.2. Main Steps in Training Neural Network

As a summary the main steps in training of RBF and GR NNs to predict optimal design and maximum deflection of the structure are as follows:(1)configuration processing of the selected space structures employing Formian,(2)selection a list of available tube sections from the standard lists,(3)implementation member grouping,(4)generation of some structures, based on span and height, to produce training set,(5)static analysis of the structures,(6)designing for optimal weight by SPSA according to AISC-ASD code,(7)training and testing RBF and GR to predict optimal design and maximum deflection,(8)improving generalization of the neural networks if it is necessary.

##### 5.3. Flowchart of the Methodology

The flowchart of the proposed methodology is shown in Figure 4*.* This flowchart includes three main blocks: data generation, optimization, and NN training. The data generation block includes the optimization block. In these two blocks the data needed for neural network training is produced. The mentioned data are stated through (5.4) to (5.7).

#### 6. Numerical Results

Typical topology of the RBF and GR neural networks to predict the optimal design and maximum deflection of the double layer grids is shown in Figures 5 and 6, respectively.

To find the optimal spread in the RBF and GR networks the minimum distance between training set and test set errors are employed [29]. The spread values in RBF networks trained to predict the optimal design and maximum deflection are 11.5 and 11.75 and for GR are 12.5 and 10.25, respectively. The results of RBF for predicting the optimal cross-sectional areas are shown in Figure 7.

The errors of RBF for predicting the maximum deflections are shown in Figure 8. The results of GR for predicting the optimal cross-sectional areas are shown in Figure 9. The errors of GR for predicting the maximum deflections are shown in Figure 10. Maximum and mean of errors of RBF and GRNN in approximation of optimal designs and maximum deflection are given in Tables 1 and 2, respectively.

The numerical results demonstrate that the generality of the GR is better than that of the RBF neural network in prediction of optimal design and maximum deflection of the double layer grids.

#### 7. Conclusion

In this investigation, an innovative methodology is proposed to predict the optimal design and maximum deflection of the square-on-square double layer grids. This methodology consists of three stages. In the first stage, a number of the double layer grids with random spans and heights are generated. In the second stage the generated double layer grids are optimized by SPSA algorithm. Also, the maximum deflections of the optimal structures are saved. In the third stage, RBF and GR neural networks are trained to predict the optimal design and maximum deflection of the double layer grids.

By concerning the following points, it can be observed that the proposed methodology is novel and innovative.(1)It is the first study based on employing the SPSA optimization algorithm to optimize double layer grids with variable geometry.(2)Application of the RBF and GR neural networks to predict the optimal design and maximum deflection of the double layer is achieved for the first time in this study. (3)The main advantage of the proposed methodology is to predict the optimal design and maximum deflection of the double layer grids with high speed and trivial errors in comparison with the traditional methods.

#### References

- M. L. Maher, D. Sriram, and S. J. Fenves, “Tools and techniques for knowledge based expert systems for engineering design,”
*Advances in Engineering Software*, vol. 6, no. 4, pp. 178–188, 1984. View at Publisher · View at Google Scholar · View at Scopus - R. Davis and J. J. King, “An overview of production systems,” in
*Machine Intelligence 8: Machine Representations and Knowledge*, E. Elcock and D. Michie, Eds., pp. 300–332, John Wiley & Sons, New York, NY, USA, 1977. - C. N. Kostem and M. L. Maher, Eds.,
*First Symposium on Expert Systems in Civil Engineering*, ASCE, Seattle, Wash, USA, 1986. - C. E. Shannon, “Programming a computer for playing chess,”
*Philosophical Magazine*, vol. 41, series 7, no. 314, pp. 256–275, 1950. View at Zentralblatt MATH - F. Erbatur, O. Hasançebi, I. Tütüncü, and H. Kinodot;liç, “Optimal design of planar and space structures with genetic algorithms,”
*Computers and Structures*, vol. 75, no. 2, pp. 209–224, 2000. View at Publisher · View at Google Scholar · View at Scopus - S. Rajasekaran, “Optimization of large scale three dimensional reticulated structures using cellular genetics and neural networks,”
*International Journal of Space Structures*, vol. 16, no. 4, pp. 315–324, 2001. View at Publisher · View at Google Scholar · View at Scopus - C. S. Krishnamoorthy, V. P. Prasanna, and R. Sudarshan, “Object-oriented framework for genetic algorithms with application to space truss optimization,”
*Journal of Computing in Civil Engineering*, vol. 16, no. 1, pp. 66–75, 2002. View at Publisher · View at Google Scholar · View at Scopus - A. Tashakori and H. Adeli, “Optimum design of cold-formed steel space structures using neural dynamics model,”
*Journal of Constructional Steel Research*, vol. 58, no. 12, pp. 1545–1566, 2002. View at Publisher · View at Google Scholar · View at Scopus - E. Salajegheh and J. Salajegheh, “Continuous-discrete variable optimization of space structures with sizing variables using approximation methods,”
*International Journal of Space Structures*, vol. 16, no. 4, pp. 237–251, 2001. View at Publisher · View at Google Scholar · View at Scopus - A. Kaveh and M. R. Dehkordi, “Neural networks for the analysis and design of domes,”
*International Journal of Space Structures*, vol. 18, no. 3, pp. 181–193, 2003. View at Publisher · View at Google Scholar · View at Scopus - A. Kaveh, Y. Gholipour, and H. Rahami, “Optimal design of transmission towers using genetic algorithm and neural networks,”
*International Journal of Space Structures*, vol. 23, no. 1, pp. 1–19, 2008. View at Scopus - A. Kaveh and H. Servati, “Design of double layer grids using backpropagation neural networks,”
*Computers and Structures*, vol. 79, no. 17, pp. 1561–1568, 2001. View at Publisher · View at Google Scholar · View at Scopus - E. Salajegheh and S. Gholizadeh, “Optimum design of structures by an improved genetic algorithm using neural networks,”
*Advances in Engineering Software*, vol. 36, no. 11-12, pp. 757–767, 2005. View at Publisher · View at Google Scholar · View at Scopus - A. Kaveh, B. F. Azar, and S. Talatahari, “Ant colony optimization for design of space trusses,”
*International Journal of Space Structures*, vol. 23, no. 3, pp. 167–181, 2008. View at Publisher · View at Google Scholar · View at Scopus - S. Gholizadeh, E. Salajegheh, and P. Torkzadeh, “Structural optimization with frequency constraints by genetic algorithm using wavelet radial basis function neural network,”
*Journal of Sound and Vibration*, vol. 312, no. 1-2, pp. 316–331, 2008. View at Publisher · View at Google Scholar · View at Scopus - S. Gholizadeh and E. Salajegheh, “Optimal design of structures subjected to time history loading by swarm intelligence and an advanced metamodel,”
*Computer Methods in Applied Mechanics and Engineering*, vol. 198, no. 37–40, pp. 2936–2949, 2009. View at Publisher · View at Google Scholar · View at Zentralblatt MATH · View at Scopus - A. H. Alavi and A. H. Gandomi, “Prediction of principal ground-motion parameters using a hybrid method coupling artificial neural networks and simulated annealing,”
*Computers and Structures*, vol. 89, no. 23-24, pp. 2176–2194, 2011. View at Publisher · View at Google Scholar - S. Gholizadeh, A. Pirmoz, and R. Attarnejad, “Assessment of load carrying capacity of castellated steel beams by neural networks,”
*Journal of Constructional Steel Research*, vol. 67, no. 5, pp. 770–779, 2011. View at Publisher · View at Google Scholar · View at Scopus - A. H. Gandomi and A. H. Alavi, “Applications of computational intelligence in behavior simulation of concrete materials,” in
*Computational Optimization and Applications in Engineering and Industry*, X. S. Yang and S. Koziel, Eds., vol. 359, chapter 9, pp. 221–243, Springer, New York, NY, USA, 2011. - S. Gholizadeh and S. M. Seyedpoor, “Shape optimization of arch dams by metaheuristics and neural networks for frequency constraints,”
*Scientia Iranica*, vol. 18, no. 5, pp. 1020–1027, 2011. View at Publisher · View at Google Scholar - S. Talatahari, A. H. Gandomi, and G. J. Yun, “Optimum design of tower structures using firefly algorithsm,”
*The Structural Design of Tall and Special Buildings*. In press. View at Publisher · View at Google Scholar - S. Gholizadeh and A. Barzegar, “Shape optimization of structures for frequency constraints by sequential harmony search algorithm,”
*Engineering Optimization*. In press. View at Publisher · View at Google Scholar - A. H. Gandomi, S. Talatahari, X. S. Yang, and S. Deb, “Design optimization of truss structures using cuckoo search algorithm,”
*The Structural Design of Tall and Special Buildings*. In press. View at Publisher · View at Google Scholar - S. Gholizadeh and F. Fattahi, “Design optimization of tall steel buildings by a modified particle swarm algorithm,”
*The Structural Design of Tall and Special Buildings*. In press. View at Publisher · View at Google Scholar - S. Talatahari, M. Kheirollahi, C. Farahmandpour, and A. H. Gandomi, “A multi-stage particle swarm for optimum design of truss structures,”
*Neural Computing & Applications*. In press. View at Publisher · View at Google Scholar - J. C. Spall, “An overview of the simultaneous perturbation method for efficient optimization,”
*Johns Hopkins APL Technical Digest*, vol. 19, no. 4, pp. 482–492, 1998. View at Scopus - P. D. Wasserman,
*Advanced Methods in Neural Computing*, Prentice Hall, D. Van Nostrand, New York, NY, USA, 1993. - The Language of Technical Computing,
*MATLAB*, Math Works, 2006. - S. S. Sadat Hosseini and A. H. Gandomi, “Short-term load forecasting of power systems by gene expression programming,”
*Neural Computing and Applications*, vol. 21, no. 2, pp. 377–389, 2012. View at Publisher · View at Google Scholar · View at Scopus