Abstract

The artificial neural network reduces humanity and society’s burden to solve complex problems highly efficiently. Artificial neural networks resemble brain activities based on the acquired training samples used for various applications such as classification, regression, prediction, smart grid, natural language processing, image processing, medical diagnosis, and so on. This paper illustrates the different artificial neural network architectures, types, merits, demerits, and applications. Therefore, this paper provides valuable information to students and researchers to enrich their knowledge about an artificial neural network and research it. This paper also proposed a multilayer-perceptron-neural-network-based solar irradiance forecasting model, an improved backpropagation neural network-based rainfall forecasting model, and an Elman neural network-based temperature forecasting model. The performances of the proposed neural network-based forecasting models are analyzed with various hidden neurons and validated using the acquired real-time meteorological data. The proposed neural network forecasting models achieve rigorous results with reduced errors for the considered applications and aid sustainability.

1. Introduction

In the modern world, ANN is actively replacing the existing methods; this motivated us to address the issue of ANN and made interest in the quest of ANN and provided the complete guideline to the reader about the ANN types, architecture, and applications of ANN. The network that resembles or mimics the biological human brain functions to accomplish a given task is an artificial neural network. In a neural network, one neuron to the other neuron connection exists with some strength known as weight or synaptic weight. The on and off state of a neuron is decided by the threshold function. The perceptron concept was introduced in 1958 by Frank Rosenblatt [1], which is the ability to learn with the single-layer network. The limitation is if the data points are not linearly separable, it cannot solve the problem. Still, many research activities are required to address the perceptron network linear separability issue. Inputs usually are binary, bipolar, and the real-time value from the environment. Many forecasting models are attempted in the literature, but simple, feasible, easy to implement, and accurate forecasting is one of the thrust research fields. In neural network-based forecasting, an interpretable machine learning tool is important [2, 3], and feature selection/extraction is a preprocessing method used to extract important relevant input features [4], data-driven, hybrid, ensemble, and deep neural network aid effective solutions [510]. Recently, some researchers performed data-driven-based forecasting [1113]. In forecasting applications, variability is caused because of the measurement shift and noise. The uncertainty about measurement shift and noise can be overcome by proper commissioning, data evaluation, quality check, sensor calibration, and usage of on-site measurement data [9].

1.1. Comparison between the Human Brain and AI (Artificial Intelligence)

(i)The machine system involves step-by-step procedures and instructions, but humans have fewer processing steps because of the massively parallel operation. In this aspect, humans are ahead of artificial intelligence.(ii)Regarding the size and complexity aspect, the human biological brain has 1011 numbers of neurons approximately and 1015 numbers interconnecting with that brain size; this is many neurons. Interconnection is highly impossible based on AI; hence, complexity exists in the human brain both outside of dendrites and inside of cell body computation, but it delays the artificial intelligence process.(iii)Regarding the strength (or) the synaptic weight of interconnection, information has been stored in the machine having replaceable storage, but the brain has an adaptable storage system.(iv)The brain has a much high fault tolerance ability compared to a computer and artificial intelligence machine.(v)Control mechanisms: the retrieval of corrupted information is complicated for the human brain than the machine, so the control mechanism is more difficult for the human brain than the machine.

1.1.1. Motivation

Practical examples: babies or kids can differentiate fruit; guava and green apple both are green in color and shape, but with respect to experience and unique features, the baby or kid determine green apple and guava if we include green mango, and green apple baby can identify similarly ANN-based acquired knowledge about the trained data sets. It has the generalization ability to identify the unknown data sets correctly.

The learning or training process is the stage; the network can acquire knowledge about the situation or environment. The acquired knowledge is stored in the synaptic weights. According to the structural interlinkages of neurons (computational elements), activation function, and weight computation process (learning algorithm), the artificial neural networks can be classified into various types, starting from Mcculloch and Pitts network to present hybrid neural networks.

In a single-layer neural network model, the hidden layer does not appear, so only single connection linkages exist from input to output. A multilayer neural network consists of more than one connection linkage from the input to the output, which can solve challenging and complex problems.

The network does not have feedback, and information can flow from the input layer to the output layer via one or more hidden layers known as a feedforward neural network. Examples of feedforward neural networks are backpropagation, multilayer neural network, radial basis neural network, etc.

The neural network consists of feedback, and information can flow from the input layer to the output layer via one or more hidden layers, and vice versa known as a feedback neural network. Examples of feedback neural networks are the Hopfield network, Elman network, and so on.

1.2. Contribution

This paper has the following contributions:(i)Give a clear understanding of the comparison between the human brain and artificial intelligence.(ii)Discuss the development background of an ANN, artificial neural network generalized procedural steps with diagrammatic explanation, typical structure, and applications of artificial neural networks.(iii)Propose various types of artificial neural networks, like multilayer perceptron neural network, improved backpropagation neural network, and Elman neural network for solar irradiance forecasting, rainfall forecasting, and temperature forecasting, respectively, to aid sustainability.

1.3. Highlights of This Paper

(i)Acquire knowledge of the primary artificial neural network types and applications(ii)Discuss various ANN architectures, types, and applications(iii)Propose a solar irradiance forecasting model using MLPNN (Multilayer perceptron neural network)(iv)Suggest a rainfall forecasting model using IBPNN (Improved Backpropagation Neural Network)(v)Present a temperature forecasting model using ENN (Elman neural network)(vi)Carry out various hidden neuron-based analyses(vii)Propose proved validity of forecasting models in real-time data(viii)Achieve rigorous forecasting outcomes with reduced errors regarding the proposed forecasting model(ix)Analyze the different neural network models and familiarize the reader with the forecasting applications

1.4. Development Background of Artificial Neural Network

The motivation of Artificial Neural Network (ANN) is the biological system’s parallel and distributed processing. In 1986, McClelland et al. [14] developed an intelligent machine with artificial intelligence, but searching for the solution is the problem with this model. Hence, many heuristic searches address the task accomplishment, and the rule-based approach addresses the representation problem. Table 1 represents the ANN development in the literature for more clarity.

1.4.1. Second Generation Neural Networks

The second generation neural networks are as follows:(i)Perceptron: for specific learning rules, the network can learn with known target values (supervised learning rule)(ii)Backpropagation: based on the various learning methods, the network can learn and adapt the learning data set to known target values (supervised learning rules)(iii)Kohonen neural network or Self-Organizing Map (SOM): the network learns without knowing the target values (unsupervised learning rule)(iv)Radial basis neural network (supervised learning rule)(v)Adaptive resonance theory (unsupervised learning rule)(vi)Elman neural network (supervised learning rule)(vii)Hopfield neural network (unsupervised learning rule)(viii)Special networks (unsupervised and supervised learning rules) like support vector machine

1.4.2. Third Generation Neural Network

Spiking neural network: in this neural network, the limitation of MLP, like cycle firing, has been avoided. This model acquires the firing of biological neurons based on spikes.

1.4.3. Fourth Generation Neural Network

The fourth-generation neural network can be classified into two types as follows:(i)Deep learning neural network: the deep learning neural network can overcome the gradient vanishing problem over one number of hidden layers(ii)Hybrid neural network: using both artificial neural network and optimization algorithms, a combination of physical and statistical methods, and so on, makes a hybrid neural network that overcomes the individual network’s limitations

2. Generalized Algorithm for Artificial Neural Network

The algorithm in neural networks is nothing but a step-by-step procedure to accomplish a specific task. We showed the generalized algorithm of an artificial neural network in Figure 1, which comprises the following steps as follows:

Step 1. Start the neural network design phase, choose the appropriate neural network model of the artificial neural network (feedforward or feedback, or special neural network, or hybrid model).

Step 2. After selecting a neural network model design, the proposed model includes the number of input parameters, the number of hidden layers, the number of hidden neurons, and so on.

Step 3. After completion of a network design process, initialize the proposed model.

Step 4. After initialization, for the chosen application, collect the data set and perform the data normalization process to eliminate the data discrepancy and missing data and improve the output accuracy. Meanwhile, divide the collected data set into two portions: the training phase and the testing phase.

Step 5. After the data collection and normalization process, learn the proposed neural network with a training data set.

Step 6. Check the proposed neural network’s output and whether there is a minimal lead error and verify if the performance is acceptable or not. If the proposed neural network model’s performance during the training phase is satisfactory with minimal error, then go for the testing phase. Else, the performance was not up to the mark; then again go to Step 2. The process continues until a match with the set goal.

Step 7. After completing the training phase, perform the proposed neural network model’s testing process on the testing data set (the testing data sets are unseen raw data specified during the training phase).

Step 8. After completion of the testing phase, check if the proposed neural network model can achieve generic performance and generalize it well or not. If it generalizes well, record the output of the proposed neural network model; else again perform the remodeling of the neural network to achieve the generalized outcome (both training and testing phases lead to an optimal result with minimal error).

Step 9. Stop the process.

3. Applications of Artificial Neural Networks

Artificial neural network applications are not confined to the specific domain; it has a wide variety of applications. An artificial neural network has scope for various applications. Some applications are tabulated in Table 2. The artificial neural network has multiple applications, but not limited content, and it can cover all fields. Hence, it is an interdisciplinary field.

Virtual reality, decision support system (medical science and engineering fields), control engineering, data mining, computer vision, image, pattern recognition, human-machine interface, and few ANN applications. A system that can efficiently and intelligently solve the problem with computation ease is known as an expert system. It poses the ability to learn the environment, think, and apply the gained experience to complete the given task without assistance from the human being.

The system with many acquired knowledge based on many examples, wholly deriving the description of patterns and acquiring knowledge, is a complex problem in pattern recognition, natural language processing, speech, and computer vision applications. The human recalls the pattern, but the machine recall of the data pattern can be in the form of handwriting speech even though the sound of communication is different at various levels in the case of appropriate human identification . The pattern was not clearly defined but also was based on the knowledge humans identified; why not a machine can identify? This question arises from the leading research for the development of existing models in the field of ANN. The human identifies and recognizes the pattern or input based on various samples and examples’ continuous learning ability.

In medical practice, ANN is used for medical disease diagnostics because identifying the disease is a challenging task for the medical practitioner and doctor because there will be an overlap of symptoms of various illnesses. Hence, there are no specific guidelines about disease identification based on the medical practice’s experience and knowledge; doctors are suggesting the appropriate treatment for the patients. Sometimes, due to human error, patients were affected and they suffered unfair treatment because of the diseases’ improper identification. To mitigate the above-said problems, nowadays, an artificial neural network occurs with a vital role in diagnosing patients’ conditions.

A game-playing, self-regulated vehicle, self-control expert system, natural language processing, robotics, etc., are some thrust research fields of AI.

Forecasting, regression, classification, and diagnosis of diseases is based on the medical field’s symptoms, computer vision, natural language processing (NLP), engineering, and science applications; the artificial neural network is widely used nowadays because of the promising solution to the challenging problem.

4. Proposed Various Artificial Neural Network for Various Forecasting Applications

Although artificial neural networks are suitable for various applications, this paper carries out modeling and analyzes artificial neural networks’ effectiveness in forecasting applications like solar irradiance forecasting, rainfall forecasting, and temperature forecasting to aid sustainability. The roadmap of the proposed forecasting model is shown in Figure 2.

4.1. Proposed Multilayer Perceptron Neural Network for Solar Irradiance Forecasting

This paper proposed five meteorological input parameters based on a multilayer perceptron neural network [27]. The multilayer perceptron neural network (MLPNN) consists of one or over one hidden layer, which performs better in computational efficiency than a single-layer perceptron neural network. It belongs to the feedforward neural network and is associated with a supervised learning rule to explore synaptic weight values, and it has a complex problem-solving ability.

The proposed multilayer perceptron neural network is arranged into an input layer, one or more hidden layers, and an output layer. Regarding the hidden layers and nonlinear transfer function, the multilayer perceptron neural network can handle linear and nonlinear relationships between the input and output vectors. In the hidden layer, the hyperbolic tangent sigmoid activation function (nonlinear transfer function) is adopted, and the proposed neural network is trained by employing the backpropagation gradient descent learning rule. The proposed multilayer neural network induces sped-up convergence because it is a fully connected network.

The proposed multilayer perceptron neural network-based forecasting model for solar irradiance forecasting considers the solar irradiance impacting parameters as the inputs such as Solar Irradiance , Temperature , Wind Speed , Dew Point , and Cloud Cover . These five meteorological input parameters are of much impact on solar irradiance. Hence, these parameters are accounted as the input parameters for the proposed neural network. The forecasted solar irradiance is viewed as the output neuron in the single output layer.

The proposed five input-based multilayer perceptron neural networks aim to achieve the best solar irradiance forecast to reduce the minimum error values. The proposed five input-based multilayer perceptron neural networks for solar irradiance structural design are depicted in Figure 3. Table 3 presents the proposed multilayer perceptron neural network design parameters. For the proposed neural network, model-independent computation is performed for each of the layers present in the neural network with respect to the received data. The computed outcomes are transferred to the next immediate layer as an input. Then, the neural network output is then obtained from the output layer; this process flow is clearly understood by Figure 3. During the training process, the involved computations are given as follows.

The proposed MLPNN input vector:

The proposed MLPNN output vector:

The proposed MLPNN synaptic weight vectors between the input vector to the hidden vector:

The proposed MLPNN obtains the net input of the hidden layer and subsequently the output of the hidden layer,where is the input vector, is the synaptic weights between the input layer and hidden layer, and is the number of hidden neurons.

The proposed MLPNN synaptic weight vectors between the hidden to output vector:

The proposed MLPNN obtains the net input of the output layer and subsequently its neural network final output:where is the synaptic weight between the hidden layer and output layer, and “f” is the activation function, namely, the tangent sigmoidal activation function.

4.2. Proposed Improved Backpropagation Neural Network for Rainfall Forecasting

The multilayer feedforward neural network with a momentum-based backpropagation learning algorithm is known as the proposed Improved Backpropagation Neural Network (IBPNN) [28]. IBPNN can balance between generalization and the network’s memorization. The proposed improved backpropagation neural network is arranged into the input layer, hidden layer, and output layer. The proposed improved backpropagation neural network training stages are classified into three phases: feedforward stage, error computation stage, weight modification, and update stage. The processing elements present in the proposed feedforward neural networks perform an independent computation based on a considered set of input data and synaptic weights with a continuous differential activation function, and the obtained outcomes are passed to the successive layers, and then a final output is achieved from the output layer, which is compared with the target for error computation. Evaluated error is propagated backward via the output layer-hidden layer-input layer to achieve minimal error.

For the proposed improved backpropagation neural network based on the given set of training inputs and target pairs, the synaptic weights changed and updated, leading to accurate rainfall forecasting with minimal error. The structural design of the proposed five-input-based improved backpropagation neural network for rainfall forecasting is illustrated in Figure 4.

The rainfall is influenced by various variables such as temperature, precipitation of water content, wind speed, and relative humidity. Hence, the proposed neural networks consider these variables as the neural network’s inputs to overcome the variance in the atmosphere. The proposed improved backpropagation neural network learning algorithm includes a momentum factor, making the neural network get faster convergence.

 = (Rainfall, Precipitation of Water Content, Temperature, Relative Humidity, Wind Speed: Forecasted Rainfall).where is the forecasted rainfall.

The proposed IBPNN input vector:

The proposed IBPNN output vector:

The proposed IBPNN synaptic weight vectors from the input layer to the hidden layer:

The proposed IBPNN net input of the hidden layer:

The proposed IBPNN output of the hidden layer:where is the input of IBPNN, is the synaptic weights between the input layer and hidden layer, and is the number of hidden neurons.

The proposed IBPNN synaptic weight vectors from the hidden layer to the output layer:

The proposed IBPNN net input of the output layer:

The proposed IBPNN output:where is the synaptic weight between the hidden layer and the output layer and is the activation function.

The proposed IBPNN computed error in the output layer:where is the derivative of the net input of the output layer.

The evaluated error (E) is propagated back to the hidden layer.

In the proposed IBPNN, each hidden neuron sums its delta inputs from the output layer neurons:

The proposed IBPNN error in the hidden layer:where is the derivative of the net input of the hidden layer.

The computed error is propagated back to the input layer to minimize the error during the backpropagation stage:

The proposed improved backpropagation neural network mathematical equations for the synaptic weight updating process were as follows:where is the learning rate and is the momentum factor.

The synaptic weights are updated and changed using the mathematical equations (21) and (22). The learning stages of the proposed neural network and weight updating process are continued in the proposed IBPNN until attaining the stopping condition (i.e., set value).

4.3. Proposed Elman Neural Network for Temperature Forecasting

The feedback neural network has advantages over the feedforward network; with feedback from the output, the neural network stability and performance can be improved. In a feedback neural network, Elman neural network (ENN) is a famous feedback neural network, which Elman suggested in 1990 [2931]. Because of the superior performance, Elman network can be used for various applications such as forecasting, speech recognition, modeling, and control. Like the feedforward neural network arranged into the input layer, hidden layer, and output layer, one more layer is also added into the feedback neural network: the feedback layer or recurrent layer. The feedback storage and memory retaining have been done with the help of the recurrent layer. It also copies the one-step delay in the hidden layer. With the help of the internal connection proposed, Elman neural network dynamic characteristics are achieved. The proposed Elman neural network-based forecasting model hidden layer is associated with the hyperbolic tangent sigmoid activation function, and the output layer is associated with the purelin activation function.

The proposed temperature forecasting model is developed using an Elman neural network with five inputs such as Temperature , Dew Point , Solar Irradiance , Wind Speed , and Relative Humidity . The proposed Elman neural network complexity is reduced to a single hidden layer with various hidden neurons and one output layer with a single output neuron, i.e., forecast temperature. The proposed Elman neural network’s objective is to achieve the accurate forecasting of temperature with reduced convergence time and minimal error. The structural design of the Elman neural network is shown in Figure 5.

During the training process, the involved computations are given as follows:

 = (Temperature, Dew Point, Solar Irradiance, Wind Speed, Relative Humidity: Forecast Temperature).

, where is the Forecast Temperature in Degrees.

Let be the synaptic weights between the context layer and the input layer.

Let be the synaptic weights between the input layer and the hidden layer.

Let be the synaptic weights between the hidden layer and the recurrent link layer.

Let be the synaptic weights between the hidden layer and the output layer.

Let be the activation function, namely, the hyperbolic tangent sigmoid activation function adopted for the hidden layer.

Let be the activation function, namely, the purelin activation function which is adopted for the output layer.

Figure 5 infers that the proposed Elman neural network comprising each layer is performing independent computations on receiving data. The obtained output is passed to the next successive layer, and after that, finally, from the output layer, the neural network output is computed. The proposed Elman neural network has the ability that previous input influences the current input responses. The input passes through the hidden layer that multiplies the synaptic weight with the association of the hyperbolic tangent sigmoid activation function. In addition to the previous state output , the current input was also added, which aids the proposed feedback neural network to efficiently learn the function. The value is passed through an output layer multiplied by the synaptic weight with the association of the purelin activation function.

The proposed ENN input vector:

The proposed ENN output vector:

The proposed ENN synaptic weight vector between inputs to the hidden vector:

The proposed ENN synaptic weight vector of the recurrent link layer vector:

The proposed ENN synaptic weight vector of the output layer vector:

The proposed ENN Synaptic weight vector of the recurrent link layer with the input vector:

The proposed ENN input:

The proposed ENN output:

The proposed ENN input of the recurrent link layer:

4.4. Proposed Artificial Neural Network Experimental Implementation

In the current year, forecasting is a vital tool because people can plan according to it. Solar irradiance forecasting is a crucial factor in the solar energy system. Based on the forecasted solar irradiance, solar energy can be estimated to operate better than the power system and the possibility of extracting maximum solar power. It carried out many research works in solar irradiance forecasting [3234]. Although a more accurate solar irradiance forecasting model is still needed, it motivates the authors to develop multilayer perceptron neural network-based solar irradiance forecasting. The meteorological center tries to forecast the precise prediction of weather reports (Temperature) and rainfall. The predicted output does not match the current values because these values are highly influenced by the atmosphere variables like wind speed, cloud cover, precipitation of water content, etc. Some researchers endeavor to forecast the rainfall [3537] and temperature [3840] because of the volatile nature outperforming the generic forecasting model required. This fact motivates the development of the high accurate forecasting model for rainfall and temperature forecasting based on the proposed feedforward neural network (i.e., improved backpropagation neural network) and feedback neural network (i.e., Elman neural network), respectively.

The design and development stage, training stage, and testing stage are the three stages of the proposed artificial neural networks. The input-related data for the chosen applications are gathered, normalized into the range of zero to one through min-max normalization (variance eliminated), and then the training and testing process of the designed neural network is carried out with the acquired training and testing data sets, respectively. The validity of the proposed neural network-based forecasting model has been proven and analyzed based on the computed error validation process. It requires the proper design in neural network modeling because the improper selection of design parameters leads to poor performance.

4.4.1. Data Collection

For the considered solar irradiance forecasting applications, rainfall forecasting and temperature-related real-time input parameters are acquired from the National Oceanic and Atmospheric Administration, United States.

For the considered solar irradiance forecasting, the related inputs (Solar Irradiance , Temperature , Wind Speed , Dew Point , and Cloud Cover ) data samples are acquired from the period of January 2014 to December 2019, which comprise 175200 total number of data samples of each input.

For the considered rainfall forecasting, the related inputs (Rainfall , Precipitation of Water Content , Temperature , Relative Humidity , and Wind Speed ) are acquired from the period of April 2014 to April 2019, which consist of 52560 data samples of each input.

For the considered temperature forecasting application, the related input parameters (Temperature , Dew Point , Solar Irradiance , Wind Speed , and Relative Humidity ) are acquired from the period of March 2012 to April 2019, which consists of 1051200 data samples of each considered input.

4.4.2. Data Normalization

The data normalization is required because the data collected from the resource center are real-time data that possess the variance with respect to various ranges and various units to remove the variance of the acquired real-time data. The data normalization process, collecting data irrespective of multiple ranges and different units, classifies the data in the range from 0 to 1. In data normalization, various methods are available for this proposed work; min-max normalization adopts the proposed artificial neural network-based forecasting model. The proposed artificial neural network’s numerical computation and accuracy can be improved by employing data normalization. The following transformational mathematical equation is used for the normalization of the real-time collected data.where is the collected real-time input data, is the minimal input data, is the maximum input data, is the minimal target value, and is the maximum target value.

4.4.3. Proposed Artificial Neural Network Modeling

The designed parameters of the proposed artificial neural networks are presented in Table 3. The proposed various artificial neural networks modeling parameter dimensions such as the number of input neurons, hidden neurons, output neurons, the number of epochs, learning rate, momentum factor, and the threshold value are tabulated in Table 3.

The implemented multilayer-perceptron-neural-network- (MLPNN-) based solar irradiance forecasting model inputs are passed to the hidden layer that is multiplied by synaptic weight (SV) with hyperbolic tangent sigmoid activation function. The hidden layer’s output is passed to the output layer that is multiplied by synaptic weight (SW) with the purelin activation function. We use the Levenberg-Marquardt training algorithm for the proposed MLPNN-based solar irradiance forecasting model training process.

The proposed improved-back-propagation-neural-network- (IBPNN-) based rainfall forecasting model inputs are transmitted to the hidden layer multiplied by the synaptic weight (SV) utilizing a hyperbolic tangent sigmoid activation function. The hidden layer’s output is transmitted to the output layer multiplied by the synaptic weight (SW) with the tangential sigmoid activation function. The training algorithm used for IBPNN is Levenberg- Marquardt backpropagation training algorithm. The momentum factor is included in the learning algorithm, which leads to speed-up convergence.

The designed Elman-neural-network- (ENN-) based temperature forecasting model and the synaptic input weights (SV) are interconnected to the hidden layer using the hyperbolic tangent sigmoid activation function. The hidden layer’s output is interconnected to the output layer with synaptic weight (SW) using the purelin activation function. As a result of training, the previous inputs get reflected in the Elman neural network. The training algorithm used for the proposed Elman neural network is gradient descent with momentum and an adaptive linear backpropagation training algorithm.

For all proposed artificial-neural-network-based forecasting models, training and testing are done through the normalized data set. The validation process is continued until the stopping condition is reached.

4.4.4. Selection of Number of Hidden Neurons in the Proposed Artificial Neural Networks

The most challenging process in the artificial neural network is selecting the required number of hidden neurons to place in the artificial neural network [4145]. There is no generalized formulation and criterion available for selecting hidden neurons in an artificial neural network. The random selection and trial-and-error methods also take much more time. If the hidden neurons are too low and too high, both condition neural networks do not achieve optimal results. Hence, the proposed artificial neural network, namely, multilayer neural network, improved backpropagation neural network, and Elman neural network with a single hidden layer, is preferred because the neural network with a single hidden layer can solve the problem with less computational difficulty. In that single hidden layer, the hidden neurons are varied from one (1) to fifteen (15). The designed neural network is validated for each considered hidden neuron, and the obtained results are tabulated. According to Tables 46, appropriated numbers of hidden neurons are identified for the proposed artificial neural network based on the minimal error and minimal convergence time.

4.4.5. Training and Testing of the Proposed Artificial Neural Network Performance

The proposed solar irradiance forecasting model is built using the training data set. The proposed MLPNN-based solar forecasting model performance is verified based on the testing data set. The acquired data are classified into two sets, like training and testing. The training set comprises 70 percentages of the obtained data samples, and the testing set comprises the remaining unseen 30 percentages of acquired data samples. For the solar irradiance forecasting, the acquired data samples are 175200 real-time data samples, data samples (122640) are used for the training stage, and the unseen data samples (52560) are used for the testing stage of the proposed neural network.

The proposed rainfall forecasting model is built on the training data set, and the proposed IBPNN based rainfall forecasting model performance is verified based on the testing data set. The acquired data are classified into two sets, training and testing, respectively. The training set comprises 70 percentages of the obtained data samples, and the testing set consists of the remaining unseen 30 percentages of obtained data samples. For rainfall forecasting, the acquired data samples are 52560 real-time data samples, data samples (36792) are used for the training stage, and the unseen data samples (15768) are used for the proposed testing stage neural network.

The proposed temperature forecasting model is built on the training data set, and the proposed ENN-based temperature forecasting model performance is verified based on the testing data set. The acquired data are classified into two sets, like training and testing. The training set consists of 70 percentages of the obtained data samples, and the testing set consists of the remaining unseen 30 percentages of acquired data samples. For temperature forecasting, the acquired data samples are 1051200 real-time data samples, of data samples (735840) are used for the training, and of the unseen data samples (315360) are used for the testing stage of the proposed neural network.

The proposed artificial neural network performance is validated based on the training and testing set. The error qualifiers like R, MAPE, MSE, MAE, RMSE, MRE, and Time are used to verify the proposed artificial neural network performance. The number of hidden neurons that leads to reduced error is fixed as the optimal number of hidden neurons in the proposed artificial neural network.

4.4.6. Error Qualifier of the Proposed Artificial Neural Networks

The performance of the proposed various artificial neural networks such as multilayer perceptron neural network, improved backpropagation neural network, and Elman neural network is verified by the evaluated error qualifiers, namely, Correlation Coefficient (R), Mean Absolute Percentage Error (MAPE), Mean Square Error (MSE), Mean Absolute Error (MAE), Root Mean Square Error (RMSE), Mean Relative Error (MRE), and Time in Minutes. The proposed artificial neural network effectiveness is evaluated based on the error qualifiers equations (33)–(38). The following mathematical equations are used for the error computation:where N is total number of data samples, is target output, is average target output, and is forecasted output.

4.5. Validation Experimental Results of the Proposed Artificial Neural Network for Various Forecasting Applications

The proposed artificial-neural-networks-based forecasting models are experimentally simulated in MATLAB platform version 2013 running on Acer computers with Pentium (R) Dual-Core processor running at 2.30 GHz with 2 GB of RAM.

4.5.1. Experimental Results of the Proposed Multilayer-Perceptron-Neural-Network-Based Solar Irradiance Forecasting

The designed multilayer-perceptron-neural-network-based solar irradiance forecasting model performance is validated experimentally, and the got results with various numbers of hidden neurons (1–15) are tabulated in Table 4 and analyzed. From the analysis of Table 4, it is observed that the proposed multilayer-perceptron-neural-network-based solar irradiance forecasting model performs better for various hidden neurons.

The proposed forecasting model output is changed drastically by varying the hidden neuron in the proposed multilayer perceptron neural network’s hidden layer. The hidden neuron in the hidden layer increases further and makes the neural network unstable, while decreasing the hidden neuron further also leads the neural network to become unstable. Based on the achieved solar irradiance forecasting results among 1 to 15 hidden neurons, the proposed multilayer perceptron neural network contains a single hidden layer that possesses 14 hidden neurons that provide the best outputs with minimal error and reduced convergence time.

Therefore, the proposed multilayer perceptron neural network with five inputs, a single hidden layer, 14 hidden neurons in the hidden layer, and a single output neuron structure is identified as the best framework. The obtained output plots based on this structural framework-associated forecasting model are shown in Figure 6. The number of data vs. solar irradiance is presented in Figure 7. The real-time target solar irradiance compared with forecasted solar irradiance is illustrated in Figure 8. Error vs. number of data is shown in Figure 9. Relationship between forecast solar irradiance and real-time target solar irradiance, respectively. Because of the space limitation, portions of the obtained results are shown in Figures 69. The 14 hidden neuron-based developed neural network (MLPNN) models forecasting solar irradiance are much matched with the real-time target solar irradiance. Hence, the error values are reduced to the minimal, clearly understood from Figures 7 and 8, respectively.

Figure 8 shows that the proposed neural network with 14 hidden neurons results in a minimal error on the considered data samples. The relationship between forecasted solar irradiance and real-time target solar irradiance is a linear relationship that illustrates that the forecasted solar irradiance accurately matches the real-time target. Hence, the proposed neural network (MLPNN) proved its validity, which is noticed in Figure 9.

4.5.2. Experimental Results of Proposed Improved Backpropagation Neural Network-Based Rainfall Forecasting

The proposed improved backpropagation neural network-based rainfall forecasting model performance is validated experimentally using the collected data, and the obtained results of the proposed neural network with various numbers of hidden neurons (1–15) are tabulated in Table 5. According to the obtained results in Table 5, it is noticed by careful analyses that the proposed rainfall forecasting model based on the improved backpropagation neural network performs very poorly for the hidden neurons 1 and 2 except that for the remaining hidden neurons results in the good output. The IBPNN-based forecasting model achieved outcome is changed drastically with the hidden neurons changes in the proposed improved backpropagation neural network’s hidden layer.

In a neural network, the neural network stability and performance are highly affected by hidden neurons. Based on the analysis of the obtained results, it is observed that the proposed improved backpropagation neural network with a single hidden layer and seven hidden neurons in the hidden-layer-based design neural network (IBPNN) achieves superior performance in terms of reduced error and reduced convergence speed compared among other hidden neurons based on design neural networks.

Therefore, the proposed improved backpropagation neural network with five inputs, a single hidden layer, seven hidden neurons in the hidden layer, and a single output layer with one output neuron has been identified as the optimal structural framework of the proposed neural network. The obtained rainfall forecasting plots with seven hidden neurons-based design IBPNN are depicted in Figure 10. Rainfall in mm vs. data samples is shown in Figure 11. Original target rainfall compared with forecast rainfall is illustrated in Figure 12. Evaluation error metric vs. the number of data samples is presented in Figure 13, the relationship between forecast rainfall and original target rainfall, respectively. Because of the space limitation, portions of the obtained results are shown in Figures 1013.

The considered data sample for the validation of the neural network is represented in Figure 10. In the seven-hidden-neurons-based developed backpropagation neural network model, where the forecast rainfall is a relative value compared with the original target rainfall, the forecasting accuracy is better which is understood from Figure 11. It is observed that the designed IBPNN achieves minimal errors for more clarity; the evaluation error with respect to the data samples is depicted in Figure 12. The relationship between forecast rainfall and original target rainfall is a linear relationship, which illustrates that the forecast rainfall is higher, accurately matched with the original target, which is clearly observed in Figure 13.

4.5.3. Experimental Results of the Proposed Elman-Neural-Network-Based Temperature Forecasting

The proposed temperature forecasting models based on Elman neural network validated with the acquired data samples and achieved experimental outputs based on different numbers of hidden neurons from 1 to 15 are tabulated in Table 6. According to the achieved results in Table 6, it is observed that the proposed Elman-neural-network-based forecasting model performs well for all hidden neurons. The proposed Elman-neural-network- (ENN-) based temperature forecasting model output is changed drastically due to the number of hidden neurons varying in the proposed Elman neural network’s hidden layer.

In a feedback neural network, hidden neurons profoundly influence the aspects of neural network stability and performance convergence. From the result analysis of the obtained outputs, it is noticed that the designed Elman neural network with a single hidden layer and 11 hidden neurons in the hidden layer achieves better performance in terms of reduced error and reduced convergence speed compared among other numbers of hidden neurons based on the designed neural networks. Therefore, the proposed temperature forecasting model based on Elman neural network with five inputs, a single hidden layer, 11 hidden neurons in the hidden layer, and a single output layer with one output neuron has been identified as the useful structural framework of the proposed neural network.

The obtained temperature forecasting plots with respect to the 11 hidden neuron-based design ENNs are shown in Figure 14. Temperature vs. data samples is presented in Figure 15. Comparison between target and forecasted temperature is presented in Figure 16. Error vs. number of data is shown in Figure 17, the relationship between target and forecast temperature, respectively. Due to the space limitation, portions of the obtained results are shown in Figures 1417.

The 11 hidden neurons associated with the single hidden layer Elman neural network forecast temperature match the target values. Hence, the error values are the least; it is clearly understood from Figures 15 and 16, respectively. The relationship between forecast temperature and the target temperature is linear; it is noticed from Figure 17.

5. Conclusion

Nowadays, human expectations and needs are increasing widely. All are interested in artificial intelligence to make their work easy and effective. This paper discusses the history of artificial neural networks, the generation of artificial neural networks, the generalized process involved in artificial neural networks, the various types, structural design, and artificial neural network applications that are elucidated in a detailed manner. The artificial neural network can address multiple applications, but this paper forecasting application is considered for performance analysis.

The highlights of the differences between the proposed models and the existing ones are as follows:(i)The proposed models were developed with five years, five years, and seven years’ data sets, respectively, for solar irradiance, rainfall, and temperature forecasting applications. Thus, we overcome the interannual variability-based uncertainty.(ii)The proposed forecasting models possess minimal design complexity, are feasible to implement, and result in minor error qualifiers.

This paper proposed two feedforward neural networks, such as multilayer perceptron neural network (MLPNN) and improved backpropagation neural network (IBPNN), which could be used for forecasting applications like solar irradiance and rainfall forecasting, respectively. Moreover, the proposed one-feedback neural network, such as Elman neural network (ENN), can be used for the temperature forecasting application. The proposed artificial neural networks (MLPNN, IBPNN, and ENN) performances are statistically analyzed with various hidden neurons. The designed neural network–based forecasting model effectiveness is successfully validated using the acquired real-time training and test data set. Error qualifiers are used to analyze the performance of the proposed neural networks. According to the obtained results from the proposed artificial neural network–based forecasting model, it is observed that the proposed neural network–based forecasting model outperforms in all considered applications with much minimal error and reduced convergence time. The proposed multilayer perceptron neural network, which comprises 5 inputs, single hidden layer, and 14 hidden neurons achieves the minimal errors like R = 1, MAPE = 2.3034e-04, MSE = 5.9791e-07, MAE = 5.0093e-04, RMSE = 7.7324e-04, MRE = 2.3034e-06, and Time = 1.32 minutes for solar irradiance forecasting application. The suggested improved backpropagation neural network, which comprises five inputs, a single hidden layer, and seven hidden neurons, achieves minimal errors like R = 1, MAPE = 0.0158, MSE = 0.0042, MAE = 0.0320, RMSE = 0.0645, MRE = 1.5841e-04, and Time = 1.24 minutes for rainfall forecasting application. Similarly, the proposed Elman neural network, which comprises five inputs, a single hidden layer, and 11 hidden neurons, achieves minimal errors like MAPE = 0.1023, MSE = 0.0011, MAE = 0.0230, RMSE = 0.0332, MRE = 0.0010, and Time = 22 sec for temperature forecasting application. Hence, the proposed neural network–based forecasting models proved their validity, and they assist sustainability.

5.1. Proposed Forecasting Model Limitation and Future Work

To efficiently handle big data is one limitation of the proposed forecasting model. Thus, the authors can implement improved intelligent model-based forecasting in future work.

5.2. Recommendation of Future Direction and Research Scope

The building blocks of ANNs are neurons, linkages with weighted connection, activation function, and learning algorithms. Still, many research works focused on neural network performance improvement. The appropriated architecture selection was lacking in the field of artificial neural networks. There are no general guidelines available for the architecture framework. Artificial neural networks can effectively handle nonlinearity but obtain a feasible solution that is not generic. It can be overcome by the optimization algorithm associated with an artificial neural network (hybrid model). The readers can focus their research attention on the deep learning artificial neural network. However, ANN provided a promising result. Still, it has inefficiencies in some other applications like smart grid, natural language processing, speech recognition, computer vision, and so on, which lead to the quest to identify the optimal modeling of ANN.

The barrier to Growth in Artificial Neural Networks:(i)Unique and privacy rights of the human being lost by an artificial neural network(ii)Scarcity of job opportunities(iii)Possibility to endanger humans and the environment

Artificial neural networks have many unique features and advantages; meanwhile, it has some barriers as well. Therefore, the advent of science and advancement should be healthy to benefit society, improve the economy and sustainability, and safeguard the environment and other living things.

Data Availability

We derived these data sets from the following domain resources https://www.noaa.gov/ upon request links to access their ordered data from an FTP site such as third-party right authors who did not provide the data openly.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The authors convey heartfelt thanks to the National Oceanic and Atmospheric Administration, the United States, for supporting real-time acquired data sample provision. Therefore, the authors validated the proposed research work successfully.