#### Abstract

We investigate the inverse problem in the nonhomogeneous heat equation involving the recovery of the initial temperature from measurements of the final temperature. This problem is known as the backward heat problem and is severely ill-posed. We show that this problem can be converted into the first Fredholm integral equation, and an algorithm of inversion is given using Tikhonov's regularization method. The genetic algorithm for obtaining the regularization parameter is presented. We also present numerical computations that verify the accuracy of our approximation.

#### 1. Introduction

Inverse heat conduction problems (IHCPs) appear in many important scientific and technological fields . Hence analysis, design implementation, and testing of inverse algorithms are also great scientific and technological interest. In this paper we will examine some inversion techniques in order to estimate the initial temperature distribution of an inverse nonhomogeneous heat conduction problem; this problem is called the backward heat problem, the backward Cauchy problem, or the final value problem. As is known, the nonhomogeneous problem is severely ill-posed; that is, solutions do not always exist, and in the case of existence, these do not depend continuously on the given data. In fact, from small noise contaminated physical measurements, the corresponding solutions have large errors. It makes it difficult to do numerical calculations. Hence, a regularization is in order. Lattes and Lions, in , regularized the problem by adding a “corrector” to the main equation. Gajewski and Zacharias  considered a similar problem. Nakamura et al.  used transformation techniques to solve the initial inverse problem in heat conduction and Al-Khalidy  dealt with the problem numerically. For a comprehensive review of the literature and summary of various approaches in the field of inverse heat conduction problems, one can consult the books by Beck et al.  and by Hensel . Although there are many papers on the linear homogeneous case of the backward problem, we only find a few papers on the nonhomogeneous case, such as in [13, 14].

The plan of this paper is as follows. In Section 2, we formulate a one-dimensional IHCP. In Section 3, we show that this problem can be converted into the first Fredholm integral equation, the numerical algorithm is derived, and an algorithm of inversion is also given using Tikhonov's regularization method. We use the genetic algorithm for the choice of the regularization parameter in Section 4. Section 5 presents numerical experiments to show the efficiency of the proposed method. Section 6 ends this paper with a brief discussion on some numerical aspects.

#### 2. Mathematical Model

##### 2.1. The Direct Problem

The direct (forward) problem consists of a transient heat conduction problem in a slab with adiabatic boundary condition and initially at a temperature denoted by .

The mathematical formulation of this problem is given by the following nonhomogeneous heat equation: where (temperature), (source term), (initial condition), (spatial variable), and (time variable) are dimensionless quantities and denotes the dispersion coefficient.

For the direct problem where the initial condition is specified, the problem given by (1) and (2) is concerned with the determination of the temperature distribution in the interior region of the solid as a function of time and position.

##### 2.2. Inverse Problem

For the inverse problem, the initial condition is regarded as being unknown. In addition, an overspecified condition is also considered available. To estimate the unknown coefficient , the additional information, is given at time , over a specified space interval . We note that the measured overspecified condition should contain measurement errors. Therefore the inverse problem can be stated as follows: estimate the unknown function by utilizing the above-mentioned measured data.

#### 3. Algorithm Analysis

The solution of the direct problem for a given initial condition is explicitly obtained using separation of variables for . Consider the initial boundary problems (1) and (2); we look for solutions in the form . We look at the following eigenvalue problem: The eigenvalues and the corresponding eigenfunctions are Now we set Formally computing and and substituting into (1), we get Hence expanding and into the following Fourier series: where we get The uniqueness of the Fourier expansion leads to the family of ODE's: In addition, so that Solving (11)–(13), we find where . is an infinite series; numerically we cannot handle infinite sums. Limit the sum to a finite number of expansion terms 200 which guarantees the convergence of the series. So Thus initial inverse problem is reduced to solving integral equation of the first kind: where where . Since is a double integral, it is difficult to determine the exact value. The authors try to seek its approximation by complex trapezoid formula.

Given that the interval is equally subdivided into subintervals, , with equally spaced sampling points where . Also, assume that the interval is equally subdivided into subintervals, , with equally spaced sampling points where . The composite Trapezoidal rule is Approximation of can be obtained by the above formula.

The first step in the numerical treatment used in this research consists in discretization of (16) by the quadrature formula. The interval can be subdivided into equal intervals of width . Let and , and since the variable is either or , let , , and (i.e., ). Also denote as , as , and as . Now if the trapezoid rule is used to approximate the given equation, then or, more tersely, There are values of , as . Therefore the equation becomes a set of equations in : which give the approximate solution to at . This may also be written in matrix form: where is the matrix of coefficients,

is the matrix of solutions, and is the matrix of the nonhomogeneous part:

The problem (22) is ill-posed in the sense that the inverse operator of exists but is not continuous. Hence, although the problem (22) has a unique solution, solving it directly will not give a right solution. Indeed, the linear operator is so badly conditioned that any numerical attempt to directly solve (22) may fail.

In this note we describe the Tikhonov regularization for finding a stable approximate solution to a linear ill-posed problem represented in the form of an operator equation: where, instead of the exact data , noisy data is available with In order to find a solution in stable manner, Tikhonov proposed to solve the following: where is the regularization parameter. The computation of the approximate solution consists in solving the Euler equation corresponding to the functional . This equation has the following form: where is the adjoint operator of and is the identity operator. The regularization solution obviously depends on the parameter , the value of which directly affects the degree of approximation and the stability of the solution. From the viewpoint of the approximation degree, the smaller the better for the value of ; however, from the viewpoint of the solution stability, the larger the better. So the key point for solving the problem is getting the value of .

In order to obtain an effectively approximate solution to the original ill-posed problem, it is particularly important to choose suitable parameter. If we do not have the information of  , we have to use an “a posteriori” parameter choice, where one uses results of the actual computation like the residual for computing a regularization parameter. The most widely used method is Morozov’s discrepancy principle, where one takes such that

On how to solve the equation , some researchers have developed some theories and methods to determine [13, 1519]. In the present paper, the genetic algorithm for obtaining the regularization parameter is presented.

#### 4. The Genetic Algorithm

Genetic algorithms belong to the general category of stochastic global optimization methods . They have their philosophical basis in a process found in nature related to the evolution of the different biological species. Based on Darwin's theory of evolution where the fittest individuals survive, while the less fit are eliminated, a simple genetic algorithm is basically composed of three operators: (i) selection, (ii) crossover, and (iii) mutation. The computational implementation of a genetic algorithm starts with random generation of a population of individuals where each individual is actually a chain composed of representations of binary numbers, sequentially, random values for design variables. Each individual in the population corresponds to a value of the objective function (fitness function), which is associated with a function of adaptability that is minimized.

With the fitness function values calculated for each individual in the population, the probability that this individual will be selected for crossover is determined. This probability is given by where   is the fitness function, defined in (31), and is the population size. The fittest individuals have higher probability of being selected for crossover. After selecting a pair of individuals, crossover is done by switching between the parts of the binary strings of these individuals. The population size remains constant through all the generations. This exchange-digit binary string corresponds in nature to the exchange of genetic information between parents, which represents the process by which it promotes the improvement of the species. The crossover for each pair of selected individuals is also subject to a probability to be performed. Mutation is a safeguard against losses of useful genetic material during reproduction and crossover. The process of mutation is simply to choose a few members from the population based on a probability of mutation and switch a 0 to a 1 and vice versa at a randomly selected site on the selected member. The process consisting of selection, crossover, and mutation is then repeated by the total number of generations, and at the end the most fitted individual is selected, that is, the one that leads to lower value of the objective function (31). The specific implementation process is shown in Figure 1.

#### 5. Numerical Examples

In this section, we are going to demonstrate some numerical results for determining in the inverse problems (1)–(3). We use the error norm and the relative error to measure the difference between the numerical and analytical solutions. The error norm is defined by and the relative error (RE) is defined by where are test points and is the total number of uniformly distributed points on . is the exact solution and is the numerical solution.

In our computations, we take . The noisy data were assumed to contain some random errors. Assume the observed data has the following noised form: where , “randn” is a normally distributed random variable with zero mean and unit standard deviation, and dictates the level of noise. “” returns an array of random entries that has the same size as .

Example 1. In this example let us consider the following inverse problem: The overspecified condition is as follows: The analytical solution of this example is and , .

The regularization parameter is chosen using the genetic algorithm. In the genetic algorithm implemented, a tournament selection scheme by two individuals was used. A uniform crossover operator is applied with a probability of 0.95 and a uniform mutation operator where each individual has a probability 0.05 of being mutated. Population size is 20, number of generations is 100. The search domain is for . The objective function , the error norm, and relative error RE are presented in Tables 1, 2, 3, and 4 corresponding to , and 1.2, respectively. Especially when , the corresponding errors between the analytical values and the estimated results by function are listed in Table 5   and Table 6  .

From this numerical example, it can be seen that the numerical results are quite satisfactory. As shown in Tables 16, when the noise level of input data with , and 1.2, the numerical solutions obtained using our method are of high degree of accuracy. Even with the noise level of input data up to , the numerical solutions are still in good agreement with the exact solutions.

#### 6. Conclusion

This paper deals with the effective algorithms for solving the backward heat problem and the following results are obtained.(1)The present study successfully applies the numerical method involving Tikhonov's regularization method in conjunction with the first Fredholm integral equation to the inverse nonhomogeneous heat conduction problems.(2)From the illustrated example, it can be seen that the proposed numerical method is efficient and accurate to estimate the initial condition .

#### Conflict of Interests

The authors declare that there is no conflict of interests regarding the publication of this paper.

#### Acknowledgments

The work of the authors is supported by the Special Funds of the National Natural Science Foundation of China (nos. 51190093 and 51179151). The authors would like to thank the referees for constructive suggestions and comments.