Research Article  Open Access
Soumya Pandit, Chittaranjan Mandal, Amit Patra, "A Methodology for Generation of Performance Models for the Sizing of Analog HighLevel Topologies", VLSI Design, vol. 2011, Article ID 475952, 17 pages, 2011. https://doi.org/10.1155/2011/475952
A Methodology for Generation of Performance Models for the Sizing of Analog HighLevel Topologies
Abstract
This paper presents a systematic methodology for the generation of highlevel performance models for analog component blocks. The transistor sizes of the circuitlevel implementations of the component blocks along with a set of geometry constraints applied over them define the sample space. A Halton sequence generator is used as a sampling algorithm. Performance data are generated by simulating each sampled circuit configuration through SPICE. Least squares support vector machine (LSSVM) is used as a regression function. Optimal values of the model hyper parameters are determined through a grid searchbased technique and a genetic algorithm (GA) based technique. The highlevel models of the individual component blocks are combined analytically to construct the highlevel model of a complete system. The constructed performance models have been used to implement a GAbased highlevel topology sizing process. The advantages of the present methodology are that the constructed models are accurate with respect to real circuitlevel simulation results, fast to evaluate, and have a good generalization ability. In addition, the model construction time is low and the construction process does not require any detailed knowledge of circuit design. The entire methodology has been demonstrated with a set of numerical results.
1. Introduction
An analog highlevel design process is defined as the translation of analog systemlevel specifications into a proper topology of component blocks, in which the specifications of all the component blocks are completely determined so that the overall system meets its desired specifications optimally [1โ3]. The two important steps of an analog highlevel design procedure are highlevel topology generation/selection [4, 5] and highlevel specification translation [6]. At the highlevel design abstraction, a topology is defined as an interconnection of several analog component blocks such as amplifier, mixer and filter. The detailed circuitlevel implementations of these component blocks are not specified at this level of abstraction. The analog component blocks are represented by their highlevel models.
During the past two decades, many optimizationbased approaches have been proposed to handle the task of topology generation/selection [7โ11]. These approaches involves the task of topology sizing, where the specification parameters of all the component blocks of a topology are determined such that the desired system specifications are optimally satisfied. The two important modules for this type of design methodology are a performance estimation module and an optimization engine. The implementation of the design methodology is based upon the flow of information between these two modules.
The performance models that are used in the highlevel design abstraction are referred to as highlevel performance models. An analog highlevel performance model is a function that estimates the performance of an analog component block when some highlevel design parameters of the block are given as inputs [12, 13]. The important requirements for a good highlevel performance model are as follows. (i) The model needs to be low dimensional. (ii) The predicted results need to be accurate. The model accuracy is measured as the deviation of the model predicted value from the true function value. The function value in this case is the performance parameter obtained from transistor level simulation [12]. (iii) The evaluation time must be short. This is measured by the CPU time required to evaluate a model. (iv) The time required to construct an accurate model must be small, so that the design overhead does not become high. As a rough estimate, the construction cost is measured as where the terms are selfexplanatory. There exists a tradeoff between these requirements since a model with lower prediction error generally takes more time for construction and evaluation.
In this work, we have developed the performance models using least squares support vector machine (LSSVM) as the regressor. The transistor sizes of the circuitlevel implementations of the component blocks along with a set of geometry constraints applied over them define the sample space. Performance data are generated by simulating each sampled circuit configuration through SPICE. The LSSVM hyper parameters are determined through formal optimizationbased techniques. The constructed performance models have been used to implement a highlevel topology sizing process. The advantages of this methodology are that the constructed models are accurate with respect to real circuitlevel simulation results, fast to evaluate and have a good generalization ability. In addition, the model construction time is low and the construction process does not require any detailed knowledge of circuit design. The entire methodology has been demonstrated with a set of experimental results.
The rest of the paper is organized as follows. Section 2 reviews some related works. Section 3 presents the background concepts on least squares support vector machines. An outline of the methodology is provided in Section 4. The model generation methodology is described in detail in Section 5. The topology sizing process is described in Section 6. Numerical results are provided in Section 7 and finally conclusion is drawn in Section 8.
2. Related Work
A fairly complete survey of related works is given in [14]. An analog performance estimation (APE) tool for highlevel synthesis of analog integrated circuits is described in [15, 16]. It takes the design parameters (e.g., transistor sizes, biasing) of an analog circuit as inputs and determines its performance parameters (e.g., power consumption, thermal noise) along with anticipated sizes of all the circuit elements. The estimator is fast to evaluate but the accuracy of the estimated results with respect to real circuitlevel simulation results is not good. This is because the performance equations are based on simplified MOS models (SPICE level 1 equations). A power estimation model for ADC using empirical formulae is described in [13]. Although this is fast, the accuracy with respect to real simulation results under all conditions is off by orders of magnitude. The technique for generation of posynomial equationbased performance estimation models for analog circuits like opamps, multistage amplifiers, switch capacitor filters, and so forth, is described in [17, 18]. An important advantage of such a modeling approach is that the topology sizing process can be formulated as a geometric program, which is easy to solve through very fast techniques. However, there are several limitations of this technique. The derivation of performance equations is often a manual process, based on simple MOS equations. In addition, although many analog circuit characteristics can be cast in posynomial format, this is not true for all characteristics. For such characteristics, often an approximate representation is used. An automatic procedure for generation of posynomial models using fitting technique is described in [19, 20]. This technique overcomes several limitations of the handcrafted posynomial modeling techniques. The models are built from a set of data obtained through SPICE simulations. Therefore, full accuracy of SPICE simulation is achieved through such performance models. A neural networkbased tool for automated power and area estimation is described in [21]. Circuit simulation results are used to train a neural network model, which is subsequently used as an estimator. Fairly recently, support vector machine (SVM) has been used for modeling of performance parameters for RF and analog circuits [22โ24]. In [25], SVMoptimized by GA has been used to develop a soft fault diagnosis method for analog circuits. In [26], GA and SVM has been used in conjunction for developing feasibility model which is then used within an evolutionary computationbased optimization framework for analog circuit optimization.
2.1. Comparison with Existing Methodologies
The present methodology uses nonparametric regression technique for constructing the highlevel performance models. Compared with the other modeling methodologies employing symbolic analysis technique or simulationbased technique, the advantages of the present methodology are as follows. (i) Full accuracy of SPICE simulations and advanced device models, such as BSIM3v3 are used to generate the performance models. The models are thus accurate compared to real circuitlevel simulation results. (ii) There is no need for any a priori knowledge about the unknown dependency between the inputs and the outputs of the models to be constructed. (iii) The generalization ability of the models is high. (iv) The model construction time is low and the construction process does not require any detailed circuit design knowledge.
The EsteMate methodology [21] using artificial neural network (ANN) and the SVMbased methodology discussed in [22, 23] are closely related with the present methodology. The methodology that we have developed, however, has a number of advantages over them. These are as follows. (1)In the EsteMate methodology, the specification parameters of a component block constitute the sample space for training data generation. The specification parameters are electrical parameters and there exists strong nonlinear correlations amongst them. Therefore, sophisticated sampling strategies are required for constructing models with good generalization ability in the EsteMate methodology. On the other hand, in our method, the transistor sizes along with a set of geometry constraints applied over them define the sample space. Within this sample space, the circuit performance behavior becomes weakly nonlinear. Thus simple sampling strategies are used in our methodology to construct models with good generalization ability.(2)In EsteMate, for each sample, a complete circuit sizing task using a global optimization algorithm is required for generation of the training data. This is usually prohibitively time consuming. On the other hand, in our method, simple circuit simulations using the sampled transistor sizes are required for data generation. Therefore, the cost of training data generation in our method is much less compared to that in the EsteMate methodology [21]. With the EsteMate methodology, the training sample points are so generated that performances such as power is optimized. On the other hand, in our methodology, the task of performance optimization has been considered as a separate issue, isolated from the performance model generation procedure. Our strategy is actually followed in all practical optimizationbased highlevel design procedures [1, 27].(3)The generalization ability of the models constructed with our methodology is better than that generated through the EsteMate methodology. This is because the latter uses ANN regression technique. Neural networkbased approaches suffer from difficulties with generalization, producing models that can overfit the data. This is a consequence of the optimization algorithms used for parameter selection and the statistical measures used to select the โbestโ model. SVM formulation, on the other hand, is based upon structural risk minimization (SRM) principle [28], which has been shown to be superior to traditional empirical risk minimization (ERM) principle, employed by the conventional neural networks. SRM minimizes an upper bound on the expected risk, as opposed to ERM that minimizes the error on the training data. Therefore an SVM has greater generalization capability.(4)The SVMbased methodology, as presented in [23], uses heuristic knowledge to determine the model hyper parameters. The present methodology uses optimization techniques to determine optimal values for them. GAbased methodology for determination of optimal values for the model hyper parameters is found to be faster compared to the grid search technique employed in [22].
3. Background: Least Squares Support Vector Regression
In recent years, the support vector machine (SVM), as a powerful new tool for data classification and function estimation, has been developed [28]. Suykens and Vandewalle [29] proposed a modified version of SVM called least squares SVM. In this subsection, we briefly outline the theory behind the LSSVM as function regressor.
Consider a given set of training samples where is the input value and is the corresponding target value for the th sample. With an SVR, the relationship between the input vector and the target vector is given as where is the mapping of the vector to some (probably highdimensional) feature space, is the bias, and is the weight vector of the same dimension as the feature space. The mapping is generally nonlinear which makes it possible to approximate nonlinear functions. The approximation error for the th sample is defined as The minimization of the error together with the regression is given as with equality constraint where denotes the total number of training datasets and the suffix denotes the index of the training set, that is, th training data, is the regularization parameter.
The optimization problem (4) is considered to be a constrained optimization problem and a Lagrange function is used to solve it. Instead of minimizing the primary objective (4), a dual objective, the socalled Lagrangian, is formed of which the saddle point is the optimum. The Lagrangian for this problem is given as where sโ are called the Lagrangian multipliers. The saddle point is found out by setting the derivatives equal to zero: By eliminating and through substitution, the final model is expressed as a weighted linear combination of the inner product between the training points and a new test object. The output is given as where is the kernel function. The elegance of using the kernel function lies in the fact that one can deal with feature spaces of arbitrary dimensionality without having to compute the map explicitly. Any function that satisfies Mercer's condition can be used as the kernel function. The Gaussian kernel function used in the present work is defined as and is commonly used, where denotes the kernel bandwidth. The two important parameters, kernel parameter , and the regulation parameter as defined in (4) are referred to as hyper parameters. The values of these parameters have to determined critically in order to make the network efficient.
4. An Outline of the Methodology
The highlevel performance model of an analog component block is mathematically represented as where is a set of performance parameters and is a set of specification parameters. The input specification parameters are referred to as the highlevel design parameters. It is to be noted that out of various possible specification parameters, only the dominant parameters are to be considered as inputs. The selection of these is based upon the designer's knowledge [12]. These highlevel design parameters describe a space referred to as the sample space. This sample space is explored to extract sample points through suitable algorithms. The numerical values of the sample points (both inputs and outputs of the performance model to be constructed) are generated through SPICE simulations. The data points so generated are divided into two sets, referred to as the training set and the test set. A least squares SVM network approximating a performance model is constructed by training the network with the training set. The test dataset is used to validate the SVM model. Suitable kernel functions are selected for constructing the SVM. An initial SVM model is constructed through some initial values of the hyper parameters. An iterative process is then executed to contruct the final LSSVM so as to maximize its efficiency through optimal determination of the hyper parameters. An outline of the process for constructing the performance model of a single component block is illustrated in Figure 1(a).
(a)
(b)
(c)
For a complex system, consisting of many component blocks, the highlevel performance model of the complete system is constructed at the second level of hierarchy, where the highlevel models of the individual component blocks are combined analytically (see Figure 1(b)). The constructed performance models are used to implement a highlevel topology sizing process. For a given unsized highlevel topology of an analog system, the topology parameters (which are the specification parameters of the individual blocks of the highlevel topology) are determined such that the desired design goals are satisfied. The entire operation is performed within an optimization procedure, which in the present work is implemented through GA. The constructed LSSVM models are used within the GA loop. An outline of the sizing methodology is shown in Figure 1(c).
The following two important points may be noted in connection with the present methodology. First, the highlevel performance model of a complete system is generated in a hierarchical manner. The major advantage of this hierarchical approach is reusability of the highlevel model of the individual component blocks. The highlevel model of the component blocks can be utilized whenever the corresponding component blocks are part of a system, provided the functionality and performance constraints are identical. This generally happens. The issue of reusability of the component block level highlevel models is demonstrated in Experiment 3, provided later. However, this advantage comes at the cost of reduced accuracy of the model of the complete system. This tradeoff is a general phenomenon in analog design automation process. It may, however, be noted that it is possible to construct the highlevel performance model of a complete system using the regression technique discussed here. For some customized applications, this may be done. Second, the requirement of low dimensionality of the models must be carefully taken care of. The scalability of our approach of model generation is not high, compared to analytical approach. However, compared to other blackbox approaches like ANNbased, the scalability of our SVMbased approach is high. In addition, many of the global optimization algorithms suffer from the problem of โcurse of dimensionality.โ For a topology sizing procedure, employing highdimensional model the design space in which to search for optimal design points becomes too large to be handled by simple optimization algorithms. Therefore, while selecting the inputs of the model, only the dominant specification parameters need to be considered.
The detailed operations of each of the steps outlined above are discussed in the following sections and subsections.
5. HighLevel Performance Model Generation
In this section, we describe the various steps of the performance model generation procedure in detail.
5.1. Sample Space Definition, Data Generation, and Scaling
In (10), both and are taken to be functions of a set of geometry parameters (transistor sizes) of a component block, expressed as and represents the mapping of the geometry parameters to electrical parameters. This is illustrated in Figure 2. The multidimensional space spanned by the elements of the set is defined as circuitlevel design space . The sample space is a subspace within (see Figure 3), defined through a set of geometry constraints. These geometry constraints include equality constraints as well as inequality constraints. For example, for matching purpose, the sizes of a differential pair transistors are equal. The inequality constraints are determined by the feature size of a technology and conditions that the transistors are not excessively large. With elementary algebraic transformations, all the geometry constraints are combined into a single nonlinear vector inequality, which is interpreted element wise as Within this sample space, the circuit performance behavior becomes weakly nonlinear [27, 30]. Therefore, simple sampling strategies are used to construct models with good generalization ability. In the present work, the sample points are extracted through Halton sequence generation. This is a quasirandom number generator which generates a set of uniformly distributed random points in the sample space [31]. This ensures a uniform and unbiased representation of the sample space. The number of sample data plays an important role in determining the efficiency of the constructed LSSVM model. Utilizing a separate algorithm, it is possible to determine an optimum size of the training sample data such that models built with smaller training set than this optimum value will have lower accuracy than the models built with optimum number of training sample and models built with larger training data than the optimum number will have no significant higher accuracy. However, in the present work, in order to make the sampling procedure simple, the number of sample data is fixed which is determined through a trial and error method.
The training data generation process is outlined in Figure 4. For each input sample (transistor sizes) extracted from the sample space , the chosen circuit topology of a component block is simulated using SPICE through Cadence Spectre tool using the BSIM3v3 model. Depending upon the selected inputoutput parameters of an estimation function, it is necessary to construct a set of test benches that would provide sufficient data to facilitate automatic extraction of these parameters via postprocessing of SPICE output files. A set of constraints, referred to as feasibility constraints are then applied over the generated data to ensure that only feasible data are taken for training.
The generated inputoutput data are considered to be feasible, if either they themselves satisfy a set of constraints or the mapping procedures through which they are generated satisfy a set of constraints. The constraints are as follows [30]. (1)Functionality constraints : these constraints are applied on the measured node voltages and currents. They ensure correct functionality of the circuit and are expressed as For example, the transistors of a differential pair must work in saturation. (2)Performance constraints : these are applied directly on the inputoutput parameters, depending upon an application system. These are expressed as For example, the phase margin of an opamp must be greater than 45ยฐ.
The total set of constraints for feasibility checking is thus . It is to be noted that through the process of feasibily checking, various simulation data are discarded. This at a glance may give an impression about wastage of costly simulation time. However, for an analog designer (who is a user of the model), this is an important advantage. This is because, the infeasible data points will never appear as solution whenever the model is used for design characterization/optimization. Even from the model developer's perspective, this is not a serious matter considering the fact that the construction process is in general a onetime process [24]. The feasibility constraints remain invariant if the performance objectives are changed. Even if the design migrates by a small amount, these constraints usually do not change [27]. This, however, demands an efficient determination of the feasibility constraints.
Data scaling is an essential step to improve the learning/training process of SVMs. The data of the input and/or output parameters are scaled. The commonly suggested scaling schemes are linear scaling, log scaling, and twosided log scaling. The present methodology employs both linear scaling as well as logarithmic scaling depending upon the parameters chosen. The following formula are used for linear and logarithmic scaling within an interval [32]: where is the unscaled th data of any parameter bounded within the interval . Linear scaling of data balances the ranges of different inputs or outputs. Applying log scale to data with large variations balances large and small magnitudes of the same parameter in different regions of the model.
5.2. LSSVM Construction
In this subsection, we discuss the various issues related to the construction of the LSSVM regressor.
5.2.1. Choice of Kernel Function
The first step of construction of an LSSVM model is the selection of an appropriate kernel function. For the choice of kernel function , there are several alternatives. Some of the commonly used functions are listed in Table 1, where , , , and are constants, referred to as hyper parameters. In general, in any classification or regression problem, if the hyper parameters of the model are not well selected, the predicted results will not be good enough. Optimum values for these parameters therefore need to be determined through proper tuning method. Note that the Mercer condition holds for all and values in the radial basis function (RBF) and the polynomial case, but not for all possible choices of and in the multilayer perceptron (MLP) case. Therefore, the MLP kernel will not be considered in this work.

5.2.2. Tuning of Hyper Parameters
As mentioned earlier, when designing an effective LSSVM model, the hyper parameter values have to be chosen carefully. The regularization parameter , determines the tradeoff cost between minimizing the training error and minimizing the model error. The kernel parameter or defines the nonlinear mapping from the input space to some highdimensional feature space [33].
Optimal values of the hyper parameters are usually determined by minimizing the estimated generalization error. The generalization error is a function that measures the generalization ability of the constructed models, that is, the ability to predict correctly the performance of an unknown sample. The techniques used for estimating the generalization error in the present methodology are as follows. (1)Holdout method: this is a simple technique for estimating the generalization error. The dataset is separated into two sets, called the training set and the test set. The SVM is constructed using the training set only. Then it is tested using the test dataset. The test data are completely unknown to the estimator. The errors it makes are accumulated to give the mean test set error, which is used to evaluate the model. This method is very fast. However, its evaluation can have a high variance. The evaluation may depend heavily on the data points that end up in the training set and on those which end up in the test set, and thus the evaluation may be significantly different depending on how the division is made.(2) โโfold crossvalidation method: in this method, the training data is randomly split into mutually exclusive subsets (the folds) of approximately equal size [33]. The SVM is constructed using of the subsets and then tested on the subset left out. This procedure is repeated times. Averaging the test error over the trials gives an estimate of the expected generalization error. The advantage of this method is that the accuracy of the constructed SVM does not depends upon how the data gets divided. The variance of the resulting estimate is reduced as is increased. The disadvantage of this method is that it is time consuming.
Primarily there are three different approaches for optimal determination of the SVM hyper parameters: heuristic method, local search method and global search method. The value is related to the distance between training points and the smoothness of the interpolation of the model. A heuristic rule has been discussed in [34] for estimating the value as where is the minimum distance (nonzero) between two training points and is the maximum distance between two training points. The regularization parameter is determined based upon the tradeoff between the smoothness of the model and its accuracy. The bigger its value the more importance is given to the error of the model in the minimization process. Choosing a low value is not suggested while using exponential RBF to model performances which are often approximately linear or weakly quadratic in most input variables. While constructing LSSVMbased analog performance model, heuristic method has been applied for determining the hyper parameters in [23]. The hyper parameters generated through heuristic method are often found to be suboptimal as demonstrated in [12]. Therefore, determination of hyper parameters through formal optimization procedure is suggested [33].
The present methodology employs two techniques for selecting optimal values of the model hyper parameters. The first one is a grid search technique and the other one is a genetic algorithmbased technique. These are explained below considering the RBF as the kernel function. For other kernels, the techniques are accordingly used.
(1) Grid Search Technique
The basic steps of the grid searchbased technique is outlined below.
(1)Consider a grid space of , defined by and , where and define the boundary of the grid space. (2)For each pair within the grid space, estimate the generalization error through holdout/fold crossvalidation technique. (3)Choose the pair that leads to the lowest error. (4)Use the best parameter to create the SVM model as predictor.
The grid search technique is simple. However, this is computationally expensive since this is an exhaustive search technique. The accuracy and time cost of the grid method are tradeoff depending on the grid density. In general, with the increase in grid density, the computational process becomes expensive. On the other hand, sparse density lowers the accuracy. The grid search technique is therefore performed in two stages. In the first stage, a coarse grid search is performed. After identifying a better region on the grid, a finer grid search on that region is conducted in the second stage. In addition, the grid search process is a tricky task since a suitable sampling step varies from kernel to kernel and the grid interval may not be easy to locate without prior knowledge of the problem. In the present work, these parameters are determined through trial and error method.
(2) Genetic AlgorithmBased Technique
In order to reduce the computational time required to determine the optimal hyper parameter values without sacrificing the accuracy, numerical gradientbased optimization technique can be used. However, it has been found that often the SVM model selection criteria have multiple local optima with respect to the hyper parameter values [28]. In such cases, the gradientbased method have chances to be trapped in bad local optima. Considering this fact, we use a genetic algorithmbased global optimization technique for determining the hyper parameter values.
In the GAbased technique, the task of selection of the hyper parameters is same as an optima searching task, and each point in the search space represents one feasible solution (specific hyper parameters). Each feasible solution is marked by its estimated generalization ability, and the determination of a solution is equal to determination of some extreme point in the search space.
An outline of a simple GAbased process is shown in Figure 5. The chromosomes consist of two parts, and . The encoding of the hyper parameters into a chromosome is a key issues. A realcoded scheme is used as the representation of the parameters in this work. Therefore, the solution space coincides with the chromosome space. In order to produce the initial population, the initial values of the designed parameters are distributed in the solution space evenly. The selection of population size, is one of the factors that affects the performance of GA. The GA evaluation duration is proportional to the population size. If the population size is too large, a prohibitive amount of time for optimization will be required. On the other hand, if the population size is too small, the GA can prematurely converge to a suboptimal solution, thereby reducing the final solution quality. There is no generally accepted theory for determining optimal population size. Usually, it is determined by experimentation or experience.
During the evolutionary process of GA, a model is trained with the current hyper parameter values. The holdout method as well as the fold crossvalidation method are used for estimating the generalization error. The fitness function is an important factor for estimation and evolution of SVMs providing satisfactory and stable results. The fitness function expresses the users' objective and favours SVMs with satisfactory generalization ability. The fitness of the chromosomes in the present work is determined by the average relative error (ARE) calculated over the test samples. The fitness function is defined as
Thus, maximizing the fitness value corresponds to minimizing the predicted error. The ARE function is defined as
Here , , and are the number of test data, the SVM estimator output, and the corresponding SPICE simulated value, respectively. The fitness of each chromosome is taken to be the average of five repetitions. This reduces the stochastic variability of the model training process in GAbased LSSVM.
The genetic operator includes the three basic operators such as selection, crossover, and mutation. Roulette wheel selection technique is used for the selection operation. The probability of selecting the th solution is given by
where is the size of the population. Besides, in order to keep the best chromosome in every generation, the idea of elitism is adopted. The use of a pair of realparameter decision variable vectors to create a new pair of offspring vectors is done by the crossover operator. For two parent solutions and , the offspring is determined through a blend crossover operator. For two parent solutions and , such that , the blend crossover operator (BLX) randomly picks a solution in the range . Thus, if be a random number in the range (0,1) and , then the following is an offspring:
If is zero, this crossover creates a random solution in the range . It has been reported for a number of test cases that BLX0.5 (with ) performs better than BLX operators with any other value. The mutation operator is used with a low probability to alter the solutions locally to hopefully create better solutions. The need for mutation is to maintain a good diversity of the population. The normally distributed mutation operator is used in this work. A zero mean Gaussian probability distribution with standard deviation for the th solution is used. The new solution is given as
The parameter is userdefined and dependent upon the problem. Also, it must be ensured that the new solution lies within the specified upper and lower limits. When the difference between the estimated error of the child population and that of the parent population is less than a predefined threshold over certain fixed generations, the whole process is terminated and the corresponding hyper parameter pair is taken as the output.
It may be mentioned here that there is no fixed method for defining the GA parameters, which are all empirical in nature. However, the optimality of the hyper parameter values is dependent upon the values of the GA parameters. In the present work, the values of the GA parameters are selected primarily by trial and error method over several runs.
5.3. Quality Measures
Statistical functions are generally used to assess the quality of the generated estimator. The ARE function defined in (17) is one such measure. Another commonly used measure is the correlation coefficient (). This is defined as follows: The correlation coefficient is a measure of how closely the LSSVM outputs fit with the target values. It is a number between 0 and 1. If there is no linear relationship between the estimated values and the actual targets, then the correlation coefficient is 0. If the number is equal to 1.0, then there is a perfect fit between the targets and the outputs. Thus, the higher the correlation coefficient, the better it is.
6. Topology Sizing Methodology Using GA
The topology sizing process is defined as the task of determining the topology parameters (specification parameters of the constituent component blocks) of a highlevel topology such that the desired specifications of the system are satisfied with optimized performances. In this section, we discuss a genetic algorithmbased methodology for a topology sizing process employing the constructed LSSVM performance models.
An outline of the flow is shown in Figure 6. A highlevel topology is regarded as a multidimensional space, in which the topology parameters are the dimensions. The valid design space for a particular application consists of those points which satisfy the design constraints. The optimization algorithm searches in this valid design space for the point which optimizes a cost function. The optimization targets, that is, the performance parameters to be optimized and system specifications to be satisfied are specified by the user. The GA optimizer generates a set of chromosomes, each representing a combination of topology parameters in the given design space. Performance estimation models for estimating the performances of a topology of the entire system are constructed by combining the LSSVM models of the individual component blocks through analytical formulae. The performance estimation models take each combination of topology parameters and produce an estimation of the desired performance cost of the topology as the output. A cost function is computed using these estimated performance values. The chromosomes are updated according to their fitness, related to the cost function. This process continues until a desired cost function objective is achieved or a maximum number of iterations are executed.
7. Numerical Results
In this section, we provide experimental results demonstrating the methodologies described above. The entire methodology has been implemented in MATLAB environment and the training of the LSSVM has been done using MATLAB toolbox [35].
7.1. Experiment 1
A twostage CMOS operational transconductance amplifier (OTA) is shown in Figure 7. The technology is 0.18โฮผm CMOS process, with a supply voltage of 1.8โV. The transistor level parameters along with the various feasibility constraints are listed in Table 2. The functional constraints ensure that all the transistors are on and are in the saturation region with some userdefined margin. We consider the problem of modeling input referred thermal noise , power consumption , and output impedance as functions of DC gain , bandwidth , and slew rate . From the sample space defined by the transistor sizes, a set of 5000 samples is generated using a Halton sequence generator. These are simulated through AC analysis, operating point analysis, noise analysis, and transient analysis using SPICE program. Out of all samples, only 1027 samples are found to satisfy the functional and performance constraints listed in Table 2.

The estimation functions are generated using LSSVM technique. The generalization errors are estimated through the holdout method and the 5fold crossvalidation method. The hyper parameters are computed through the grid search and the GAbased technique. In the grid search technique, the hyper parameters are restricted within the range and . The grid search algorithm is performed with a step size of 0.6 in and 10 in . These parameters are fixed based on heuristic estimations and repeated trials. The determined hyper parameter values along with the quality measures and the training time are reported in Tables 3 and 4 for the holdout method and the crossvalidation method, respectively. From the results, we observe that the average relative errors for the test samples are low (i.e., the generalization ability of the models is high) when the errors are estimated using the crossvalidation method. However, the crossvalidation method is much slower compared to the holdout method.


For GA, the population size is taken to be tentimes the number of the optimization variables. The crossover probability and the mutation probability are taken as 0.8 and 0.05, respectively. These are determined through a trial and error process. The hyper parameter values and the quality measures are reported in Tables 5 and 6. From the results the above observations are also noted.


A comparison between the gridsearch technique and the GAbased technique with respect to accuracy (ARE), correlation coefficient , and required training time is made in Table 7. All the experiments are performed on a PC with PIV 3.00โGHz processor and 512โMB RAM. We observe from the comparison that the accuracy of SVM models constructed using the grid search technique and the GAbased technique are almost the same. However, the GAbased technique is at least tentimes faster than the grid search method. From (1), we conclude that the construction cost of the GAbased method is much lower than the grid searchbased method, since the data generation time is same for both the methods.

The scatter plots of SPICEsimulated and LSSVM estimated values for normalized test data of the three models are shown in Figures 8(a), 8(b), and 8(c), respectively. These scatter plots illustrate the correlation between the SPICE simulated and the LSSVMestimated test data. The correlation coefficients are very close to unity. Perfect accuracy would result in the data points forming a straight line along the diagonal axis.
(a) Scatter plot of estimated and original values for the noise model with normalized test data
(b) Scatter plot of estimated and original values for the power model with normalized test data
(c) Scatter plot of estimated and original values for the impedance model with normalized test data
7.2. Experiment 2
The objective of this experimentation is to quantitatively compare between our methodology and the EsteMate [21]. The power consumption model is reconstructed using the EsteMate technique. The specification parameter space is sampled randomly. A set of 5000 samples is considered. For each selected sample, an optimal sizing is performed and the resulting power consumption is measured. The sizing is done with a simulated annealingbased optimization procedure and standard analytical equations relating transistor sizes to the specification parameters [36] following the EsteMate procedure. Of these, 3205 samples are accepted and the rest are rejected. The determination of the training set took 10 hours of CPU time. The training is done through an artificial neural network structure with two hidden layers. The number of neurons for the first layer is 9, the number of neurons for the second layer is 6. The holdout method is used for estimating the generalization ability.
A comparison between the two methodologies is reported in Table 8. From the results, we find that the data generation time is much less in our method compared to the EsteMate method. In addition, we find that the accuracy of our method is better than the EsteMate method. The experimental observations verify the theoretical arguments given in Section 2.1.

7.3. Experiment 3
The objective of this experimentation is to demonstrate the process of constructing highlevel performance model of a complete system and the task of topology sizing.
System Considerations
We choose a complete analog systemโinterface electronics for MEMS capacitive sensor system as shown in Figure 9(a). In this configuration, a halfbridge consisting of the sense capacitors is formed and driven by two pulse signals with 180ยฐ phase difference. The amplitude of the bridge output is proportional to the capacitance change and is amplified by a voltage amplifier. The final output voltage is given by
where is the nominal capacitance value, is the parasitic capacitance value at the sensor node, is the amplitude of the applied ac signal, and is the gain of the system, depending upon the desired output voltage sensitivity. The topology employs a chopper modulation technique for low noise purpose.
The desired functional specifications to be satisfied are (i) output voltage sensitivity (i.e., the total gain, since the input sensitivity is known) and (ii) cutoff frequency of the filter. The performance parameters to be optimized are (i) inputreferred thermal noise, (ii) total power consumption, and (iii) parasitic capacitance at the sensor node . The functional specifications and design constraints for the system are based on [37] and are listed in Table 9.

(a) Voltage sensing configuration of the interface electronics for MEMS capacitive sensor
(b) OTAC realizations of amplifier and filter
Identification of the Component Blocks and the Corresponding Performance Models
The synthesizable component blocks are the preamplifier (PA), inverter (IN) of the phase demodulator, lowpass filter (LF), and the output amplifier (OA). These are constructed using OTAs and capacitors. Figure 9(b) shows the implementations of the amplifier and the filter blocks using OTAs and capacitor [38, 39].
Highlevel performance models for the synthesizable component blocks corresponding to the performance parametersโ(i) input referred thermal noise, (ii) power consumption, and (iii) sensor node parasitics are constructed. The specification parameters which have dominant influence on the first two performances as well as on the functional specification, that is, the output voltage sensitivity and the cutoff frequency are the transconductance values of all the OTAs involved. On the other hand, for the last performance parameter, that is, sensor node parasitics, transconductance value of the first OTA of the preamplifier block is the single design parameter. Thus the values of the OTAs are considered as highlevel design parameters. In summary, we construct three performance models, input referred thermal noise, power consumption, and sensor node parasitics as functions of the values of the OTAs.
Construction of Performance Models for the PA Block
The geometry constraints and the feasibility constraints for the PA block of the topology are tabulated in Table 10. Similar types of constraints are considered for the other component blocks also. The inputoutput parameters of the models to be constructed are extracted through techniques discussed earlier. The sensor node parasitic capacitance is measured utilizing the halfbridge circuit shown in Figure 9(a), with only one amplifier block. Considering โfF, โfF, a square wave signal with amplitude โmV is applied and transient analysis is performed. Measuring the signal at the node , is calculated using (22).
Table 11 shows the hyper parameter values, percentage average relative error, and correlation coefficient of the constructed performance models for the preamplifier, with respect to SPICEsimulated value.


Reusability of Models and Construction of HighLevel Model for the Complete System
The performance models corresponding to the noise and the power consumption for the PA block are reused for the other component blocks. This is because all the component blocks have topological similarities and each of them is constructed from OTA circuits, as demonstrated in Figure 9(b). The issue of reusability of individual highlevel models in a complete system is thus applied here.
The highlevel models of the PA, IN, LF and OA blocks are combined analytically to construct the model of the complete system. The input referred noise and power consumption of the total system is given by
is the gain of the preamplifier. is the thermal noise model for the PA block, is that for the IN block of the phase demodulator, and so on. It is to be noted that need not be constructed again. It is same as . This is true for and . This reusability principle is applied for the power consumption model of all the blocks. The sensor node parasitics is the same as the input parasitics of the preamplifier. It is to be noted that while constructing the highlevel performance model of a complete system, the interactions between the transistors are taken care of while constructing the componentlevel performance model utilizing SPICE simulation data and the coupling between the blocks are considered through analytical equations.
Optimization Problem Formulation and Results
With these, the optimization problem for the topology sizing task is formulated as follows:
where are the associated weights.
The target output voltage sensitivity of the system (i.e., the total gain of the system) is taken as 145โmV/g and the cutoff frequency is taken as 35โKHz. The synthesis procedure took 181 seconds on a PIV, 3.00โGHz processor PC with 512โMB RAM. The crossover and the mutation probability are taken as 0.85 and 0.05, respectively. These are determined through a trial and error process. Table 12 lists the synthesized values of the topology parameters, as obtained from the synthesis procedure.

Validation
To validate the synthesis procedure, we simulate the entire system at the circuitlevel using SPICE. Exact values of are not achievable often. In such cases, the nearest neighbouring values are realized. An approximate idea about the transistor sizes required to implement the synthesized values are made from the large set of data gathered during the estimator construction. A comparison between the predicted performances and simulated values is presented in Table 13. We observe that the relative error between predicted performances and simulated performances in each case is acceptable. However, for the output sensitivity and the cutoff frequency, the error is high. This is because the circuitlevel nonideal effects have not been considered in the topology sizing process while formulating the final cost function and constraint functions. Following conventional procedure, this has been done purposefully in order to make the functions simple and the process converge smoothly [1, 27]. The acceptability and feasibility of the results are ensured to a large extent, since the utilized model is based on SPICE simulation results. The robustness of the results, however, could be verified by process corner analysis [27].

8. Conclusion
This paper presents a methodology for generation of highlevel performance models for analog component blocks using nonparametric regression technique. The transistor sizes of the circuitlevel implementations of the component blocks along with a set of geometry constraints applied over them define the sample space. Performance data are generated by simulating each sampled circuit configuration through SPICE. Least squares support vector machine (LSSVM) is used as a regression function. The generalization ability of the constructed models has been estimated through a holdout method and a 5fold crossvalidation method. Optimal values of the model hyper parameters are determined through a grid searchbased technique and a GAbased technique. The highlevel models of the individual component blocks are combined analytically to construct the highlevel model of a complete system. The entire methodology has been implemented under MATLAB environment. The methodology has been demonstrated with a set of experiments. The advantages of the present methodology are that the constructed models are accurate with respect to real circuitlevel simulation results, fast to evaluate and have a good generalization ability. In addition, the model construction time is low and the construction process does not require any detailed knowledge of circuit design. The constructed performance models have been used to implement a GAbased topology sizing process. The process has been demonstrated by considering the interface electronics for an MEMS capacitive accelerometer sensor as an example. It may be noted that multiobjective optimization algorithms [40] can also be used in the proposed approach for solving (25).
Acknowledgment
The first author likes to thank the Department of Science and Technology, Government of India for partial financial support of the present paper through Fast Track Young Scientist Scheme, no. SR/FTP/ETA063/2009.
References
 E. S. J. Martens and G. G. E. Gielen, HighLevel Modeling and Synthesis of Analog Integrated Systems, Springer, New York, NY, USA, 2008.
 G. G. E. Gielen, โCAD tools for embedded analogue circuits in mixedsignal integrated systems on chip,โ IEE Proceedings: Computers and Digital Techniques, vol. 152, no. 3, pp. 317โ332, 2005. View at: Publisher Site  Google Scholar
 S. Y. Lee, C. Y. Chen, J. H. Hong, R. G. Chang, and M. P.H. Lin, โAutomated synthesis of discretetime sigmadelta modulators from system architecture to circuit netlist,โ Microelectronics Journal, vol. 42, pp. 347โ357, 2010. View at: Publisher Site  Google Scholar
 E. Martens and G. G. E. Gielen, โClassification of analog synthesis tools based on their architecture selection mechanisms,โ Integration, the VLSI Journal, vol. 41, no. 2, pp. 238โ252, 2008. View at: Publisher Site  Google Scholar
 S. Pandit, C. Mandal, and A. Patra, โAn automated highlevel topology generation procedure for continuoustime ฮฃฮ modulator,โ Integration, the VLSI Journal, vol. 43, no. 3, pp. 289โ304, 2010. View at: Publisher Site  Google Scholar
 S. Pandit, S. K. Bhattacharya, C. Mandal, and A. Patra, โA fast exploration procedure for analog highlevel specification translation,โ IEEE Transactions on ComputerAided Design of Integrated Circuits and Systems, vol. 27, no. 8, pp. 1493โ1497, 2008. View at: Publisher Site  Google Scholar
 G. G. E. Gielen, โModeling and analysis techniques for systemlevel architectural design of telecom frontends,โ IEEE Transactions on Microwave Theory and Techniques, vol. 50, no. 1, pp. 360โ368, 2002. View at: Publisher Site  Google Scholar
 J. Crols, S. Donnay, M. Steyaert, and G. G. E. Gielen, โHighlevel design and optimization tool for analog RF receiver frontends,โ in Proceedings of the IEEE/ACM International Conference on ComputerAided Design (ICCAD '95), pp. 550โ553, November 1995. View at: Google Scholar
 F. Medeiro, B. PerezVerdu, A. RodriguezVazquez, and J. L. Huertas, โVertically integrated tool for automated design of ฮฃฮ modulators,โ IEEE Journal of SolidState Circuits, vol. 30, no. 7, pp. 762โ772, 1995. View at: Google Scholar
 H. Tang and A. Doboli, โHighlevel synthesis of ฮฮฃ modulator topologies optimized for complexity, sensitivity, and power consumption,โ IEEE Transactions on ComputerAided Design of Integrated Circuits and Systems, vol. 25, no. 3, pp. 597โ607, 2006. View at: Publisher Site  Google Scholar
 Y. Wei, A. Doboli, and H. Tang, โSystematic methodology for designing reconfigurable ฮฮฃ modulator topologies for multimode communication systems,โ IEEE Transactions on ComputerAided Design of Integrated Circuits and Systems, vol. 26, no. 3, pp. 480โ495, 2007. View at: Publisher Site  Google Scholar
 S. Pandit, C. Mandal, and A. Patra, โSystematic methodology for highlevel performance modeling of analog systems,โ in Proceedings of the 22nd International Conference on VLSI DesignโHeld Jointly with the 7th International Conference on Embedded Systems (VLSID '09), pp. 361โ366, January 2009. View at: Publisher Site  Google Scholar
 E. Lauwers and G. G. E. Gielen, โPower estimation methods for analog circuits for architectural exploration of integrated systems,โ IEEE Transactions on Very Large Scale Integration (VLSI) Systems, vol. 10, no. 2, pp. 155โ162, 2002. View at: Publisher Site  Google Scholar
 R. A. Rutenbar, G. G. E. Gielen, and J. Roychowdhury, โHierarchical modeling, optimization, and synthesis for systemlevel analog and RF designs,โ Proceedings of the IEEE, vol. 95, no. 3, pp. 640โ669, 2007. View at: Publisher Site  Google Scholar
 A. NunezAldana and R. Vemuri, โAn analog performance estimator for improving the effectiveness of CMOS analog systems circuit synthesis,โ in Proceedings of Design, Automation and Test in Europe Conference and Exhibition (DATE '99), pp. 406โ411, Munich, Germany, 1999. View at: Google Scholar
 A. Doboli, N. Dhanwada, A. NunezAldana, and R. Vemuri, โA twolayer librarybased approach to synthesis of analog systems from VHDLAMS specifications,โ ACM Transactions on Design Automation of Electronic Systems, vol. 9, no. 2, pp. 238โ271, 2004. View at: Publisher Site  Google Scholar
 M. Del Mar Hershenson, S. P. Boyd, and T. H. Lee, โOptimal design of a CMOS opamp via geometric programming,โ IEEE Transactions on ComputerAided Design of Integrated Circuits and Systems, vol. 20, no. 1, pp. 1โ21, 2001. View at: Publisher Site  Google Scholar
 P. Mandal and V. Visvanathan, โCMOS opamp sizing using a geometric programming formulation,โ IEEE Transactions on ComputerAided Design of Integrated Circuits and Systems, vol. 20, no. 1, pp. 22โ38, 2001. View at: Publisher Site  Google Scholar
 W. Daems, G. G. E. Gielen, and W. Sansen, โSimulationbased generation of posynomial performance models for the sizing of analog integrated circuits,โ IEEE Transactions on ComputerAided Design of Integrated Circuits and Systems, vol. 22, no. 5, pp. 517โ534, 2003. View at: Publisher Site  Google Scholar
 X. Li, P. Gopalakrishnan, Y. Xu, and L. T. Pileggi, โRobust analog/RF circuit design with projectionbased performance modeling,โ IEEE Transactions on ComputerAided Design of Integrated Circuits and Systems, vol. 26, no. 1, pp. 2โ15, 2007. View at: Publisher Site  Google Scholar
 G. van der Plas, J. Vandenbussche, G. G. E. Gielen, and W. Sansen, โEsteMate: a tool for automated power and area estimation in analog topdown design and synthesis,โ in Proceedings of the IEEE Custom Integrated Circuits Conference (CICC '97), pp. 139โ142, May 1997. View at: Google Scholar
 X. Ren and T. Kazmierski, โPerformance modelling and optimisation of RF circuits using support vector machines,โ in Proceedings of the 14th International Conference on Mixed Design of Integrated Circuits and Systems (MIXDES '07), pp. 317โ321, June 2007. View at: Publisher Site  Google Scholar
 T. Kiely and G. G. E. Gielen, โPerformance modeling of analog integrated circuits using leastsquares support vector machines,โ in Proceedings of Design, Automation and Test in Europe Conference and Exhibition (DATE '04), pp. 448โ453, February 2004. View at: Publisher Site  Google Scholar
 M. Ding and R. Vemuri, โA combined feasibility and performance macromodel for analog circuits,โ in Proceedings of the 42nd Design Automation Conference (DAC '05), pp. 63โ68, June 2005. View at: Google Scholar
 H. Li and Y. Zhang, โAn algorithm of soft fault diagnosis for analog circuit based on the optimized SVM by GA,โ in Proceedings of the 9th International Conference on Electronic Measurement and Instruments (ICEMI '09), pp. 41023โ41027, August 2009. View at: Publisher Site  Google Scholar
 M. Barros, J. Guilherme, and N. Horta, โAnalog circuits optimization based on evolutionary computation techniques,โ Integration, the VLSI Journal, vol. 43, no. 1, pp. 136โ155, 2010. View at: Publisher Site  Google Scholar
 H. E. Graeb, Analog Design Centering and Sizing, Springer, New York, NY, USA, 2007.
 V. Vapnik, Statistical Learning Theory, Springer, New York, NY, USA, 1998.
 J. A. K. Suykens, T. V. Gestel, J. D. Brabanter, B. D. Moor, and V. J. Vandewalle, Least Squares Support Vector Machines, World Scientific, 2002.
 H. Graeb, S. Zizala, J. Eckmueller, and K. Antreich, โThe sizing rules method for analog integrated circuit design,โ in Proceedings of the IEEE/ACM International Conference on ComputerAided Design (ICCAD '01), pp. 343โ349, San Jose, Calif, USA, November 2001. View at: Publisher Site  Google Scholar
 G. P. Box, W. G. Hunter, and J. S. Hunter, Statistics for Experimenters: An Introduction to Design, Analysis and Model Building, Wiley, New York, NY, USA, 1978.
 Q. J. Zhang, K. C. Gupta, and V. K. Devabhaktuni, โArtificial neural networks for RF and microwave designโfrom theory to practice,โ IEEE Transactions on Microwave Theory and Techniques, vol. 51, no. 4, pp. 1339โ1350, 2003. View at: Publisher Site  Google Scholar
 K. Duan, S. S. Keerthi, and A. N. Poo, โEvaluation of simple performance measures for tuning SVM hyperparameters,โ Neurocomputing, vol. 51, pp. 41โ59, 2003. View at: Publisher Site  Google Scholar
 G. Rubio, H. Pomares, I. Rojas, and L. J. Herrera, โA heuristic method for parameter selection in LSSVM: application to time series prediction,โ International Journal of Forecasting, vol. 27, pp. 725โ739, 2010. View at: Publisher Site  Google Scholar
 โLSSVM Toolbox,โ February 2003, http://www.esat.kuleuven.ac.be/sista/lssvmlab/. View at: Google Scholar
 P. E. Allen and D. R. Holberg, CMOS Analog Circuit Design, Oxford University Press, 2004.
 J. Wu, G. K. Feeder, and L. R. Carley, โA lownoise lowoffset capacitive sensing amplifier for a 50ฮผg/$\sqrt{Hz}$ monolithic CMOS MEMS accelerometer,โ IEEE Journal of SolidState Circuits, vol. 39, no. 5, pp. 722โ730, 2004. View at: Publisher Site  Google Scholar
 E. SรกnchezSinencio and J. SilvaMartรญnez, โCMOS transconductance amplifiers, architectures and active filters: a tutorial,โ IEE Proceedings: Circuits, Devices and Systems, vol. 147, no. 1, pp. 3โ12, 2000. View at: Publisher Site  Google Scholar
 R. Schaumann and M. E. Van Valkenburg, Design of Analog Filters, Oxford University Press, 2004.
 I. GuerraGรณmez, E. TleloCuautle, T. McConaghy et al., โSizing mixedmode circuits by multiobjective evolutionary algorithms,โ in Proceedings of the 53rd IEEE International Midwest Symposium on Circuits and Systems (MWSCAS '10), pp. 813โ816, August 2010. View at: Publisher Site  Google Scholar
Copyright
Copyright ยฉ 2011 Soumya Pandit et al. This is an open access article distributed under the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.