Abstract

In the era of big data, the efficient use of idle data in reinforced concrete structures has become a key issue in optimizing seismic performance evaluation methods for building structures. In this paper, based on the evaluation method of structural displacement seismic performance and based on the characteristics of high scalability and high fault tolerance of the cloud platform, the open source distributed and storage features of the Hadoop architecture cloud platform are introduced as a subproject of Apache Nutch project, Hadoop cloud platform. With features such as high scalability, high fault tolerance, and flexible deployment, the storage platform is secure, stable, and reliable. From the evaluation of the seismic performance of newly-built buildings and existing damaged buildings, according to the structural strength-ductility theory of the structure, the building structure resists earthquakes with its strength and ductility and buildings are divided into four categories. Due to the influence of time or seismic damage on the structure of reinforced concrete frame structures, their material properties are often deteriorating. Using the distributed computing design concept to efficiently process big data, a dynamic evaluation model for the seismic performance of reinforced concrete frame structures is established. A project of a 10-story reinforced concrete frame structure was selected for calculation and analysis; the engineering example was used to verify the accuracy and efficiency of the model, and the seismic performance of the floor was analyzed. It can be seen that the initial stiffness index of the structure is not sensitive to the damage location of the structure. The platform based on the concept of distributed computing big data processing can effectively improve the efficiency and accuracy of the evaluation of reinforced concrete frame structures.

1. Introduction

The seismic evaluation theory for building structures with reinforced concrete structure performance is based on the investigation and statistics of a large number of seismic damage data and scientific analysis. With the advent of the era of big data, the accuracy, comprehensiveness, and efficiency of the evaluation of seismic performance have also increased. He et al. estimated the seismic performance of rural residential buildings in Guangzhou from the aspects of building structure, building age, and building fortification and put forward the countermeasures and suggestions for earthquake prevention and disaster reduction of rural residential buildings. The flaws of the seismic evaluation theory and the seismic design theory gradually exposed. The American scholar J. P. Moehle [1] first proposed the concept of structural seismic design thought based on structure displacement but did not use abundant building big data. This article introduces the open source distributed and storage features of the Hadoop architecture cloud platform as a subproject of the Apache Nutch project. The Hadoop cloud platform features high scalability, high fault tolerance, and flexible deployment. The storage platform is secure, stable, and reliable. MapReduce distributed computing is one of the core components of the Hadoop platform. It was first introduced by Google. It is a model and framework for computing task processing in the context of big data [2]. MapReducer’s design ideas can be divided into three levels: (1)First of all, the parallel computing concept is adopted to divide large-scale and massive data into multiple small data blocks, which is independent of other data blocks, and uses the same calculation method to perform parallel processing on the resources, and finally, the combined results are processed(2)The unified system architecture is used to hide the implementation details of the system, so that programmers do not have to care about distributed details to use the platform for algorithm processing and improve the development efficiency. MapReduce can autonomously complete the following aspects of the system’s underlying aspects of processing: automatic data segmentation and storage, automatically divide computing tasks and computing resources, synchronization of the node’s computing tasks and data resources, merge, load balance optimization of the overall system performance, detect all nodes in the system, and handle the failure(3)The model abstraction of distributed computing is mainly divided into map and reduce. MapReduce draws on the ideas in the Lisp functional language; Lisp defines the overall operation of the list elements and provides similar map and reduce operations. The map and reduce functions provide an abstract parallel programming interface for the system, which can quickly parallelize the data [3]. MapReduce abstracts computational processing into two operations map and reduce. The map operation is responsible for recording and repeating calculations on a group of data. The reduce operation is responsible for further integrated calculation and output of the calculation results of the map operation. It provides an open source computing platform for the data of reinforced concrete structures in building big data and realizes the efficient evaluation of reinforced concrete structures in building

2. Hadoop Dynamic Evaluation System for Seismic Performance

2.1. Hadoop Architecture

In general, the typical Hadoop architecture cloud platform is assumed to operate in a trusted environment. The typical Hadoop architecture cloud platform includes Hadoop cluster core nodes and auxiliary terminal interfaces. The cluster core nodes include: NameNode, backup NameNode, DataNode, and JobTracker members [4]. Therefore, the architecture of the Hadoop-based cloud platform is divided into three layers: the data storage layer, the middleware control layer, and the application node layer, as shown in Figure 1.

Among them, DataNode is the storage resource and computing resource of the Hadoop architecture cloud platform. Its main role is to store distributed system data and perform map-reduce tasks. Hadoop architecture cloud platform data information is stored on the DataNode terminal device. When the cloud platform receives a MapReduce job task, the NameNode node allocates the storage or computational resources of the current DataNode node through the resource scheduling policy, depending on the current task completion. For Hadoop-based cloud platforms with distributed file systems, the JobTracker plays a central role in controlling the NameNode node. It is responsible for submitting user tasks to the service running on each DataNode node of the TaskTracker, so this node also plays the role of resource application (please see Table 1).

2.2. HDFS Distributed File System

HDFS is fundamental to the unified management of stored data in distributed computing. Its features such as good fault tolerance, high reliability, high availability, high throughput, and strong expansion capability provide reliable storage protection for extracting information from a vast array of data. It has brought a lot of convenience to the research and excavation processing of massive data collection (please see Figure 2). HDFS is mainly designed to solve the following problems: (1)Inevitable System Errors. People equip the Hadoop distributed file system with a file system that can be executed on general hardware conditions. Therefore, in the course of daily use, mistakes are commonplace. Hadoop distributed file system is usually composed of hundreds of server nodes. Each node stores part of the information in all databases, and any part of this system has the probability that errors will occur at any time. Therefore, fault monitoring and rapid active restart are the main building concepts of the Hadoop distributed file system(2)Streaming Data Access. Programs executed on Hadoop distributed file systems are usually not acquired on a one-time basis but are often obtained gradually, in batches. Therefore, HDFS must support high throughput of data to better serve streaming data access(3)Mass Data Mining. With the rapid increase in the amount of data, the data to be processed is usually at or above the GB level. Therefore, a distributed file system needs to be able to provide a function of saving large files and can give a substantially higher amount of data for data transfer, which can be increased to hundreds of nodes in one cluster. Any distributed file system should be able to carry large-scale file resources. Because the process of excavating large amounts of data determines when the amount of data is very large, moving the execution process to copy it is much cheaper than migrating the data that will be analyzed. When the amount of information is very large, the effect becomes more pronounced. The overhead of migrating data is greater than the cost of the mobile computing process. This can reduce the possibility of network congestion and increase the throughput of the entire platform. In other words, it is a relatively good choice to move the calculation to the vicinity of the data, which is better than moving the data to the program execution place. Hadoop distributed file system gives the corresponding interface, so that the application can be actively moved to the data storage location to run

2.3. MapReduce Distributed Computing Model

MapReduce distributed computing is one of the core components of the Hadoop platform. It was first introduced by Google. It is a model and a framework for computing task processing in the context of big data [5]. MapReduce has changed the organization of large-scale computing and realized large-scale model abstraction of computing at the level of large-scale server clusters.

Similar to HDFS, MapReducer’s architecture is also the most commonly used master-slave structure. This master-slave architecture makes the system more scalable and robust. Introduce MapReducer’s architecture from four aspects: Job Tracker, Task Tracker, Task, and Client: (1)Job Tracker. The main track of the progress of all jobs in the system, the use of resources in the MapReduce, and the operating information submitted to the Task Scheduler. Tasks submitted to the system will be assigned to the Task Tracker to execute the corresponding job and resources and to monitor and maintain the job’s operation status. The failed job will be assigned to other Task Tracker for execution at any time(2)Task Tracker. Task Tracker is mainly responsible for job tasks allocated by Job Tracker; through the monitoring of local node resources, resources are allocated to different tasks according to usage. At the same time, Task Tracker communicates with the Job Tracker through Heart Beat timing, returns information to the Job Tracker, and accepts Job Tracker’s task scheduling(3)Task. Task is a specific calculation task, including Map Task and Reduce Task. The Map Task mainly refers to the processing in the map stage. The input file is parsed into <key, value> key-value pairs, and the key-value pairs are sorted and merged. The result is stored to the local disk and is partitioned by the partition function. The corresponding Reduce Task refers to the processing operation in the reduce stage, sorting the intermediate results after map processing and processing the Reduce function to generate a new <key, value> key-value pair and storing it in HDFS(4)Client. The client is used to interactively communicate with the MapReduce computing model. The user submits the task request through the client. The client submits the requested job to the Job Tracker. At the same time, the user can also view the running status of the task through the client

3. Evaluation Rules for Seismic Performance of New-Build Reinforced Concrete Frame Structures

3.1. Seismic Capacity Division of Building Structure
3.1.1. The Meaning of Earthquake Resistance

According to the theory of structural strength and ductility, the building structure resists earthquakes with its strength and ductility. When the earthquake occurs, the building structure is first resisted by the strength, the acceleration of the ground surface increases to make it yields, and then the ductility is used to resist the acceleration of the ground motion [6]. When the ductility is exhausted, the building structure will be destroyed. At this time, the acceleration of ground motion that causes structural damage is defined as the seismic capacity of the building. Therefore, in the evaluation of the seismic capability of reinforced concrete frame structures, the strength and ductility of the reinforced concrete frame structure are calculated according to the size of the actual structure of the building structure and the reinforcement, combined with the internal forces of the rod components obtained by the elastic seismic analysis of the frame. Calculate the seismic capacity of the upper and lower layers of each floor of a building structure, that is, the acceleration of ground motion sustained during the destruction .

3.1.2. Evaluation of Earthquake Resistance

In China’s newly published “General Principles of Design for Seismic Performance of Construction Projects (Applicable)” (CECS160:2004) according to the function of the building, the buildings are divided into the following four categories [7]: (1)IV class: use buildings that cannot interrupt or store large quantities of dangerous or toxic substances during or after an earthquake. Once the earthquake has caused the release and alienation of these items (such as toxic gases, explosives, and radioactive materials), it will cause unacceptable harm to the public(2)III class: postearthquake use of buildings or densely populated construction sites where functions must be restored in the short term or are critical for postearthquake operations, such as hospitals, schools, fire stations, police stations, communication centers, emergency control centers, and disaster relief centers(3)II class: all buildings and facilities except types IV, III, and I belong to this category(4)I class: when destroying construction earthquakes that do not endanger human life and do not cause serious property damage, such as general warehouses

3.2. Structural Failure Mode and Ductility Ratio Analysis

According to the material property and section size of the element, the strength of the structural element can be calculated according to the design code of concrete structure. However, the test results show that the bearing capacity of the element under repeated loading is lower than that under monotonic loading. Therefore, the bearing capacity of the beam and column elements calculated in this section should be multiplied by the corresponding reduction factor, where the reduction factor is 0.8 [810].

3.2.1. Moment Strength at the Yield of Beams and Columns

The column section bears axial force and bending moment at the same time. In the known conditions of column section size and reinforcement, the axial force-bending relation diagram (curve) is known, as shown in Figure 3. When the point corresponding to the internal force () of the column unit falls within the curve, the unit will not be destroyed. When the column element’s internal force () is at any point on the curve, the element breaks. However, under the specific earthquake action, the failure of the unit should correspond to a certain point on a curve.

It can be seen from Figure 3 that under the action of the gravity load representative values, the generated axial force and bending moment are represented by and , respectively. Under small earthquakes (corresponding earthquake acceleration is ), the axial force and bending moment generated in the column section are expressed by and , respectively. It is assumed that when the earthquake is getting larger, the total axial force and total bending moment increase slowly along the straight line AB. When the ground motion acceleration reaches, the axial force and bending moment just intersect with the axial force and one bending moment (-) intensity map at point C, assuming that point C is the point where the unit failure is affected by the earthquake. At this time, the column section yields and the bending bearing capacity is . Because the axial force of the beam is very small, its axial force is set to zero.

The flexural bearing capacity of beam elements is calculated according to the “Code for Design of Concrete Structures” (GB50010-2002) [9]. The load-bearing capacity of the column section is assumed to maintain the plane according to the section strain, and the stress-strain relationship between the steel and concrete specified in the concrete specification is used for calculation.

From Figure 4, the axial equilibrium equations are listed, i.e., . In the available form, the sum of the moments of the center point and the external force of the section is equal to zero, i.e., , available formula is given as follows:

In the formula, is the column axial force at the time of destruction, is the bending moment of the column at the time of destruction, is the height of concrete under pressure, is the area of the pressed steel bar and the yield stress, is tensile stress and area of tensile reinforcement, is the thickness of the concrete protective layer, is the concrete compressive stress, is the effective height of the column section, , and is the height of the concrete column section.

3.2.2. Shear Strength of Beam and Column Sections

The shear capacity of beam and column sections is generally considered to be provided by concrete and shear reinforcement . In the plastic hinge area, due to the repeated stress, the cracking of the concrete is very serious, so the part decreases with the increase of the ductility ratio. In addition, is also related to the existence of axial compressive stress. If the axial compressive stress is large, is effective. When the axial compressive stress is less than , can be regarded as zero. turns out to be , When the ductility ratio reaches the ductile capacity, it is reduced to . The value of is calculated as follows: where is the full cross-sectional area and is the axial force at which the section yields. Under the repeated action of the ground motion, the column unit may be stressed at a certain moment and may be pulled at another time, especially for the upper column unit of the structure. In order to reduce the influence of uncertainties, only the influence of the axial force of the column element on the shear capacity of the structure under the action of the gravity load representative value is considered. For ease of analysis, separate formulas for the shear capacity of beams and columns in the code are written separately. That is, represents the contribution of concrete to shear capacity and represents the contribution of reinforcing steel to shear capacity. Combined with the “Code for Design of Concrete Structures” (GB50010-2002) [10], the formulas for rectangular beams and columns and are as follows. (1)For beams

For beams that are subjected to concentrated loads (including the case where there are multiple loads in which the concentrated load is greater than 75% of the total shear value due to the shear force at the support section or the edge of the joint), shear bearing capacity is calculated as follows: where is the concrete compressive strength, is to configure the entire cross-sectional area of each stirrup in the same section, is the design value of stirrup tensile strength, is the curved rebar cross-sectional area in the same plane, is the angle between the longitudinal axis of the reinforcing bar and the component, and is the tensile strength of the platter. (2)For the columnwhere , is the column height, and is the axial force of the column element under the gravity load representative value. When , is the cross-sectional area of the component. When is pulling force, . When is pressure, .

(1) Correction of Ductile Capacity. When the bending moments of beams and columns are broken and if the stirrups of the plastic hinge area meet the requirements of the code, the ductile capacity when the beams break is defined . Ductile capacity at column failure . Otherwise, calculate the percentage of ductile reduction according to the actual constraint stirrup amount . At this point, the ductile capacity becomes , but not less than 1.0.

value is calculated as follows: where is reinforcement ratio, is calculated according to the “Building Seismic Design Code” (GB50011-2001) calculated by the amount of the stirrup reinforcement ratio, is a plastic hinge length, , is the length of the joint from the bending point to the beam column, and is the diameter of the main bar.

(2) Determines the Failure Mode and Ductility Ratio of Single Beam and Column. There are three failure modes for beams and columns. The first situation is shear failure. In the second case, the shear failure occurs before the ductile capacity is reached although the yield of the bending moment occurs. The third situation is the destruction of ductile capacity reaching , and at this time no shear failure has occurred.

In the first case, the rod end bending moment is . Since the yield has not yet occurred, the ductility ratio is taken as 1.0. In the second failure situation, the bending moment yield occurs, but before the ductile capacity is reached, the shear failure occurs and the ductility ratio is corresponding. In this case, the beam and column bear the shear force: . Its size is plus a certain multiple of . The corresponding rod end bending moment is plus this multiple times . In the third case, the ductile capacity has been fully used, so the ductility ratio is .

3.2.3. Nodal Strength Comparison, Column Bearing Shear, and Ductility

At the node where the beam and column intersect, whether the column is destroyed first or the beam first destroyed depends on the respective bending moments of the beam and column. If is greater than , it belongs to weak column strong beam. When calculating the seismic resistance of the upper or lower half of the node, the shearing force at the failure of the column is equal to the shear force at the failure of the column above, and the ductility ratio is also based on the ductility ratio obtained in the above section. If is less than , it is a strong beam. When calculating the seismic resistance of the half-layer of the node, under the shear force assumed when the column breaks: where is above the node, is below the node, is the shear force of the column under the gravity load representative value, and is the shear force of the column caused by small earthquakes. The ductility ratios corresponding to the upper and lower columns of the node are calculated by weighting the ductility ratios of the left and right beams, and the moments of the earthquakes assumed when the beam is destroyed are weighted. The corresponding ductility ratios for the upper and lower columns are calculated as follows: where and are the bending moment when the left and right beams break, respectively, and are the bending moment of the left and right beams caused by the gravity load representative value, respectively, and and are the ductility ratio of the left and right beams, respectively.

3.2.4. Acceleration of the Yielding Ground Motion of Each Half Layer

The acceleration of the yielding ground motion of each half layer is times the seismic acceleration corresponding to the small shock . The coefficients corresponding to the upper and lower halves of the -th layer and can be calculated as follows:

In the formula, and are the shear force of the upper and lower columns of a node under the action of small earthquakes, respectively, and are the shear force taken on the failure of the upper and lower columns of the node, respectively, and and are the shear force of the upper and lower columns under the action of the representative value of gravity load, respectively.

3.2.5. Seismic Reduction Coefficient in the Structural System

Seismic reduction coefficient () is the multiplier of seismic acceleration which can be increased when the structure is yielding. The ductility ratio of the system reaches the permissible ductility capacity () at this time. is related to the cycle of building structure, the type of site, and the . The relationship between the permissible ductility capacity () and the ductility capacity () is . Combined with the code for seismic design of buildings, can be calculated as follows:

In the formula, is a characteristic cycle (s) and is the cycle of building structure (s). When , the upper computing is 1.

3.2.6. Yield Ground Motion Acceleration and Seismic Capacity of Half Stories in Frame Structures

The ground acceleration of the upper and lower layers of the layer is and multiplied by the seismic acceleration corresponding to the small earthquake [1113]. The ductility capacity of the upper and lower half of the node must be the weighted average of the ductility capacity corresponding to the failure of each node. It can be calculated as follows:

In addition, the allowable ductility capacity of each half layer can be calculated as follows:

With allowable ductility capacity, the seismic reduction factor and of the structural system can be calculated according to the period and site type of the building structure.

The seismic capacity and of the upper and lower floors of the level nodes can be calculated as follows:

In the formula, and are the yield ground motion acceleration of the upper and lower half of the node, respectively. The values of aseismic capacity for each half story may be different. For safety consideration, the minimum value of all values is taken as the aseismic capacity of the whole building structure. Up to now, the aseismic capacity of the frame structure has been obtained.

4. Dynamic Evaluation of Seismic Performance of Reinforced Concrete Frame Structures at Present

The performance of building reinforced concrete frame structure deteriorates gradually with time or earthquake damage [14]. First, the constitutive relation of the material deterioration, the corrosion rate-based constitutive relation of reinforcement, and the slip constitutive relation considering the corrosion of the reinforced concrete are introduced. Then, it is embedded in the seismic capacity model of reinforced concrete frame structure, and a dynamic evaluation model for seismic performance is established.

4.1. Deterioration Model of Reinforced Concrete Frame Structure
4.1.1. The Constitutive Relation of Steel Deterioration Based on the Corrosion Rate

It is shown that after the corrosion of the reinforced concrete, the mechanical properties of the reinforced concrete structure will decrease sharply and the damaged form of the corroded reinforced concrete structure has obvious brittleness characteristics, especially the reinforced concrete structure which has been damaged by the earthquake [1519]. The concrete protection layer caused by structural damage is cracked or exfoliated, which makes the reinforced bar contacts with the external environment, which greatly increases the probability of corrosion of the structure during the service period of the service. Therefore, the constitutive relationship of reinforced bars should be newly determined for seismic evaluation of deteriorated reinforced concrete structures [20]. By studying the corrosion test of a group of reinforced concrete beams, Yingshu et al. gave the constitutive relation of corroded steel bars, as shown in formula (14)–(16) and assumed that the elastic modulus of steel bars remained unchanged before and after the corrosion.

When ,

When ,

In the formula, is the corrosion rate of steel, is the yield strength of corroded steel bar, is the elongation rate of corroded steel, is the ultimate strain of corroded steel bar, is the yield strength of not corroded steel bar, is the elongation of not corroded steel bar, and is the ultimate strain of noncorroded steel bar.

By observing the formula (17) and (18), in order to determine the constitutive relation of corroded steel bars, first, it is necessary to determine the corrosion rate of the steel bar () and the corrosion rate can be expressed by the depth of corrosion of the steel bar (). Jin Weiliang’s formula for calculating the amount of steel corrosion in the literature is given as follows:

In the formula, is the corrosion depth of steel bar and and are, respectively, the corrosion sites of reinforcement and the environmental correction factors of reinforced concrete structural members. and are the annual average temperature (°C) and annual average humidity (%) near the surface of reinforced concrete structures, respectively. is the thickness of the concrete protective layer (mm), is the compressive strength of concrete cubes (MPa), and and are, respectively, the service life of the reinforced concrete structure and the time when the steel bar starts to rust. The initial corrosion time of reinforcement in reinforced concrete structural members can be determined by the deterioration criterion expressed as follows:

In the formula, is a carbonization parameter (mm) and it is a random variable [21, 22]. In order to simplify the calculation, its mean can be determined by the following formula:

4.1.2. Slip Constitutive Relation of Reinforced Concrete Connections Based on the Corrosion Rate

Xu Shanhua studied the constitutive relationship between the bond stress and the slip between the reinforced concrete and the concrete, and the relationship between the bond strength and the corrosion rate of the corroded reinforced concrete members was obtained by the regression analysis of the experimental data by mathematical regression analysis [23], as shown in the following formula:

In the formula, is the bond strength between concrete and steel bar for noncorroded reinforced concrete members and is the reduction coefficient of the bond strength of corroded reinforced concrete members, as shown in the following formula:

Due to the surface properties of steel bars, the bond and slip of concrete and steel bar will be greatly affected. The expression of bond strength () is generally determined by the formula recommended in CEB-FIP Model Code 1990: Design Code, as shown in formula (25) and (26): (1)Glossy steel bar(2)Ribbed steel bar

In the formula, is the thickness of concrete protective layer (mm) and is the diameter of steel bar (mm).

4.2. Dynamic Evaluation of Seismic Behavior of Reinforced Concrete Frame Structures
4.2.1. Hadoop Framework of Reinforced Concrete Frame Structure

The Apache Hadoop library is a framework that extends a single server to thousands of servers; each of which provides local computing and storage rather than hardware to achieve high availability, mainly including the following modules [24]: (1)Hadoop common: commonly used utilities supporting other Hadoop modules(2)Hadoop distributed file system (HDFS): a distributed file system that provides high-throughput access to application data(3)Hadoop YSRN: framework for cluster job scheduling and resource management(4)Hadoop MapReduce: based on the YARN system, parallel processing large data sets

4.2.2. Construction of the HDFS File System for Aseismic Performance of Reinforced Concrete

HDFS is a master-slave architecture. If the client wants to access HDFS, it needs to complete the interaction between Name Node and Data Node. Name Node is responsible for managing the entire file system, and multiple Data Nodes is the unit responsible for storing data. When the client needs to operate the data in the file system, it communicates with the Name Node, but the actual reading and writing operation is done directly with the Data Nodes communication.

Name Node, the master control server, is responsible for processing the Namespace of the file system, recording the location of each file data block, and undertaking the work of coordinate and scheduling and also in charge of the changes in the Namespace. Data Node is responsible for actually storing the data needed by users. The Namespace of the distributed file system is open, so that users can store data according to the style of the file. Usually each data block occupies 64 MB, and the data stored on the HDFS is usually written only once, but it will be read many times in the later use. In general, the files stored on the Hadoop distributed file system will be divided into multiple blocks based on the size of the typical block, and each block will be divided as far as possible and stored separately in a separate Data Node. Name Node is responsible for handling all kinds of Namespace operations in the file system, such as Openfiles, Closefiles, Rename, and mapping data blocks to Data Node. Data Node is responsible for executing the I/O operation of users. According to the commands of Name Node, Create, Copy, Delete, and other actions of data blocks are carried out. If the client wants to access the file, the first step needs to get all the information of the data block address from the Name Node, obtains the specific location of the data block, and then interacts directly with the corresponding node to obtain the required data.

The most basic configuration of a Hadoop distributed file system is important, which can be summarized as running Name Node on a dedicated computer and executing Data Node on other nodes, respectively. The idea of just executing a Name Node on a machine cluster largely makes the structure of the system simpler (please see Figure 5).

In Name Node, a lot of transaction logs (Edit Log) are stored. These Edit Log save various operations and updates in the distributed file system and save Namespace with Image File, including mapping and file attribute information. Edit Log and Image File retain the local disk of the main node. Each time you start, you need to get Image File and Edit Log from the disk, then read the records in the Edit Log on the Image File, and rewrite the local new Image File, which can cut out the old Edit Log, and this process is called the checkpoint. The Hadoop distributed file system also sets up a Secondary Name Node, which helps Name Node handle Image File and Edit Log. When the Name Node starts running, the Image File and Edit Log are merged, Secondary Name Node will be copied to the temporary directory on time from Name Node, and then, it will be uploaded again. Name Node updates the Image File and processes Edit Log to control the value of Edit Log under the standard.

4.2.3. Distributed Calculation of Seismic Performance Evaluation of MapReduce

First, the MapReduce function splits the input file into split shards. Each shard is split into approximately 16 to 64 megabytes. Then, the shards are parsed into key-value pairs. The entire task is based on split shards. The MapReduce distributed computing model schedules multiple job tasks for computation through the core master program. The mapper will input the split assigned to it from HDFS based on the cut point of the cut. In the map phase, each split will create a corresponding map task and assign the task to the corresponding worker program. The worker program will call the map function in the system to perform the calculation and sort the intermediate results and then use the key value of the same key value. Pairs are grouped together to form a list, then the result set is divided into partitions and stored locally. In the reduce stage, the data of different map data will be integrated. After the sort processing, a new key-value pair will be formed. Finally, the result set will be processed through the reduce function to obtain the latest result (please see Figure 6).

Finally, the output of the reduce phase will now form the output file of master and finally output to HDFS. When all the map and reduce tasks of the whole process are completed, the reducer’s task will output files, and these files can continue to be passed as input files to other MapReduce computing processes. Or these files can be invoked by other distributed system applications as resources.

5. Example Analysis

5.1. Engineering Survey

In order to systematically explain the application process of this method in seismic performance evaluation of existing reinforced concrete structures and to check its effectiveness, a project is selected in this paper: a 10-story reinforced concrete frame structure is calculated and analyzed. The project is built in 2009, class C building. In the process of analysis, a side frame is first calculated. The load arrangement and reinforcement are shown in Figure 7. The height of the 1 layer is 4.2 m, the height of the 2~10 layers is 3.6 m, and the cross section size of the frame column of the 1~3 layers is 750 mm × 750 mm; the section size of the frame column of the 4~10 layers is 700 mm × 700 mm, and the cross section of the frame beam is 300 mm × 600 mm. The standard values of the floor and live loads are 5 K and concrete, respectively, and the concrete strength grade girder and pillar takes C35; the longitudinal force bar is HRB400, and the hoop is HPB300. The concrete section hoop information is shown in Table 2, and the basic acceleration 0.20 g is designed.

In the process of building design, the reinforced concrete hoop method, that is, the material information in the hoop process, is shown in Table 2.

In order to verify the effectiveness and accuracy of the seismic performance evaluation method of this paper when the structural damage appears on different floors, the example project was built in 2009 and under normal environment to determine the degree of concrete carbonization and reinforcement corrosion of its structure during its subsequent service life. Considering the constitutive relation of the material deterioration, the ductility curve of the members of the project is mainly determined by the stiffness and strength degradation coefficient caused by the deterioration of the structure, as shown in Table 3.

5.2. Numerical Simulation

In this paper, the framework of the reinforced concrete frame platform is constructed by using Apache Hadoop software library and the seismic performance evaluation model of the new building structure and deterioration structure data is introduced. The client has access to HDFS through the interaction between Name Node and Data Node; a write operation and multiple read operations are performed through the master control server Name Node. Then, with Hadoop as the large data processing platform, the numerical simulation of the seismic fortification intensity of 8 degrees, and the design of the basic acceleration 0.20 g, the changes in the seismic capacity of the different floors with the floor displacement and the interlayer displacement angle change, using the MapReduce function to divide into split segments, and the output results will now form the output file of master. The final output to HDFS is shown in Figures 810.

From Figures 8 and 9, it is known that the corrosion of steel bar and the connection slip of reinforced concrete have great influence on the floor displacement. Floor displacement angle of the floors above 4 layers has little influence on the floor displacement angle of the top layer above 8 layers. That is, the influence of the deterioration of the reinforced concrete structure on the seismic performance of the building is mainly distributed in the 4 to 8 layers. The comparison shows that the floor displacement angle increases with the increase of floor displacement distance when the floor is below 6 stories and decreases with the increase of floor displacement distance when the floor is above 6 stories. When the deterioration structure enters the yield, the stiffness degradation is more sensitive to the damaged parts of the structure. When the number and damage degree of the structural damage component are both at the same time, the lower the location of the structural damage is, the greater the degree of the stiffness degradation after the yield of the damage structure. As shown in Figure 10, the evaluation of reinforced concrete frame structure based on Apache Hadoop software library and Hadoop for large data processing platform is compared with the traditional displacement seismic evaluation method. The Hadoop seismic assessment method under large data background has high efficiency and accuracy.

6. Conclusion

Based on the seismic performance evaluation method of structural displacement, this paper follows the idea of large data processing with distributed computing. Based on the large reinforced concrete data of high-rise buildings and the seismic performance evaluation of new buildings and existing buildings, an efficient evaluation model for seismic performance of reinforced concrete frame structures is established and the seismic characteristics of floors are analyzed. The following conclusions can be drawn from the analysis of the seismic behavior of the floor. (1)Through the large data processing of reinforced concrete structure, it is found that the initial stiffness index of the structure is not sensitive to the identification of the damage location of the structure. The initial stiffness changes before and after the structural damage can only identify the seismic capacity of the structure, but cannot determine the location and size of the structural damage(2)The calculation results of the Hadoop seismic assessment model underestimate the interlayer displacement angle of the structure on the low floor and overestimate the interlayer displacement angle at the higher floor, but the overall error is smaller than the real value, and the smaller the upper level error is, the seismic performance evaluation results are ideal(3)The efficiency and accuracy of the platform based on the idea of distributed computing and big data processing can effectively improve the evaluation of RC frame structures. Efficient use of large building data improves the seismic performance evaluation method on the one hand and on the other hand improves the reliability of seismic performance evaluation results

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

We declare that there is no conflict of interest regarding the publication of this paper.