Abstract

The speed of earthquake emergency web document data cleaning is one of the key factors affecting emergency rescue decision-making. Data classification is the core process of data cleaning, and the efficiency of data classification determines the speed of data cleaning. This article is based on earthquake emergency Web document data and HTML structural features, combined with TF-IDF Algorithm and information calculation model, improves the word frequency factor and location factor parameters, and proposes the weighted frequency algorithm P-TF-IDF for earthquake emergency Web documents. To filter out less frequent words and optimize the FastText model, N-gram Feature word vectors effectively improve the efficiency of Web document data classification; for text classification data, use missing data recognition rules, data classification rules, and data repair rules to design an artificial intelligence-based earthquake emergency network information data cleaning framework to detect invalid data sets value, complete data comparison and redundancy judgment, clean up data conflicts and data errors, and generate a complete data set without duplication. The data cleaning framework not only completes the fusion of earthquake emergency network information but also provides a data foundation for the visualization of earthquake emergency data.

1. Introduction

Earthquake emergency work is a paramilitary operation, and the key to success or failure lies in whether scientific and reasonable decisions can be made and put into action in the shortest possible time. Scientific and reasonable decision-making requires comprehensive, accurate, and timely disaster information support. Accurate disaster information is also a key factor in guiding public opinion involving earthquakes. The integration of artificial intelligence technology and earthquake emergency network information has become possible to quickly obtain massive network disaster information after the earthquake, and the basis for emergency response and timely rescue decision support is becoming more and more comprehensive and rich.

At present, the construction of emergency response information system is lagging behind, and emergency rescuers have a heavy workload for information processing of sudden earthquake events, which seriously hinders the efficiency of emergency response. Especially for the massive amount of earthquake disaster information that cannot be collected and processed in a timely and effective manner. Taking network resources as an example, there is special network information of the earthquake industry, news media websites, microblogs, forums, etc., and a large amount of useful disaster information is mixed in various forms. Among the massive data, in addition, the information often contains a large amount of “dirty data” (such as information duplication, error, incompleteness, inconsistency of information, and logic errors). Therefore, how to use artificial intelligence technology to quickly detect the “dirty data” in the disaster data, and effectively clean and organize them is important, so that the earthquake emergency scientific research personnel can carry out the next analysis and research, which is one of the key problems to be solved in the earthquake emergency work at the present stage, especially after the acquisition of large earthquake disaster.

Organize data before data cleaning is particularly important, among which data classification is one of the important means of organize data. Recently, Dr. Mario et al. proposed a completely innovative fuzzy classifier based on fuzzy similarity calculation, and implemented a low-cost analysis system including probe on hardware, which effectively improves the efficiency of airline maintenance of aircraft safety standards [1]. The research on data cleaning first appeared in the United States, starting from the correction of the national social security number error, mainly by introducing data mining methods, using backtracking ideas, analyzing the characteristics of data sources, and analyzing data according to the corresponding business rules. Analyze and inspect each link, summarize the corresponding data cleaning rules and strategies, and extract data cleaning strategies, algorithms, and frameworks that meet the actual needs based on the existing cleaning algorithms and models, and finally on the data set. Apply these algorithms to restore the data to the data that can be used by the user. For example, the clustering method is used to detect abnormal records, the model method finds records that do not conform to the existing pattern, and then the association rule method is used to find abnormal data to detect and eliminate abnormal and approximate duplicate records. At the same time, with the development of Internet information, the research on data cleaning technology has been promoted. For example, in terms of network data cleaning, it is mainly divided into two categories: web page level data cleaning and page internal element level data cleaning. The former is represented by the PageRank algorithm proposed by Google and the HITS algorithm of the Clever system of IBM. The idea of the latter is mainly reflected in the VIPS algorithm, which is one of the core technologies of MSN Search engine [25].

In 2006, He Qing, an associate researcher of the Intelligent Information Processing Laboratory of the Institute of Computing Technology, Chinese Academy of Sciences, established a noise removal strategy for web page data cleaning based on the interference of web page noise data on web mining. This strategy established a noise discrimination model based on HTML markup information and the noise discrimination model based on data redundancy and the noise discrimination model based on the information entropy of the marked text perform noise removal. In 2010, Xiangfeng Meng, Air Force Radar Academy, proposed a similar repeated record detection method based on genetic neural network in order to effectively solve the detection problem of similar duplicate records in the field of web data cleaning. This method calculates the similarity between the corresponding fields of two records, builds a neural network-based detection model, uses genetic algorithm to optimize the weight of the network model, and uses genetic neural network to combine similarities on multiple fields to detect similar duplicate records. In 2017, Associate Researcher Jiangong Song of Beijing University of Aeronautics and Astronautics proposed an Internet information processing method for earthquake emergency response, and designed a set of web cleaning algorithm EESRC, which defines the priority of data attributes and calculates the corresponding weight; the data is weighted. The highest attribute fields are clustered to form several small data sets. The LP algorithm is used to cluster and merge them into a set, and then all the classes are individually matched for multiple times. For repeated records in different groups, choose to include high relevance to topic words. The attribute value records select the most recent web page update time and keep them, and delete the rest to generate a complete data set without duplication [612].

Earthquake emergency network information provides strong data support for the acquisition and analysis of earthquake disasters. However, there are few researches on the cleaning of “dirty data” related to earthquake emergency networks. Although some experts have conducted some research on data cleaning and analysis in the seismic field, they all have certain limitations, especially after earthquake emergencies. The need for earthquake emergency needs with high timeliness requirements, the existing related research results are even more stretched, losing the advantage of massive data, making the role and value of data greatly discounted, and not being able to provide well for emergency decision-making by relevant departments support.

Therefore, artificial intelligence technology is applied to the earthquake emergency network data cleaning, and the artificial intelligence + earthquake emergency network information data cleaning framework is constructed. First of all, according to the disaster assessment report issued by the earthquake emergency expert group, based on the earthquake emergency information corpus and the machine learning accuracy and recall rate evaluation algorithm, construct the data cleaning quality evaluation index; second, the earthquake emergency network Web document dynamically collected by the web crawler technology Data-based, based on HTML structural features, based on TF-IDF Algorithm and Information calculation model, Design the weighted frequency algorithm of earthquake emergency network Web document P-TF-IDF, Filter out single words with low frequency, optimize FastText algorithm N-gram. The feature word vector realizes the rapid text classification of the earthquake emergency network Web document data; third, the data pattern of each attribute of the text classification data is used to construct missing data recognition rules, data classification rules, and data repair rules to detect invalid data in the data set Value, complete data comparison and redundancy judgment, clean up data conflicts and data errors, and generate a complete data set without duplication. The data cleaning framework not only completes the integration of earthquake emergency network information, but also provides a data basis for the visualization of earthquake emergency data, so that massive earthquake data can give full play to its value and significance and provide effective and sufficient data guarantee for decision support to minimize the loss of people’s lives and property.

The first section of this paper introduces the background of the project and the advanced concepts and basic structure of the articles of experts in related technical fields. Section 2 takes data cleaning technology as a clue and introduces how to improve the classification speed of data classification model without losing accuracy. Section 3 introduces the technical process of data cleaning framework for earthquake emergency network information. Section 4 introduces the evaluation index of data cleaning quality and P-TF-IDF optimization algorithm and compares and analyses the data cleaning framework of earthquake emergency network information in many ways with the data cleaning rules. Section 5 puts forward some issues and summaries worth further discussion in the data cleaning framework of the current earthquake emergency network information.

2. Data Cleaning Technology

2.1. Data Cleaning Technology Based on Artificial Intelligence

Data cleaning is the process of identifying and removing “dirty data” in the data. Take the network data cleaning process as an example, obtain the Web document data set M through dynamic web crawler technology, and combine the data quality requirements R; after the corresponding data cleaning, find the Web document data set M′, M′ not only meets all the data quality requirements, but also minimizes the time consumption during the cleaning process and the cleaning cost. In different data application fields, the data quality requirements R are used in different forms. In the earthquake emergency network data cleaning requirements R′, it is necessary to ensure the timeliness, completeness, and correctness of the data content. Timeliness is mainly reflected in 5 minutes, 10 minutes, half an hour, one hour, 2 hours, and other time nodes after the earthquake occurred. The relevant information of the earthquake event should be continuously summarized; the completeness is mainly reflected in: Web document data. The main data information in the set M′ cannot be empty or repeated, and the relevant data meets the constraints of the earthquake emergency data type; the correctness is mainly reflected in: avoiding wrong information in the Web document data set M′ during the emergency rescue process to prevent it from causing wrong emergency decision. It is not difficult to see that the core of earthquake emergency needs is to obtain more accurate data in a relatively short period of time to serve emergency decision-making.

The core of the earthquake emergency data cleaning process is the data classification process. Under the requirements of the seismic emergency network data cleaning requirement R′, the web document data cleaning is mainly reflected in how to complete the data cleaning in less time and consume the least resources. There are a variety of machine learning classification algorithm models used in the stage, including BP neural network, k-nearest neighbor (KNN), decision tree (Decision Tree), support vector machine, and FastText. BP neural network has a strong advantage in nonlinear mapping, but in order to ensure stability, its learning efficiency is low, and the model training speed is slightly slower than the other algorithms; the k-nearest neighbor (KNN) algorithm has higher accuracy, but its Computational complexity and space complexity are also high, and resource consumption is high; Decision Tree (Decision Tree) Computational complexity lower and better comprehensibility, but excessive phenomena will occur in the matching process; support vector machines are better to implement, but the classification accuracy is poor; FastText supports fast text classification and is compared with the above classification algorithms. To speed up the training speed and test speed, you can use N-gram to train the word vector yourself. Since the earthquake emergency network data is mostly Web documents, the earthquake emergency network information data cleaning framework adopts FastText Research on the algorithm model [1316].

2.2. FastText Model

FastText is a machine learning model open source by Facebook in 2016. The model can be applied in the direction of effective text classification of supervised learning, and it can also be applied in the direction of word vector learning of unsupervised learning, and it is mainly used in the direction of text classification. FastText uses Hierarchical SoftMax and N-gram to optimize the model. It quickly improves the model training speed and test speed while ensuring high accuracy, and its performance is not inferior to deep learning. The FastText model is mainly divided into three layers, the input layer, the hidden layer, and the output layer. It uses text as the input layer data, and finally the output layer outputs the classification corresponding to the text. The architecture of the FastText model is shown in Figure 1.

The process of FastText model for text classification is as follows: enter the relevant word sequence (where n is n words) of the text to be classified in the input layer, and then use the word vector model N-gram (Here, N is a multi-element model, commonly used are binary Bi-Gram and ternary Tri-Gram) for the word sequence converted into feature word vectors, and then these feature word vectors are accumulated and averaged, and then mapped to hidden layer, go through Hierarchical SoftMax Perform normalization processing, and finally the FastText model outputs the predicted text classification label. FastText model uses Hierarchical SoftMax and the complexity has been reduced from N to log N, so how to improve N-gram in the execution efficiency of the model is particularly important.

2.3. N-Gram Model

N-gram is an algorithm based on statistical language models. Among them, N is a multi-element model. When N = 1, it is a single-element model, when N = 2, it is a dual-element model, and when N = 3, it is a three-element model, etc. The two-element Bi-Gram model and three Meta’s Tri-Gram model are commonly used, such as “The earthquake happened,” when expressed in word granularity, there are only three words in its vocabulary: “the,” “earthquake,” “happened.” At this time, the number of gram represented by the binary Bi-Gram model is , namely, “The The,” “The earthquake,” “The happened,” “earthquake The,” “earthquake earthquake,” “earthquake happened,” “happened The,” “happened earthquake,” “happened happened.” When N in N-gram is a specific value, as shown in Table 1.

As shown in Table 1. Gram number statistics table, when the number V of the vocabulary is large, there are more feature word vectors that need to be calculated, which can be used in TF-IDF. The algorithm calculates the frequency of words and sorts them, removes some less frequent words, reduces the number of vocabulary V, and improves N-gram in the execution efficiency of the model.

2.4. TF-IDF Algorithm

TF-IDF algorithm is to calculate the word frequency (TF) and the inverse document frequency (IDF the product of) is a statistical method. By counting the frequency of a certain word in a file, it describes the importance of the word in the file and returns the frequency value of the corresponding word. When there are multiple words, the TF-IDF algorithm returns the frequency value of all words. Using TF-IDF, the algorithm calculates the word frequency values of all words in the earthquake emergency network Web document collection, filters out words with smaller word frequency values, and uses important words as the input of the N-gram model to perform the feature vector conversion. Although using TF-IDF the algorithm optimizes the N-gram input, which can achieve the speed improvement of the FastText model, but the location factors of keywords such as title, keyword, summary, and hyperlink in the Web document collection of earthquake emergency network are ignored. The location attributes of related words are missing, and the importance of words in the current collection of earthquake emergency network Web documents cannot be fully expressed, resulting in a decrease in the accuracy of label classification. Therefore, this article is based on TF-IDF Algorithm, design the location-weighted frequency algorithm of the feature words of the earthquake emergency network P-TF-IDF. Optimization is performed to ensure that the accuracy is not lost under the premise of improving the classification speed of the FastText model.

3. Technical Route

The main process of earthquake emergency network information data cleaning: according to the disaster assessment report issued by the earthquake emergency expert group, based on the earthquake emergency information corpus and the machine learning accuracy and recall rate evaluation algorithm, the data cleaning quality evaluation index is constructed, and the data is dynamically collected based on the web crawler technology. Earthquake emergency network Web document data, refer to HTML structure characteristics, set different positions in the Web document (<title>The title within the label,<meta the abstract described in name = “description” content = “”>, <meta name = “keywords” content = “”>Keywords described in, hyperlinks between <a> and </a>, paragraph text between <p> and </p>, etc.) the position weight of related words W, based on the TF-IDF algorithm to improve the word frequency factor Parameters, combined with the information calculation model to improve the location weight W, design the location-weighted frequency algorithm P-TF-IDF of the characteristic words of the earthquake emergency network, filter out the words with less word frequency in the earthquake emergency network Web document data, and filter the Web with good word frequency. The document data is used as the input of the FastText model, and the N-gram feature word vector in the FastText algorithm is optimized to realize the fast text classification of the seismic emergency network Web document data, and generate the seismic emergency network Web document classification data with Label. After the data recognition rules and data Classification rules, data repair rules, detect invalid values in the data set, complete data comparison and redundancy judgment, clean up data conflicts and data errors, and generate a complete data set without duplication. Finally, the data cleaning quality evaluation index is used to evaluate the quality of the data after the earthquake emergency network Web document cleaning. The data cleaning framework not only completes the integration of the earthquake emergency network information, but also provides a data basis for the visualization of the earthquake emergency data, enabling massive seismic data. It can give full play to its value and significance, and provide effective and sufficient data protection for decision-making support to minimize the loss of people’s lives and property. The main process of data cleaning for earthquake emergency network information is shown in Figure 2.

4. Seismic Emergency Network Information Data Cleaning Framework

4.1. Data Cleaning Quality Evaluation Index

The quality of earthquake emergency information is the key to the success or failure of the earthquake emergency process. Therefore, it is necessary to establish a standardized evaluation standard to evaluate the quality of data cleaning. Based on the disaster evaluation report produced by the earthquake emergency industry as the evaluation index, the data cleaning quality evaluation index is constructed based on the earthquake emergency information corpus and the machine learning accuracy rate and recall rate evaluation algorithm. Cleaning quality mainly refers to the accuracy and comprehensiveness of cleaning. Commonly used evaluation indicators include accuracy, recall, and comprehensive evaluation index F value [2], as shown in formula.

When using the accuracy rate, recall rate, and comprehensive evaluation index to measure the accuracy of data cleaning, improve the earthquake emergency information corpus, that is, given a database K instance with mode R, and a database instance that meets all data quality requirements . The specific process is as follows:(1)Based on the disaster assessment report issued by the earthquake emergency expert group, use basic data models such as post-earthquake building damage assessment, damage quantity assessment, natural environment change assessment, secondary disaster investigation, disaster area calculation, rescue resettlement selection and setting R establish a statistical table of casualties, building damage, lifeline engineering damage, secondary disaster damage statistics, special engineering damage statistics, and improve the earthquake emergency information corpus K.(2)According to the disaster assessment report issued by the earthquake emergency expert group, fill in the earthquake emergency information corpus K, and deliver it to experts in the earthquake emergency field to verify each piece of data, and finally modify it to generate a clean data generation example database.(3)Based on the earthquake emergency information corpus, use the earthquake emergency information structured data cleaning framework for data cleaning, and fill the cleaned data into the database instance. Use and to calculate the corresponding accuracy rate as formula (2), the recall rate is as shown in formula (3).

Bring the accuracy rate equation (2) and recall rate equation (3) into equation (1):

After the data cleaning is completed, the data cleaning quality evaluation index is used to evaluate the quality of the data after the cleaning of the earthquake emergency network Web document.

4.2. Design P-TF-IDF Algorithm and Optimize FastText Model

Based on the web crawler technology to dynamically collect the seismic emergency network web document data, after preliminary data preprocessing, remove the noise data such as advertisements, refer to the HTML structure features, and structure the title, keywords and abstracts, text, hyperlinks, and other data. The text retains the <p> tag. Design the word frequency optimization algorithm P-TF-IDF for location information, including word frequency factor and location factor.

4.2.1. Word Frequency Factor

In the entire Web document, the frequency of words can be divided into high and low. Words with high frequency can better reflect the main purpose of the document. However, the proportion of the number of words in the document to the total number of words cannot fully meet the current needs of Web document analysis. For Web document analysis needs, this article constructs word frequency factor function based on TF-IDF algorithm as formula:where is the word frequency factor of the word in the entire document, is the word frequency value of the word in the text, is the number of titles in the entire document, is the number of natural paragraphs in the entire document (a paragraph between a <p> and </p>), is the number of hyperlinks in the entire document, and is the sum of the number of titles, natural paragraphs, and hyperlinks that contain words in the text.

4.2.2. Location Factor

In view of the relevant characteristics of the earthquake emergency network Web documents (the positions of some important words in the Web document titles, keywords, abstracts, hyperlinks, etc. are shown in Table 2. Position factor description table), it loses the position attributes of related words. It is unable to fully express the importance of words in the current collection of earthquake emergency network Web documents, resulting in a decrease in the accuracy of label classification.

In order to better calculate the frequency value of words in Web documents and better reflect the words in different positions in Web documents, the following is based on the information calculation model H proposed by Wiener (one of the founders of information theory) as in formula (6) [17], and the structure can be as the information content function describing the word boundary as shown in formula (7), and the realization of the position factor function is as formula (8).

In formula (6), H represents the “information amount;” P represents the “probability of the occurrence of this event among all possibilities;” and M represents the “total number of all the possibilities of this event.”

Equation (7) represents the amount of information of the word in different positions; represents the frequency of the word in the Web document, represents the paragraph frequency of the word in the Web document, and represents the word length of the word , where , , and l are both greater than or equal to 1.

Equation (8) represents the position factor of the word in different positions, where and represent the information amount coefficient at different positions in the Web document. After a lot of experiments and the coefficient value setting is as shown in Table 3. Web document different position information coefficient and value.

Based on the above-mentioned word frequency factor and location factor, the P-TF-IDF function of the word frequency optimization algorithm is constructed as the following equation.

Among them, represents the word frequency value of the word in the Web document, represents the word frequency factor value of the word in the Web document, and represents the position factor value of the word in the Web document. Through the function , the word frequency value of all words in the Web document can be calculated and removed. Part of the words with small word frequency values are classified, and the remaining words are classified as the input of the FastText model of the classification model to generate tagged seismic emergency network Web document data, which provides a data basis for further completing the Web data cleaning.

4.3. Establish Data Cleaning Rules

After classification based on the FastText model, the tagged seismic emergency network web document data is established to identify missing data, data classification rules, and data repair rules, detect invalid values in the data set, complete data comparison and redundancy judgment, and clean up data conflicts, Data errors, and generate a complete data set without duplication.

4.3.1. Missing Data

Missing data means that some attribute values in the database instance are missing or contain invalid values. Missing data recognition Rule: for invalid values beyond the property field, invalid values in the data set are detected by the data mode of each property. For example, in the statistical table of destruction of secondary disasters in Table 4 [Type] is “N/A.” It is worth noting that [Range of influence] is also a missing value because “Null” is not a valid expression of the scope of influence. The detection of missing data is relatively simple. It only needs to check whether the attribute value that is not allowed to be empty is empty, “Null” Or “N/A.”

4.3.2. Data Redundancy

Data redundancy means that the same data appears multiple times in a database instance, that is, there is duplication between data. Establish data classification rules: group according to tags, query redundant data in the same group, and select the last record according to the data release time. For example, tuples T1 and T3 in the statistical table of secondary disaster damage . They all represent the secondary disaster damage information of Jiuzhaigou Scenic Area, Sichuan, and the last record of the time is selected for preservation.

4.3.3. Data Conflict and Data Error

There is a data conflict between the two pieces of data that cannot satisfy the integrity constraint. Establish repair rules, as shown in formula (10), where: X is a subset of attribute set Q, is a possible value of attribute set X, representing the evidence value; attribute Y belongs to Q\Y, is a set of constants in the domain of Y, representing the wrong value of Y, and belongs to the domain of Y, but does not belong to means the correct value of Y.

Take Table 4 damage statistics of secondary disasters as an example, Q = {“Place name,” “Types of,” “Destruction,” “Sphere of influence”}, X = {“Place name”}, Y = {“Types of,” “Destruction,” “Sphere of influence”}; Settings  = {“Jiuzhaigou Scenic Area, Sichuan”},  = {(“landslide,” “severe,” “null”), (“landslide,” “null,” “null”), (“Barrier lake,” “null,” “null”)…},  = {“Barrier lake,” “severe,” “Jiuzhaigou Scenic Area”}, When the data tuple t = T2 when using repair rules for data cleaning, is equal to , , T2 is the noise data, delete the noise data T2 .

After classification based on the FastText model, generate tagged seismic emergency network web document data, data identification rules, data classification rules, and data repair rules, and perform data cleaning to generate a complete data set without duplication.

4.4. Cleaning Results and Analysis
4.4.1. Data Preprocessing

To better verify the cleaning effect, select “Yushu Earthquake,” “Tibet Earthquake,” “Wenchuan Earthquake,” “Jiuzhaigou Earthquake,” “Ludian Earthquake”, and other major earthquakes after the earthquake industry dedicated network information, news media, etc. Website 13,644 Web data such as HTML, used for FastText model training and test, go through data preprocessing.Pretreatment data as is shown in Table 5.

4.4.2. P-TF-IDF Algorithm Optimization FastText Model

For 13644 web data such as HTML after preprocessing, use P-TF-IDF's word frequency factor and location factor, compute network Web Data per word of Weights data value, and filter according to the weight data value lose partial word frequency and location nun important of Words, and then perform model training, and the training results are shown inTable 6. The comparison of the accuracy rate of different numbers of words input after filtering according to the weight is shown in Figure 3, the comparison of the training time of different numbers of words input after filtering according to the weight is shown in Figure 4.

4.4.3. Data Cleaning

After optimization using P-TF-IDF algorithm, FastText model data is classified and carried out in accordance with the rules of the data cleaning framework, using “Yun nan Da li Yang bi earthquake” 15871 strip data and “Qing hai Guo luo zhou Ma duo earthquake” 7961 strip data conduct cleaning, combined data cleaning quality evaluation index F conduct a comprehensive evaluation, conducted an evaluation and comparison, the results are shown in Table 7.

The experimental results show that through the P-TF-IDF algorithm optimization FastText model to take the Bi-Gram model as an example, when filtered out 50% Words (input 50%, word quantity for 91711). Compared with the case where word frequency factor and position factor are not added (input 100%, number of words is 191297), the time is 7.57 seconds faster and the accuracy rate is increased by 0.76%; compared with the case where 25% words are filtered (input 75%, number of words is 132660), the time is 2.83 seconds faster and the accuracy rate is increased by 0.76%; compared with the case where 75% words are filtered (input 25%, number of words is 47071), the time is 6.83 seconds faster and the accuracy rate is increased by 0.71%. This article chooses Add to Word frequency factor and location factor, and 50% of words are reserved for model training, which guarantees a high accuracy rate under the premise of effectively improving the model training time. Using the earthquake emergency network information data cleaning framework, 15871 pieces of data of “Yun nan Da li Yang bi earthquake” and 7961 pieces of data of “Qing hai Guo luo zhou Ma duo earthquake” are cleaned. From the results, it is not difficult to see that only the word frequency factor is used to optimize the model. Compared with the traditional FastText model for data classification, the time of “Yun nan Da li Yang bi earthquake” is shortened by 1.98 seconds. The comprehensive evaluation index increased by 13.08%, the time of “Qing hai Guo luo zhou Ma duo earthquake” was shortened by 1.94 seconds, and the comprehensive evaluation index increased by 15.41%. After adding the location factor, although the time increased slightly, the comprehensive evaluation index of “Yun nan Da li Yang bi earthquake” increased by 18.21%, and the comprehensive evaluation index of “Qing hai Guo luo zhou Ma duo earthquake” increased by 19.06%. In the location factor, different locations have different impacts on the comprehensive evaluation indicators of the data cleaning framework. Titles, hyperlinks, and paragraph text have a greater impact. It is not difficult to see that some websites have abstracts and keywords described in HTML. But inconsistent with the theme of the Web document, the abstract and keywords have little effect on the accuracy of cleaning.

5. Conclusion

In the context of the advent of the big data era and the development of emerging technologies, the integration of artificial intelligence technology and earthquake emergency network information ensures the comprehensiveness and integrity of the data, so that the earthquake emergency work can be comprehensive and accurate in the shortest time, timely disaster information, make scientific and reasonable decisions and put them into action. This article is based on Web document data of earthquake emergency network dynamically collected by web crawler technology, and use TF-IDF Algorithm design word frequency factor, use information calculation model to calculate the amount of information of each word, describe the boundary of each word through the amount of information and design location factor, combining word frequency factor and location factor, and designed a Location Weighted Frequency Algorithm P-TF-IDF in Web Documents, P-TF-IDF reduces the number of words in the input layer of n-gram in FastText model and improves the classification efficiency of FastText model and outputs labeled classification data, using data identification rules, data classification rules, and data repair rules to complete data cleaning. Under the premise of ensuring the cleaning effect of the earthquake emergency network information data cleaning framework, the calculation time is shortened. The cleaning framework is suitable for loosely structured Web document page data and other text data with standardized formats and has certain practicability. Currently, the data cleaning framework is only a preliminary cleaning of the earthquake emergency network web documents, and the cleaning results need to be verified multiple times to check the correctness of the algorithm. Therefore, further research is needed.

Data Availability

No data were used to support this study.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This work got Earthquake Science and Technology Spark Program Youth Project Research support for special funds (XH21057Y) for business expenses.