Abstract

Using 100 random questionnaires, 89 people said they were increasingly dependent on the Internet for information, while the remaining 11 said they were not dependent much. Nowadays, the development of Internet technology is more and more mature, and the scale of the Internet is more and more large. At the same time, with the gradual deepening of global economic integration, English has become one of the indispensable language methods for international communication and cooperation. The development of network technology has been applied more and more widely in the process of English teaching; especially, the construction, research, and practical application of corpus have ushered in a broad development prospect. Based on web crawler technology, this paper focuses on the construction of web English corpus, which lays a foundation for English learning. Experiments show that crawler technology can effectively solve the collection and recognition of big data in English corpus.

1. Introduction

In recent years, with the in-depth development of my country’s education reform, Chinese higher education is developing in the direction of international trade, communication, and information technology. The transformation and development of higher education has begun to pay attention to the use of information network media. The university websites of Chinese universities are gradually becoming more and more important. At the same time, most college websites have both Chinese and English versions. College English websites have become an important medium for their outreach, exchange programs and partnerships, finding different students and countries, and raising their profile globally [1]. There is no doubt that English language websites play an important role in international educational exchanges and participation, international school participation, and international student education. At present, most Chinese universities have opened English websites. Since then, good translation, good publicity, good communication, website construction, and internationalization of college English websites have also become new concepts studied by scholars. In this context, the development of online English corpus is particularly important. Therefore, this paper focuses on the development of online English corpus with the help of web crawler technology. On this basis, a web crawler graph is first created [2], as shown in Figure 1.

2. Literature Review

Buts and Jones said that before the big data revolution, the government and enterprises could not save all data for a long time, nor could they efficiently manage and analyze such a huge data set [3]. Ukraine said that under the traditional technology, the data storage is limited, the management is backward, and the cost is expensive [4]. In a big data environment, the most powerful new process is the collection, cleansing, and analysis of multiple files to ensure efficient, flexible, and efficient operations. According to Wang and He, from the government, industry, and all walks of life, big data becomes important for them to see new ideas and provide self-help [5].

At present, Azazil has few research results on the application of corpus to college English websites [6]. Elgibreen et al., a corpus-based comparative analysis of English profiles of professors on English websites of Chinese and American universities, conducted a study on English profiles of Chinese and American professors from the perspective of evaluation theory [7]. In order to study the distribution of evaluation resources in the introduction discourse of English website professors in Chinese and American universities, Zhang et al. built a corpus and made a comparative analysis of the distribution of evaluation resources in the three subsystems of attitude, judgment, and grade difference under the framework of evaluation theory [8]. Kim and Davies selected 18 English profiles of Chinese and American university websites and built a micro corpus to study the public selection of high-frequency words of English profiles of Chinese and American university websites based on corpus [9]. Ding et al. conducted a corpus-based study on the core theme words of English profile of Chinese and American college websites. From minimal translation selections to full texts, these studies have started using the tools used to develop college English websites and completed some studies [10].

This paper makes a preliminary discussion on the translation of school website from the perspective of translation work. There are great differences in language and culture between English and Chinese college web pages. A full understanding of these differences is of great significance to the English translation of profiles. It is proposed that interpretative addition, modification or reorganization, and zero translation can be used as effective strategies for the translation of web profiles. Using the methods of case study and comparative study, this paper analyzes the English language news updated by four “985” colleges and universities in a city in 2021 and discusses the difficulties and solutions encountered in the network communication of college English websites. The audiences of college English websites mainly include students from various countries, overseas media, and brother colleges [11]. English websites must meet the needs of the above three groups of people in order to attract audiences and achieve the expected communication effect; this paper analyzes the problems existing in the translation of college English website propaganda and puts forward countermeasures and suggestions for the construction of college English websites from the following two aspects: on the one hand, through the comparison of parallel texts, analyze the similarities and differences in language and structure of global college English websites, and find out possible language translation errors, cultural translation errors, and functional translation errors; on the other hand, taking the text of college English website as the corpus, this paper investigates the translation initiator, translator, and audience and makes appropriate adjustments to the content and presentation of the source language, so as to make the translation meet the needs of the audience and make the college English website really play the role of external publicity.

3. Method

3.1. Overview of Web Crawler Technology

A web crawler (also known as a web spider and web bot) is a program or script that receives information from around the world through certain websites under certain laws. Web crawlers can skip a website’s standard hyperlinks to search and store information. It starts with one page of the website, reads the content of the page, looks at other hyperlinks in the page, and then sees the next page through these hyperlinks [12]. Continue until all web pages on the Internet are captured. The broad classification of web crawler technology is shown in Figure 2.

Crawling means moving slowly in one direction. Technically speaking, web crawlers are tools used for data collection in search engines. They are called web crawlers, web spiders, or web robots. With the continuous development of technology, web crawlers are becoming more and more mature, which has gone beyond the definition of just a tool for search engines to collect data [13]. Generally speaking, a basic web crawler should have a set of seed URLs as input and a set of crawled web pages as output. The specific workflow is as follows: (1)Firstly, several initial URLs are selected as the starting position of web crawler according to the target(2)Put the seed URL in the URL line to get it(3)Read the URL through the URL line, resolve the DNS, and get the IP of the host. To pull the web page relative to the URL, identify the desired page, or remove the URL from the web page and place it on a line with the entry, and provide an entry for the URL in the URL field(4)Compare the URL extracted from the web page with the crawled URL to remove the duplicate. Finally, put the deduplicated URL into the URL to be crawled, enter the cycle, and stop the web crawler when the stop condition is reached

3.2. Framework Comparison

Web crawler is the core tool for big data industry to obtain data [14]. Generally, when it comes to the collection task with small amount of data and simple capture logic, you can use self-made web crawler code for data collection. However, when it comes to the collection task with large amount of data and complex capture logic, if we continue to use the way of self-made web crawler for data collection, it will greatly increase the code development time, thus increasing the cost and time of the project. At present, there are many web crawler frameworks in the market, and different programming languages correspond to multiple web crawler frameworks. Table 1 gives some open source web crawler frameworks corresponding to programming languages in the current market.

Although there are many web crawler frameworks in the market at present, there are not many open source web crawler frameworks that are popular with developers and often used. The specific comparison can be seen in Table 2.

Scrapy is also an open source web crawler framework developed by python programming language [15]. Scrapy framework crawler has extremely powerful functions, high crawling efficiency, many related extension components, and very high degree of configurability and scalability and can flexibly customize crawling data. Scrapy can be used to download websites and extract output files from files without any issues on the page (Scrapy also provides users with options (links more like lxml) and cannot complete HTML code and simply delete files) and does a good job. For data mining, monitoring and automated testing, Scrapy can be easily customized as needed. Scrapy can be easily customized as needed. It supports the generation of crawler files from built-in definition templates, speeds up the creation of crawler code, and ensures that the code remains unchanged in large projects. In addition, scrapy also provides a variety of data export formats (JSON, CSV, XML, etc.), which can facilitate the connection with the database and data transmission in the project pipeline. The scrapy web crawler framework community has a large number of people and complete documents. It can deal with almost all current anticrawling websites. It is the most widely used web crawler framework in Python at present, as shown in Figure 3.

GitHub sets three options for open source project code: watch, star and fork, which can generally be used to indicate the activity and attention of the open source project code. The following table shows the comparison of watch, star and fork data of GitHub, an open source web crawler framework commonly used in the market so far (in descending order of star number, see Table 3) [16].

3.3. Algorithm Comparison

The most commonly used feature selection algorithms are data frequency, data gain, mutual data, and access statistics.

3.3.1. Document Frequency

The number of documents containing a word in corpus training is the frequency data for that word. The basic idea of this method is that words with low frequency often carry little information, so they cannot distinguish the categories well. Therefore, words with low frequency can be deleted, which can not only reduce the feature dimension but also improve the accuracy of classification.

3.3.2. Information Gain

Information gain () is to calculate the difference between the amount of information carried by the system when a feature appears and does not appear. For text classification, the difference between document frequency with and without feature word represents the value of feature word . value adopts the following formula, as shown in

3.3.3. Chi Statistics

Chi statistics is often called square statistics, which is used to test whether two variables are independent. On the premise that the two variables are independent of each other, the deviation degree between the actual observed value and the theoretical value of the sample is calculated and expressed as chi value [17]. The larger the chi value, the two variables tend to be correlated; on the contrary, the two variables tend to be independent. The correlation between feature words and document categories can also be measured in this way. First, assume that the entry is independent of a category. The larger the chi value of the entry calculated on this basis, the greater the deviation between the result and the assumption, and the more relevant the entry is to the category. Therefore, the process of feature selection in this method is to calculate the chi value of each entry and category and sort it from large to small, and the top value is the feature. The calculation formula of chi value of word for category CI is shown in

In the model, represents all the data in the body, represents the data that contains and is , represents the data that contains but does not contain , represents the number of data that does not contain but belongs to , and represents no word or data from .

3.3.4. Mutual Information

In data theory, interpersonal data (MI) refers to the amount of data provided by two events. The greater the amount of mutual information, the greater the correlation. The mutual information calculation formula of word and category is shown in

The meaning of variables in formula (3) is consistent with that in formula (2).

3.4. Final Feature Extraction

Based on the candidate feature set, extract the final feature set. The specific steps are shown in Figure 4. (1)Calculate the information gain value. The information gain value of each feature in the candidate feature set is calculated in the training corpus(2)Get features. According to the decreasing order of information gain value, select some of the top features(3)Get the final feature. Add the features obtained from the entity information to the feature set obtained in step (2), and take this feature set as the final feature set

3.5. Classifier Based on Naive Bayesian Algorithm

The classifier is constructed by naive Bayesian algorithm [18]. Suppose that each instance is represented by a feature set, and class takes a value from a finite set . A training example set and a test example are provided.

The target of the instance to be classified is to obtain the class tag () of the instance , as shown in

What we need to do now is to estimate the two probability values in equation (4) based on the training example set. Naïve Bayes classifiers assume that attribute values are conditionally independent of each other when a class tag is given. That is, the joint probability is exactly the product of each individual feature probability. The specific formula is shown in

Substituting into formula (4), the classification formula of naive Bayesian classifier can be obtained, as shown in

It can be simply calculated by calculating the occurrence frequency of different classes and eigenvalue combinations in the training example set. The specific formula is shown in

Obviously, this approach leads to an underestimation of the results when the value of the zero-frequency property is present. In more severe cases, some values will be 0, making all numbers calculated by equation (6) to be 0. Laplace estimation is often used for smoothing to avoid the above problems. Equations (7) and (8) are rewritten, as shown in

The workflow of naive Bayesian classifier based on entity link is as follows: (1)The feature extraction method based on entity link is used for feature extraction [19](2)According to the obtained feature set, a naive Bayesian classifier is constructed and trained(3)Preprocess the web page captured by the crawler, including topic information extraction, word segmentation, and other preprocessing, and then, quantify the web page(4)The classifier is used to recognize the theme of the web page after vectorization processing. If the web page belongs to the theme class, the web page is saved to the theme page library; otherwise, the page is discarded [20]

4. Results and Analysis

For the evaluation of subject recognition effect, three indexes are mainly used: accuracy (P), recall (R), and F value. Accuracy is the proportion of the number of texts related to the topic accurately identified; recall rate is the ratio of the number of texts related to the topic accurately identified to the number of texts related to all topics in the training set; F value is a comprehensive evaluation index [21]. Suppose: in the training corpus, the text related to the subject and determined to be related to the subject is , the number of texts unrelated to the subject but determined to be related to the subject is , and the number of texts related to the subject but not determined to be related to the subject is ; then, the calculation formula of the three evaluation indexes is shown in

Accuracy:

The recall rate is shown in

The F value is shown in

A total of 1443 military (587) and nonmilitary (856) articles were selected from Sogou news corpus as training corpus. The method based on entity link proposed in this chapter is used to construct naive Bayesian classifier for experiment [22]. The experimental results are shown in Figure 5.

From the experimental results, compared with the traditional naive Bayesian classifier, the introduction of entity link technology to improve it can achieve better results. As the entity link-based topic recognition algorithm can achieve good recognition effect, it lays a technical foundation for the construction of English corpus, because only efficient and accurate corpus recognition can build English corpus with richer content and more complete functions.

Corpus collection is the first step in the construction of a special English corpus for college English websites. In order to connect with international famous universities as soon as possible, improve the internationalization of Chinese university English website construction, standardize the classification system of Chinese university English website construction, and further improve the information content of Chinese university English website, we should focus on the current international first-class university English website construction, learn advanced experience, collect relevant corpus, and classify as a whole. So as to improve the Chinese expression of Chinese English and Chinese College English websites.

Use Baidu Encyclopedia, Wikipedia, and global college websites to collect information on the top ten colleges and universities in the United States. Classify and sort out the source language (English) texts of global university websites. Taking the English texts of global university websites (including website introduction, school brochures, and English version of teaching materials) as the research object, classify and sort out the corpus of Chinese and American university websites according to common terms and special terms, and construct “English original university website corpus” and “Chinese university English website corpus,” respectively. This paper takes the special English of global college English websites as the research object, which mainly includes the Chinese and English versions of special English in two categories and ten subcategories of global college English websites. It mainly includes the Chinese and English versions of special English in two categories and ten subcategories of global college English websites. Each group of special texts should be no less than 10,000 words (English), and each special character text should be no less than 2,000 words (English). The specific classification is shown in Table 4.

From the results, a major problem in the development of English websites in Chinese colleges and universities is that there are too many Chinglish and Chinglish languages, and the language is incorrect. There are many news publicity materials and few practical application materials. There are many propaganda terms for Chinese audiences, but few actual contents for international scholars and international students. In particular, the construction of English websites of subordinate colleges and departments of Chinese universities, English websites of research institutions, and English websites of international students lags far behind that of the official website of Chinese universities.

On the basis of corpus collection, this paper describes the language characteristics of the English texts of “English original university website corpus” and “Chinese university English website corpus,” respectively, and summarizes the unique language characteristics of the two kinds of corpus. Based on the two corpora established by the project, the world original vocabulary and vocabulary with Chinese characteristics are produced. The characteristics of vocabulary, syntax, and stylistic structure of college English websites are studied. Then, according to the characteristics of vocabulary, syntax, and stylistic structure of English websites, this paper explores typical sentence patterns and translation skills in the construction of college English websites. Through the comparative study, we find the similarities and differences between the English text of college websites translated from Chinese and the original English text and then put forward the similarities and differences between the two functions, so as to provide a real and objective basis for the subsequent improvement of the language quality of college English websites and the translation research of college English websites. Through corpus description and text analysis, we find that there are major differences between American college English official websites and Chinese college English websites in language style, discourse structure, vocabulary syntax, cultural connotation, and so on.

Generally speaking, the discourse structure of Chinese college English websites is influenced by Chinese discourse, focusing on parataxis and paying attention to the integrity, richness, and literary grace of the discourse. The English discourse of American college websites embodies the characteristics of English, focusing on hypotaxis and focusing on the form and logic of the discourse. Taking the above general introduction of the Tsinghua University and Princeton University as an example, it is not difficult to see that Tsinghua University advances linearly in chronological order, from the initial historical evolution to the introduction of the current school development in the middle and then to the final expression of the future vision and objectives of the school. On the whole, it is close to the structure of relevant Chinese texts on the school Chinese website, and the length is relatively long. The official website of the Princeton University first makes an overall introduction to the school in simple language to give readers a clear impression, and then, the detailed information classification in the website makes it easy for readers to find the practical information they want to know. In other words, a whole long text is divided into several short sub texts, which are introduced by classification, and the language introduced is mostly phrases and short sentences, which are concise and easy to understand.

5. Conclusion

The collection and recognition of corpus data is the foundation of English corpus construction, and web crawler technology has obvious advantages in data collection. Through experiments, this paper proves that web crawler technology can effectively identify and collect English corpus data, thus solving the core technical problems of English corpus construction. In addition, the construction of online English corpus based on web crawler technology is feasible, and it is very efficient in the collection and identification of corpus materials.

Data Availability

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The author declares that there are no conflicts of interest.

Acknowledgments

This work was supported by the fund project: The Phased Achievements of the Provincial Teaching Research Project (Research on School-Based Construction and Practice of Online and Offline Hybrid “First-Class Courses” of College English. No. 2020699) of Higher Education in Hubei Province in 2020 and Key Project of Science Research Plan of Hanjiang Normal University (Research on Information Literacy Based on Mobile Assisted Language Learning, No. XJ2020a01).