Abstract

In order to explore the relationship between intelligent image recognition technology and the mentality and quality of life of the elderly, this paper combines intelligent image simulation technology to identify the behavior of the elderly, protect the safety of the elderly, and provide timely feedback on the adverse conditions of the elderly. Moreover, this paper improves the traditional intelligent image recognition algorithm, verifies the research method of this paper through experimental research, and puts forward corresponding suggestions. Through investigation and research, we can see that the level of health literacy of elderly patients with chronic diseases is low. Therefore, in the future health education, we should strengthen health education for elderly patients with chronic diseases, use different mass media to propagate health knowledge, and promote the formation of healthy lifestyles and behaviors for elderly patients with chronic diseases. At the same time, the experiment also verified that the intelligent image recognition technology proposed in this paper has a positive effect in improving the mentality and quality of life of the elderly.

1. Introduction

With the rapid economic development, all walks of life in our country have taken on a new look; people’s living standards have also been rapidly improved; and the overall development has shown a prosperous situation. However, we also have to admit that in the process of development, there are still many problems that need to be summarized and resolved, such as environmental protection issues, population structure issues, urban-rural gap issues, development balance issues, industrial structure adjustment issues, and so on. Among them, the problem of population structure is particularly prominent, and as an important part of our country’s population, the problems of the elderly are worthy of in-depth study. In particular, with the increasing aging of our country, the resulting problems and actual impacts have become more and more obvious so that social development has been restricted on many occasions. Therefore, it is urgent to solve the aging problem [1].

In today’s society, a common phenomenon is that families are miniaturized, there are more and more only children, and the burden of caring for disabled elderly people is getting heavier. Therefore, children urgently need new technologies and new things that can help them take care of the elderly, and disabled elderly people urgently need new things and new technologies that can assist them in their daily lives. If the control of external devices can be combined with the EEG signals of the elderly, it will not only regain their confidence in life and reduce the burden on children but also make meaningful explorations and attempts for the application and promotion of EEG research and development in real life [2]. Health literacy is the ability of individuals to acquire, process, and understand basic health care knowledge and services needed to make appropriate health decisions. Moreover, as an important part of health education and health promotion, health literacy has gradually attracted attention from domestic and foreign scholars in recent years [3].

The framework of this article is as follows:

With the rapid economic development, all levels of Chinese society have new faces; people’s living standards are rising rapidly; and the overall development is developing in a prosperous environment. But we must also admit that there are many problems in the development process, which need to be summarized and solved, such as environmental problems, demographic structure, differences between cities and network problems of balanced development, industrial structure changes, and other especially important parts of the population problem of the elderly population, worthy of in-depth study. In particular, as our country ages, problems and practical consequences become increasingly apparent, and social development is in many cases hindered. Therefore, the solution to the aging problem is inevitable (1).

A common phenomenon in today’s society is the miniaturization of families, the increasing number of children, and the increasing burden of caring for the elderly. Therefore, children desperately need new technologies and things to help them care for the elderly, while disabled people need new and new technologies to help them get through their daily lives. If combined with the control of external devices and electronic brains of the elderly, these indicators not only rebuild their lives and reduce the burden on the children’s confidence but also can be related to the research and development of electronic brain attempts in real life quality (2). Health is a person’s ability to access, treat, and understand the basic knowledge and services required for appropriate decision-making in health care. In addition, as an important part of health education and health promotion, health self-improvement has gradually attracted the attention of scientists at home and abroad in recent years.

In the past research work, people mostly used traditional sensing methods such as wearable devices and cameras to recognize indoor movements. Among them, the more typical example is the use of accelerometers, gyroscopes, or their fusion system to recognize action behavior [4]. The advantage of this type of wearable device is that it is highly robust to changes in the external environment. Secondly, researchers also use various types of cameras, such as depth cameras, to track and reconstruct human nonrigid motion [5]. This method can capture a large amount of external information, and the upper limit of performance is very high. The investigation in literature [6] shows the rejection of invasive devices by the elderly. Moreover, the camera has high requirements for lighting conditions, is easy to leak privacy, and is not suitable for use in an environment with high privacy requirements such as home. The main principle of using radar for action recognition is the micro-Doppler feature [7]. The action of the target will change the characteristics of the radar echo, and the activities of various parts of the body are different when different actions occur. If the human body is regarded as a collection of a series of reflection points, the overall frequency distribution finally reflected by these reflection points will correspondingly differ around the main modulation band. Researchers used to extract the required micro-Doppler features mainly based on the time-Doppler map (TDM). The most common are artificially predefined physical characteristics, including changes in the time-frequency spectrum envelope caused by the physical characteristics of the action [8]. This physical feature is easily affected by changes in the external environment (such as human height, weight, etc.) and often requires experts in related fields to design, which is time-consuming and labor-intensive. In order to solve these problems, a series of feature definition methods in other similar fields are transplanted into this task, including classic wavelet-based features [9] and Mel-frequency cepstrum coefficient (MFCC) [10] commonly used in speech analysis. The existence of these features simplifies the step of manually defining features but relatively also makes a certain compromise in classification performance. Since the different representations of radar data [11] can also be regarded as a category of generalized images, advanced deep learning techniques in the image field can be used to automatically extract low-dimensional abstract features from it. Literature [12] used an unsupervised pretrained sparse autoencoder to extract features from TDM and achieved an 87% recognition rate in fall detection. Literature [13] used a convolutional autoencoder to replace the traditional sparse autoencoder, borrowed the idea of inception to design the network structure, and obtained a classification accuracy of 94.2% on 12 types of actions. In addition, there is also a more comprehensive use of optimized networks for human-action-recognition-related work [14]. However, one of the problems in these studies is that the signal representation domains used exist in a two-dimensional form, which cannot simultaneously characterize the time, space, distance, and Doppler frequency information contained in the radar signal, and it is bound to lose part of the information during action recognition.

In order to accomplish the goal of behavior inference, the most direct method is to install multiple sensors in the home environment. At present, the existing representative work includes the AAL behavior inference system that uses the depth camera to recognize actions and analyze context information [15] and the scheme of synchronizing the depth camera and wearable inertial sensor data proposed in literature [16]. Literature [17] used a variety of sensors (passive infrared sensors, door frame magnetic strips, and flow sensors) placed in various locations in the house to perceive the behavior of the monitored person in the entire indoor environment. This solution combines the prior knowledge of the sensor installation position to determine the behavior. For example, the doorframe magnetic stripe data can be used to infer the target’s going out behavior, and the bathroom flow sensor data can be used to infer the target’s bathing behavior. At the same time, some work focuses on the use of data mining methods to analyze mixed time series from multiple sensors and establish a person’s daily behavior characteristics through the time changes of several quantitative or qualitative parameters [18]. Although such programs have achieved good results, their respective shortcomings are also obvious. For example, information-rich cameras and wearable sensors can cause privacy and intrusive issues when used for behavior inference and motion recognition. Moreover, the multisensor fusion scheme is very cumbersome to install in practical applications and requires specific protocols to coordinate with each other. This has led to this type of research can only be limited to the laboratory environment and far from being applied and promoted in the actual family [19].

3. Intelligent Image Simulation and Recognition Algorithm Applied in Medical Field

In practical applications, two convolution templates in horizontal and vertical directions are mostly used. Common ones are Sobel, Prewitt, and Scharr’s horizontal convolution template and vertical convolution template . Prewitt operator can suppress the influence of noise to a certain extent, but its detection ability is not strong. The Sobel operator considers that the importance of neighboring pixels to the current pixel is related to the spatial distance of the current pixel. Generally speaking, the closer the distance, the stronger the impact. The Scharr operator also assigns different importance to adjacent pixels according to the spatial distance relationship with the current pixel.

For a given behavioral image , its gradient magnitude can be expressed as follows:

where represents the gradient magnitude of the behavior image I(x, y) at the position (x, y). The direction of the gradient is the direction where the function F (x, y) changes the fastest. When there is an edge in the image, there must be a large gradient value. On the contrary, when there is a smooth part in the image, the gray value changes little, so the corresponding gradient is also small.

We can get the gradient image by computing the gradient of each pixel in the image. and are the horizontal and vertical gradient amplitude components of the behavior image at the point (x, y), respectively. The calculation method is as follows [20]:

The direction where the greatest change occurs in the gradient intensity field is called the gradient direction. From the horizontal and vertical gradient amplitude components, the gradient direction of this point can be defined as follows [21]:

If the D(x,y) of a pixel is different from its neighborhood, it means that the gradient direction of the pixel has changed.

The gradient magnitude of the behavior image can accurately extract the structural features of the behavior image, and the structure information of the behavior image is the human visual.

The principle of Shader is to use the threshold and UV to calculate the degree of transparency and then use LERP control to change the range of transparency, avoiding conditional judgment. Instead of creating a material for each image, you can use a C# script to dynamically create a material for Shader, which is also easier to use, more user-friendly, and easier to animate.

Why does the gradient information, which has been repeatedly successful in the field of natural behavior image quality evaluation, fail in the face of screen behavior image quality evaluation? This is a question worthy of our consideration. Different from natural behavior images, screen behavior images have the following characteristics:(1)The noncontinuous tone content in the screen behavior image is directly generated by the computer without the need for a collection process. Therefore, the area is completely noise-free.(2)Screen behavior images usually contain various forms of content, such as text, graphics, and natural behavior images.(3)The text area in the screen behavior image usually has many sharp edges and a large number of smooth areas, such as PowerPoint, e-magazines, and web pages.

As shown in Figure 1, we counted and compared the statistical characteristics of the gradient amplitude of the screen behavior image and the natural behavior image. It can be observed from the figure that the gradient amplitude probability of the natural behavior image has a high degree of Kurtosis in the logarithmic domain, that is to say, it has a sharp peak at the zero point and a heavy-tailed distribution on both sides of the zero point (heavy-tailed distribution). The statistical characteristics of the gradient amplitude of the natural behavior image can be modeled using a generalized Laplace distribution. In comparison, the gradient amplitude of the screen behavior image exhibits severe oscillations in the “tail” part of the heavy tail distribution.

Although the existing gradient-based behavioral image quality evaluation models have general performance in evaluating screen behavioral image quality, these algorithms have a common defect: they only consider the gradient magnitude information and ignore the effect of gradient direction on behavioral image quality. The importance of evaluation: the gradient magnitude is a scalar, which can only represent the brightness change of a local area of the behavior image, and the gradient direction carries rich spatial information even in a smooth area where the brightness changes slowly.

The gradient direction information, in the form of gradient histogram, effectively solves the problem of object detection in the field of computer vision and digital behavior image processing. However, the gradient direction is important for the evaluation of behavioral image quality. In particular, the value of screen behavior image quality assessment has not been deeply explored. The traditional gradient direction calculation method is shown in formula (4), but this calculation method is unstable. To this end, we design a new gradient direction calculation method based on the local information of the gradient amplitude of the behavioral image. This method can better describe the change of gradient direction characteristics caused by behavioral image distortion (see Figure 2).

As shown in Figure 3, we have considered a total of 12 angles from О to x at equal intervals, that is, , where . Therefore, for a specific screen behavior image, we can obtain 12 gradient patterns D in different directions. The gradient pattern of each fixed direction can be obtained by convolution operation between the convolution kernel L1 in the i-th direction and the gradient amplitude M(x, y) of the screen behavior image [22].

3.1. Test Method

So what is recognition? The so-called image recognition, as the name implies, is to make a variety of image processing, analyze, and finally identify the target we want to study. As we see in many pictures in lif, we humans can recognize planes, people, cars, traffic signs, and so on, and we can do much more with the information we collect.

“Computer image recognition” and “human image recognition” in principle and there is no essential difference, but “human image recognition” information processing is more complex, the computer lack of human perception and visual impact. In fact, we recognize objects not only by the memory of the whole image, but by recognizing the features of the image itself. The first is the image, which is recognized by every kind of characteristic image, but a lot of times we don’t realize this. For example, when we see a rose, how do we know it is a rose? When we see a flower, we can tell from its features whether it has thorns or not, based on what we learned as children from our parents, teachers, books, and so on. Is it shaped like the petals of a rose? What is the color? What is the shape of the leaves? It goes on to identify whether the flower is a rose or some other kind of flower.

Up to now, there are many different technology methods of image recognition, which can be roughly divided into traditional image recognition method and fusion neural network algorithm recognition based on traditional image recognition. Neural network image recognition technology is a relatively new image recognition technology; the neural network here refers to the artificial neural network, that is to say, this neural network is not the real neural network of the animal itself but the artificial creation of a human imitation of animal neural network. In neural network image recognition technology, a deep learning model based on a convolutional neural network is a new star in the field of artificial intelligence and has made remarkable progress in many fields of artificial intelligence, especially in the field of image recognition.

The specific calculation steps are as follows:

For the specific screen behavior image I(x, y), the gradient pattern of the i-th direction is calculated as follows:

where M(x, y) represents the gradient magnitude of the screen behavior image I(x, y) at the position (x, y) and can be expressed as follows:

where the symbol represents the absolute value operation, and and represent the directional derivatives of the screen behavior image I(x, y) in the horizontal and vertical directions, respectively, and can be expressed as follows:

As shown in Figure 3, the convolution kernel L in each fixed direction can be obtained by rotating L0 by . Through the calculation method shown in formula (5), After a set of 12 convolution kernels are convolved with the gradient amplitude M(x, y) of the screen behavior image, a corresponding set of 12 different directions of gradient pattern Di is obtained. Compared with the method shown in formula (4), the proposed method takes into account a larger range of neighboring pixel values, so it has better stability, especially in the smooth area without sharp fluctuations.

After obtaining the gradient direction map of the 12 different directions of the screen behavior image I(x, y), the final gradient direction is by selecting the direction with the strongest response among the 12 directions of each pixel as the final gradient direction of the point [23]:

The gradient directions of the reference and distorted screen behavior images are extracted, and the gradient direction similarity is defined as follows:

where and represent the gradient pattern of the reference and distorted screen behavior images, respectively, and C is a positive number to prevent instability when the denominator is 0. In this experiment, is used.

In order to generate a single quality index of the distorted screen behavior image, the partial quality map DS(x, y) needs to be merged. Since different areas have different effects on the overall perception of behavioral image quality, we use the standard deviation of DS(x, y) as the quality indicator of the distorted screen behavioral image:

where represents the collection of all pixels in the entire screen behavior image, N is the total number of pixels in the collection, and is the average value of DS(x, y). It should be pointed out that the larger the value of DSS, the deeper the distortion of the behavior image, which indicates that the perceived quality of the behavior image is worse.

Considering that the gradient magnitude can effectively describe the structural information of the behavior image, we use the gradient magnitude as an auxiliary feature and jointly consider the gradient direction and the gradient magnitude to describe the perceptual quality of the screen behavior image. Similarly, the similarity of gradient magnitude is defined as follows [24]:

where and are the gradient amplitude maps of the reference and distorted screen behavior images, respectively, and CM is also a positive number to ensure the stability of the score. In this experiment, .

As mentioned earlier, DS(x, y) and MS(x, y) describe the similarity between reference and distorted screen behavior images from different aspects. In order to fully describe the similarity between the reference and distorted screen behavior images, we combine DS(x, y) and MS(x, y) together as follows:

where and are both positive numbers, which are used to adjust the relative importance of the similarity of the gradient direction and the similarity of the gradient amplitude in the proposed method. In particular, when , only the gradient magnitude is used for quality evaluation. When , only the gradient direction should be used for quality assessment. The gradient direction and the gradient magnitude have the same importance for the screen behavior image quality evaluation, so we set in this algorithm.

Finally, the standard deviation of the joint similarity is used as the behavioral image quality indicator:

where is the average value of GS(x, y). According to the definition of GSS, the smaller the value of GSS, the greater the similarity between the reference and the distorted screen behavior image, and the better the quality of the distorted behavior image.

4. The Relationship between Intelligent Image Simulation and Recognition Technology and the Mentality and Quality of Life of the Elderly

Similar to biological systems, the main purpose of computer vision is to obtain machine behavior through possible “techniques” and “processing methods.”

Only the combination of complex “information processing tasks” and the “reasonable control and coordination” of these tasks can achieve image understanding.

The main purpose of computer vision is to use technically feasible “techniques” and “processing methods” to achieve machine behavior similar to the behavior of biological systems.

Figure 4 shows the main architecture of the active nursing system for the elderly in this paper. The architecture adopts the C/S mode. The server provides services such as data storage and system interface. The client is used to view and monitor the activity information of the elderly under guard.

In the previous section, the system model of this paper has been constructed and applied to the real-time monitoring of the elderly, and the improvement of the quality of life of the elderly after the application of the model has been calculated. The effectiveness of the method in this paper is verified through experiments combined with questionnaire surveys.

The average score of health literacy of elderly patients with chronic diseases is (57.60 ± 21.96), and the scores of each dimension are shown in Table 1.

Depression is significantly negatively correlated with basic activities of daily living and functional activities of daily living and is significantly negatively correlated with health literacy. Basic and functional daily living abilities are significantly positively correlated with health literacy, as shown in Table 2.

The first step is the regression analysis of depression on daily activity ability to obtain the path coefficient c. The second step is the regression analysis of depression on health literacy to obtain the path coefficient a. In the third step, health literacy is included, and the regression analysis of depression and health literacy on daily activity ability is performed to obtain path coefficients b and c' (Table 3). The results found that depression and health literacy can significantly predict the ability of daily living. After adding health literacy, the R2 adjusted by the model increased from 0.324 to 0.521. The above results show that health literacy satisfies the conditions of the intermediary variables.

Finally, the model of this paper is evaluated through expert evaluation, which mainly evaluates the effect of improving the mental health of the elderly, the effect of improving the quality of life, and customer satisfaction. The results are shown in Table 4 and Figure 5.

It can be seen from the above experimental research that the method proposed in this paper can effectively improve the mental health of the elderly and improve the quality of life of the elderly.

5. Conclusion

Due to the imperfect elderly care system in my country, the trend of aging, empty nesting, and chronic disease among the elderly has intensified. Moreover, symptoms such as hemiplegia and cerebral thrombosis are common in the elderly, which makes the long-term care of the elderly population more and more serious. In particular, the care of disabled elderly people needs to be solved urgently. Currently, elderly patients with chronic diseases have low levels of health literacy. Therefore, in the future health education, we should strengthen health education for elderly patients with chronic diseases, use different mass media to propagate health knowledge, promote the formation of healthy lifestyles and behaviors of elderly patients with chronic diseases, and fundamentally improve the health literacy level of elderly patients with chronic diseases. In addition, we can also use intelligent simulation recognition technology to recognize the behavior of the elderly and improve the monitoring effect of the elderly, so as to quickly respond to emergencies in a timely manner. Through experimental research, it can be known that the method proposed in this paper has good results.

Data Availability

There is no analytical permission from the data provider because of trade confidentiality.

Conflicts of Interest

The authors declare that they have no conflicts of interest.