#### Abstract

In order to take better pictures, a method of composition proportion in photography based on digital information technology is proposed. Feature extraction is carried out from the captured pictures, and the modular customized design of digital information technology is adopted. The research is carried out in the two fields of image blur and approximate repeated image de redundancy. An improved image blur algorithm based on bag of words (bag of words) kernel function is proposed. The experimental results show that the coarse matching time is the time spent in matching a query image with thousands of images in the database, and the fine matching time is the time spent in matching a query image with eight candidate images with orb. The total query time is about 0.7 seconds, while it takes about 500 seconds to match thousands of images only with orb algorithm. In general, we combine coarse matching and fine matching, which can balance the accuracy and speed well. It is proved that the digital information technology proposed in this paper has a good effect in the calculation and function of composition proportion in photography.

#### 1. Introduction

At present, the research and practice of modular customized design using digital information technology are still in the development stage. In the research stage, modular customization design applies the design method of digital information technology, but the concept of module is not complete; the modular customized design in the practical stage still has development potential in form diversity and space quality. In recent years, with the vigorous development of information technology and Internet applications, multimedia massive data are growing at an unimaginable speed. As a common information carrier, digital image has an indispensable position [1]. With the popularity of social software such as microblog and WeChat and the popularity of intelligent camera, everyone can easily capture, process, and spread images, so the image data on the Internet are increasing in an unpredictable way. While digital image technology enriches our life, some problems are becoming increasingly prominent. Shooting angle refers to the angle presented by the relative position between the camera and the subject. In life, people observe things with their eyes at a certain angle. The choice of this angle usually represents some behavior habits and psychological characteristics of people. Therefore, the choice and change of shooting angle evolved from people’s lives and habits. The choice of angle means the starting point of looking at things. The choice of angle can reflect the photographer’s psychology and represent the photographer’s position [2]. Angle has subjectivity and distinct personality. It can strengthen and highlight some things but also weaken and dilute the performance of some things. The choice of different angles has different emphases and expressiveness. In terms of photography art, the expression theme is different, and the angle selection is different. The composition has different ideas and different angles. The change of angle is complex and changeable. Shooting a thing or theme has countless shooting points to choose from in the three-dimensional space around it. Students need to learn, think, and master how to make rational use of these shooting points regularly and methodically and find out the best shooting angle. The complex changes of shooting angle can be divided into three categories: changes in shooting direction, changes in shooting height, and changes in shooting distance. The angles in the shooting direction can be summarized as front, side, oblique side, and back. The angle of shooting height is flat shot, overhead shot, and overhead shot. The change of shooting distance is the change caused by changing the focal length of the lens or the object distance between the shooting point and the subject, that is, the five categories of long-range, panoramic, medium range, close-up, and close-up. These angles and scenes can be used singly, and there are also repeated and overlapping collocations, which reflect a strong complementary relationship [3]. Through the reasonable selection and application of these changes, we can better express the theme, more skillfully render the various relationships between people, characters, and the environment, and better help readers understand the intention of the picture. The choice of angle should be novel and unique. On the basis of logic, we should break the traditional viewing mode and find another way. Find unusual shooting angles from common scenes. The shooting angle is novel, and the content and form of expression are unique. Only in this way can we be unconventional and make the picture unconventional and refreshing to the viewer.

#### 2. Literature Review

JM Davidson proposed two noise sensitivity estimation methods called different focus evaluation functions in the subsequent study of automatic focusing methods [4]. Of course, these two methods need to calculate all the focusing methods in the process of selecting the best focusing method. For some occasions with high real-time requirements, the application of this selection method is limited. Despite some limitations, this paper also makes many new attempts to quantitatively evaluate the auto focus function. In the study by Clero et al., according to the basic imaging principle of the imaging system, a defocus blur estimation method based on the point spread function is proposed on the basis of the point spread function, and the image restoration filtering technology is also applied to the automatic focusing system, and a better effect is obtained [5]. Reichardt et al. used the optical autocollimation method to develop the automatic focusing system for integrated circuit lithography [6]. Shavkat et al. used analog circuit to detect the high-frequency component in the circuit signal and apply it in the automatic focusing of camera [7]. In the study by Hong et al., the teaching and research office of optical instruments has developed the automatic focusing system in satellite photography equipment by using the autocollimation method [8]. Jin et al. studied the automatic focusing system of image detection band cutting differential comparison CCD. Shanghai Institute of laser technology has studied the automatic focusing technology of the microscopic system of optical disc recorder and player by using the astigmatism method [9]. Liang et al. studied the automatic focusing of the microscopic system of the linewidth measuring instrument by using the eccentric beam method and reached the international advanced level [10]. Song et al. used CCD camera technology and image processing method to automatically screen biological pictures [11]. Mark et al. and others mainly focus on the research of image gray gradient and put forward the sum of image gray gradient vector modulus square, Roberts gradient, and Laplace operator, as well as the autofocus evaluation function of digital image processing. Using the above three mathematical models of focus evaluation function, the experimental results of defocus blur and motion blur are given, compared, and analyzed. The following conclusions are obtained: these focus evaluation functions have the characteristics of good unbiased, unimodal, and high sensitivity and can be applied to the evaluation of defocus blur and motion blur, and the modulus sum of gray gradient vector is the most ideal focus evaluation function [12]. Dai et al. and others analyzed the application of gradient function, spectrum function, and entropy function in autofocusing technology, respectively, and studied and analyzed the performance of the evaluation function of autofocusing together. In the experiment, a large number of comparisons are made between the energy gradient function and the spectrum function, and it is concluded that the spectrum function is better than the energy gradient function in sensitivity. However, in terms of processing time, the processing time of energy gradient function is shorter than that of spectrum function, so the real-time performance of energy gradient function is better than that of spectrum function [13].

Based on the current research, this paper proposes a method of composition proportion in photography based on digital information technology. Feature extraction is carried out from the captured pictures, and the modular customized design of digital information technology is adopted. The research is carried out in the two fields of image blur and approximate repeated image de redundancy, and an improved image blur algorithm based on bag of words is proposed.

#### 3. Image Blur Algorithm Based on Word Bag Model

Then, we use a novel feature point detection algorithm based on concentric rings to obtain the feature points. Next, we use the improved Gabor filter-based feature description algorithm to describe each feature point. After feature extraction of all images in the image database, all feature point descriptors of all images are clustered into a “visual dictionary” by the *K*-means algorithm. Finally, this “visual dictionary” is used to obtain the final “visual signature” of each image and save it for later matching (Figure 1).

Blurring stage of photographed image: as shown in Figure 2, when an image to be queried is obtained, we first extract feature points and obtain descriptors of each feature point using the same method mentioned in the database building stage. Then, use the “visual dictionary” saved in the database building stage to generate the “visual signature” of this figure. Then, the “visual signature” will be roughly matched with the “visual signature” of all images saved in the database to obtain a group of candidate similar images [14].

##### 3.1. Feature Point Detection Method Based on Concentric Ring

From the data in Table 1, it can be seen that although the number of detected feature points has changed, this transformation is small for the number of feature points of the whole image and can not be considered. Therefore, the feature point extraction algorithm based on concentric rings can resist image selection, noise, and scale transformation and has strong robustness.

##### 3.2. Feature Point Description Method Based on Gabor Filter

In terms of computer vision, Gabor feature is very suitable for processing camera images and has the following advantages:(1)Because Gabor kernel function filters the DC component of the filtered image, it can resist the transformation caused by local illumination and is very suitable for processing camera images.(2)Gabor filtering results can describe the gray distribution information of images in different scales and directions. Therefore, it can not only describe the global information in a large scale to reduce the impact of noise but also describe the local details in a small scale. Firstly, the object of Gabor filtering is the contour map, so it only filters the lines in the contour, and the local lines contain little useful information. Secondly, Gabor filters the image and then extracts local small blocks, which is very time-consuming; Finally, Gabor does not resist rotation transformation. In view of the above shortcomings, this paper improves the Gabor and applies the improved Gabor to the ambiguity of the image taken by the camera. The specific extraction process after the improvement is as follows:(1)For each feature point extracted by the feature point extraction algorithm based on concentric rings, calculate its local coordinate system, take the connecting direction between the center point *o* of the rectangular box and as the ordinate, and take the straight line passing through and perpendicular to the ordinate as the abscissa . As shown in Figure 3, each feature point has its own local coordinate system.(2)Draw a small block of N around Pi on the gray image corresponding to feature point *P* instead of the previous local line, as shown in Figure 3 [15].(3)Because the region may contain multiple pixels, we calculate that the average value of the filtering results of the region in the direction is . The average values of the filtering results of the four regions in the 8 directions together form the feature vector F, as shown in Figure 4.

The improved Gabor changes the filtering object from the original local line to the corresponding gray image, making full use of more information on the image. Due to the use of the local coordinate system, the improved Gabor can resist rotation transformation. At the same time, the filtering area does not filter the whole image but only the small blocks around the feature points, so the filtering time has been greatly improved. Table 2 shows the efficiency of changing the filtering area from the whole image to small patches around the feature points, *D* is the improved parameter above, and the parameter *t* is the ratio of the actual filtering area of the local patches around the filtering feature to the filtering area of the entire image; it can be seen from Table 2 that when the number of feature points is between 100 and 1000, the value of *T* is between 1% and 3%; that is, the actual filtering area only accounts for 1% to 3% of the overall image.

Figure 5 is a group of images rotated from the original image through various angles. We use the images to test the antirotation performance of our algorithm. We use our improved feature extraction based on the word bag model (including the improved Gabor algorithm) to extract a final descriptor from each image and then use the kernel regression nonuniform interpolation algorithm of structural tensor to calculate their fuzziness. Their differences are listed in Table 3. According to the data in the table, our algorithm can resist rotation.

##### 3.3. Kernel Regression Nonuniform Interpolation Algorithm Based on Structural Tensor

Kernel regression is a nonparametric regression based on statistics. Consider the random variables *X*, *y*, hypothesiswhere

And, when *1 ≠**j*, , and are uncorrelated, the joint distribution is *f* (*x*, *y*), and the marginal distribution of *X* is *f*(*x*). In general, *m*(*x*) can be considered as the conditional expectation of *Y* when *X* = *x*, that is,

Use the following formula to estimate *m* (*x*):

Function *K* (●) is called kernel, which is actually a weighted average function in essence. The density function *f* (*x*) is estimated as follows:

Among them, *h* is called bandwidth, and the size depends on the data. We call this as method of estimating m(*x*) kernel estimation. The regression model is called kernel regression.

When applied to image interpolation, can be regarded as the gray value of the sampling point, is the sampling point, *m* (*x*) is the image gray value function to be obtained, expressed by *Z* (*x*), is the error, and N is the number of samples. At this time, the kernel regression model can be written as follows:

From (7), it can be seen that the kernel function actually plays a role of weighted average. In image interpolation, the core idea of the interpolation method is how to select the kernel. Many interpolation methods can be regarded as kernel regression model. The kernel functions of some interpolation methods are given.

The kernel function of the nearest interpolation is as follows:

The kernel function of the bilinear interpolation method is as follows:

The kernel function of cubic convolution interpolation is as follows:

The kernel function of *B*-spline interpolation is as follows:

It can be seen from (11) that the kernel function is actually the weight, so the kernel function must meet certain conditions [16]. Generally, the kernel function should satisfy the following:(1)*K* (*x*) is nonnegative(2)*K* (*x*) is symmetric(3)*K* (*x*) is unimodal and reaches its maximum at zero, namely, *C* is a constant.

There are many kinds of kernel functions that meet the above conditions; at present, there are four types of kernel functions that are most used.(1)Linear kernel function:(2)Polynomial kernel function:(3)Gaussian kernel function:(4)Sigmoid kernel function:

Gaussian kernel function is widely used because of its separability and locality. The kernel function used in this paper is Gaussian kernel function.

##### 3.4. Rough Feature Matching Based on Gaussian Kernel Function

A scale transformation invariant is proposed to correct the wrong matching problem caused by Euclidean distance, and this method is used to make a rough matching between the image to be matched and the image in the database, which is called rough matching, to obtain a group of candidate images. The specific method is shown as follows.

Assuming that the image signature of an image extracted by the word bag model is , we assume that the image signature after taking photos becomes ; the relative relationship between the image signature before and after taking photos is unchanged; that is, when the number of one feature before taking photos is more than that of another feature, it is still the same after taking photos. As shown in Figures 6 and 7, there is a linear relationship between the image signatures before and after photographing, and this relationship is caused by scale transformation.

We define the linear relationship of image signature before and after photographing as , where a is the coefficient representing the linear relationship between and *x*. We define an invariant coefficient F, as shown in the following formula:

Let the invariant coefficient of the image after photographing be , and formula (18) proves that the invariant coefficient is fixed before and after photographing. proves that there is an invariant between the images before and after taking pictures. We use this invariant to correct the error caused by Euclidean distance matching. When the image to be matched is matched with the image in the database by Euclidean distance, a group of similar candidate images will be generated, including both correct and wrong results [17].

#### 4. Experimental Results and Analysis

In order to test the effectiveness of the algorithm in this chapter, we collected thousands of home images as our test database. Then, process the database, extract the visual dictionary, calculate the visual signature of each image, and save it for use. The size of the visual dictionary we use is 180 dimensions, so each visual signature saved in the database is a 180-dimensional histogram.

In order to evaluate the effectiveness of our algorithm, we use the short name of mean reciprocal rank (MRR) to evaluate the algorithm of Gaussian kernel function; as a mechanism to evaluate the advantages and disadvantages of ambiguous algorithms, MRR has certain universality in the world. It mainly measures whether the original image corresponding to the image to be matched in the database is blurred and arranged in front [18]. For a group of *Q* to be queried, the reciprocal formula of average sorting is as follows:

The following example illustrates MRR. Suppose we have a fuzzy algorithm, which has been used for three queries to obtain the furniture with the corresponding color. For each query, three furniture are returned as results, the first of which is the best matching result considered by the algorithm. Through these three queries, we can calculate the MRR value of this algorithm: (1/3 + 1/2 + 1)/3 = 11/18, which is 0.61.

In order to test and verify the effectiveness of our algorithm, we carried out simulation experiments with a computer with 2.8GhZ CPU and 4G memory, and we did 15 groups of experiments, each containing 100 query images. For each group of experiments, we calculated the MRR value. Figure 8 shows the comparison results of our algorithm with the algorithms of Eits, Horace, and Li Fei Fei. It can be seen from the figure that our algorithm combining coarse matching and fine matching has the best accuracy [19–21].

The time consumed by our algorithm includes two parts: the rough matching time (RMT) and the fine matching time (FMT). As shown in Figure 9, the coarse matching time is the time spent matching a query image with thousands of images in the database, and the fine matching time is the time spent matching a query image with eight candidate images with orb. The total query time is about 0.7 seconds, while it takes about 500 seconds to match thousands of images only with the orb algorithm. In general, we combine coarse matching and fine matching, which can balance the accuracy and speed well. Figure 10 shows the results of comparing the time performance of our algorithm with other algorithms. From the table, we can see that our algorithm has the best time performance [22–24].

#### 5. Conclusion

This article improves the properties of Gabor. There are three main improvements. First, the image filter as a whole is modified by filtering only small local blocks around the feature points, which effectively reduces the filter area and improves the filtering speed. Second, the filter object is changed from the original inner line to gray feature grade mapping enriched the filter information, and third a local coordinate system is introduced to improve the nonreverse rotation characteristics of Gabor. Scale invariance is proposed to update the appropriate results of the structural tensor core regression nonuniform interpolation algorithm, which effectively improves the matching accuracy. Experiments have shown that our algorithm can accurately and quickly capture camera images from blurred images, making it an ideal program for extracting image features.

#### Data Availability

No data were used to support this study.

#### Conflicts of Interest

The author declares that there are no conflicts of interest with any financial organizations regarding the material reported in this manuscript.