Abstract

This study aims to discuss the application value of KMC algorithm optimized by heuristic method in basketball big data analysis and visual management. Because the data in basketball big data is too complicated and incomplete, the extraction of information is not direct and effective enough. Based on the metaheuristic K-Means clustering (KMC) algorithm, the weights and genetic algorithm are introduced to optimize it, and the University of California at Irvine (UCI) data set is applied to analyze the big data clustering performance of the optimized KMC algorithm. The 2018-2019 season National Basketball Association (NBA) shooting guards are selected as the research objects, and the optimized KMC algorithm is used to process the data and analyze the NBA scoring functional factors. It is found that the number of clusters increased from 2 to 16. After optimization, the Between-Within Proportion (BWP) value of the KMC algorithm only drops by 0.35, and the improved BWP (IBWP) value only drops by 0.288, which shows the smallest drop among all the algorithms. When the number of nodes is 4, the running time of the optimized KMC algorithm for processing the COVTYPE data set is 1922 s after optimization, and the running time for processing the IRIS data set is the shortest (113 s). When the number of parallel nodes is 10, the speedup ratio of the optimized KMC algorithm for processing COVTYPE data set is 4.16, and the maximal expansion rate is 0.81. The clustering accuracy of traditional KMC algorithm is 89.33%. After optimization, the clustering accuracy of KMC algorithm is 98.67%. The leader factor, offensive contribution factor, shooting stability factor, and passing ability factor in the core grouping are all at the maximum, which are 0.59, 0.51, 0.47, and 0.43, respectively. The optimized KMC algorithm has been shown to reduce the number of iterations, reduce convergence time, and improve clustering accuracy. The optimized KMC algorithm has been shown to reduce the number of iterations, reduce convergence time, and improve clustering accuracy. The conclusion of this study can provide reference basis for big data clustering and visual management.

1. Introduction

For data research or data application requirements, data visualization is to present specific data in the form of statistical charts and information. Big data analytics refers to the process of extracting potentially valuable information from a large amount of noisy and incidentally incomplete application data [1]. Big data analytics is a poorly multidisciplinary methodology. The main areas are neural networks, pattern recognition, spatial data analysis, image databases, signal processing, artificial intelligence, knowledge base systems, data acquisition, and bioinformatics [2, 3]. Big data analysis has concept description, association analysis, classification and prediction, cluster analysis, external analysis, and evolutionary analysis [4]. Clustering is an unsupervised classification method that automatically divides big data into multiple classes or clusters according to a certain standard. Cluster analysis can preprocess the data by observing the characteristics of each class or concentrating on a certain type of valuable data for further analysis and processing [5]. Cluster analysis is widely used in data analysis, image segmentation, pattern recognition, and other fields [6]. Currently, the common clustering method is the K-Means Clustering (KMC) algorithm based on the heuristic algorithm. The KMC algorithm is widely used in data statistics, data analysis, and machine analysis due to its short and fast properties [7]. The KMC method based on heuristic algorithm shows significant advantages in small- and medium-scale data analysis. However, when the large-scale data sets are clustered, it is necessary to manually determine the number of clusters, the clustering results are unstable, and the misselecting noise and abnormal points will eventually lead to inefficient data processing and poor clustering quality [8].

The data analysis process is based on a large amount of data, and the ability of human brain to absorb and process information is limited. Visualization technology can transform scientific data into graphic image information that changes with time and space through computer and image processing technology and finally achieve the interactivity, visibility, and multidimensionality of the data [9]. Researchers can analyze the data and its changing trends through graphs and images. Data visualization speeds up data processing and increases the utilization of effective data. Data visualization has been widely used in various fields such as natural sciences, engineering technology, finance, communications, and commerce [10]. Basketball has become a popular sport because of its features such as simplicity, fun, fitness, and education. The depth of basketball is measured by the game. Basketball statistics can make an objective analysis of the data and unearth potential actual combat information. However, there are few studies on applying cluster analysis methods to basketball big data analysis.

In summary, the KMC method based on heuristic algorithm for processing big data has to be further optimized, and there is limited research on applying the clustering data analysis method to basketball data analysis. In this study, the KMC algorithm in the heuristic method is optimized and applied to basketball big data analysis to provide a reference for basketball big data clustering and visual management.

2. Materials and Data

2.1. The Cluster Analysis Methods of Big Data

Big data cluster analysis is the process of grouping a collection of physical or abstract objects into multiple classes composed of similar objects, clustering a collection of data objects in the same cluster. Big data analysis is not a postprocess that obtains effective results after simple analysis of input data. It needs to go through the continuous repetition of a multistep complex process to obtain accurate results. For n vectors in the a-dimensional space Ra, they are assigned to one of the c clusters, so that the distance between each vector and its cluster center is the smallest. Then, the distance between the vectors Xi and Xj can be expressed as follows:

Cluster analysis mainly includes two kinds of data matrix and discrepancy matrix [11]. They differ from the matrix diagram method in that they are not filled with symbols on the matrix diagram but filled with data to form a matrix for analyzing the data. The data matrix is a matrix in which d data objects of the entire data set are described with l attributes, and the final data object set is regarded as a matrix. The data matrix can be expressed as follows:

The difference matrix refers to the degree of similarity between any two data points in the overall data object set [12], which can be expressed as follows:

In (3), n represents the number of data points, and the d(i, j) in the matrix represents the difference degree calculated according to the specified degree of similarity of the data points i and j in the data object collection. The larger the d(i, j) value, the greater the degree of difference between the data objects.

The core of cluster analysis is to obtain the degree of similarity among different data objects [13]. At present, the Minkowski distance calculation method, the Euclidean distance calculation method, and the Chebyshev distance calculation method are commonly used for evaluation [14]. Among them, the data obtained by the Euclidean distance calculation method is not affected by coordinate translation and rotation changes, and it is a commonly used distance similarity measurement method [15]. The calculation method of Euclidean distance is given as follows:

In the above equation (4), d(i, j) represents the Euclidean distance between two data points, which satisfies the conditions , , and .

The similarity factor is mainly used to gauge the similarity among data points [16]. The angle cosine method is a commonly used similarity coefficient calculation method. The value range of the similarity coefficient is [−1, 1]. When the orthogonal value is 0, it means that the two vectors are completely dissimilar. The calculation method of the similarity coefficient of the angle cosine method is as follows:

The correlation coefficient method represents the degree of correlation between two data vectors [17], and its value range is [−1, 1]. 0 means that they are not correlated, 1 means that positive correlation is found, and −1 means that negative correlation can be seen. The correlation coefficient method can be expressed as follows:

Appropriate criterion function in cluster analysis can further improve the quality of clustering [18]. The criterion functions commonly used in cluster analysis are as follows: squared margin of error, squared weighted mean value distance sum, and interclass distance sum [19]. The error sum of squares is often used for data analysis with dense samples and little difference between samples [20]. The error sum of square (Ja) can be expressed as follows:

In the equation above, mj is the average value of the class Ck, and . nj refers to the number of objects in the class Ck.

The interclass distance and criterion (Jb) calculates the distance sum of every clustering epicenter to the global epicenter. The higher the similarity of the research data, the less obvious the clustering result, and the results making Jb the largest result have to be found.

The weighted average squared distance (Jc) is applicable to data objects with a large disparity in the number of samples, and its calculation method can be expressed as follows:

In the above two equations, refers to the average squared distance between samples within a class, and Pj is the prior probability.

2.2. Establishment of Cluster Analysis Method Based on Metaheuristic Algorithm

KMC is the most classic and most widely used clustering method in the metaheuristic algorithm. The kinetic Monte Carlo method (KMC) is simple in principle and highly adaptable, so it is the first choice of researchers in many cases. This method takes Euclidean distance as the correlation measure, and the error sum of squares criterion (Ja) as the criterion function to minimize the evaluation index. The KMC algorithm divides the data set A into the closest classes, and its cluster center is . The calculation method of each cluster center point is shown in equation (11), in which and ni was the number of data objects in the class Ci.

The traditional KMC method has a great dependence on the selection of the initial clustering center point, and it is susceptible to the interference of local noise data. The different feature weights assigned to the attributes of each data point can improve the KMC results greatly. The feature weights of variable patterns were assigned to data points, which were named KMC based on density, DK-Mean. The attribute feature weight value of the j-th dimension is assigned to the object data. The calculation method is expressed as follows:

In (11), aj is the ratio of the distance between the classes of the attribute and the distance within the classes, and (db refers to the distance between classes, and ; di refers to the distance within the class, and ). mi represents the mean value of the data set on the j-th dimension attribute; K is the number of clusters, and j is the number of attribute bits. Then, the weighted Euclidean distance calculation equation can be written as follows:

The KMC method relies on the cluster center point, which is easy to cause local optimal clustering. Based on the density, the choice of the original aggregation centers is improved accordingly in this study. The clustering criterion function can be denoted as the following equation:

In (13), represents the distance within the class, and . Kbe represents the distance between classes, and . Then, the density D(x) at sample point X can be expressed as the following equation:

In (14), Di represents the weighted Euclidean distance, and r is the specified radius.

Cluster analysis method can solve such problems; cluster analysis method is an exploratory analysis method, which can analyze the inherent characteristics and laws of things and is a commonly used technology in data mining. Genetic algorithm shows good applicability and scalability and can reduce the initialization requirements of traditional clustering algorithms in cluster analysis. The genetic algorithm is introduced further based on the DK-Mean algorithm to increase the accuracy of the clustering algorithm in this study. The genetic algorithm search can minimize the Jk value, and then the fitness feature can be represented as formula (16):

The probability of an individual being selected can be expressed as follows:

In the above equation, f(xi) is the fitness value, and .

The crossover operation is performed on two individuals x1, and x2, and the new individuals produced by them can be expressed as equations (18)∼(19), in which is the uniform arithmetic crossover parameter.

The improved DK-Mean algorithm calculates the data gap matrix and initializes the target eigenvalues to obtain new cluster centers. The specific process of the DK-Mean algorithm is shown in Figure 1.

2.3. Visual Data Analysis and Visualization Based on Clustering Algorithm

The biggest difference between visual analysis and visualization lies in the analysis of this point, the process of visualizing data for business simulation, correlating multidimensional business data to form a more comprehensive data result, and providing users with auxiliary decision-making process, which is called visual analysis. Parallel coordinate method has the characteristics of mapping high-dimensional data to low-dimensional space and can interact with users at the same time, and it is a commonly used method of visual data analysis at present. In this study, a visual data analysis model based on the KMC algorithm is established based on the optimized KMC algorithm and the parallel coordinate method.

It is assumed that G is a collection of n-dimensional data objects, and , of which is an n-dimensional collection ; the basic coordinate axis corresponds to the attribute of the i-th dimension, and each n-dimensional vector can be expressed as . The polyline H of the n-dimensional data using linearly independent equations is given as follows:

According to the mapping principle from the midpoint of the coordinate system to the parallel coordinate, the following equation can be obtained:

In the equation, mi is the slope, and bi represents the intercept on the axis in parallel coordinates .

The technology and process of data analysis are applied in the basketball data visualization management system, which can be data processing automation. The data analysis visualization process based on the optimized KMC algorithm is shown in Figure 2. After the sample data is processed through selection operations and cross operations, a new visualization population can be formed.

2.4. The Evaluation Indicators of Cluster Validity

The main indicators in the cluster analysis evaluation are as follows: Accuracy, Precision, Recall, and F1 value, four commonly used indicators. The ideal clustering result should reflect the internal structure of the data set as much as possible, so that the sample similarity between classes is the smallest, and the samples within the class are the most similar. In this study, the Between-Within Proportion (BWP) and improved BWP (IBWP) indicators were adopted to analyze the clustering results and performance, where BWP is the ratio of the clustering deviation distance to the clustering distance, and its calculation method is given as follows:

In the equation, c(i, j) represents the interclass distance of the object i in the j-th class, and c(i, j) represents the intraclass distance of the object i in the j-th class. Among them, , n is the number of data objects, which can be the number of divided clusters, na represents the number of elements of the class a, and j represents the class label. The larger the BWP(i, j) value, the more effective the clustering of sample objects.

The IBWP indicators can evaluate the clustering effectiveness of a single data object very well, and its calculation method is shown as follows:

In the (21), ic(i, j) and represent the interclass distance and the intraclass distance of the object i in the j-th class. The larger the value of IBWP index, the more effective the clustering of individual points of the sample.

The speedup ratio and the expansion rate are used to evaluate the parallelization effect and performance of the clustering algorithm in analyzing data. The calculation method of speedup ratio and the expansion rate is given as follows:

In the above two equations, T1 is the data processing time for a single node; Tm is the data processing time for m nodes. N is the size of the processed data; Tn is the time for data to be processed on a child node; and Tmn is the time for data of size n to be processed on m nodes.

2.5. Data of Testing Dataset of the Model

In this study, the IRIS, WINE, SEED, ABALONE, LETTER, and COVTYPE data sets in the UCI database are undertaken as the validation sets to verify the algorithm model established. Table 1 shows the total number of samples, dimensions, and categories of the data set. Data analytics is the process of analyzing data sets in order to make decisions about the information they hold, to be used in the business industry to enable organizations to make business decisions.

2.6. Research Objects and Methods of Basketball Big Data

The NBA shooting guards in the 2018-2019 season were selected as the research objects. The relevant indicators included in the study were analyzed statistically using literature data method, logical analysis method, mathematical statistics method, video analysis method, and comparative analysis method. The technical statistical data of the season finals are collected on related websites such as Tencent Sports Video, Hupu NBA, the control video, and official statistics, which were repeatedly confirmed to ensure the authenticity and reliability of the data source. These raw data were adopted for statistical analysis of basketball technical indicators. Data analytics can help businesses better understand their customers, improve their advertising campaigns, personalize their content, and improve their bottom line.

2.7. Analysis on Influencing Factors of Basketball Scoring Based on KMC Algorithm

The 2018-2019 season NBA scoring guards were selected as the research objects. Based on the NBA data query website (https://www.basketball-reference.com/), 17 basic pieces of data such as player scores, rebounds, assists, and steals, as well as advanced data such as passing ability, defensive contribution, and offensive contribution are selected as the original data. The original data removes rebounds, blocks, and fouls and reduces the original data from 36 dimensions to 22 dimensions, including Field goal attempts (FGA), Field goals (FG), Free throws (FT), Free throw attempts (FTA), Assists (AST), Steals (STL), and Points (PTS).

3. Results

3.1. Analysis on Results Based on Metaheuristic Clustering

Using the traditional KMC algorithm to cluster the data in the IRIS data set, the clustering results are divided into 4 clusters, but the clustering results of some data overlap (Figure 3(a)), and the clustering results are significantly different from the real data. The optimized KMC algorithm is used to cluster the data in the IRIS data set. The clustering results are divided into 4 clusters, the data clustering results are of high quality, and there is no crossover phenomenon of different types of data (Figure 3(b)). In addition, there is no difference between the clustering results and the real data.

3.2. Comparison on Classification Indicators of Different Clustering Algorithms

The traditional KMC algorithm is compared with DKMC, the optimized KMC algorithm, self-organizing feature map (SOM) algorithm, quantum evolutionary clustering algorithm (QEAM), and k-medoids algorithm in terms of BWP values (Figure 4(a)). As the number of clusters increases, the BWP values of different algorithms show a downward trend all, and the number of clusters increased from 2 to 16. After optimization, the BWP value of the KMC algorithm only drops by 0.35, which is the smallest drop among all algorithms. The BWP value of an optimized KMC algorithm with the same number of clusters is higher than that of other algorithms. The IBWP values of the different algorithms are compared, and the results are shown in Figure 4(b). As shown in the figure, the IBWP values of the various algorithms show a decreasing trend as the number of clusters increases. The number of clusters increases from 2 to 16. After optimization, the IBWP value of the KMC algorithm only decreases by 0.288, which is the smallest decrease of all algorithms. The IBWP value of an optimized KMC algorithm with the same number of clusters is higher than that of other algorithms.

3.3. Execution Time of Clustering Algorithm

The optimized KMC algorithm was used to perform cluster analysis for six datasets in the UCI database (Figure 5). As the number of nodes increases, the time taken by the KMC algorithm to collect 6 different data sets decreases. If the number of nodes is 4, the COVTYPE dataset output time is 1922 s, and the IRIS dataset output time is 113 s.

3.4. Performance Analysis of Clustering Algorithm in Parallel Data Processing

The speed ratios of the optimized KMC algorithm between the six data sets were analyzed and compared, and the comparison results are shown in Figure 6(a). As the number of parallel nodes increases, the speed ratio of the KMC 6 algorithm when processing the data set shows an increasing trend. If the number of parallel nodes is 10, the maximum speed ratio of the COVTYPE data set is 4.16. The expansion ratios of the KMC-optimized algorithm between the six data sets were analyzed and compared, and the comparison results are shown in Figure 6(b). As the number of parallel nodes increases, the expansion ratio of the KMC algorithm for processing the six data sets appears to decrease. If the number of parallel nodes is 10, the maximum expansion ratio of the COVTYPE data set is 0.81.

3.5. Analysis on Test Performance Rate of Clustering Algorithm

The group accuracy of the different cluster algorithms across the different datasets is compared, and the results are shown in Figure 7(a). There are three different algorithms for the accuracy of data grouping. After optimization, the accuracy of the KMC algorithm when processing the six data sets was clearly higher than the other two algorithms (). The convergence times of the different clustering algorithms in the different data sets are compared, and the results are shown in Figure 7(b). The convergence time of the optimized KMC algorithm in the different data sets was shorter than that of the other algorithms, and the difference was statistically significant (). The number of iterations of different clustering algorithms in different datasets is compared, and the results are shown in Figure 7(c). The number of iterations of the optimized KMC algorithm across the different data sets was lower than that of the other algorithms, and the difference was statistically significant ().As the convergence time increases, different clustering algorithms are proportional to the data iteration effect.

3.6. Analysis on Cluster Visualization Result

The traditional KMC algorithm and the optimized KMC algorithm are performed with the cluster analysis on the Luanweihua data set in the IRIS data set. The cluster visualization analysis is performed on three types of Luanweihua data in this study. After cluster analysis, all data are divided into 3 clusters, with 50 groups of data in each cluster of original data. After clustering using the traditional KMC algorithm, a total of 16 sets of data have been misclassified, and the clustering accuracy is 89.33%, as illustrated in Figure 8. After clustering using the optimized KMC algorithm, there are two sets of data that are misclassified, and the clustering accuracy is 98.67% (as given in Figure 9). Through the clustering visualization analysis of the traditional KMC algorithm, the traditional KMC clustering accuracy is about 10% lower than that of the optimized KMC.

3.7. Analysis on the Statistic Results of Basketball Technical Indicators

According to the data from the NBA official website, the technical indicators of the top 10 Eastern teams in the 82 regular seasons in the 2018-2019 season are counted, and the results are shown in Figure 10. The figure illustrates that, except for the significant differences between the lost points and the score items, there is little difference in other indicators. The field goal score of Indiana Pacers is 25.4, which is lower than the first place (Milwaukee Bucks, 38.2 scores), showing a difference of 12.8 between the two. The comparison on the indicators of the first and tenth teams shows that the scores for shots, hits, and rebounds of Milwaukee Bucks are 5.8, 2.8, and 3.4 higher than those of the Miami Heat.

The technical indicators of the top 10 Midwestern teams in 82 regular season games are compared, and the results are given in Figure 11. The free throw of Houston Rockets is as high as 45.4, which is obviously higher than that of other teams. The scores in rebounds and assists of Los Angeles Clippers are 22.6 and 28.5, respectively, which are much higher than those of other teams. The shooting percentage of Los Angeles Lakers (the 10th) is 0.7, which is much lower than that of the San Antonio Spurs (0.82) and the Golden State Warriors (0.8).

A cluster analysis is performed in 20 teams in regular games (Figure 12), which shows that 20 teams in the East and West are clustered into 7 categories. Among them, the Golden State Warriors and the Milwaukee Bucks, the first place in the East and West teams, are grouped into the same category. It shows that the top NBA teams have similar characteristics to a certain extent.

The technical indicators of the Golden State Warriors and the Milwaukee Bucks team in the East and West teams are compared, as given in Figure 13. The scores in shooting, attempts, free throws, rebounds, assists, and blocks of the Golden State Warriors and the Milwaukee Bucks are higher than the average scores of all teams. The Milwaukee Bucks and the Golden State Warriors have rebound scores of 49.8 and 46.2, respectively, which are higher 4.6 points and 1 point than the average scores of all teams, respectively, which shows that the rebounding technical indicators of the Milwaukee Bucks have a significant advantage. The scores in assists of the Milwaukee Bucks and the Golden State Warriors were 26 and 29.4, respectively, showing that the score in assists of the Golden State Warriors has a significant advantage.

3.8. Analysis on the Result of the Influencing Factors of Basketball Scoring

The coefficients of the NBA basketball score scheme in the different groups were compared based on the optimized KMC algorithm. As shown in Figure 14, the cluster boundary factor first increases and then decreases as the number of clusters increases. If the number of clusters is 7, the cluster boundary factor reaches a maximum value of 0.24.

Based on the optimized KMC algorithm, the functional factors of basketball NBA scores are analyzed, and the matrix of different factors after coordinate translation is shown in Table 2. All factor coefficients are close to 0 or 1.

The influencing factors of basketball NBA score are analyzed based on the optimized KMC algorithm, and the distribution of different factor cluster centers is shown in Figure 15. The leader factor, offensive contribution factor, shooting stability factor, and passing ability factor in the absolute core grouping are all the maximum values, which are 0.59, 0.51, 0.47, and 0.43, respectively.

4. Discussion

In this study, the KMC algorithm in the metaheuristic clustering method is optimized, its clustering and visualization performance are analyzed, and it is applied to the analysis of basketball NBA score functional factors. It is found that the clustering results of traditional KMC algorithm have the overlapping of some data clustering results, and there is a big difference with the real data. The optimized KMC algorithm does not have the crossover phenomenon of different types of data, and the clustering results are closer to the real data. Such results prove that the optimized KMC algorithm shows improved quality of clustering results. The clustering centers of the traditional KMC algorithm are randomly selected, which leads to errors in the clustering results, and the clustering analysis of the traditional KMC algorithm requires multiple iterations, so the clustering results are greatly different from the true data distribution. The optimized KMC algorithm has coordinate rotation to select the cluster center, which reduces its randomness, so the initial cluster center can be determined more accurately. The number of clusters increased from 2 to 16. After optimization, the BWP value of the KMC algorithm only drops by 0.35, and the IBWP value only drops by 0.288, which is the smallest drop of all algorithms. Such results suggest that the optimized KMC algorithm shows better clustering results. The BWP and IBWP values of the optimized KMC algorithm are greater than those of other algorithms, indicating that the optimized KMC algorithm shows higher clustering accuracy on the samples. As the number of nodes increases, the time for the KMC algorithm to cluster 6 different data sets shows a downward trend. When the number of nodes is 4, the optimized KMC algorithm can process the COVTYPE data set for a maximum of 1922 s, and the shortest running time for processing the IRIS data set is 113 s. The sample size of the IRIS dataset is observably lower than that of the COVTYPE dataset. Such results indicate that the optimized KMC algorithm takes longer time to process low sample size data. This is because each operation needs to start the Map and Reduce tasks, which takes a certain amount of time, so when the task start time dominates, the small samples are processed. Xu pointed out that as the number of sample nodes increases, the running time of clustering decreases. The larger the data size, the better the acceleration ratio of the algorithm, and the better the algorithm’s ability to handle large data. If the number of parallel nodes is 10, the maximum speed ratio of the KMC algorithm optimized for processing the COVTYPE data set is 4.16. He found that the optimized KMC algorithm has several advantages in big data processing. As the number of parallel nodes increases, the level of expansion of the KMC algorithm for processing the six datasets appears to decrease. If the number of parallel nodes is 10, the maximum expansion rate of the COVTYPE data set is 0.81. This is because the amount of communication between each node increases as the number of nodes increases. Some studies have shown that as data size increases, parallel uptime increases, which is similar to the results of this study. This is because the optimized KMC algorithm introduces weights in the calculation of the Euclidean distance, which increases the Euclidean distance between the abnormal point and the cluster center and makes the algorithm iteration result closer to the real data, thereby reducing the number of iterations and convergence time and improving clustering accuracy. Yin pointed out that the optimized KMC algorithm has improved the clustering efficiency, and its clustering accuracy has not changed greatly. The reason is that the study did not consider the impact of each sample data on the entire clustering result, so the Euclidean distance calculation method was not optimized. The clustering accuracy of the traditional KMC algorithm is 89.33%. After optimization, the clustering accuracy of the KMC algorithm is 98.67%, and the clustering accuracy is improved by 9.34%.

Effective clustering can correctly display the player’s status, which is helpful for the rational operation of the team. The research results of this study show that as the number of clusters increases, the cluster contour coefficients first increase and then decrease. When the number of clusters is 7, the cluster contour coefficient reaches the maximum value of 0.24. The leader factor, offensive contribution factor, shooting stability factor, and passing ability factor in the absolute core grouping are all the maximum values, which are 0.59, 0.51, 0.47, and 0.43, respectively. These results show that the absolute core group has an important influence on the team’s score. The main influence factors of absolute core are leader factor, offensive contribution factor, shooting stability factor, and passing ability factor.

5. Conclusion

The KMC algorithm in metaheuristic clustering is optimized and applied to the analysis of NBA scoring functional factors in this study. The results of statistical analysis of basketball technical indicators show that there are significant differences in the gain and loss of scores, and other differences are not significant. It turns out that the optimized KMC algorithm reduces the number of iterations and convergence time and improves the clustering accuracy. The leader factor, offensive contribution factor, shooting stability factor, and passing ability factor are functional factors of NBA scoring. However, there are still some shortcomings in this study. Only a preliminary analysis of the functional factors of basketball NBA scores has been carried out, and the clustering results of different players of different teams have not been analyzed and verified. Therefore, it will further increase the sample size and perform cluster analysis to verify the different players of the team in future. In short, this study provides a reference basis for big data clustering and visual management.

Data Availability

The data used to support the findings of this study can be obtained from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.

Acknowledgments

This work was supported by 2022 Key Scientific Research Projects of Higher Education Institutions in Henan Province, project no. 22A890001.