Abstract

The current business environment has no room for inefficiency as it can cause companies to lose out to their competitors, to lose customer trust, and to experience cost overruns. Business processes within the company continue to grow and cause them to run more complex. The large scale and complexity of business processes pose a challenge in improving the quality of process model because the effectiveness of time and the efficiency of existing resources are the biggest challenges. In the context of optimizing business processes with a process mining approach, most current process models are optimized with a trace clustering approach to explore the model and to perform analysis on the resulting process model. Meanwhile, in the event log data, not only the activities but also the other resources, such as records of employee or staff working time, process service time, and processing costs, are recorded. This article proposes a mechanism alternative to optimize business processes by exploring the resources that occur in the process. The mechanism is carried out in three stages. The first stage is optimizing the job shop scheduling method from the generated event log. Scheduling the time becomes a problem in the job shop. Utilizing the right time can increase the effectiveness of performance in order to reduce costs. Scheduling can be defined as the allocation of multiple jobs in a series of machines, in which each machine only does one job at a time. In general, scheduling becomes a problem when sequencing the operations and allocating them into specific time slots without prolonging the technical and capacity constraints. The second stage is generating the resource value that is recorded in the event log from the results of analysis of the previous stage, namely, job shop scheduling. The resource values are multivariate and then clustered to determine homogeneous clusters. The last stage is optimizing the nonlinear multipolynomials in the homogeneous cluster formed by using the Hessian solution. The results obtained are analyzed to get recommendations on business processes that are appropriate for the company’s needs. The impact of long waiting times will increase service costs, but by improving workload, costs can be reduced. The process model and the value of service costs resulting from the mechanism in the research can be a reference for process owners in evaluating and improving ongoing processes.

1. Introduction

The current business environment has no room for inefficiency as it can cause companies to lose out to their competitors, to lose customer trust, and to experience cost overruns. Therefore, companies must now focus on continuous monitoring and supervision and continuously make adjustments to their business operations to ensure that the company is working at an optimal level [1]. By monitoring the business processes carried out, the company will gain insights, be able to identify bottlenecks, anticipate problems, record policy violations, and recommend preventive actions, and be able to streamline processes to make them more efficient [2]. Real-time data processing retrieval is very important. How to understand data processing? Process mining becomes an important mechanism in improving the efficiency of company business processes. In this context, process mining is an enabler that can uncover the root causes of process inefficiency by reconstructing and visualizing current business processes as they are with their various variations [3]. In principle, process-mining techniques use event log data to find the process model and check the suitability of the process model with predetermined procedures and the ability to improve the model with a variety of information about bottlenecks, decisions, and resource use [4]. In business processes, the resource use will certainly affect the related processes whether they can run smoothly or not [5]. The initial mechanism of the process mining is the process discovery. The patterns of the process model obtained are explored to uncover situations that lead to errors and inefficiencies or exacerbate the risks of ongoing processes. This mechanism is often used to perform an analysis related to the resulting process model. The optimization of this model mostly focuses on activity trace that is recorded in the event log and focuses on optimization of the trace to obtain an optimal process model.

In the last 1.5 decades, the concept of process modelling has undergone significant development, and process mining has become a concept and its implementation is an effort to evaluate business processes. In several studies, it is stated that the largest part of real-life processes is complex and ad hoc and produces unstructured events [6]. Various efforts have been made to carry out the process mining for these conditions. The existing workflow discovery techniques had difficulty in dealing with event logs that were generated from highly flexible environments because the models that were mined from those logs often suffered from problems of high complexity. This condition then gave rise to various new techniques to obtain the expected process model. One technique that is still widely used is the trace clustering [7]. By using this technique, the existing event log is broken down into homogeneous subsets based on the activity trace, and then, a process model is created for each subset. Several studies use the same basic concepts, such as studies in [822], in which the optimization method used is trace clustering with a procedural approach. In general, the thing to do is to use a trace profile and then use a reference matrix profile to do clustering based on the size of the distance. Then, there is also a variation in the use of trace clustering with a model approach as done in [2328]. These studies present an approach that aims to overcome the difficulties experienced by extracting only useful data and presenting it in an understandable way. The mechanism is divided into two stages, namely, preprocessing and activity sequence clustering. In general, the steps taken are the active trace-clustering approach, generating a cluster trace based on a machine learning approach with a number of iterations until a cluster is obtained and then measuring the clustering results.

There are other studies that divide traces by using the time-based clustering approach [29]. The clustering approach consists of two steps: The first step is extracting the distance of activity pair for each log-trace. The second step is partitioning the logs given into subsequent trace clusters that have a similar structure. There is also a similar study but with a different time approach that is conducted in [30]. In their research, it is stated that when business processes are based on time, they are not always stable, in which certain times can affect the process activities. The trace clustering is based on the identification of similar times. This means that the clusters are formed both in structural similarity and in temporal proximity. There is another interesting approach in terms of modelling optimization with a trace clustering approach but by improving the quality of the resulting clusters as done in research studies by [3140].

There is also an approach that carries out a two-stage clustering mechanism, in which, in the first stage, the clustering is based on the employability indicators, and in the second stage, there is a grouping of the clusters obtained by using the AXOR algorithm that is based on a trace profile [41]. There is also a mechanism that combines trace clustering and text mining [42]. The use of text mining aims to improve the process discovery technique to extract more useful insights from data processing. The latest approach to track clustering is a study conducted in [43] by detecting the business areas that are most correlated with the clusters found.

Based on the explanation above, it can be seen that most of the trace clustering is based on vectors, then focuses on the model, and then on the context. Thus, in general, the clustering in previous studies exhibits clustering that is centered on flexible, complex, and specific process behaviors based on strict sequences of events. Basically, the trace clustering mechanism carried out is to obtain optimization of the resulting process model. Several studies that have started to use alternatives to the trace clustering approach are research studies by [44, 45] and [46], in which, in these studies, the process mining is carried out to use multiple perspectives. In their research studies, it is said that a multiperspective view of the process is needed, by looking at other things beyond the control perspective that determine the activity sequence of a process. The multiperspective view consists of control flow perspective, resources perspective, data perspective, time perspective, and function perspective. These five perspectives are often considered in the literature on business process management, process model, and process mining [44]. In the event log data, not only the activities but also the other resources, such as records of employee or staff working time, process service time, processing costs, and workload, are recorded. These indicators become important parts in measuring the processes that occur within the company [44]. In measuring process efficiency, it is not enough to just model the process through process discovery, but other indicators are needed that allow the process evaluation results to be more reliable [47].

Therefore, this study proposes an alternative in optimizing the process based on the resources recorded in the event log based on process mining, especially process discovery. The resources recorded in the event log are resources items that define the patterns and relate to the processes. This is in line with the concept stating that several interacting process perspectives, especially activities, data, resources, time, and other necessary indicators, can be considered further together. These things are not owned by the process model resulted from the trace clustering. The business process optimization is carried out in three stages. The first stage is optimizing the job shop scheduling method from the generated event log. The second stage is generating the resource value that is recorded in the event log from the results of analysis of the previous stage, namely, job shop scheduling. The resulting resource value, which is in the form of multivariate, is then clustered to determine a homogeneous cluster by using K-means clustering. The final stage is to perform nonlinear optimization on homogeneous clusters formed using the Hessian approach. The Hessian approach is used as a nonlinear multivariate solution approach to obtain a minimum service cost value. The use of three stages is based on the fact that appropriate steps are needed to optimize resources. The job shop model is used to analyze time based on scheduling that occurs. Utilization of the right time can improve performance effectiveness in order to reduce costs. Then, the use of multivariate resources clustering aims to explore the resources recorded in the event log.

The remainder of this article is structured as follows: Section 2 explains materials and methods including experimental setup, dataset, and our proposed method. Section 3 presents the results and discussion of the analysis obtained from the three approaches used. Section 4 evaluates the results and discussion. Finally, Section 5 contains the discussion of the analysis, and Section 6 contains conclusions and opportunities for this research in the future.

2. Materials and Methods

2.1. Process Mining and Process Discovery

Process mining is a relatively new science in the last decade. Basically, process mining is a technique that helps companies analyze and monitor internal processes to identify their activities and resources. In the past, companies would manage this process through a series of interviews that produced an overall ideal summary of the business. Process mining allows companies to automatically monitor and track business processes in real-time without human intervention [48]. The process mining system uses various forms of technology, including data mining, deep process analysis, and artificial intelligence (AI)-based business intelligence, to provide actionable insights into the enterprise processes [49]. Another important aspect of process mining is that it enables business processes to identify process risks and bottlenecks, to determine potential improvements that can improve efficiency, and to monitor data to predict future events [50]. The process mining mechanism starts by capturing event logs that are obtained from the daily system run by the company. Then, in the event log, a process-mining mechanism is carried out, which consists of three stages, namely, process discovery, conformance checking, and enhancement [48]. The three stages are carried out according to the needs of the analysis, and they do not have to be sequential but can be random.

Process discovery is a data-based automated mechanism for finding, mapping, and documenting existing business process activities. Then, an analysis of the data that are obtained automatically is carried out so that it can be recommended automatically in the modelling process or in the workflow [51]. The discovery and analysis of a company’s business processes can be used to identify key problem areas, not only at the start of a digital transformation initiative but also when improving the performance of existing processes. Along with business process modelling, process discovery is an important part in improving the quality of business process management [52].

2.2. Production Process Analysis and Scheduling

Production is one of the most important parts of a company related to the transformation of various inputs into outputs in accordance with the predetermined quality standards. Meanwhile, the production process is a series of steps used to convert inputs into outputs. Basically, the production process flow is based on the concept of scheduling. Scheduling can be divided into three types, namely, job shop, flow shop, and project [53]. Apart from that, there is also the problem of scheduling open shop, mixed shop, and general shop [54]. Scheduling is the process of managing resources to complete tasks that involves work, resources, and time. According to [55], scheduling is the allocation of resources to carry out a series of tasks based on time. Flow shop scheduling or so-called flow shop is a production system that produces its products with similar process flow or sequence. The overall process flow of the product is constant. Resource management follows the process flow of the product. Flow shop has two types, namely, continuous flow shop and intermittent flow shop [56]. In the continuous flow shop, materials continuously move (in and/or out) in one process without waiting for the process to finish. The example of continuous flow shop can be found in the chemical industry. Meanwhile, in the flow shop, new material moves from one process, when the process is finished, the material moves to the next process.

In job shop scheduling, the operations performed by each job are often on a different route. In this case, [57] identified several objectives of scheduling activities as follows:(1)Increasing resource usage or reducing waiting time(2)Reducing the number of jobs waiting in the queue(3)Reducing some work delays that have a deadline for completion in order to minimize late fees (cost of delay)(4)Assisting in making decisions regarding enterprise capacity planning and the type of capacity required

In addition to the above discussed concepts, to complete the scheduling analysis, a dispatching rule approach is needed. Dispatching rule is a method used for both static and dynamic job shop scheduling cases [58]. Priority rules provide guidance for the sequence of work to be carried out [59]. The rules are particularly applicable for process-focused facilities. Priority rules are applied to reduce completion time, number of jobs processed in one unit of time, and processing delays due to resource availability. The scheduling problems that will be used in this study are the flow shop and job shop. This is due to the case context that is used more to the flow shop and job shop conditions.

2.3. Method

The flow of the study conducted is shown in Figure 1.

The research steps shown in Figure 1 are described in detail.

2.3.1. Generating Event Log

The first step is generating an event log based on system activity records according to the explored business processes. In this study, a dataset about repair services is used and has been adapted to the needs of the analysis [60]. The dataset used is an event log data about the repair process. The business process of improvement services is the focus of process modelling in this study.

2.3.2. Preprocessing

After getting the event log, a preprocessing step is carried out to tidy up the event log obtained in preparation for generating indicators.

2.3.3. Determining the Number of Activity Classes

In this step, activities are grouped by class. In the event log, different classes of activity can occur from one case to another. The initial step in the framework of this study is determining the number of activity classes. The number of activity classes determines the process path based on the number of activities. In the process that is recorded in the event log, the number of classes may have different activities depending on the system record. This step can be visually seen in Figure 2. The step is necessary to see how many forms of activity paths occur.

After determining the number of activity classes, the next step is generating a resource indicator from the event log that has been divided by the activity class. The resource indicator is based on the objectives desired by the analyst. In this study, the basic indicators produced are workload and waiting time. These two indicators are important things that affect the process [61, 62]. Then, as a complement, another indicator is selected, namely, service cost [63]. Thus, there are three indicators of concern in this study, namely, workload, waiting time, and service cost.

2.3.4. Production Process Analysis

In this step, a job shop analysis is performed in the event log by the activity class.

2.3.5. Enriching Indicators from the Event Log

In this step, the event log indicators are generated, which are resulted from the production process analysis. The steps are as follows:(a)Event log indicator generation: In the previous step, it has been stated that after determining the activity class, the indicator generation is carried out. The indicator generation is based on need. In this study, the indicators generated are workload, waiting time, and service costing. These 3 things are the reasons for using related resources as multivariate variables in research. This is reinforced by several studies that always focus on these three things alternately, such as the research studies conducted in [6466] and [67]. This means that in this case there are very few studies that simultaneously use all the three.(b)The basic dimensions of the indicators generated in this study are workload, waiting time, and service costing. Activity indicator is one of the important indicators because the process that occurs in each case can be determined by the number of activities carried out. From one case to another, the number of activities in the process may differ. It all depends on whether the case is complicated or not.

Workload analysis measures the number of activities performed by each resource in a given period. Workload analysis shows the number of activities started, activities completed, and activities in progress. Using the activity frequency, workload analysis provides information about how much is the workload of each resource. There are many ways to find the workload value in each case in the event log. One way is using the definition from [68] that workload is the number of activities that has been carried out during a certain period. The definition of “how busy” the resource is generally means that it only looks at the resource quantitatively not qualitatively. This is because it only looks at the amount of work completed in a certain period, and it is not based on a standard assessment, in which the workload assessment can differ from one company to another. Another method is stated in Gabriel’s research that uses the factor evaluation system (FES) [69]. The FES mechanism uses clear measurements because it is based on predetermined standards and weights. Therefore, the assessment is based on achievements in the workload. This formula was developed and adapted in this study.

There are many definitions related to waiting time in different contexts that depend on the case being discussed. Waiting can be caused by a person or machine that is required for the process, even if the person or machine is working at full capacity. This is shown in detail in Figure 3.

Figure 3 shows that the green color shows the working time of item or document, while the red color shows the waiting time that the work item or document will be worked on in the next process. In an event log process, there may be different waiting times between one case and the other. Thus, in general, the waiting time calculation equation is as follows:

In general, the total service cost is obtained based on the combined sum of the workload cost equation ( and the waiting time cost equation (. This can be seen in Figure 3. The calculation of service costs is based on the calculation of each workload cost and waiting time. The calculation concept is based on the project triangle approach [70]. In the project triangle concept, implementation always experiences problems, namely, cost, time, and quality. It is impossible in its implementation that a product or service will be able to apply these three components. Only two components apply out of three options. If a product or service wants a fast time and has good quality, then the value of the goods will be expensive. Otherwise, if you want cheap goods and fast time, the desired quality of goods will not be obtained [71]. In this study, quality becomes the default feature so that only two components are compared. This means that, in this case, cost and time are two inversely proportional components. Thus, in compiling the equation, we apply the inverse ratio between cost and time as the basis of the equation. The following is a detailed calculation of each equation.

(1) Activity cost: The activity cost equation is calculated in each case. The value is based on the standard cost of each activity that has been set. This means that the activity cost depends on the number of activities in each case because there may be differences in activity costs from one case to another. The activity cost equation is represented as follows:

(2) Workload cost: The workload cost value is calculated as follows:

(3) Waiting time cost: The impact of waiting on cost-effectiveness is highly scenario dependent. The scenario used in this study is a scenario where the cost per unit time decreases with the waiting time. Thus, in this context, the longer the waiting time (meaning that the longer the service duration), the lower the service fee. The simple logic used in this case is process cost versus waiting time, which is represented as follows :

(4) Service costing: The determination of service costing is based on the concept of quality. In principle, of course, it is not true that if the waiting time experienced by customers in completing the process is very long, the costs are also large. Supposedly, a short service process will have high costs consequences, while a long service process results in low costs. This is the basis for this research. Based on the project triangle concept, it will never be possible that there will be product quality along with fast turnaround times and low costs [70]. What happens is that there are only 2 possibilities for the 3 elements [71]. This means that by determining that the process being carried out prioritizes product quality, the service cost is inversely proportional to the speed of the process completion time. Based on the calculation of these resource components and the basic concept, the total cost per case can be calculated as follows:

Total service costing (TSCi) is an accumulation of activity costs, workload costs, and waiting time costs.

2.3.6. Multivariate Clustering of Resources

Clustering the event log with the indicator has been selected. In this case, after obtaining the indicator value from the generated event log, multivariate clustering will be carried out. In this context, the approach used is K-means. K-means clustering algorithm provides easy implementation, very flexible, and easy to use adaptation [72, 73]. In addition to the time needed to do learning faster, it also uses simple principles that can be explained in nonstatistical terms. This is the reason for using K-means clustering compared to the other method approaches.

2.3.7. Optimized Resource Based on Multivariate Clustering

After getting the clustering results, the last step is optimizing the selected clusters by using the Hessian method and then performing a comparison analysis of the selected clusters between class activities. The reason for using the Hessian method is because the objective function formed is a nonlinear multipolynomial function. Therefore, a nonlinear programming problem solving method is needed to get the optimum solution. The method to get solutions to nonlinear problems is by direct or indirect solutions. One of the direct solutions in the context of this study is using the Hessian solution approach without constraints. This is done because of the characteristics of the resulting data [74].

2.3.8. Evaluation

The most commonly used performance evaluation metrics such as accuracy, sensitivity, and precision are calculated and compared for the resulting clusters. Sensitivity and precision are used to ensure the performance of K-means clustering. The formulation is as follows:where tp, tn, fp, and fn are true positive, true negative, false positive, and false negative, respectively.

2.3.9. Result Analysis

In this step, the analysis of the results obtained is carried out.

3. Results and Discussion

The first step in this research is to do a sampling on the event log to facilitate the simulation process [75]. The dataset of event log obtained is shown in Table 1.

Table 1 shows the event log in which activities are in the repair process, such as registering, analyzing defects, repairing (simple), and repairing (complex). Each case ID shows that the number of activities may differ from one case to another. In terms of determining the activity class from the event log, there are many types of variants or classes. The type of activity class can be detected through ProM as shown in Table 2.

The number of events in each case causes different variants in each service process provided. The number of activities between one customer and another customer may vary depending on the importance and designation of the repair services carried out. In principle, each variant is a normal activity. The difference in the number of activities does not mean that there is an abnormal activity, although in the process modelling, it is often found that there are “abnormal” activities. The idea in this study is to explore the process model that occurs through multiperspectives. Thus, in this case, the number of classes is grouped to facilitate exploration. Therefore, the next step is grouping Case ID in the class based on the number of activities as shown in Tables 36.

The first analysis carried out as a part of the mechanism in this study is analyzing the flow shop and job shop of the production system concept. The basis for doing this optimization is to see if in the repair process, there may be queues that can cause process delays. Therefore, if an analysis of the production system is performed, it can be seen that the queuing process can be improved through a job shop analysis approach, in order to see the optimal scheduling time as an improvement in providing future improvement processes. The first step in this stage is to determine the job and machine in the scheduling analysis based on the event log used. In job shop scheduling, every process that must be passed is considered as a machine. This is illustrated in Table 7.

Then, based on the machine notation that has been set, the next step is arranging input in the form of a job shop, as shown in Table 8. The analysis shown is the treatment carried out in class 1, which has a total of 5 activities (in this study, all treatments are only presented in class 1, while the overall results of the treatment are shown at the end).

The next step is analyzing the production system in the event log as shown in Table 9, which has been compiled by setting the rules. The collection of data processing obtained is data resources, process flow, and the processing time of each job. The rules used in the process are as follows:(i)The company operates every day for 12 hours(ii)The completion of the order process is based on the FPTO (finish production time of order) schema. Finish production time is met when all the items ordered have passed the processing stage.(iii)An activity exactly visits one machine once(iv)Each machine can only process one activity at a time and is always ready at all times without any breakdowns or under repair(v)An activity cannot be interrupted. In other words, once an activity has taken place, it must be completed before any other activity is processed on the same machine.(vi)After an activity has been processed on a machine, the activity will be transferred directly to the next machine(vii)Each activity has a certain type of activity, which is carried out by a certain machine and in a certain processing time(viii)Human and machine resources are set to be available 12 hours per day for 7 days per week(ix)No machine breaks down or stops working(x)Service processes are based on orders and are carried out sequentially(xi)The standard fee used for one full service is 100,000 (in this case, we modify it in Indonesian rupiah)

Many methods can be used to analyze the job shop scheduling performance, such as FCFS (first come-first served), SPT (short processing time), LPT (long processing time), and EDD (earliest due date). The choice of method is very dependent on the needs. Each method has advantages and disadvantages. In the analysis of scheduling carried out, the three methods used are FCSC, SPT, and LPT, as shown in Table 9.

In addition, it can also be seen in Figure 4.

Based on Table 9, it is found that the job shop scheduling with SPT rule has a shorter flow time compared to other rules. This can also be seen in Figure 4. In principle, the selection of which scheduling to use is left to the user because each priority rule has its own advantages and possibilities. Different evaluation results on different order data may occur. However, in this study, the SPT rule is chosen. The graphs show that the SPT rule minimizes makespan 28818 and average flow time 873.27. Therefore, the next step is using the event log based on the job shop scheduling analysis and the selected SPT rule. These results are in accordance with the research studies conducted in [76, 77]. The next mechanism is performing clustering in the event log by previously generating resource values for workloads, waiting times, and service costing. The values are shown in Table 10, which are based on the calculation results of equations (3)–(5).

The process of performing clustering is based on the fact that the resulting multiple perspectives or resources may vary or are not homogeneous. Before doing clustering, the data have been tested by using one of the indicators of the classical assumption (the heteroscedasticity test) with a 5% confidence level. The data having been tested show that the data have heteroscedasticity or nonhomogeneous conditions. The clustering method uses K-means with the Davies Bouldin approach. Davies Bouldin’s value obtained is shown in Table 11.

Based on Davies Bouldin’s value, the number of the selected cluster is 4. Cluster number 4 is selected because in the 4th iteration, Davies Bouldin’s value drops to 0.410, which means that it stops at the 3rd iteration and that the event log is divided into four clusters. The results of K-means clustering show that the order of clusters from the best cluster is cluster 1, cluster 2, cluster 3, and cluster 0. This is shown in Tables 1215.

From each table, it can determine the objective function model with the Hessian solution approach. The objective function model is generated by using an approach in the form of multi-nonlinear polynomial. After going through the linearity test, by using a nonlinear multipolynomial equation approach, the objective function for cluster 0 in class 1 is obtained as equation (6). Cluster 0 is chosen because it is the best cluster compared to other clusters in class 1.

In this context, the resulting objective function has no constraints. Therefore, the solution approach that can be used to find the minimum service costing value is the Hessian optimization solution approach. By using the Hessian optimization solution for nonlinear multipolynomial equations, the minimum cost function value is 102363.07 at  = 320 and at  = 57.27.

The same process is carried out on the other clusters, and the results are shown in Table 16.

Table 16 shows the results of the comparison of the minimum costs for each class.

4. Evaluation

In this section, measurements are made to see the performance of K-means clustering. Table 17 shows the performance of the 4 classes produced in this study.

The performance results show that the F-measure value reaches 0.92 in almost all classes. This shows that the clustering performance is good enough.

5. Discussion

Based on the simulation, it is found that the job shop schedules with SPT rule as an initial reference to generate resource value from the process. This mechanism shows that the patterns of running services can be seen from the types of transactions that occur. This can provide input for the process owner to consider which type of transaction can take precedence, in order to get the desired time optimization. From the analysis results of the job shop, the optimal time sequence is obtained, and the values (workload, waiting time, and service costs) needed in the next analysis are generated and then are clustered based on the number of classes. The calculation results show that cluster 0 is the cluster that has the most optimal objective function value compared to other clusters in class 1. From four classes, four process models (that are accompanied by a minimum cost value) are obtained, which are described by using a Petri net with inductive miner algorithm. Figures 58 show the process model of each class.

The results obtained indicate that the model in Figure 7 has the lowest cost compared to other models. When referring to the minimum service cost, the process owner will definitely choose the model in Figure 7. However, in this case, further analysis is needed to determine which process model is better. Note that the described process model comes from 4 classes. Figure 5 is a process model generated from class 1. Class 1 has an optimal cost value of 102363.07 with  = 320 and  = 57.21. This value can be optimized in the process by eliminating the restart repair activity. While in Figure 7, which comes from class 3, the repair (complex) activity is eliminated. This causes the process cost to be optimal at 109008.58 at points  = 322.93 and  = 19.55. In addition, what can be used as a comparison is that the process model in Figure 7 has a lower waiting time value compared to the process model in Figure 6. The consequence of a low waiting time results in a larger workload value. This has the consequence that service costs become more expensive. While, the models in Figures 6 and 8 have the same number of activities. The difference lies in the last activity, where the model in Figure 6 eliminates the repair archive activity, while the model in Figure 8 eliminates the restart repair activity. The process model in Figure 6 has an optimal process value at service costs of 104610.99 which has a value of  = 151.04 and  = 31.20. While, the process model in Figure 8 has an optimal service cost of 106708.98. With this analysis, the process model that occurs in the cluster can be a reference for process owners in evaluating and improving ongoing processes.

The weakness of this research is that almost half of the simulation activities are still carried out manually in Excel, while the rest use applications, such as POM-QM, ProM, DISCO, and RapidMiner. To overcome this, it is necessary to build an algorithm as a basis for compiling a special analytical tool, so that it can be a solution to the optimization problem in this study. Then, the case study used is a study that allows job shop analysis where the process flow does not occur constantly or each operation can be carried out by more than one agent. In addition, we have limitations in terms of available datasets. Apart from that, we also have limitations in conducting simulations such as the dispatching rule mechanism. Therefore, further research refinement steps are needed.

6. Conclusions and Future Works

The study that has been conducted provides good alternative results because it makes a major contribution to the analysis of the process in the case used in it. The mechanism is carried out in several stages, namely, production system analysis, multivariate clustering on class activities, and cluster optimization that provides a new perspective in analyzing the process. The production system analysis stage is able to produce a more optimal scheduling analysis in handling processing time, in which this is not a concern in most studies on the process mining. The resulting job shop scheduling analysis provides notes for the process owner to pay attention to which activity sequence that causes the process time to be more optimal with the SPT Rule. This will be an input in the management planning process within the company.

Then, the process of generating resource indicators becomes an advantage in the stages carried out. Generating the value of resource indicators makes the analysis stronger because it not only focuses on the sequence of activities but also pays attention to multiperspective aspects. The process that occurs is rich in information, so it must be a concern. Multivariate clustering stages on resource indicators are able to cluster the event log to be more homogeneous based on the resulting indicators. The resulting information can add indicators of the analysis, in addition to clustering based on the sequence of activities.

The last stage in this research is the optimization of clusters resulting from the K-means clustering process. With the results of clustering performance which is quite good at 0.92, resource indicator data can be used as the basis for optimization calculations. Optimization results provide information on which processes are optimal based on service cost indicators. Cost optimization becomes a significant force in seeing which process is the most optimal. It provides meaningful input into the overall process analysis. The approach taken in this research provides a new perspective in viewing and analyzing processes within the company. The collaboration between multiperspective aspects of process mining and production systems will open up new insights and perspectives on related research studies. There are still many multiperspective aspects of the process that have not been touched upon and require further study. The notational explanation of all the formulas in this article is shown in Table 18.

Data Availability

The event log data used to support the findings of this study are deposited in the Mendeley Data repository (DOI: 10.17632/w9hpjv84k9.1).

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

The authors would like to thank their colleagues from the Computer Science Doctoral Program with scheme no. IF186401.at the FTEIC Intelligent Information Management laboratory, Institut Teknologi Sepuluh Nopember, Surabaya, as pleasant colleagues during the research discussion and PPM unit of Telkom University that has provided assistance and support in this research process.