Abstract
This studyemploys classification and clustering methodologies on datasets derived from digital transformation and Internet of Things (IoT) initiatives within the cable and automotive sectors. The analytical procedures are conducted utilizing the KNIME platform, employing Support Vector Machines (SVM) and K-Means algorithms. The results indicate that SVM exhibits superior accuracy rates compared to K-Means within both industries. The data collection methodology facilitated by the Mert Software IoT platform is identified as reliable and efficacious. The primary objective of this article is to augment decision-making precision in digital transformation software and contribute to the scholarly discourse within this domain.
Keywords: Machine Learning, Classification, ClusterAnalysis, Industry 4.0.
1.Introduction
Digital transformation has emerged as a significant topic in factories in the world over the recent years. The concept of the Internet of Things (IoT) represents all entities connected to the internet through a network [1]. Digital transformation begins by strategically determining the items that will undergo production on assembly lines within a factory. Subsequently, it encompasses the identification of the staff scheduled for duty during production, specifying their cycle times, and recognizing instances of downtime in the course of the process. Digital transformation assumes a pivotal role in these dimensions, given its paramount importance for industry economies. These changes are imperative for the future and have garnered increased prominence in light of technological advancements. Detecting downtime during production is a critical factor that directly affects manufacturing. When a machine fails, this directly impacts the whole factory, and the length of duration is often unpredictable. The software solutions aim to provide real-time notifications to reduce downtime. The ultimate goal is to increase production quantities. In pursuit of this goal, the data collected by Mert Software has been transformed into knowledge discovery toolthrough classification and clustering analyses. Clustering analysis is used to classify data based on similarities, particularly for data with an unknown number of groups and unclassified data. It is a technique that groups data into discrete clusters based on similarities with respect to units or variables [2]. This study concentrates on two distinct industries: the automotive industryand the cable production industry.The rationale behind choosing two disparate manufacturing sectors lies in the distinct operational characteristics inherent to each. Within the cable production industry, a singular product is manufactured sequentially utilizing a solitary machine. In contrast, the automotive industry permits the simultaneous production of multiple products. A pivotal demarcation lies in the StockId field within the data sets, exhibiting variation for each record. This study endeavors to assess the efficacy of identical algorithms by applying them to these sectors characterized by disparate production methods.
Klein M., in her article titled "Scenarios of Digital Transformation in Enterprises -A Conceptual Model Proposal," underscores the imperative for businesses to formulate comprehensive digital transformation strategies encompassing processes, business models, customer relationships, and management. Within her study, she delineates the trailblazers of diverse digital transformation methodologies, foreseeing their role in aiding businesses in the development of robust digital transformation strategies [3]. In their article titled "Formation of Digital Factories with Production Tracking Systems and Conceptual Data Analysis," published in 2022, Mustafa and Halil discussed the significance of production tracking, including planning on workstations, raw material inputs, inventory tracking, quality control processes, and maintenance processes, as well as the importance of Key Performance Indicators , (KPIs) for businesses. Their study delved into the relationship between Manufacturing Execution Systems , (MES) and Enterprise Resource Planning (ERP) systems, highlighting the advantages of digitizing every moment within businesses through these processes. They found that this digitization would significantly reduce production costs, increase labor productivity, decrease production expenses, and facilitate inventory tracking [4].
Altuntaş, conducted a quantitative study involving 114 corporate executivesin the article titled "The Impact of Digital Transformation Practices on Corporate Brand Value". The study revealed that 89% of the executives had engaged in digital transformation initiatives, and 34% reported increased revenue and improved customer relationships as a result. Altuntaşargued that the Industry 4.0 revolution would yield positive outcomes for every sector. Furthermore, the author anticipated that digital investmen ts, particularly those enabling brands in the fast-moving consumer goods and retail sectors to offer personalized services to customers, would strengthen brand loyalty andcontribute to an increase in brand value [5]. In the article titled "The Impact of the Internet of Things (IoT) Technology on Businesses in the Digital Transformation Process," published by Çarkin 2020, the author conducted a study to understand the influence of the Internet of Things (IoT) technology on businesses and to assess the existing literature on the subject. The study utilized content analysis as its methodology and reviewed data from the Web of Science (WoS), Ulakbim, and DergiPark data bases. Within the context of the digital transformation process, Çarkprovided recommendations for preparing and adapting individuals, organizations, and society as a whole to Industry 4.0 technologies. These recommendations aimed to facilitate a healthy transition and adaptation to the digital transformation era [1].
Gürkan's thesis, titled "Development of an Intelligent Factory Management and Information System Supported by Industry 4.0 and Digital Transformation Technologies," introduces a novel approach[6]. The primary goal is to establish a high-tech automation system infrastructure in industry while minimizing human interventi on. This approach aims to reduce error rates during the production phase to a minimum and produce high-quality products. It also focuses on enabling factories, especially in terms of production performance, to self-optimize via network connectivity. The objective is to maximize production speed in factories and minimize production costs. In the course of this research, Gürkandeveloped anAndroid-based IoT application and implemented it, using electronic cards, specifically in smart marble factories, particularly in the marble drying phase. As a result of this study, several advantages were realized, including the ability to facilitate planned production, contribute to high-quality production, accelerate mass production processes, and minimize production losses [6].Kaynarand his colleagues, in their 2016 article titled "Sentiment Analysis with Machine Learning Methods," conducted sentiment analysis on datasets containing movie reviews from IMDB using classification analyses, specifically the Support Vector Machines , (SVM) algorithm and Multilayer Perceptrons , (MLP) algorithms. The study revealed that the SVM analysis had a significantly higher accuracy rate compared to other methods [7].
2.Materials and Methods
2.1. KNIME Platform
KNIME is a platform that processes data and enables reporting through relationshi ps between nodes. Being open source, it is open to further development by programmerswho can add additional features to the system. It is generally used in data analysis applications in business intelligence processes. It has various components, and these components are called nodes. Analyses are performed through nodes without the need for coding. The outputs of each node can be viewed and interpreted separately. It is widely used in research related to data analytics. With its powerful features and open system architecture, it is becoming increasingly popular [8].
2.2. Support Vector Machines (SVM)
Classification refers to the process of appropriately distributing data into predefin ed classes within a dataset. Classification algorithms are used to learn the characteristics of classes from training data and to predict to which class incoming test data belongs. Among classification algorithms, the most commonly used ones in the literature include Naive Bayes, Artificial Neural Networks (ANN), Support Vector Machines (SVM), K-Nearest Neighbors (KNN), and the KStar algorithm. In this study,SVM was employed. SVM is generally usedfor pattern recognition and classification problems. It trains a support vector classifier using a multi-term kernel. It normalizes all attributes with predefined data [9]. The SVMalgorithm is based on Lagrange multiplier equations. Its primary objective is to find the optimal separating hyperplane that best divides data points into different classes. Support Vector Machines accommodate two scenarios: linear and non-linear. Linear SVM is applied to problems that are linearly separable. The structure of linear SVM is illustrated in Figure 1(a).
When there is no linear separability, moving the data to a higher-dimensional space can be considered as a solution. This is the fundamental theory behind non-linear Support Vector Machines. SVM achieves these transformations using kernel functions. In this case, a dataset of dimension 𝑎is transformed into a new dataset of dimension 𝑏, where 𝑏> 𝑎. The structure of this SVM is presented in Figure 1(b). There are numerous kernel functions developed for SVM, and in this study, a radial basis kernel function is utilized. The equation of this function is presented in Equation 1.[13]
2.3. K-meansAlgorithm
Cluster analysis is a method that groups units under investigation in a research study into specific categories based on their similarities, enabling classification, revealing common characteristics of units, and making general descriptions related to these classes [10]. In cluster analysis, grouping is done based onsimilarities and differences. The inputs involve similarity measures or the necessity of calculating which similarities can be applied to the data. Depending on the purpose and field of use, the objectives of cluster analysis are as follows:
1.Identifying the correct types.
2.Building models.
3.Predictions based on groups.
4.Hypothesis testing.
5.Data exploration.
6.Hypothesis generation.
7.Data reduction.
In cluster analysis, distances are calculated between rows of the data matrix. In the formulas, "i" and "j" represent rows of the data matrix, "k" represents columns, "x_ik" represents data in the "i"th row and "k"th column, and "p" represents the total number of variables.The K-Means clustering method is an effective technique for evaluating many commercial datasets due to its efficiency in clustering large datasets. Being widely used in practical applications for over fifty years, K-Means has become the preferred clustering method in this study for several reasons, such as its speed in comparison to hierarchical clustering methods when there is an expectation of forming a low number of clusters, and its ease of implementation.
The K-Means clustering method is a simple yet effective algorithm for creating clusters based on the available data. The application steps for this method can be outlined as follows [14]:
Step 1: Determination of the number of clusters to which the dataset needs to be partitioned.
Step 2: Random assignment of initial cluster centers for k clusters.
Step 3: Identification of the nearest cluster center for each cluster data point.
Step 4: Calculation of the cluster centroid for k clusters and updating the position of each cluster center based on the new centroid value.
Step 5: Iteration of the processes between steps 3 and 5 until convergence or termination is achieved.In the third step, the distance from each cluster data point to the nearest cluster center is determined using the Euclidean Distance formula:
3.Dataset
In manufacturing facilities where production processes occur, signals generated by machinery, including production and stoppage signals, are conveyed to operator panels through an electronic card known as IOCARD. The IOCARD serves the function of converting signals originating from the machinery into readabledata strings, which are subsequently transmitted to the panel software. The panel program, developed in C#, undertakes the reception of this data and parses it based on station-specific definitio ns. For instance, upon receipt of a signal indicating an increase in production count, the program increments the production accordingly. Similarly, if a signal denoting operational status is received, the program sets the status to 'operational.' In the event of a stoppage signal, the system state is modified to a predetermined 'stop' status. Notably, the system accommodates the establishment of approximately 40 distinct definitions.In this study, the accuracy of the production data collected by Mert Software IoT platformwas used for classification and clustering. This study was conducted using datasets collected in the automotive and cable industries. In Table 1, an example dataset used in the cable industryis provided. In Table 2, an example dataset used in the automotive industryis provided. Table 3 explains the variables, and Table 4 specifies the data types.
4.Evaluation
The process of collecting data from production and reporting the classified and clustered analyses of the collected data is as follows:
-Determination of points to receive signals from the machine.
-Installation of SQL SERVER for recording data collected from the machine. Activation of the system.
-Analysis of the collected data. In this phase, the system conducts initial checks before system analyses, verifying the existence of shift definitions. If a shift definition exists, the system initiates operations and performs clustering analysis for Personnel, Production, and the Manufactured product. The system operates synchronously, conducting Classification analysis for Production and Time.
-Following these analyses on the raw data, the results are utilized in reporting and dashboard tools for faster and analyzed data presentation.
4.1Confusion Matrix
It is a metric used to evaluate the performance of an algorithm in classification problems. The confusion matrix visualizes the number of correct and incorrect classifications by comparing the actual class labels with the predicted class labels. The confusion matrix is typically presented in the form of a 2x2 table, as shown in Figure 3, but it can be larger for multi-class classification problems. The 2x2 confusion matrix includes the following four elements:
True Positive (TP): The number of true positives. TP increases when the actual class label is positive, and the predicted class label is also positive.True Negative (TN): The number of true negatives. TN increases when the actual class label is negative, and the predicted class label is also negative.False Positive (FP): The number of false positives. FP increases when the actual class label is negative, but the predicted class label is positive. It is also known as a false alarm.False Negative (FN): The number of false negatives. FN increases when the actual class label is positive, but the predicted class label is negative. It represents the cases that were missed.
Confusion matrix is used to calculate various performance metrics using these four elements. Among these metrics, accuracy, precision, recall, and F1 score values are included. The confusion matrix is an important tool for understanding the performance of classification algorithms, assessing the strength of the model, and understanding classification errors [11].4.2Performance MetricsTo obtain reliable accuracy results, some measurements are made using the values in the confusion matrix. These measurements are achieved with the accuracy, precision, recall (sensitivity), F1 score, and specificity formulas [11]
4.2.1 Accuracy
Accuracy represents the accuracy value, which is the ratio of correct classifications to the total number of classifications.
4.2.2 Precision
Precision gives the ratio of correctly classified data to all the positives. Here is the formula for precision:
4.2.3Sensitivity (Recall)
Sensitivity provides the ratio of data correctly classified as positive to the actual positive data. Here is the formula for sensitivity (recall):
4.2.4 F1 Score
F1 Score is a value calculated by taking the harmonic mean of precision and recall values. Here is the formula for calculating the F1 Score:
4.2.5 Specificity
Specificity is a value calculated by taking the ratio of data correctly classified as negative to the actual negative data. Here is how you can calculate specificity:
5.Results
5.1 KNIME Workflow
The workflow developed on the KNIME platform is illustrated in Figure 4. Initially, data is read using the "Excel Reader" node, and then irrelevant elements in the dataset are removed using the "Column Filter" node. Elements such as primary key fields like "CompanyId" and "ID" are among those removed from the dataset. To facilitate learning, the dataset is divided into 80% for training and 20% for testing using the "Partitioning" node. Data is normalized for classification analysis. The "Normalizer" node is used to normalize the dataset, applying min-max normalization to the fields "PstopId," "PID," and "QTY," scaling values to a range between 0 and 1.Subsequently, the dataset is split for training and testing through the "Partitioning" node. The training set provides insights into how well the model explains information in the target variable, while the test set indicates the model's performance with unseen observations. In the modeling phase, learning and prediction nodes for Artificial Neural Networks are added to the model.Following the Normalizer node, for classification analysis, the data is linked to the "SVM Learner" node. This node selects the "Status" field for classification. Then, the output of the Normalizer node and the SVM Learner node are combined in the "SVM Predictor" node. This node aligns the learned data from the SVM Learner with the test data input. Subsequently, it connects to the "Scorer (JavaScript)" node for evaluation. Simultaneously, for clustering analysis, the "k-Means" node is employed, receiving input from the Partitioning node based on the fields "QTY" and "QTY2." This node generates two classes with a defined iteration count of 90. The output from the k-Means node, the Clustering Model, is linked as input to the "Cluster Assigner" node. Following the clustering analysis, to interpret the cluster names as "Working" and "Stopping," the "String Replacer" node is connected. This node replaces the name of cluster_0 with "Working" and cluster_1 with "Stopping." Subsequently, it is connected again to another "String Replacer" node, where cluster_1 is named "Stopping." Finally, the workflow connects to the "Scorer (JavaScript)" node and the "Scorer" node forevaluation.
5.2 Suppor t Vector MachineResults
After the model completes its job, the Scorer node produces results, and as shown in Figure 5, an Overall Accuracy of 74,14% and an Overall Error of 25,86% were obtained.
6. Discussion and Conclusion
In this study, an attempt has been made to classify and cluster the production, downtime, and staffdata of companies involved in digital transformation projects by Mert Software. Machine Learning methods such as classification and clustering analyses were employed. In the data sets we used in this study, the SVM algorithm produced more successful results than the K-means algorithm in both the cable industry andthe automotive industry.The study revealed that the method employed by Mert Software for digital transformation projects achieves high accuracy rates.
A major problem with industrial data is that sometimes operators make decisions that can lead to incorrect feedback. This affects the accuracy of the data and causes problems. In this study, we selected examples from factories where signals about production and downtime are collected automatically so that operator errors do not affect data quality. We believe that this study can contribute to the decision-making accuracy of digital transformation software and contribute to scientific research in this field
publication address : https://journals.orclever.com/oprd/article/view/280/190