Zum Hauptinhalt springen

The least sample size essential for detecting changes in clustering solutions of streaming datasets.

Atif, M ; Farooq, M ; et al.
In: PloS one, Jg. 19 (2024-02-20), Heft 2, S. e0297355
Online academicJournal

The least sample size essential for detecting changes in clustering solutions of streaming datasets  1 Introduction

The clustering analysis approach treats multivariate data tuples as objects and groups them into clusters based on their similarities or dissimilarities within the dataset. However, in modern world, a significant volume of data is continuously generated from diverse sources over time. In these dynamic scenarios, the data is not static but continually evolves. Consequently, the interesting patterns and inherent subgroups within the datasets also change and develop over time. The researchers have paid special attention to monitoring changes in cluster solutions of evolving streams. For this matter, several algorithms have been proposed in the literature. However, to date, no study has examined the effect of variability in cluster sizes on the evolution of cluster solutions. Moreover, no guidance is available on determining the impact of cluster sizes on the type of changes they experience in the streams. In the present simulation study using artificial datasets, the evolution of clusters is examined concerning the variability in cluster sizes. The findings are substantial because tracing and monitoring the changes in clustering solutions have a wide range of applications in every field of research. This study determines the minimum sample size required in the clustering of time-stamped datasets.

The clustering approach is an unsupervised learning problem that considers multivariate data tuples as objects and partitions them into a prespecified number of clusters. Entities in each cluster are comparably similar to one another than entities belonging to different clusters. Commonly, the similarity is measured in terms of a distance function between each pair of objects [[1]]. Some well-known similarity measures include Euclidean, squared Euclidean, Manhattan, cosine, and Chebyshev distance functions. The notion of a cluster is not specifically defined, and the standards can vary significantly from case to case [[2]–[4]]. Some algorithms aim to minimise the intracluster variation, while others identify clusters as the dense region in the feature space. From the perspective of cluster definition, the traditional algorithms are divided into five categories namely partitioning, hierarchical, density-based, grid-based, and model-based clustering algorithms [[5]].

In today's world, a bulk of information is continually generated by different sources over time. In these applications, the data are not stationary but rather evolves. Consequently, the interesting patterns and natural subgroups in the datasets do not remain stagnant; they change and evolve over time. In such a dynamic environment, the entire training dataset is not available to the learning algorithm at once. Consequently, a sequence of cluster solutions needs to be generated at some discrete points in time [[6]]. This phenomenon raises an important question: are these cluster solutions static over time or do they experience any transitions? Significant work has been conducted in monitoring and tracking changes in these clustering solutions obtained at successive time points in all these years.

Over the past two decades, practice of monitoring and tracing the evolution of clustering solutions within data streams has gained considerable importance across various domains. The clusters undergo certain changes, and comprehending the type of transitions that occur can provide a significant advantage. The elements within the clusters may migrate from one cluster to another, leading to the disappearing, merging, and splitting of the clusters at later time points. Similarly, the surviving clusters can adopt internal transitions, such as changes in size, location, and cohesion [[7]]. Applications of monitoring changes in cluster solutions span a diverse range of fields and scenarios. Some key applications include, businesses utilize cluster monitoring to cluster customers based on purchasing behavior, detecting sudden shifts in cluster solutions can help identify anomalies in data streams, monitoring cluster changes can reveal evolving patterns in patient data, Monitoring cluster evolution in social networks helps uncover changing community structures, tracking changes in species distribution through clustering helps assess ecosystem health and detect potential threats, changes in product data clusters can reveal production variations and help maintain quality standards, etc. Atif et al. [[9]] provided a comprehensive review of the literature on the various techniques used to monitor and trace the development of clustering solutions over time. The discussion has made it abundantly clear that tracking and monitoring changes in a dynamic environment are essential for forecasting the future and formulating policies. Nevertheless, segmentation studies are inherently exploratory and are greatly influenced by the number of variables and the sample size. A comprehensive simulation study is crucial to estimate the minimum sample size necessary to ensure statistical validity without the need to reduce the number of variables.

The novelty in estimating the minimum sample size necessary for monitoring changes in cluster solutions lies in its pioneering approach. This research breaks new ground by addressing the often-overlooked aspect of sample size determination in the context of evolutionary cluster analysis. By focusing on this critical element, it offers a fresh perspective on ensuring the validity of results, and adapt to evolving data patterns.

The key contribution of the present study is to establish sample size requirements for data-driven segmentation analyses. This provides data analysts with a valuable tool to assess whether the available sample for monitoring and tracking changes in clusters is adequate, considering the number of variables within the segmentation base. It helps analysts determine whether they should collect additional data or reduce the number of variables used in their analysis to ensure robust and meaningful results. A comprehensive review of the literature has been conducted, and the algorithms proposed for tracing the cluster solutions are discussed in the next section.

2 Related work

The evolutionary clustering framework was first introduced by Chakrabarti et al. [[10]]. This framework generate a series of clustering solutions {ξ1, ξ2, ⋯, ξn} at consecutive discrete time points. In other words, it provides a clustering solution ξi = {X1, X2, ⋯, Xk} for each time point ti in the stream. The algorithm works by simultaneously maximising two criteria. First, each clustering solution must directly represent the dataset at the associated time point as closely as possible. Second, the clustering solution should not significantly deviate from the result obtained at the adjacent time point. The evolutionary clustering framework was initially developed for k-means and agglomerative hierarchical clustering techniques. The spectrum clustering problem is expanded within the evolutionary paradigm by Chi et al. [[11]], producing more reliable and consistent outcomes. Spectrum evolutionary clustering is more stable against long-term drift and less sensitive to short-term noise. However, these algorithms are incapable of detecting the emergence of new clusters and the aging of existing clusters. To address this issue, Zhang et al. [[12]] combines the concept of evolutionary clustering with density-based algorithms to prevent arbitrary groups from developing and disappearing in dynamic social networks. These techniques greatly improve the evolutionary clustering literature; however, they cannot capture a variable number of clusters over time. The hierarchical Dirichlet process and the hidden Markov model are combined by Xu et al. [[13]] to handle this problem and significantly boost performance.

Using self-organising maps, Denny and Squire [[7]] proposed a method for identifying structural changes in cluster solutions of temporal datasets. This approach involved comparing clustering results at various points in time and monitoring how the new results differed from the old one. This research work uses world development indicators to assess the performance of the proposed algorithm. However, this algorithm failed to detect the newly emerged and disappearing clusters. This issue was addressed by the Relative Density Self-Organising Map (ReDSOM), a visualization-based approach developed by Denny et al. [[14]]. The algorithm could recognize various changes in cluster structures, including the formation of new clusters, disappearance, splitting, merging, expansion, centroid movement, and changes in cohesiveness.

The MONIC framework, used for modeling and monitoring changes in clustering solutions of cumulative data streams over time, was introduced by Spiliopoulou et al. [[8]]. This framework compares the results obtained at two subsequent time periods to track the structural changes in the clusters. The changes adopted by the clusters can generally be categorized into two groups: external transitions and internal transitions. External transitions encompass the survival, merging, splitting, disappearance, and re-emergence of clusters. Internal transitions involve adjustments in the size, cohesion, and location of the surviving clusters. The overlap, a non-symmetric matrix that serves as the foundation for the MONIC framework, is expressed by the following expression:

Graph

Overlap(Xi,Yj)=|XiYj||Xi|,i=1,2,,k1,j=1,2,,k2 (1)

where Xi is a member of the set of clusters produced by the first clustering and Yj is a member of the set of clusters produced by the second clustering. A matrix of order k1*k2 is produced, where k1 and k2 represent the number of clusters from the first and second clustering, respectively. The similarity index between clusters is represented by the value on the appropriate element of the matrix and acts as a marker for tracking the external transition. The cluster membership is assessed to track the internal transition of the clusters that have survived.

The Monitoring Clusters Transition (MClusT) algorithm, developed by Oliveira and Gama [[15]], visualizes the transition of clusters on a bipartite graph by utilizing conditional probabilities as edge weights. MClusT incorporates a tracking technique based on graph theory, a transition detection algorithm, and a taxonomy of transitions. For each pair of clusters obtained from the stream at successive time points, the algorithm calculates conditional probabilities. These conditional probabilities act as indicators for monitoring the cluster solutions. For more in-depth information on real-time detection of changes in clusters using density-based algorithms, interested readers can refer to articles [[16]–[18]]

3 Methodology

This paper aims to demonstrate the influence of variability in cluster sizes on clusters' temporal evolution. Additionally, it seeks to identify the minimum sample size required in dynamic streams. For this purpose, we implement the MONIC framework in R-software, which can be downloaded from the URL https://CRAN.R-project.org/package=clusTransition [[19]]. A comprehensive literature review reveals the introduction of several algorithms designed for monitoring the changes in cluster solutions of streaming datasets. However, to our knowledge, no research work has been conducted to examine the effect of variability in cluster sizes on their evolution in streams. Consequently, there was no guidance available to choose the performance measures and assess how cluster size influences the changes adopted by the clusters over time.

3.1 Methods

In supervised learning, the performance of the model is evaluated by comparing the predicted class with the true class labels of the outcome attribute. However, in unsupervised learning problems, the true class labels are not provided to the learning algorithms. Hence, assessing the performance of unsupervised learning algorithms is quite challenging [[21]]. Given the lack of pertinent literature, identifying performance indicators in this research study proved to be a formidable challenge. Generally, the consistency of cluster solutions over time and the accurate identification of changes in the stream represent critical performance indicators. In this study, we applied both the conventional logistic regression and the generalized additive logistic regression models to fit the binary response variable. The logistic regression model utilizes the logistic function to model the probability of a specific event. In this research, we consider survival of the smallest cluster as event of interest and record the dichotomous response variable yi used in these models as:

Graph

yi={1,Ifthesmallestclustersurvive(withprobabilityπ(x))0,Ifthesmallestclusterexperiencetransition(withprobability1-π(x)) (2)

It is computed as follows: In each iteration of the simulation, the datasets generated at time point t2 have identical cluster centers to the those that evolved at time point t1. Ideally, all the clusters at time point t2 should survive, and therefore, the small cluster in the stream should not experience any external transitions. The response variable records the status of the smallest cluster, indicating whether it survives or undergoes a transition. To model this phenomenon, the conventional logistic regression model is given as:

Graph

Model1:logit(πi(x))=α+j=1pβj(xj) (3)

where xj's represents the co-variates included in the study and βj's are the corresponding regression coefficients. The predictors and their respective levels included in the study are presented in Table 1. These predictors encompass the size of the smaller cluster, the number of clusters, the number of variables, and the separation between clusters. The model includes both their main effects and second-order interaction terms. The term π(x) = P(yi = 1|X) computes the conditional probability that the smallest cluster survives at subsequent time point. We can write the binary generalised additive model structure as:

Graph

Model2:logit(πi(x))=α+S1(Size)+j=2pSj(xj) (4)

Graph

Table 1 Overview of the factors used in simulation study.

FactorsLevels
Number of variables (d)2, 3, 4, 5, 10, 15
Size of small cluster (n1)10*d*k, 20*d*k, 30*d*k, 40*d*k, 50*d*k
60*d*k, 70*d*k, 80*d*k, 90*d*k, 100*d*k
Number of clusters (k)3, 4, 5
Time points2
Separation value-0.1, 0.0, 0.1
Total number of datasets6*10*3*2*3 = 1080

1 The larger clusters in the datasets consist of 5000 observations each. The datasets were simulated at two time points and the changes in cluster solutions were monitored.

The terms S1 is the smooth non-parametric basis function that transform Size with the specific form depending on the chosen basis. On the other hand, Sj's are the linear functions capturing the effects of other co-variates.

Generalised Additive Models (GAMs), introduced by Hastie and Tibshirani [[23]], belong to a family of statistical models that utilize smoothing functions to capture the non-linear relationships between the response variable and predictors, thereby accommodating the data's complexities. GAMs offer a reliable approach for smoothly fitting unseen data while avoiding excessive model complexity. The core concept involves fitting smooth non-linear functions to a set of predictors Xi to uncover the relationships among the variables in the model [[24]]. In the case of GAMs, the relationship between individual predictors and the response variable follows some linear or non-linear smooth pattern. This pattern can add up to predict the expected value of the dependent variable.

3.2 Data generation

Clustering analysis is an unsupervised learning technique that seeks interesting patterns in datasets without pre-existing true class labels. Since the true class labels are unknown, and hence the true structure can not be predicted in the dataset [[26]]. To overcome this limitation, we use the simulated datasets to achieve the study objectives. One significant advantage of using a simulation study over real-life datasets is the availability of true class labels in simulated data. This enables the tracking of data item migration between clusters based on their actual class memberships, facilitating monitoring of the cluster evolution.

3.3 Clustering algorithm

The MONIC framework operates under the assumption that each data item is exclusively assigned to one and only one cluster. This assumption effectively eliminates the feasibility of applying density-based and model-based clustering algorithms. This constraint leads us to focus on partitioning methods for clustering. Among these methods, one of the most common and suitable choices is the traditional k-means algorithm. The true number of classes (i.e., number of clusters generated in each simulation) was used as a relevant value of k. The k-means [[28]] is one of the most widely used learning algorithms for partitioning data into a specified number of clusters k. Here, k represents the optimal number of clusters pre-specified by the analyst. The algorithm optimizes two competing criteria: objects belonging to the same cluster display high intra-class similarity, while those from distinct clusters exhibit low inter-class similarity.

4 Results and discussion

The results section is organized as follows: Section 4.1 presents a preliminary analysis mapping the impact of cluster size variability on the evolution of clusters in a data stream. Ten streams were specifically created for this purpose, with data items emerging at two successive time points in each stream. At time point t1 in each stream, the dataset comprised four distinct clusters with varying sizes (i.e. one small and three large clusters). Each of the larger clusters contained 5000 observations, while the size of small clusters in the streams were 50, 100, 150, 200, 250, 300, 350, 400, 450, and 500 respectively. At time point t2, four clusters with cluster centers similar to those that emerged at t1 were simulated, each containing 2,000 observations. The consistency and stability of the clusters in the second clustering are assessed, as the clusters at subsequent time periods have similar centers.

In section 4.2, data streams were generated at two consecutive time points. To ensure reliable recommendations different circumstances are encountered in artificial datasets, including 1) number of variables (d), 2) sample sizes, 3) number of clusters (k), 4) and separation value between neighboring clusters. The datasets were generated using genRandomClust() functions from clusterGeneration and cluster.Gen() function from clusterSim package in R [[19], [30]].

The number of respondents in the small cluster was determined based on the recommendations of Qiu and Joe [[32]]. Their recommendations indicate a linear dependence of sample size on the number of variables (d) and number of clusters (k). These two covariates can significantly influence the outcomes and behavior of unsupervised learning algorithms. The number of variables directly affects the complexity of analysis, and computation time. The number of clusters, on the other hand, directly determines the granularity of grouping within the data. A smaller k may result in clusters that are too generalized, while a larger k can lead to fine-grained clustering that might not capture meaningful patterns. This results in 10*d*k, 20*d*k, 30*d*k, 40*d*k, 50*d*k, 60*d*k, 70*d*k, 80*d*k, 90*d*k, and 100*d*k objects simulated across each data scenario. This recommendation also serves to justify the utilization of simulated datasets in unsupervised learning, as the actual number of clusters remains unknown in real datasets. Table 1 below present an overview of the factors included in the study.

4.1 Preliminary analysis

Fig 1 below demonstrates the external and internal transitions adopted by the clusters at time point t2 in relation to different sizes of the small cluster. Though the datasets at time point t2 being simulated with centers identical to the clusters evolved at t1. Yet, extremely unstable clustering solutions were obtained at succeeding time points due to variability in the cluster sizes.

Graph: The x-axis represents the smaller cluster size, while the y-axis denotes the number of clusters experiencing the corresponding transition.

Sub-plot A in Fig 1 illustrates that the algorithm identified one survived, two merged, and one split candidate at time point t2. However, as the size of the smaller cluster exceeded 150 observations, two survived, two merged, and one newly emerged candidate were detected. Moreover, once the size of the smaller cluster reached 400 observations, all four clusters survived at the subsequent time point. Similarly, sub-plot B in Fig 1 depicts the internal transition of the surviving clusters at time point t2. It can be observed that the survived cluster is more diffused than its ancestor when the size of the smaller cluster is 100 observations or less. On the other hand, when the size of the smallest cluster exceeds 400 observations, one cluster experiences a change in cohesion, becoming more compact than its previous state.

This unstable clustering solution is explained in Fig 2, highlighting the impact of cluster size variability in the stream. The color scheme represents the predicted class, whereas the symbols depict the true class labels in the stream. The clustering solution at time point t1 clearly shows that the smallest class (⋅) in the dataset is a part of its neighboring larger cluster. However, with the introduction of new data points at time point t2, the cluster splits into two daughter clusters. Meanwhile, class (+) was divided into two clusters, which then merge into a single cluster at the subsequent time. The method accurately identifies one split and two merged clusters at time point t2 due to this phenomenon. However, because the smallest cluster contains more than 150 observations, it is recognised as a recently emerging candidate. Consequently, the adjacent cluster undergoes changes in cohesiveness and location. Furthermore, as the smallest segment comprises over 400 observations, all clusters manage to survive. In light of this fact, even the smallest class is now recognised as a separate cluster.

Graph: Sub-plot A represents cluster solution at time point t1. The sub-plot B represent cluster solution at time point t2 when the size of smaller cluster is 50, whereas sub-plot C represent cluster solution at time point t3 when size of smaller cluster is 250.

4.2 Model 1: Logistic modeling

To further investigate the impact of sample size on the survival of clusters over time in streaming datasets, we analyzed simulated data using logistic regression models. According to Table 2 in the usual logistic regression model, the coefficients of eight factors (Sample size, d.3, d.4, d.5, d.10, d.15, k.5, and Separation.0.1) are found to be statistically significant. Whereas six two factor interactions (d.3:Separation.0.1, d.15:Separation.0.1, k.5:Separation.0.1, d.5:k.5, d.10:k.5, and d.15:k.5) are found significant at 1% level of significance (see details in Table 2).

Graph

Table 2 Summary of regular logistic and GAM logistic regression models.

FactorsModel 1Model 2
EstimateStd. ErrorPr(>|z|)EstimateStd. ErrorPr(>|z|)
(Intercept)-38.170955.764963.56e-11 ***-49.1299615.855100.001944 **
Sample size0.347310.049492.25e-12 ***--
d.37.979182.686480.002977 **1.149722.594550.657674
d.412.253832.979233.90e-05 ***-1.00243.670160.784747
d.513.422423.057801.14e-05 ***-4.30944.860850.375317
d.1018.803603.56971.38e-07 ***-2.91726.224570.639307
d.1521.875553.87681.68e-08 ***0.000000.00000NA
k.44.298162.204650.051225-3.37501.900110.075693
k.52.841532.770073.56e-06 ***0.000000.00000NA
Separation.02.662062.185510.2232052.616642.357640.267061
Separation.0.110.83642.577682.62e-05 ***7.018572.77470.000424***
d.3:Separation.0-5.141932.453500.036104*-4.183142.459550.088985
d.4:Separation.0-2.859702.369170.227415-2.278182.470280.356407
d.5:Separation.0-2.841622.386190.233710-2.211942.532060.382350
d.10:Separation.0-1.666732.371780.482222-0.452073.039390.881761
d.15:Separation.0-1.664302.365730.4817426.357788.879110.473968
d.3:Separation.0.1-6.352382.454490.009652 **-4.450372.622130.089652
d.4:Separation.0.1-4.085942.403810.0891741.257813.619870.728235
d.5:Separation.0.1-4.030482.376420.08988110.5344113.79740.445163
d.10:Separation.0.1-4.100042.405820.08834117.2823813.27440.192942
d.15:Separation.0.1-6.352682.454330.009644 **1.617243.135080.605957
k.4:Separation.00.871421.604540.5870620.340371.832690.852662
k.5:Separation.0-0.305211.602080.848913-1.415352.026440.48490
k.4:Separation.0.1-1.436241.598460.3689123.586272.726710.188431
k.5:Separation.0.1-4.941251.731750.004326 **-5.01202.390890.036054
d.3:k.41.691852.385240.4781411.395362.270340.538817
d.4:k.40.513882.373060.8285603.568472.963950.228606
d.5:k.40.480542.348610.8378794.452652.999510.137687
d.10:k.4-0.636942.352530.78658616.766111.30120.137923
d.15:k.4-1.786482.363300.44969515.4050813.827570.265243
d.3:k.50.463072.384560.8460232.843012.693420.291180
d.4:k.5-4.136322.435200.089403-0.068373.179330.982843
d.5:k.5-7.671372.572110.002859 **-4.976852.797580.075242
d.10:k.5-8.761432.636950.000892 ***0.3410633.981050.991992
d.15:k.5-9.961512.728080.000261 ***-2.026519.060570.823020
Smooth term---edfRef.dfp-value
Sample size---6.436.8557.16e-05 ***

Subsequently, we enhance our model to GAM by incorporating smooth functions of the explanatory variables. According to Table 2 all two factor interaction terms are found to be insignificant at 1% level of significance (Pvalue > 0.01). However, the Separation.0.1 is the only significant factor (Pvalue <.01) in the GAM model. The edf of smoothing term for sample size is 6.43 (Pvalue < 7.1e − 05) indicating that the effect of sample size on log odds is not linear.

4.2.1 Model comparison

Now, let's compare our regular logistic model with GAM logistic model using multiple performance standards. Table 3 below demonstrates some of these performance standards for model comparisons.

Graph

Table 3 Model selection standards.

ModelsAICBICPseudo R2
Model 1169.9385320.143481.3
Model 2145.3006318.807394.3

Comparing these standards indicates that it is now safe to conclude that the GAM performs considerably better than the regular logistic model. To further validate this comparison, we conducted the standard test using R-software's anova function to statistically compare the two models. The results of the analysis of deviance for model comparison are presented in Table 4.

Graph

Table 4 Analysis of deviance table.

ModelsResid. DfResid. DevDfDeviancePr(>Ch5)
Model 1505.0099.938---
Model 2499.5764.4415.429635.4971.912e-06 ***

Now, with additional statistical evidence, we can confidently propose that incorporating the non-linear relationships of the covariates significantly enhances the performance of the model.

4.2.2 Interpretation

A comprehensive analysis of the results reveals that variability in cluster sizes has a considerable impact on the evolution of clusters. To provide adequate recommendations regarding minimum size required for detecting the correct transitions, smooth functions are plotted. Fig 3 below represents the smooth function of sample size on the log-odds of the dependent variable, as explained in Eq 2, considering all levels of the covariates. The impact of an increased sample size on the accurate detection of the transition is highly significant. The curve demonstrates a sharp ascent initially, followed by a gradual flattening once the size of the smallest class exceeds 70. In summary, for correct detection of transitions in aggregate settings, the recommended sample size is determined to be 70 * k * d.

Graph: The standard error limits are shaded.

However, Separation.0.1 is also significant in the GAM model, which suggests that separation between neighboring clusters also plays a vital role. Fig 4 below demonstrates the smooth function of sample size and its effect on the log-odds of dependent variable for different values of separation. Clearly, if there is no obvious structure in the dataset, i.e., if neighboring clusters are close to one another, then a higher sample size is required to achieve significantly improved results. In such cases, the impact of increasing the sample size of the smallest cluster is particularly strong. Thus, if there is no obvious structure, higher sample sizes are required. On the other hand, if clusters are well separated from each other, i.e., Separation.0.1, then the function is almost horizontal. This suggests that there is no additional effect of sample size once the size of the smallest cluster exceeds 50*k*d. The smooth function of sample size on the log-odds of the dependent variable explained in Eq 2 over all levels of the covariates are provided in S1, S2 Figs and S1 Text available in supporting information.

Graph: The adequate sample size is mentioned with dashed lines.

Clustering methods heavily rely on the sizes of the actual classes in a dataset. Therefore, finding the appropriate structure in datasets becomes quite tricky if the cluster size are inadequate. The issue of determining minimum sample sizes required for the smallest cluster in a dataset has been explored by a limited number of studies. For example, Dolnicar et al. [[33]] performed a study using simulation techniques. They put forward a suggestion regarding the smallest sample size required for data-driven market segmentation in order to precisely detect clusters. This also plays a role in the transitions of clusters when clustering time-stamped datasets. If the clustering algorithm fails to capture the underlying pattern accurately, it can lead to highly unstable cluster outcomes at consecutive time points.

5 Conclusion

Clustering is a data organization technique that groups data points into clusters. The key characteristic is that points within the cluster are more similar to each other than to those in different clusters. However, in recent times, a substantial volume of data is generated in the form of continuous streams. Consequently, the underlying structure of data evolves over time, leading to the development of algorithms specifically designed to monitor and adapt to these changes. In recent decades, researchers have focused on tracking changes in cluster solutions for temporal datasets. However, to the best of our knowledge, no study has been conducted to explore how variations in cluster sizes affect the temporal evolution of clusters. The primary objective of this study is to determine the minimum sample size required for effectively monitoring changes in temporal datasets. Results of study reveal that increasing size of the smallest cluster within the dataset significantly improves the outcomes. Especially in situations where no distinct segmentation structure exists, larger sample sizes are required for precise detection of the changes. The findings suggest that at least 70*k*d observations are required in the smallest cluster. This implies that, for accurately detecting transitions in temporal datasets, a sample size equal to 70 times number of variables and number of clusters is deemed sufficient. However, in cases where the dataset lacks clear segmentation, significantly higher sample sizes are necessary. These findings are substantial because tracing and monitoring the changes in clustering solutions have a wide range of applications in every field of research.

6 Limitations and future work

The MONIC framework operates under the assumption that each individual object is exclusively allocated to a single cluster. This assumption restricts the study to only partitioning clustering algorithms, effectively ruling out the possibility of applying density-based and model-based clustering algorithms. In the future, this study could be expanded to encompass density-based algorithms.

Supporting information

S1 Text

Smooth function of sample size.

(TXT)

S1 Fig

Smoothing curve from GAM model in which the survival of smaller cluster has been modeled as a smooth function of the sample size.

(TIF)

S2 Fig

Smoothing curve from GAM model in which the survival of smaller cluster has been modeled as a smooth function of the sample size.

(TIF)

S1 Data

(XLSX)

Footnotes 1 The authors have declared that no competing interests exist. References Han J., Kamber M., & Pei J. (2012). Cluster Analysis: Basic Concepts and Methods. 3rd ed. Morgan Kaufmann Publishers Inc. San Francisco, United States (pp. 443–495). https://doi.org/10.1016/B978-0-12-381479-1.00010-1. 2 Estivill-Castro V. ((2002)). Why so many clustering algorithms: A position paper. ACM SIGKDD Explorations Newsletter, 4(1), 65–75. https://doi.org/10.1145/568574.568575 3 Nasraoui O., & CEB N. (2019). Clustering methods for big data analytics: Techniques, toolboxes and applications. Springer Nature Switzerland (pp. 150–180). 4 Wierzchon S., & Klopotek M. (2018). Modern algorithms of cluster analysis. Springer International Publishing Switzerland. 5 Witten I.H., Frank E., Hall M.A., & Pal C.J. (2017). Data mining: Practical machine learning tools and techniques. 4th ed. Morgan Kaufmann series. Elsevier Science. 6 Miloslavskaya N. ((2020)). Stream data analytics for network attacks' prediction. Procedia Computer Science, 169, 57–62. https://doi.org/10.1016/j.procs.2020.02.114 7 Denny., & Squire D.M. (2005). Visualization of cluster changes by comparing self-organizing maps. In: Wei D., Liu Y., Zhu F., Zang L., Zhou W., Lu Y., & Hu S. Advances in Knowledge Discovery and Data Mining: Lecture Notes in Computer Science (pp. 410–419). https://doi.org/10.1007/1143091948. 8 Spiliopoulou M., Ntoutsi I., Theodoridis Y., & Schult R. (2006). Monic: Modeling and monitoring cluster transitions. In: Wei D., Liu Y., Zhu F., Zang L., Zhou W., Lu Y., & Hu S. Advances in Knowledge Discovery and Data Mining: Lecture Notes in Computer Science (pp. 706–711). https://doi.org/10.1145/1150402.1150491. 9 Atif M., Shafiq M., & Leisch F. ((2023)). Applications of monitoring and tracing the evolution of clustering solutions in dynamic datasets. Journal of Applied Statistics, 50(4), 1017–1035. https://doi.org/10.1080/02664763.2021.2008882, 36925905 Chakrabarti D., Kumar R., & Tomkins A. (2006). Evolutionary clustering. In: Proceedings of the ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (pp. 554–560). https://doi.org/10.1007/978-0-387-30164-8271. Chi Y., Song X., Zhou D., Hino K., & Tseng B.L. ((2009)). On evolutionary spectral clustering. ACM Transactions on Knowledge Discovery from Data, 3(4). https://doi.org/10.1145/1631162.1631165 Zhang Y., Liu H., Deng B. (2013). Evolutionary clustering with dbscan. In: Ninth International Conference on Natural Computation (ICNC) (pp. 923–928). https://doi.org/10.1109/ICNC.2013.6818108. Xu T., Zhang Z., Yu P.S., & Long B. (2008). Evolutionary clustering by hierarchical dirichlet process with hidden markov state. In: Eighth IEEE International Conference on Data Mining (pp. 658–667). https://doi.org/10.1109/ICDM.2008.24. Denny., Williams G.J., & Christen P. ((2009)). Visualizing temporal cluster changes using relative density self-organizing maps. Knowledge and Information Systems, 25(2), 281–302. https://doi.org/10.1007/s10115-009-0264-5 Oliveira M., & Gama J. (2010). Bipartite graphs for monitoring clusters transitions. Berthold ed. In: Cohen P.R., Adams M.R., Advances in intelligent data analysis. Springer Berlin Heidelberg (pp. 114–124). Huang L., Wang C.D., Chao H.Y., & Yu P.S. ((2020)). MVStream: Multiview Data Stream Clustering. IEEE Transactions on Neural Networks and Learning Systems, 31(9), 3482–3496. https://doi.org/10.1109/TNNLS.2019.2944851, 31675346 Fahy C., & Yang S. ((2019)). Finding and Tracking Multi-Density Clusters in Online Dynamic Data Streams. IEEE Transactions on Big Data, 1(1). https://doi.org/10.1109/TBDATA.2019.2922969 Fahy C., Yang S., & Gongora M. ((2019)). Ant Colony Stream Clustering: A Fast Density Clustering Algorithm for Dynamic Data Streams. IEEE Transactions on Cybernetics, 49(6), 2215–2228. https://doi.org/10.1109/TCYB.2018.2822552, 29993761 R Core Team. (2020). R: A language and environment for statistical computing. R Foundation for Statistical Computing. Vienna, Austria. https://www.R-project.org/. Atif M., & Leisch F. ((2022)). clusTransition: An R package for monitoring transition in cluster solutions of temporal datasets. PLoS ONE, 17(12), e0278146. https://doi.org/10.1371/journal.pone.0278146, 36520935 James G., Witten D., Hastie T., & Tibshirani R. (2017). An introduction to statistical learning: With applications in r. tth ed. Springer International Publishing Switzerland. Basar S., Ali M., Ochoa-Ruiz G., Zareei M., Waheed A., & Adnan A. ((2020)). Unsupervised color image segmentation: A case of RGB histogram based K-means clustering initialization. PLoS ONE, 15(10), e0240015. https://doi.org/10.1371/journal.pone.0240015, 33091007 Hastie T., & Tibshirani R. ((1986)). Generalized additive models. Statistical Science, 1(3), 297–310. http://www.jstor.org/stable/2245459 Breiman L. ((2001)). Statistical modeling: The two cultures (with comments and a rejoinder by the author). Statistical Science, 16(3), 199–231. https://doi.org/10.1214/ss/1009213726 Rigby R.A., & Stasinopoulos D.M. ((2005)). Generalized additive models for location, scale and shape. Journal of the Royal Statistical Society: Series C (Applied Statistics), 54(3), 507–554. https://doi.org/https://doi.org/10.1111/j.1467-9876.2005.00510.x. Alloghani M., Al-Jumeily D., Mustafina J., Hussain A., Aljaaf A.J. (2020). A systematic review on supervised and unsupervised machine learning algorithms for data science. In: Berry M.W, Mohamed A, Yap (Eds.). Supervised and unsupervised learning for data science. Springer International Publishing (pp. 3–21). https://doi.org/10.1007/978-3-030-22475-21. Basar S., Waheed A., Ali M., Zahid S., Zareei M., & Biswal R.R. ((2022)). An Efficient Defocus Blur Segmentation Scheme Based on Hybrid LTP and PCNN. Sensors, 22(7), 2724. https://doi.org/10.3390/s22072724, 35408338 MacQueen, J. (1966). Some methods for classification and analysis of multivariate observations. 1st ed. In: Fifth berkeley symposium on mathematical statistics and probability, vol 1: Statistics (pp. 281–297). Basar S., Ali M., Ochoa-Ruiz G., Waheed A., Rodriguez-Hernandez G., & Zareei M. ((2021)). A Novel Defocused Image Segmentation Method Based on PCNN and LBP. IEEE Access, 9, 87219–87240. doi: 10.1109/ACCESS.2021.3084905 Qiu W., & Joe H. (2020). Clustergeneration: Random cluster generation (with specified degree of separation). [R package version 1.3.5]. https://CRAN.R-project.org/package=clusterGeneration. Walesiak M., & Dudek A. (2020). The choice of variable normalization method in cluster analysis. In: Proceedings of the 35th international business information management association conference (ibima) seville, spain (pp. 325–340). Qiu W., & Joe H. ((2006)). Generation of random clusters with specified degree of separation. Journal of Classification, 23(2), 315–334. https://doi.org/10.1007/s00357-006-0018-y Dolnicar S., Grun B., Leisch F., & Schmidt K. ((2014)). Required sample sizes for data-driven market segmentation analyses in tourism. Journal of Travel Research, 53(3), 296–306. https://doi.org/10.1177/0047287513496475

By Muhammad Atif; Muhammad Farooq; Mohammad Abiad and Muhammad Shafiq

Reported by Author; Author; Author; Author

Titel:
The least sample size essential for detecting changes in clustering solutions of streaming datasets.
Autor/in / Beteiligte Person: Atif, M ; Farooq, M ; Abiad, M ; Shafiq, M
Link:
Zeitschrift: PloS one, Jg. 19 (2024-02-20), Heft 2, S. e0297355
Veröffentlichung: San Francisco, CA : Public Library of Science, 2024
Medientyp: academicJournal
ISSN: 1932-6203 (electronic)
DOI: 10.1371/journal.pone.0297355
Schlagwort:
  • Sample Size
  • Computer Simulation
  • Cluster Analysis
  • Algorithms
Sonstiges:
  • Nachgewiesen in: MEDLINE
  • Sprachen: English
  • Publication Type: Journal Article
  • Language: English
  • [PLoS One] 2024 Feb 20; Vol. 19 (2), pp. e0297355. <i>Date of Electronic Publication: </i>2024 Feb 20 (<i>Print Publication: </i>2024).
  • MeSH Terms: Algorithms* ; Sample Size ; Computer Simulation ; Cluster Analysis
  • References: Sensors (Basel). 2022 Apr 01;22(7):. (PMID: 35408338) ; Stat Methods Med Res. 1995 Sep;4(3):187-96. (PMID: 8548102) ; IEEE Trans Neural Netw Learn Syst. 2020 Sep;31(9):3482-3496. (PMID: 31675346) ; J Appl Stat. 2021 Dec 7;50(4):1017-1035. (PMID: 36925905) ; PLoS One. 2022 Dec 15;17(12):e0278146. (PMID: 36520935) ; PLoS One. 2020 Oct 22;15(10):e0240015. (PMID: 33091007) ; IEEE Trans Cybern. 2019 Jun;49(6):2215-2228. (PMID: 29993761)
  • Entry Date(s): Date Created: 20240220 Date Completed: 20240222 Latest Revision: 20240222
  • Update Code: 20240222
  • PubMed Central ID: PMC10878511

Klicken Sie ein Format an und speichern Sie dann die Daten oder geben Sie eine Empfänger-Adresse ein und lassen Sie sich per Email zusenden.

oder
oder

Wählen Sie das für Sie passende Zitationsformat und kopieren Sie es dann in die Zwischenablage, lassen es sich per Mail zusenden oder speichern es als PDF-Datei.

oder
oder

Bitte prüfen Sie, ob die Zitation formal korrekt ist, bevor Sie sie in einer Arbeit verwenden. Benutzen Sie gegebenenfalls den "Exportieren"-Dialog, wenn Sie ein Literaturverwaltungsprogramm verwenden und die Zitat-Angaben selbst formatieren wollen.

xs 0 - 576
sm 576 - 768
md 768 - 992
lg 992 - 1200
xl 1200 - 1366
xxl 1366 -