Abstract
The classic Fuzzy C-means (FCM) algorithm has limited clustering performance and is prone to misclassification of border points. This study offers a bi-directional FCM clustering ensemble approach that takes local information into account (LI_BIFCM) to overcome these challenges and increase clustering quality. First, various membership matrices are created after running FCM multiple times, based on the randomization of the initial cluster centers, and a vertical ensemble is performed using the maximum membership principle. Second, after each execution of FCM, multiple local membership matrices of the sample points are created using multiple K-nearest neighbors, and a horizontal ensemble is performed. Multiple horizontal ensembles can be created using multiple FCM clustering. Finally, the final clustering results are obtained by combining the vertical and horizontal clustering ensembles. Twelve data sets were chosen for testing from both synthetic and real data sources. The LI_BIFCM clustering performance outperformed four traditional clustering algorithms and three clustering ensemble algorithms in the experiments. Furthermore, the final clustering results has a weak correlation with the bi-directional cluster ensemble parameters, indicating that the suggested technique is robust.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Avoid common mistakes on your manuscript.
1 Introduction
As one of the most commonly used data analysis methods in machine learning, data mining, and artificial intelligence, clustering divides data sets into clusters according to their features so that the sample points in the same cluster are highly similar, while those in different clusters are dissimilar [1]. Without knowing the class relations of any data, clustering is an unsupervised learning method, which is widely used in image processing [2], information security [3], and market analysis [4], and other fields since it can discover the potential values of data. According to various clustering theories, clustering can be classified into five categories [5]: partition-based clustering, density-based clustering, grid-based clustering, hierarchical clustering, and model-based clustering. Among them, partition-based clustering such as K-means [6] and FCM [7, 8] are simple and efficient. Therefore, they are broadly applied in the engineering field.
First proposed by Dunn in 1974 and different from the hard clustering of K-means, FCM [7] offers a more flexible way of clustering and introduces fuzzy membership. The fuzzy membership function is introduced in fuzzy mathematics and uncertainty theory [9, 10]. The optimal memberships and cluster centers of sample points are obtained by iteratively calculating and are finally clustered following the principle of maximal membership. Although FCM has been applied successfully in many sectors since its emergence, it still has some shortcomings, e.g., sensitivity to initial cluster centers, noise points, and boundary points, poor performance in unbalanced data sets, and the tendency for local optima during iteration.
This work proposes a bi-directional FCM clustering ensemble technique that takes local information into account (LI_BIFCM) to address the drawbacks of FCM. The suggested technique considers not just clusters diversity but also sample points local information.
The following are the paper’s significant innovations and contributions:
-
1.
A vertical clustering ensemble is used to keep the algorithm stable.
-
2.
Using multiple horizontal clustering ensembles, the technique effectively prevents border points from being misclassified.
-
3.
LI_BIFCM increases clustering performance even further by employing horizontal and vertical ensembles.
The rest of this paper is laid out as follows: the second section examines and summarizes the associated research. The core ideas of FCM are briefly described in Sect. 3. The LI_BIFCM method is discussed in detail in Sect. 4. The experimental evaluation of the suggested algorithm is discussed in Sect. 5. Section 6 wraps up this paper and looks ahead to future research.
2 Related Works
Many FCM-derived algorithms have been presented and examined in past research work to address FCM’s shortcomings.
To overcome its tendency for local optima, two improved FCM algorithms (named FCM-IDPSO and FCM2-IDPSO) were proposed by Silva [11], which dynamically adjusted the parameters based on improved particle swarm optimization and provided a better balance between exploration and exploitation. Experimental results suggested that the proposed method delivered excellent clustering results at a faster speed. To automatically identify the cluster center and initial location of FCM, an entropy-based fuzzy clustering method was presented by Yao [12] and to determine the cluster center by calculating the entropy of each sample point. A new method was also devised to estimate the initial membership functions of fuzzy sets, which has good predictive output values. Proposed by Ding et al. [13], the improved FCM algorithm combined genetic algorithm and Gaussian kernel technique, which could overcome the deficiency of FCM’s failure to determine the number of clusters and boost the clustering performance. Zou et al. [14] presented an initialization method of FCM, which extracted approximate cluster centers from samples based on grid and density, and adopted the number of approximate cluster centers to initialize the number of cluster centers. Experiments indicated that this method could effectively improve the clustering performance and shorten the clustering time. To tackle the problem of FCM’s sensitivity to initial cluster centers and noise points, the improved FCM algorithm based on the initial center optimization method, density clustering, and grid clustering was proposed by Shi [15] and its effectiveness was proved by taxi trajectory data sets. Aiming to improve FCM’s clustering performance in noisy data sets, the effective objective function with the cluster center learning method-based quadratic mean distance, entropy, and regularization terms was presented [16]. The cluster centers were updated according to the new objective function, and the strengths of this method were validated by varied data sets. Given that most of the improved FCM neglected the dissimilarity in clusters, Qamar et al. [17] introduced a dissimilarity measure between clusters, designed an objective function that generated highly dissimilar clusters and verified better performance of the improved FCM through data sets experiments. Li et al. [18] proposed a double fuzzy C-means clustering model, which comprised two interconnected and interactive FCM algorithms, and redesigned a new objective function to enhance the intra-cluster compactness and inter-cluster separation and to improve the clustering accuracy. Since FCM is easily affected by the Euclidean distance, Wang et al. [19] presented a weighted FCM algorithm based on the weighted Euclidean distance, which included feature weights into the weighted Euclidean distance. Experiments suggested that the clustering performance could be enhanced by the improved FCM. Haldar et al. [20] devised an improved FCM algorithm based on the Mahalanobis distance, which improved the clustering quality compared with the traditional FCM algorithm. Considering the varying weights of sample points, Wu et al. [21] proposed an improved FCM algorithm combining adaptive weights with SA-PSO, which could avoid local optima and effectively improve the clustering performance. To solve the fuzzy boundary value and fuzzy integrodifferential equations problem, the reproducing kernel algorithm is adopted by Arqub [22, 23]. In response to different sample weights and feature weights, Wu [24] further proposed an improved FCM algorithm, which introduced the adaptive data weight vector, the adaptive feature weights matrix, and constructed a new objective function. Experimental results demonstrated that the clustering performance of the algorithm was improved significantly. Although the above-related studies of FCM have made some progress, the performance of an individual or improved FCM clustering algorithm is limited for different data sets, and poor stability.
With the deepening of the research and the differences in the actual application of data sets, the individual FCM clustering algorithm has certain limitations. Hence, the idea of integrating different clustering algorithms has emerged to upgrade the clustering results and meet the actual application requirements [25]. As an unsupervised learning technique based on ensemble, clustering ensemble combines multiple individual clustering results into a unified robust result and can achieve higher accuracy than the individual clustering method [26]. To overcome the failure of traditional FCM to cluster large data sets effectively, Li et al. [27] proposed an FCM-based ensemble clustering algorithm for large data sets, which could improve the clustering accuracy by clustering atoms on data sets. To improve the robustness of clustering, Su et al. [28] proposed a link-based pairwise matrix method for the clustering ensemble of FCM, which adopted a fuzzy graph to represent the relationship between component clusters and then obtained the final ensemble clustering results. Experimental results illustrated that the proposed method outperformed other methods. Su et al. [29] also presented a hierarchical fuzzy clustering ensemble approach, which employed FCM and hierarchical clustering method to generate base clusters and achieve consensus functions, and verified its advantages in clustering accuracy and time efficiency on large data sets. The fuzzy C-means clustering algorithm with improved random projection was proposed [30], which improved the efficiency of clustering through singular value decomposition of the concatenation of membership matrices. To unify the fuzzy clustering partitions, Wan et al. [31] presented an FCM-based fuzzy consensus clustering framework (FCC), which redesigned the objective function and translated FCC into a weighted and segmented FCM clustering, and verified its effectiveness theoretically and experimentally. At present, much attention is paid to the generation methods of cluster membership and design of consensus function in the field of clustering ensembles, whereas little consideration is given to local information of sample points in the process of the ensemble.
In conclusion, the quality of a single FCM cluster is limited, and effective border point allocation is impossible. As a result, more research is required in response to these flaws. In contrast, our research will focus on the membership category of cluster border points and increase the clustering effect by focusing on the idea of clustering ensemble. The original FCM and the proposed algorithm are described in depth in the next section.
3 Fuzzy C-Means Clustering Algorithm
As a classical partition clustering algorithm, the FCM algorithm is mainly worked by obtaining the fuzzy membership of each sample point to all cluster centers through optimizing the objective function, to determine the category of sample points. Given the data set \(X=\{x_1,\ldots ,x_i,\ldots ,x_n\}\), it is divided into c clusters, and the center of each cluster is \(c_j\left( j=1,2,\ldots ,c\right)\).
The objective function of FCM is defined as follows:
where \(u_{ij}\) is the fuzzy membership of \(x_i\) in the \(c_j\) cluster, e is the membership factor, and \(||x_i-c_j||\) represents the Euclidean distance from \(x_i\) to \(c_j\).
In the process of fuzzy clustering, the membership \(u_{ij}\) and the cluster center \(c_j\) are iterated constantly until clustering is completed, that is when the membership does not change greatly or the number of iterations t satisfies the following equation, or when the objective function \(J_e\) reaches the local optimum (minimum):
where \(\varepsilon\) is an error threshold.
To get the minimum \(J_e\), the updating calculation method of \(u_{ij}\) and \(c_j\) is shown in the following equations:
FCM determines the final clustering result \(C_{xi}^{\mathrm{result}}\) according to the principle of maximum membership, as shown in the following equation:
The clustering algorithm of FCM is as follows:

4 The LI_BIFCM Algorithm
This section explains the LI_BIFCM algorithm’s framework and the essential steps of the proposed technique in depth.
4.1 The LI_BIFCM Algorithm Framework
Figure 1 depicts the LI_BIFCM structure, which is separated into three sections: vertical ensemble, horizontal ensemble, and final clustering ensemble. First, the parameters introduced into the system are the data set, cluster center, ensemble times m, multiple KNN parameters p (p is a percentage, normally takes 2%), and s (s is used to control the number of executions of multiple KNN). Then, the data set is executed for m times by FCM, yielding distinct membership matrices due to the randomness of the FCM cluster center, resulting in the vertical ensemble. Following that, after each cluster, multiple K-nearest neighbors are utilized to build the local membership matrix of sample points to form a horizontal ensemble, and the m times cluster can produce m horizontal ensemble clustering results. Finally, the vertical and horizontal ensembles are combined to generate the final clustering results.
4.2 Vertical Ensemble
A single FCM is used as the base cluster member in LI_BIFCM. Clustering results are different because FCM randomly generates the initial cluster center each time, and this feature is used to generate diverse and stable clustering results.
Suppose the data set has n sample points \(X=\{x_1,\ldots ,x_i,\ldots ,x_n\}\) and c clusters \(\{c_1,\ldots ,c_j,\ldots ,c_c \}\). FCM performs m times to get m membership matrices \(U=[U_1,\ldots ,U_f,\ldots ,U_m]\). Each membership matrix is shown in the following equation:
After each cluster is completed, the corresponding maximum membership matrix is calculated, as shown in the following equation:
where \(l_{x_i}^{f}\) represents the class number of the maximum membership of the sample point \(x_i\).
FCM can obtain the maximum membership class number \(l_{x_i}^{f}\) of each sample point after performing m times of clustering, and the clustering results after vertical ensemble are shown in the following equation:
4.3 Horizontal Ensemble
The multiple K-nearest neighbors approach is used to consider the local information of sample points to solve the problem of boundary point misclassification.
First, the K-nearest neighbors of each sample point is calculated, and its calculation formula is shown in the following equation:
where \(d(x_i,x_j)\) denotes the Euclidean distance between the point \(x_i\) and \(x_j\), \(NN_k (x_i)\) represents the k-th nearest neighbor of \(x_i\), and \(k\in [1,n]\).
The membership matrix is obtained after one FCM clustering, and the maximum membership matrix \(C_{x_i}^{1,k}\) of K-nearest neighbors of the sample points is calculated, as shown in the following equation:
Then, the maximum membership matrix \(C_{x_i}^{1,k+s}\) of multiple K-nearest neighbors (k takes different values) of the sample points are calculated, as shown in the following equation:
where s is used to control the number of executions of multiple K-nearest neighbors, and its range is 1 to \(n-1\).
A horizontal ensemble is carried out by combining the clustering result of one FCM and the membership matrix of multiple K-nearest neighbors to obtain its clustering result \(C_{x_i}^{1,Horizontal}\), as shown in the following equation:
Similarly, m horizontal clustering ensemble results \(C_{f}^{\mathrm{Horizontal}}\) can be obtained after performing m times of clustering.
Last, the final clustering results \(C_{x_i}^{\mathrm{result}}\) are obtained by integrating the clustering results of vertical ensemble and m horizontal ensembles, as shown in the following equation:
4.4 The LI_BIFCM Algorithm Flow
The following is the LI_BIFCM algorithm flow:
-
1.
Set the number of cluster centers as c, the fuzzy factor as e, the condition of stopping iteration as \(\varepsilon\), the maximum number of iteration as T, the number of cluster ensemble parameters m and s.
-
2.
Random initial cluster center: \({c_1,c_2,\ldots ,c_c }\) and membership matrix: \(U_0\).
-
3.
Execute the FCM algorithm m times to get membership matrix \(U=[U_1,U_2,\ldots ,U_m]\) as shown in Eq. (6).
-
4.
Calculate the maximum membership matrix \(L^f\) of each clustering by Eq. (7), and carry out a vertical ensemble by Eq. (8) to obtain the vertical ensemble clustering results \(C_{x_i}^{\mathrm{Vertical}}\).
-
5.
Calculate \({\mathrm{KNN}}(x_i)\) of each sample point \(x_i\) according to Eq. (9).
-
6.
Calculate the maximum membership matrix \(C_{x_i}^{1,k}\),\(C_{x_i}^{1,k+1}, \ldots , C_{x_i}^{1,k+s}\) of multiple K-nearest neighbors after each run of FCM according to Eqs. (10) and (11).
-
7.
In combination with the membership matrix of FCM and the maximum membership matrix of multiple K-nearest neighbors, the result \(C_{x_i}^{1,{\mathrm{Horizontal}}}\) of a horizontal clustering ensemble is obtained according to Eq. (12).
-
8.
Similarly, m times of horizontal ensemble clustering results \(C_{f}^{\mathrm{Horizontal}}\) can be obtained if FCM performs m times.
-
9.
According to Eq. (13), the final clustering ensemble result \(C_{x_i}^{\mathrm{result}}\) can be obtained by combining the clustering result of the vertical ensemble with that of m times horizontal ensemble.
LI_BIFCM’s pseudo code is given as follows based on the above analysis.

4.5 Time Complexity Analysis
The time complexity of LI_BIFCM is given when the preceding description is combined. Let n be the number of test sample sets, c be the number of cluster centers, and t be the number of FCM iterations. Our algorithm’s time complexity is determined by the following two factors: (a) the time it takes to run FCM numerous times determines the vertical clustering ensemble. Executing FCM takes O(nct) time while performing vertical ensemble takes O(mnct), where m is the number of layers in the vertical ensemble. (b) Horizontal clustering ensemble uses multiple K-nearest neighbors to calculate the time of local information. The time complexity of searching for K-nearest neighbors of a point is O(n). As a result, the time complexity of K-nearest neighbors executed s times is O(sn). In summary, LI_BIFCM has a time complexity of \(O(mnct+sn)\).
5 Experimental Results and Analyses
5.1 Experimental Settings
To evaluate the performance of the LI_BIFCM algorithm, twelve synthetic and real-world data sets are used in the experiment, which are from different fields and have different sizes, dimensions, and category numbers, as shown in Tables 1 and 2. These data sets are widely used in clustering tests, through which the clustering performance of LI_BIFCM in different application scenarios can be simulated. The experimental environment is a PC with an Intel (R) Core(TM) i7-7500 CPU @ 2.70GHz, 2.90 GHz, 12G RAM, Windows 10 64-bit OS, and the programming tool is Matlab 2015b.
Three classical clustering evaluation indexes are used in the experiment. The accuracy of clustering (ACC) describes the comparison results between the clustering labels and the real labels of the sample points [32]. The adjusted rand index (ARI) represents the overlap degree between clustering partition and actual partition [33]. The adjusted mutual information index (AMI) indicates the consistency between the clustering results and the real categories [34]. The specific calculating formulas of the three evaluation indexes are as follows:
where n represents the total number of samples, \(u_i\),\(v_i\) are the clustering labels and the real labels, respectively, and \(\delta ({u_i},{v_i})\) is a delta function.
where a means the sample logarithm of the clustering result and the real category in the same category, b means that the clustering result and the real category are the sample logarithm of different categories, and \(E[{\mathrm{RI}}]\) means the expectation of rand index RI.
where U,V represent the real labels vector and clustering labels vector, respectively, R,C represent the number of real clusters and clustering clusters, respectively, and P(.) stands for probability, and H(.) stands for information entropy.
To eliminate the difference between data dimensions, the data sets need to be standardized before the experiment and the calculation method is shown as the following equation:
where \(x_{ij}\) represents the attribute value of the sample points \(x_i\) in jth column, and max\(\left( x_j\right)\) and min\(\left( x_j\right)\) are the maximum and minimum values of the jth attribute column, respectively.
5.2 The Comparison with the Original FCM Algorithm on Synthetic Data Sets
To test the effectiveness of LI_BIFCM in processing boundary points, three two-dimensional synthetic data sets are chosen for experimental comparison with FCM. The size of these three data sets is 5000, with 15 clusters. The degree of contact between border points varies, and misclassification is common. Figures 2, 3 and 4 depict the clustering findings. Figure 2 shows that LI_BIFCM can handle the boundary points of the S1 data set successfully, whereas FCM has two issues: one is an error in the cluster center, and the other is an error in the boundary points. This is because when a cluster center error occurs in FCM, it sets off a chain reaction. Black boxes 1 and 2 are independent clusters, as illustrated in Figs. 2c, but because black box 1 has two cluster centers, the clusters in black box 2 can only be assigned to the adjacent cluster. Figure 3 depicts the clustering result of S2. The problem of boundary point misdivision and cluster center error exists in FCM as well, but not in our technique. Figure 4 depicts the experimental results of S3. Similarly, FCM splits a cluster into many clusters by mistake, but LI_BIFCM maintains excellent efficiency.
Figure 5 summarizes the clustering index results from Figs. 2, 3 and 4, showing that LI_BIFCM beats FCM in all three indicators on the synthetic data sets. FCM is easily influenced by the initial clustering center, falls into a local optimum, and is sensitive to border points for this reason. By using a vertical ensemble, our approach can discover the best clustering center. The proposed algorithm uses vertical ensemble to determine the best clustering center, and horizontal ensemble to efficiently assign boundary points, which increases clustering quality.
5.3 The Comparison with the Basic Algorithms and the Clustering Ensemble Algorithms on Real-World Data Sets
To test the LI_BIFCM algorithm’s clustering performance further, seven representative algorithms are selected for comparison, which includes basic algorithms (K-means, DBSCAN, AP, and FCM), soft clustering ensemble algorithm (CSPA [35, 36]), and hard clustering ensemble algorithms (HGPA [35, 36], and MCLA [37]).
K-means is a clustering technique based on partitions. It must determine the number of clusters to be created, choose initial cluster centers at random, and calculate the distance between each sample and each cluster center. The cluster centers are recalculated once each sample is assigned to the cluster that is closest to it. This step is repeated until the algorithm reaches a particular termination condition.
DBSCAN is a clustering algorithm that is based on density. The number of clusters does not need to be specified in advance, but it does have two parameters: domain radius (Eps) and core point threshold (MinPts). The clustering process starts from a selected core point and continuously expands to the area where the density is reachable to obtain a maximum area containing the core point and the boundary point, and any two points in the area are connected by density.
AP is a clustering algorithm that works by transferring information between sample points. It uses all samples as network nodes, iteratively calculating the information (responsibility and availability) of each network edge until several high-quality exemplars are obtained and the remaining points are assigned to the appropriate clusters.
CSPA is a cluster-based similarity partition algorithm that starts by defining a new similarity between any two sample points, then calculates a \(n * n\) similarity matrix, and finally clusters the data using a pairwise similarity-based clustering algorithm.
HGPA is a kind of hypergraph partition algorithm. It starts by creating a hypergraph with sample points at the vertices. Each cluster in each cluster member is a closed hyperedge that includes all of the cluster’s vertices, and the clustering results are obtained using the hypergraph partition procedure.
MCLA is a meta-clustering algorithm. It measures the similarity of two clusters using the Jaccard coefficient, then obtains meta-clusters using the METIS method, and lastly allocates samples to the most relevant meta-clusters.
As a result, using these algorithms to conduct comparative trials is extremely sensible. Each basic algorithm experiment is conducted ten times to get the average value of the clustering findings, and each clustering ensemble algorithm is run until 10 ideal cluster member sets are produced. The clustering accuracy of eight algorithms on nine test data sets is shown in Table 3, in which the best clustering evaluation index among eight clustering algorithms is indicated in bold. Table 3 reveals the average ACC of LI_BIFCM ranks first among eight algorithms, achieves the highest clustering accuracy in six of the nine data sets and ranking top two in the other data sets (Dermatology and Iris). Only in the Ionosphere data set, LI_BIFCM ranks 4th.
To see if there are statistically significant differences between the LI_BIFCM method and the other seven algorithms in Table 3. With a confidence level of 0.95, the Aligned Friedman test [38] is used. The specific method is the clustering accuracy of a certain algorithm on a certain data set minus the average clustering accuracy of all algorithms on this data set to participate in the sorting. The clustering accuracy of eight algorithms on nine data sets ranging from 1 to 72 is shown in Table 4, and the numbers in bold indicate the best ACC sorting. Table 4 shows that LI_BIFCM achieved the best-ranking result with an average of 16.8, MCLA ranked second with an average ranking value of 24.1, FCM ranked third with an average ranking value of 28.6, and the ranking of the remaining algorithms are as follows: AP, K-means, CSPA, DBSCAN, and HGPA.
The Friedman aligned rank test [38] is then applied, with the calculation method depicted as the following equation:
In the eight algorithms and nine test data sets, the statistic T obeys the Chi-square distribution with seven degrees of freedom. We can observe that the p value of \(\chi ^2 (7)\) is 1.91E\(-\)04, which is substantially less than 0.05, by searching up the Chi-square distribution table. Therefore, the null hypothesis can be rejected. The results obtained by all algorithms on the nine data sets are considered to be significantly different.
Similarly, Tables 5 and 6 show that the average ARI and AMI of LI_BIFCM rank first among eight algorithms, and achieve the highest ARI and AMI in at least seven data sets, indicated by bold numbers. Then, the Friedman Aligned Rank test [38] is used in conjunction with Table 7 to compare the clustering index ARI, and calculated \(T=25.4161\). The statistical variable T obeys the chi-square distribution with seven degrees of freedom, the p value of \(\chi ^2 (7)\) is 6.40E\(-\)04, which is much less than 0.05. Therefore, the null hypothesis can be rejected, and the ARI value of the LI_BIFCM algorithm is statistically significantly better than the other seven algorithms in the ranking results. On the other clustering index AMI, the ranking results of AMI are shown in Table 8, the p value of \(\chi ^2 (7)\) is 4.25E\(-\)05, and its value is also much smaller than 0.05. The bold numbers in Tables 7 and 8 represent the best ARI and AMI rankings of the algorithm, respectively. Therefore, it is concluded that the proposed algorithm is superior to other seven algorithms.
In summary, LI_BIFCM achieved the best clustering results on the most real-world data sets for the three reasons listed below. First, numerous membership matrices can correctly locate cluster centers and ensure rather stable clustering results in the vertical ensemble. Second, the method of multiple K-nearest neighbors is employed in the horizontal ensemble process to fully use the local information for the sample points to find the optimal category for border points. Third, the attribution of sample points is clarified after the bi-directional clustering ensembles, and clustering performance is effectively improved.
5.4 Parameter Analysis
5.4.1 The Number of Vertical Ensembles Parameters m
Each additional vertical ensemble will increase the time cost using the algorithm of this paper. Experiments are carried out to test the number of vertical ensembles parameters in the algorithm on clustering performance, in which m fuzzy membership matrices can be obtained by running the FCM algorithm m times. In other words, the core of LI_BIFCM is to carry out m times of FCM to get the vertical ensemble and combine multiple K-nearest neighbors (\(s=5\)) to get the final clustering result. The average clustering evaluation index value is obtained by repeating the algorithm 10 times. The values of m are 5, 10, 15, 20, 25, and 30, respectively, and the experimental results are shown in Fig. 6, which shows that with the number of vertical ensembles changing from 5 to 10 to 15... and to 30, no obvious fluctuation is observed in all clustering evaluation indexes.
5.4.2 The Number of Horizontal Ensembles Parameters s
In a horizontal clustering ensemble, the most critical parameter s governs the number of executions of multiple K-nearest neighbors. Therefore, to study the influence of the parameter s on the clustering performance, the following experiments are carried out. The average clustering index of the algorithm conducted 10 times is utilized as the experimental result, assuming that the parameter m of the vertical clustering ensemble of LI_BIFCM is five. The s values are 4, 8, 12, 16, and 20. Figure 7 shows the experimental result, with the horizontal ensemble parameter s increasing from 4 to 8, then to 12, ... , and 20. The three clustering indexes did not show any noteworthy changes.
As a result, the experimental results of Figs. 6 and 7 can be summarized as follows. For the following reasons, LI_BIFCM is unaffected by the parameters of vertical and horizontal ensembles: FCM and KNN are reasonably stable; increasing the number of vertical and horizontal ensembles will only result in homogenized clustering; and the additional number of vertical and horizontal ensembles will have no effect on LI_BIFCM’s performance.
5.5 Run Time Comparison of 4 Algorithms
The following experiment was done to see how long LI_BIFCM takes to run. The CSPA, HGPA, and MCLA algorithms were chosen for experimental research since the single K-means, DBSCAN, AP, and FCM algorithms are much faster than the clustering ensemble method. To acquire the final clustering results, the LI_BIFCM, CSPA, HGPA, and MCLA algorithms are used to integrate the cluster members produced by running the FCM algorithm five times. Each experiment was performed 10 times under the same conditions, with the average of the 10 running times used as the result. Table 9 shows the experimental result, with the minimal running time for each data set highlighted in bold.
Table 9 shows that LI_BIFCM takes substantially less time to run than the other three algorithms in all data sets (excluding Pima). Because the CSPA method must calculate the similarity of all sample points, the HGPA algorithm must produce a hypergraph, and the MCLA must also compute the similarity of any two clusters. The execution times of HGPA and MCLA are nearly same. LI_BIFCM not only provides a better clustering effect than other clustering ensemble algorithms when combined with the preceding experimental analysis, but it also has a shorter run time.
6 Conclusion
In this paper, to increase clustering performance, we introduced a new clustering ensemble framework that considers local information (LI_BIFCM), dubbed the bi-directional FCM clustering ensemble algorithm. To achieve the best clustering results, LI_BIFCM combines vertical and horizontal ensembles. The properties of FCM random initial clustering centers are employed in the vertical ensemble, and FCM is run numerous times to acquire multiple cluster members for the vertical ensemble to ensure clustering stability. To avoid boundary point misdivision, multiple K-nearest neighbors are used to execute a horizontal ensemble after acquiring each cluster member. A comprehensive experimental analysis indicates: (1) LI_BIFCM surpasses FCM in dealing with boundary points. (2) On most data sets, LI_BIFCM outperforms a single clustering method as well as some clustering ensemble algorithms. (3) The vertical and horizontal ensembles parameters do not affect LI_BIFCM. (4) When compared to the clustering ensemble approach, the suggested algorithm takes less time to run. Base clustering optimization and its application to real-world scenarios, such as customer analytics, will receive greater attention in the future.
References
Xu, R., Wunsch, D.: Survey of clustering algorithms. IEEE Trans. Neural Netw. 16, 645–678 (2005)
Chaira, T.: A novel intuitionistic fuzzy C means clustering algorithm and its application to medical images. Appl. Soft Comput. 11, 1711–1717 (2011)
Kalyani, S., Swarup, K.S.: Particle swarm optimization based K-means clustering approach for security assessment in power systems. Expert Syst. Appl. 38, 10839–10846 (2011)
Hosseini, S., Maleki, A., Gholamian, M.R.: Cluster analysis using data mining approach to develop CRM methodology to assess the customer loyalty. Expert Syst. Appl. 37, 5259–5264 (2010)
Jain, A., Murty, M., Flynn, P.: Data clustering: a review. ACM Comput. Surv. 31, 264–323 (1999)
Jain, A.K.: Data clustering: 50 years beyond K-means. Pattern Recognit. Lett. 31, 651–666 (2010)
Dunn, J.C.: A fuzzy relative of the ISODATA process and its use in detecting compact well-separated clusters. J. Cybern. 3, 32–57 (1974)
Bezdek, J.C.: Pattern Recognition with Fuzzy Objective Function Algorithms. Plenum Press, New York (1981)
Arqub, O.A., Al-Smadi, M.: Fuzzy conformable fractional differential equations: novel extended approach and new numerical solutions. Soft Comput. 24, 12501–12522 (2020)
Arqub, O.A., Al-Smadi, M., Momani, S., Hayat, T.: Numerical solutions of fuzzy differential equations using reproducing kernel Hilbert space method. Soft Comput. 20, 3283–3302 (2016)
Silva, T.M., Pimentel, B.A., Souza, R.M., Oliveira, A.L.: Hybrid methods for fuzzy clustering based on fuzzy c-means and improved particle swarm optimization. Expert Syst. Appl. 42, 6315–6328 (2015)
Yao, J., Dash, M., Tan, S.T., Liu, H.: Entropy-based fuzzy clustering and fuzzy modeling. Fuzzy Set Syst. 113, 381–388 (2000)
Ding, Y., Fu, X.: Kernel-based fuzzy c-means clustering algorithm based on genetic algorithm. Neurocomputing 188, 233–238 (2016)
Zou, K., Wang, Z., Hu, M.: A new initialization method for fuzzy c means algorithm. Fuzzy Optim. Decis. Mak. 7, 409–416 (2008)
Shi, Y.L., Nana, J.Z.: Improved FCM algorithm based on initial center optimization method. J. Intell. Fuzzy Syst. 32, 3487–3494 (2017)
Ramathilaga, S., Leu, J.J., Huang, K.K., Huang, Y.M.: Two novel fuzzy clustering methods for solving data clustering problems. J. Intell. Fuzzy Syst. 26, 705–719 (2014)
Qamar, U.: A dissimilarity measure based fuzzy c-means (FCM) clustering algorithm. J. Intell. Fuzzy Syst. 26, 229–238 (2014)
Li, L., Wang, R.X., Li, X.C.: Double fuzzy C-means model and its application in the technology innovation of China. J. Intell. Fuzzy Syst. 31, 2895–2901 (2016)
Wang, X.Z., Wang, Y.D., Wang, L.J.: Improving fuzzy c-means clustering based on feature-weight learning. Pattern Recognit. Lett. 25, 1123–1132 (2004)
Haldar, N.A., Khan, F.A., Ali, A., Abbas, H.: Arrhythmia classification using Mahalanobis distance based improved Fuzzy C-Means clustering for mobile health monitoring systems. Neurocomputing 220, 221–235 (2017)
Arqub, O.A., Al-Smadi, M., Momani, S., Hayat, T.: Application of reproducing kernel algorithm for solving second-order, two-point fuzzy boundary value problems. Soft Comput. 21, 7191–7206 (2016)
Arqub, O.A.: Adaptation of reproducing kernel algorithm for solving fuzzy Fredholm–Volterra integrodifferential equations. Neural Comput. Appl. 28, 1–20 (2015)
Wu, Z.H., Wu, Z.C., Zhang, J.: An improved FCM algorithm with adaptive weights based on SA-PSO. Neural Comput. Appl. 28, 3113–3118 (2017)
Wu, Z.H., Wang, B.: DwfwFcm: an effective fuzzy c-means clustering framework considering the different data weights and feature weights. J. Intell. Fuzzy Syst. 37, 4339–4347 (2019)
Strehl, A., Ghosh, J.: Cluster ensembles: a knowledge reuse framework for combining multiple partitions. J. Mach. Learn. Res. 3, 583–617 (2002)
Topchy, A., Jain, A.K., Punch, W.: A mixture model for clustering ensembles. In: Proceedings of the Fourth SIAM International Conference on Data Mining. Lake Buena Vista, FL, SIAM, USA, pp. 379–390 (2004)
Li, J., Gao, X.B., Tian, C.N.: FCM-based clustering algorithm ensemble for large data sets. Fuzzy Syst. Knowl. Discov. 4223, 559–567 (2006)
Su, P., Shang, C.J., Shen, Q.: Link-based pairwise similarity matrix approach for fuzzy c-means clustering ensemble. In: 2014 IEEE International Conference on Fuzzy Systems, Beijing, Peoples Republic of China. IEEE, USA, pp. 1538–1544 (2014)
Su, P., Shang, C.J., Shen, Q.: A hierarchical fuzzy cluster ensemble approach and its application to big data clustering. J. Intell. Fuzzy Syst. 28, 2409–2421 (2015)
Ye, M., Liu, W.F., Wei, J.H., Hu, X.X.: Fuzzy c-means and cluster ensemble with random projection for big data clustering. Math. Probl. Eng. 2016, 1–13 (2016)
Wan, X., Lin, H., Li, H., Liu, G.N., An, M.B.: Ensemble clustering via fuzzy c-means. In: 2017 14th International Conference on Services Systems And Services Management, Dalian, Peoples Republic of China. IEEE, USA, pp. 1–6 (2017)
Wang, Z.C., Parvin, H., Qasem, S.N., Tuang, B.A., Pho, K.H.: Cluster ensemble selection using balanced normalized mutual information. J. Intell. Fuzzy Syst. 39, 3033–3055 (2020)
Vinh, N., Epps, J., Bailey, J.: Information theoretic measures for clustering comparison: variants, properties, normalization and correction for chance. J. Mach. Learn. Res. 11, 2837–2854 (2010)
Boudane, F., Berrichi, A.: Gabriel graph-based connectivity and density for internal validity of clustering. Prog. Artif. Intell. 9, 221–238 (2020)
Punera, K., Ghosh, J.: Consensus-based ensembles of soft clusterings. Appl. Artif. Intell. 22, 780–810 (2008)
Strehl, A., Ghosh, J.: Cluster ensembles: a knowledge reuse framework for combining multiple partitions. J. Mach. Learn. Res. 3, 583–617 (2003)
Hu, J., Li, T.R., Luo, C., Fujita, H., Yang, Y.: Incremental fuzzy cluster ensemble learning based on rough set theory. Knowl. Based Syst. 132, 144–155 (2017)
Garcia, S., Fernandez, A., Luengo, J., Herrera, F.: Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: experimental analysis of power. Inf. Sci. 10, 2044–2064 (2010)
Acknowledgements
This paper is supported by the National Key R&D Program of China (No. 2018YFB1701500 and No. 2018YFB1701502).
Author information
Authors and Affiliations
Contributions
Literature and data collection: CR; research design: CR and LS; manuscript writing: CR; manuscript review and revise: CR and LS.
Corresponding author
Ethics declarations
Conflict of interest
As far as we know, the author named in this paper has no financial or interest conflicts.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Ren, C., Sun, L. A Bi-directional Fuzzy C-Means Clustering Ensemble Algorithm Considering Local Information. Int J Comput Intell Syst 14, 171 (2021). https://doi.org/10.1007/s44196-021-00014-z
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s44196-021-00014-z