Literature DB >> 36236728

Sorting Center Value Identification of "Internet + Recycling" Based on Transfer Clustering.

Cheng Cheng1, Xiaoli Luan1.   

Abstract

As the core link of the "Internet + Recycling" process, the value identification of the sorting center is a great challenge due to its small and imbalanced data set. This paper utilizes transfer fuzzy c-means to improve the value assessment accuracy of the sorting center by transferring the knowledge of customers clustering. To ensure the transfer effect, an inter-class balanced data selection method is proposed to select a balanced and more qualified subset of the source domain. Furthermore, an improved RFM (Recency, Frequency, and Monetary) model, named GFMR (Gap, Frequency, Monetary, and Repeat), has been presented to attain a more reasonable attribute description for sorting centers and consumers. The application in the field of electronic waste recycling shows the effectiveness and advantages of the proposed method.

Entities:  

Keywords:  GFMR model; Internet + Recycling; inter-class balanced data selection; transfer clustering; value identification

Mesh:

Year:  2022        PMID: 36236728      PMCID: PMC9572044          DOI: 10.3390/s22197629

Source DB:  PubMed          Journal:  Sensors (Basel)        ISSN: 1424-8220            Impact factor:   3.847


1. Introduction

In 2015, China’s recycling pattern shifted from “manual recycling” to “Internet + Recycling” [1]. The “Internet + Recycling” process is very important to identify the value of the sorting center [2]. Take electronic waste (e-waste) recycling as an example. The company will offer information to sorting centers and receive a commission [3]. Therefore, the “Internet + Recycling” companies will design a specific strategy based on sorting center costs [4], recycling channels [5], and value dimensions [6]. In this way, the company can improve its business competitiveness and reduce churn at the sorting center. However, the data set of sorting center is small and imbalanced that is called “Absolute Rarity”, [7] which makes it difficult to carry out the task of value assessment. The traditional oversampling methods [8] may not evaluate the accurate value for sorting centers. Transfer learning is a branch of machine learning that has been shown to help solve problems with small data sets. It has been widely used in image classification [9], signal processing [10], and text classification [11]. An effective model for the target domain can be obtained by leveraging useful relative information from the source domain. However, there is still limited research on the transfer clustering problem. Jiang et al. proposed transfer spectral clustering (TSC), which could transfer knowledge from related clustering task [12]. Wang et al. extended three traditional Gaussian mixture model (GMM) to transfer clustering versions [13]. These methods are more suitable for the clustering problem that has definite boundaries. However, the amount of sorting centers is too limited to obtain boundaries. Fuzzy c-means (FCM) [14] is a clustering algorithm that could solve this problem. The algorithm are more applicable in many fields by changing the objective function of the FCM [15]. Transfer fuzzy c-means (TFCM) [16] is a transfer clustering version of FCM which has good performance on small data set clustering by transferring knowledge from the relative source domain’s cluster centers. There are plentiful customers in “Internet + Recycling” which contain useful and relative information. Thus, in this paper, we adopt TFCM to transfer the cluster centers of customers as knowledge to assist cluster sorting centers. In order to achieve accurate cluster centers, a comprehensive model that can describe the characteristics of customers is necessary. The model with clustering algorithm has been widely used in customer value identification. Pondel et al. compared three different clustering algorithms’ results of 56,237 customers who made at least 2 purchases in the online store [17]. Kumar et al. practiced the model on 127,037 business customer [18]. Many other scholars also adjust the model according to the features of customers. For example, Li et al. used an improved model with added indicators to classify 4000 customers on the e-commerce platform [19]. Fahed et al. used an enhanced model to classify 42,172 retail customers [20]. However, in the “Internet + Recycling” process, some high-value customers are the individual economy which is rare in customers. The original model could not describe a comprehensive characterization of the high-value customers. In summary, to get the accurate value identification of sorting centers, TFCM that transferring knowledge from customers is adopted to solve the problem of the small amount of data set, and inter-class balanced data selection (IBDS) is proposed to help solve the problem of imbalanced data set. In order to obtain accurate customer cluster centers for transfer, an improved model is proposed. The application in “Internet + Recycling” company proves our approach can effectively improve the accuracy of classifying sorting centers. With the accurate value identification of sorting centers, “Internet + Recycling” companies can apply their marketing strategies more precisely, which could improve business competitiveness and reduce sorting center churn.

2. Acquisition of Customer Cluster Centers

Accurate cluster centers of customers is the prerequisite for TFCM. The RFM model is a popular customer value analysis tool widely used to measure customer lifetime value as well as customer value identification and behavioral analysis. The original definition is as follows: R: recency of the last trade F: frequency of trades M: monetary value of the trades However, the original model could not identify the active customer. The R of is almost the same for a new customer and an active loyal customer. In the “Internet + Recycling” company, the top 20% customers who are active in trade devote over 60% trading volume. Therefore, it is important to divide high-value customers from others. The high-value customers of “Internet + Recycling” companies are sometimes the individual economy. They will recycle some specific goods and store them. When the price of the goods is relatively high, they will place several orders online. Therefore, their characteristics are short trade gap, high frequency, huge monetary value, and focus on particular goods. In order to strengthen the ability to identify high-value customers, this paper proposed model as follows. G: average trade gap time between two trade F: frequency of trade M: monetary value of all trade R: maximum number of repeat transactions for the same goods Especially, G is defined as in Equation (1) where denotes the statistical interval, denotes the first date the consumer traded during the statistical interval, denotes the last date the consumer traded during the statistical interval. If a customer has only traded one time during the , this paper assumes the consumer’s average trade gap is bigger than the statistical interval set . Otherwise, calculate the true trade gap. The model will reduce the effect of randomness because the four indicators all have small relation to the sampling date. The average gap will separate intensive trade consumers from the others. The frequency and monetary value will identify loyal consumers. The repeat recycling times could identify individual economies. Consequently, is more suitable for identifying the consumer value of “Internet + Recycling”. Based on model, k-means algorithm [21] is used to obtain the cluster centers. Definition of variables: : domain consisting of the standardized data of m customers K: the number of clusters : the kth cluster center of customers Steps of acquiring customer cluster centers by k-means algorithm are as followed: Step1: Randomly generate as initial cluster centers. Step2: Calculate the distance of each to as and classify the sample into the cluster corresponding to the minimal distance. Step3: Calculate the mean value of all samples within each cluster and update the . Step4: Repeat step2 and step3 until the maximum number of iterations. The final is the customer cluster centers.

3. Transfer Clustering for Sorting Centers

The data set of the introduced customer clustering study ranged from 4000 to 127,037, the amount of sorting centers is only 223 which is far from the modeling order of magnitude. In transfer learning, the domain containing a large amount of useful information is often defined as the source domain. And the domain we need to learn is defined as the target domain. In this paper, the customers data set is the source domain and the labeled sorting centers data set is the target domain . Due to the small size of the data set, there are no legible boundaries between each class. FCM could help solve this problem. The objective function of the original FCM is as follows. where K denotes the number of clusters, is the fuzzy/possibilistic partition matrix whose element denotes the membership of the ith sample belonging to the kth class, denotes the fuzzy index, is the matrix of K cluster centers whose element denotes the kth cluster center of sorting centers. Transferring knowledge from the is a must because the cannot be trained to a satisfactory model on its own. There is a transfer learning version of FCM in which the objective function is defined as followed [16]. where and are non-negative balance parameters. The learning rules based on Equation (3) are as follows: There TFCM Algorithm 1 is described below. The computational complexities of TFCM is that is the same as FCM. This method transfers knowledge from the source domain to the target domain through source domain cluster centers. Changing and could adjust the level of learning from the source domain. As is proved in the article [16], if the source domain has bad knowledge, it will have a negative influence on the clustering performance in the target domain, which is called negative transfer. The original article tried and failed to adopt appropriate parameter values for reducing the effect of the bad source domain. Therefore, in this paper we choose data selection methods rather than adopting parameter values that have been proven to be effective. In data selection, the key is to find a measurement between the source and target domain. Kullback–Leibler (KL) divergence is often used to measure the difference between two distributions. The KL divergence is defined as follows. where denotes the cross-entropy, denotes the information entropy. In , the is constant. And proved by Gibbs’ inequality that is bigger than , so is monotonic and could represent . The smaller means q is closer to p. In this paper, assume the distribution of the source domain is p which has m samples, and the distribution of the target domain is q which has n samples. In this paper, we attempt to find source domain samples that are more similar to the target domain. Turn into the math equation, the smaller is what we want. So order the source domain by , and select the relative smaller s samples to compose the subset of source domain. It is easy to obtain: which means we can measure source domain samples by , the smaller the sample is closer to the target domain. However, due to the imbalanced number of different categories, if not separate the target domain or the source domain and calculate the distribution individually, a class with a larger amount will overwrite the features of a class with a smaller amount. As is shown in Figure 1, when fitting the distribution of G of consumers. The five best-fitting distributions all neglect small samples.
Figure 1

Fitted Distributions of G of Customers.

This paper proposes an inter-class balanced data selection method(IBDS). The steps of IBDS are given in Algorithm 2. The computational complexities of IBDS is , where b is the number of features of . The proposed algorithm will increase the complexity of the algorithm to some extent, but it can significantly improve the overall accuracy of the algorithm. This paper ordered the source domain samples by calculating for each sample in a different category. Then, use the best s samples of each category to get a balanced and more similar subset that provides better cluster center TFCM. The method has proved to be effective.

4. Experimental Results

In this section, the proposed algorithm is evaluated on real-world data set. This paper collected 754,904 e-waste recycling order records of consumers and 19,703 e-waste transporting records of sorting centers from January 2021 to December 2021 of China’s “Internet + Recycling” company. 308,059 consumers and 223 sorting centers were detected. The company has developed four marketing programs targeting high-value, potential-value, stable-value and low-value sorting centers. The goal of this paper is to accurately identify 4 types of sorting center values.

4.1. Data Processing

The consumer is modeled by . Due to the trade of sorting centers, it is always counted by cars without the specific trade category. So, the model used is the model. The data is normalized by The smaller G is better. Thus G is normalized by Equation (8). F, M, and R are the bigger the batter, so they are normalized by Equation (9). , , , are the standardized variables.

4.2. Customer Value Identification

Based on the and model, results of the k-means cluster algorithm to identify the value of customer are as follows. As is shown in Table 1, the difference of R between four clusters is not obvious. Thus, the algorithm may incorrectly classify some high-value users into potential-value user groups, which leads to the high M of potential-value. Between stable-value and low-value, the difference between these two clusters is mainly the R which is caused by the random time the new customer begins to use the service. As is shown in Table 2, the difference of G between four clusters is very obvious, and the added R vivid segments high-value customers from the others. Low-value customers who trade only once are clearly separated from stable-value users who trade more actively. Thus, the result of the model is better than the model. The cluster centers are more informative for sorting centers clustering.
Table 1

Clustering customer by .

ClusterCount R F M
High-value165119.3455304.300016,427.7000
Potential-value1146124.179890.77495467.7400
Stable-value153,09495.24531.724296.5492
Low-value153,654286.72481.401376.7217
Table 2

Clustering customer by .

ClusterCount G F M R
High-value5430.9715193.99319243.0920156.0792
Potential-value57,6835.58294.6739263.00614.1112
Stable-value633776.15962.5564193.29281.7175
Low-value243,496364.99821.000056.23381.0000

4.3. Sorting Center Value Identification

However, there is still a large gap between clustering centers of the source domain and target domain, as shown in Table 3 and Table 4.
Table 3

True centers of target domain.

Cluster G F M
High-value0.99670.72580.7035
Potential-value0.97180.17710.3338
Stable-value0.59450.00730.0107
Low-value0.18650.00360.0059
Table 4

Cluster centers of .

Cluster G F M Distance
High-value0.99220.33730.28060.5742
Potential-value0.98470.00700.00670.3689
Stable-value0.79120.00300.00410.1969
Low-value0.00000.00000.00130.1866
This means that there is a great difference between the source domain and the target domain. Therefore, IBDS is practiced to find a more appropriate source domain. The cluster centers of are shown in Table 5.
Table 5

Cluster centers of .

Cluster G F M Distance
High-value0.99870.62160.41560.3062
Potential-value0.97790.14830.28470.0572
Stable-value0.69370.00210.00310.0997
Low-value0.13920.00160.00350.0473
As is shown in Table 5, the four clustering centers are all closer to the target domain. IBDS orders samples by , the smaller, the more similar to the target domain. The variation of distance between the four clustering centers and the true target domain centers with different top ratios of is shown in Figure 2.
Figure 2

Variation of distance between true centers and clustering centers.

As is shown in Figure 2, when taking 50% of the distance is minimum. If the ratio is too large, the subset will contain some samples that are not very similar to the target domain. If the ratio is too small, the randomness of the samples will also affect the similarity between the subset and the target domain. To prove the effect of IBDS, this paper also presents the accuracy of transferring the knowledge of and 50% of with in Table 6 and Table 7. A greener background color in Table 6 and Table 7 means higher accuracy, and a redder background color means lower accuracy.
Table 6

Accuracy of transferring top 50% of by TFCM.

λ 2 00.0050.10.50.711.51050100AverageMax
λ 1
030.04%31.84%25.56%83.41%77.13%90.13%3.59%2.24%2.24%2.24%34.84%90.13%
0.00529.15%24.66%25.56%83.41%77.13%89.24%3.59%2.24%2.24%2.24%33.95%89.24%
0.169.51%69.06%2.24%78.92%75.78%2.24%3.59%2.24%2.24%2.24%30.81%78.92%
0.582.96%82.51%89.69%34.53%52.47%2.24%3.59%2.24%2.24%2.24%35.47%89.69%
0.786.10%86.10%91.93%19.28%28.25%48.43%2.24%2.24%2.24%2.24%36.91%91.93%
188.79%88.79%92.83%13.00%16.14%25.56%2.24%2.24%2.24%2.24%33.41%92.83%
1.591.48%91.48%92.38%90.58%10.31%13.45%3.59%2.24%2.24%2.24%40.00%92.38%
1096.41%96.41%96.41%95.96%95.07%95.07%94.17%2.24%2.24%2.24%67.62%96.41%
5096.41%96.41%96.41%96.41%96.41%96.41%95.96%96.41%2.24%2.24%77.53%96.41%
10096.41%96.41%96.41%96.41%96.41%96.41%96.41%96.41%97.31%2.24%87.09%97.31%
Average76.73%76.37%70.94%69.19%62.51%55.92%30.90%21.08%11.75%2.24%47.76%
Table 7

Accuracy of transferring by TFCM.

λ 2 00.0050.10.50.711.51050100AverageMax
λ 1
030.04%29.60%38.57%19.28%11.21%5.38%3.59%2.24%2.24%2.24%14.44%38.57%
0.00528.70%33.18%37.67%19.28%11.21%5.38%3.59%2.24%2.24%2.24%14.57%37.67%
0.144.84%43.95%18.39%23.32%13.45%5.38%3.59%2.24%2.24%2.24%15.96%44.84%
0.578.92%79.37%71.75%39.01%26.91%2.24%2.24%3.59%2.24%2.24%30.85%79.37%
0.782.51%82.51%78.92%45.29%31.39%2.24%2.24%2.24%2.24%2.24%33.18%82.51%
184.75%85.20%89.69%56.95%41.26%18.83%4.48%2.24%2.24%2.24%38.79%89.69%
1.587.89%88.34%91.48%74.44%56.05%31.84%7.17%2.24%2.24%2.24%44.39%91.48%
1092.83%92.83%91.93%92.38%92.38%91.93%91.03%3.59%2.24%2.24%65.34%92.83%
5092.83%92.83%92.83%92.38%92.38%92.38%92.38%88.34%3.59%2.24%74.22%92.83%
10092.83%92.83%92.83%92.38%91.93%91.93%91.93%91.48%3.59%3.59%74.53%92.83%
Average71.61%72.06%70.40%55.47%46.82%34.75%30.22%20.04%2.51%2.38%40.63%
As is shown in Table 6, compared to 100% of , the accuracy of top 50% of with the same and is 95.07% which is higher than 91.03%. Compared to the original , the results of have higher accuracy in most parameters. exceed 7.13% of all kinds of parameters. When is bigger than , it usually has good accuracy because is effected by the randomness of small data set. This impact can be reduced by enhancing the learning of high quality data in the source domain. Thus, IBDS combined with TFCM effectively clusters small and imbalanced data sets. To highlight the advantages of our approach, we also compared it with FCM and CSS (clustering with stratified sampling technique) [8]. The detailed results are shown in Figure 3.
Figure 3

Comparison of different clustering approaches.

Obviously, TFCM combined with IBDS could get the result that is most close to real situation. The data set of the sorting center is small and imbalanced. Thus, the accuracy of FCM is only 60.09%. FCM classifies some high-value sorting centers into the potential-value sorting centers and classifies some potential-value sorting centers into the stable-value sorting center because setting the cluster centers in the dense samples could achieve a lower score of the Equation (2). CSS is an imbalanced data classification algorithm. The accuracy of CSS is 83.41%. CSS is more accurate in identifying low-value sorting centers. But TFCM combined with IBDS outperforms CSS in identifying high-value sorting centers. Because the small data set is easily overfitted by the way of generating samples through oversampling. Transfer learning can effectively improve the accuracy of sorting center value identification and avoid overfitting at the same time.

5. Conclusions

Considering the fact that the data set of sorting centers is small and imbalanced, TFCM combined with IBDS has been proposed to solve the value identification problem. TFCM could transfer knowledge from clustering centers of customers. The IBDS could find a subset that are more similar and balanced than the target domain. In further research, different ratios of the subset exhibits different disparities from the target domain which is caused by randomness and redundant samples. A suitable ratio that balances sample diversity and validity will have better performance. In most of the parameters, IBDS elevates the accuracy which proved the validity of the method. Compared with FCM, the value assessment accuracy of the sorting center elevated from 60.09% to 91.03%. The method in this paper is also less likely to be overfitted compared to the oversampling method. Further research will focus on automatic adjustment of the ratio to balance sample diversity and validity. In this paper, customers and sorting centers share similar characteristics. However, transferring knowledge from data sets without similar features is still a challenge.
  2 in total

1.  Transfer Learning for Class Imbalance Problems with Inadequate Data.

Authors:  Samir Al-Stouhi; Chandan K Reddy
Journal:  Knowl Inf Syst       Date:  2015-08-25       Impact factor: 2.822

2.  Stakeholder Analysis of the Waste Electrical and Electronic Equipment Internet Recycling Industry.

Authors:  Tingting Liu; Qian Zhang; Zichen Zheng; Shangyun Wu; Zhixiong Weng
Journal:  Int J Environ Res Public Health       Date:  2022-08-13       Impact factor: 4.614

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.