Abstract
Data streams online modeling and prediction is an important research direction in the field of data mining. In practical applications, data streams are often of nonstationary nature and containing outliers, hence an online learning algorithm with dynamic tracking capability as well as anti-outlier capability is urgently needed. With this in mind, this paper proposes a novel robust adaptive online sequential extreme learning machine (RA-OSELM) algorithm for the online modeling and prediction of nonstationary data streams with outliers. The RA-OSELM is developed from the famous online sequential extreme learning machine algorithm, but it uses a more robust M-estimation loss function to replace the conventional least square loss function so as to suppress the incorrect online update of the learning algorithm with respect to outliers, and hence enhances its robustness in the presence of outliers. Moreover, the RA-OSELM adopts a variable forgetting factor method to automatically track the dynamic changes of the nonstationary data streams and timely eliminate the negative impacts of the outdated data, so it tends to produce satisfying tracking results in nonstationary environments. The performances of RA-OSELM are evaluated and compared with other representative algorithms with synthetic and real data sets, and the experimental results indicate that the proposed algorithm has better adaptive tracking capability with stronger robustness than its counterparts for predicting nonstationary data streams with outliers.
Keywords
Introduction
Data streams analysis and prediction is an important issue in the data mining community, and it has been widely applied in various fields, such as business data analysis, traffic flow prediction, industrial process monitoring, network intrusion detection and so on.1,2 In these situations, the data are usually generated successively in the form of a data stream, and an online learning algorithm is more desired than a batch learning algorithm to deal with the streaming data, since the online learner can update in an incremental fashion and without the need for retraining whenever a new sample arrives.1,3
Among the existing online learning algorithms, the online sequential extreme learning machine (OSELM) 4 is an emerging and practical one. OSELM is developed on the basis of the interpolation theory and approximation theorem of the extreme learning machine (ELM), 5 and its core idea is to transform the training process of single hidden layer feedforward neural networks (SLFNs) into an equivalent problem of solving linear equations, and use a recursive least squares (RLS) method to calculate the output weights recursively so as to implement the online update of the learning model. By virtue of the powerful approximation ability of ELM and the concise recursive calculation manner of RLS, OSELM can realize an online incremental learning of sequential samples accurately and efficiently. Compared with other popular online learning algorithms, OSELM has many advantages such as fast learning speed, strong generalization ability and simple implementation, which make it very suitable for engineering applications requiring online learning. In view of the good theoretical foundation and superior online learning ability, OSELM has aroused great interests since it was proposed, and it has also achieved successful applications in the field of data streams modeling and prediction, such as rainfall forecasting, 6 wind speed forecasting, 7 streamflow prediction, 8 time series prediction,9,10 time-varying system prediction,11,12 etc.
In practical data streams applications, data samples are usually acquired by sensor devices or manual methods. Due to instrument degradation, mechanical fault, human error and other reasons, there will inevitably be a small number of outliers in the collected data streams. Outliers in data streams may seriously affect the accuracy of the modeling process and result in poor prediction results. In order to improve the robustness of OSELM in the presence of outliers, several works have been done. On the basis of bi-objective optimization theory, Huynh et al. 13 proposed a regularized OSELM algorithm (R-OSELM) to deal with the noisy data set. By using Tikhonov regularization method, the R-OSELM tends to generate a learning model with good generalization performance, and thus enhances the immunity of the model to noise and outliers. However, this algorithm is not a targeted solution for outliers problem, and its outliers’ resistance ability is not so satisfying. In Sun et al., 14 an empirical survival error potential based sequential extreme learning algorithm (ESEP-ELM) incorporating an integrated noise compensation mechanism was developed and used for predicting noisy chaotic time series. In comparison with the original OSELM, the ESEP-ELM has higher computational efficiency and stability with much better immunity to outliers, but its learning accuracy is not as good as OSELM. In recent years, researchers have put forward multiple robust ELM algorithms against outliers by combining different optimization learning methods,15–18 but these algorithms are exclusively of batch learning type and they are not applicable in the online environment.
Moreover, data streams in real world are often nonstationary and exhibit some kind of dynamic behaviors, that is to say, the internal variation trends of the streaming data often evolve with time,19,20 and accordingly, the new data are more helpful to depict the real-time status of the corresponding target system and the out-of-date data may be useless or even become noises to harm the learning model. 21 Therefore, in the online modeling process of nonstationary data streams, we should attach more importance to the recent data than to the remote ones and timely eliminate the negative impacts of the outdated data, so as to better track the dynamic changes of the nonstationary data streams. In order to enhance the tracking capability of OSELM in a nonstationary environment, the concept of forgetting factor was introduced into the OSELM and an improved OSELM algorithm named F-OSELM was proposed.22,23 Different from the original OSELM which treats all the samples coequally, the F-OSELM assigns a forgetting factor less than one as weight coefficient to the old samples to weaken the roles of them, and indirectly enhance the contributions of the new samples, so as to better track the real-time status of the nonstationary system. On this basis, the generalized regularization technique was further incorporated into the F-OSELM to solve the potential ill-posed problem of it, and then a new GRF-OSELM algorithm was developed. 24 Compared with F-OSELM, the GRF-OSELM has a better stability and reliability by avoiding the potential ill-conditioned matrix inverse, and which is more practicable in real applications. The above researches22–24 have manifested that forgetting factor is indeed an effective approach to track the intrinsic changes of the nonstationary systems, while in some complex nonstationary environments, the variable forgetting factor (VFF) strategy is usually a more attractive choice than a fixed forgetting factor for global adaptivity. Therefore, a variety of VFF-based OSELM algorithms25–30 have been put forward in recent years. Online prediction error is the most direct and effective indicator to reflect the dynamic changes of nonstationary system, so many scholars directly take online prediction error as the core observation variable and design some kind of correction function about the prediction error, so as to dynamically adjust the size of forgetting factor.25–29 These VFF methods are very concise and efficient, with almost no additional computational overhead, but the parameter settings of these methods are meticulous for different problems, and it makes these methods difficult to apply in practical applications. In Soares and Araújo, 30 the authors proposed a new directional forgetting factor based OSELM algorithm (DFF-OSELM), in which the old data samples were forgotten according to a certain direction. It has been demonstrated by experiments that the DFF-OSELM is effective for dynamic system modeling and prediction, but there are also many control parameters that need to be set manually and cautiously in this algorithm.
For the above stated outliers problem and dynamic tracking problem in data streams modeling, the individual solutions have been presented in our previous works10,12 to cope with them separately. However, in actual applications, these two problems often occur simultaneously, and a comprehensive online learning algorithm with both outliers’ resistance ability and dynamic tracking ability are much needed. With this in mind, this paper proposes a novel robust adaptive OSELM algorithm (RA-OSELM) by integrating the merits of our previous works10,12 under a unified online learning framework, so as to address the above two problems simultaneously. Different from the originate OSELM which adopts the conventional least square as learning criterion, the RA-OSELM employs a more robust M-estimation loss function to train the learning model, so as to reduce the sensitivity of the learning model to outliers. Meanwhile, a new variable forgetting factor method based on error gradient descent is derived and incorporated in the RA-OSELM to further enhance the dynamic tracking ability and adaptivity of the algorithm in nonstationary environments. The performances of the proposed RA-OSELM algorithm are illustrated with both synthetic and real data streams, and the experimental results show that the RA-OSELM has better adaptive tracking capability with stronger robustness compared with other representative algorithms for predicting nonstationary data streams with outliers.
The remainder of this paper is outlined as follows. The Preliminaries section reviews the preliminary knowledge of this paper. The Proposed RA-OSELM section explains the details of our proposed RA-OSELM. The Experiments section presents the comparison experiments for validating the performances of the RA-OSELM. Finally, the Conclusions section 5 concludes this paper.
Preliminaries
OSELM algorithm
OSELM
4
is an online implementation of the batch ELM algorithm.
5
For a given data set with N different samples
The above N equations in equation (2) can also be compactly written as
Huang et al.
5
have mathematically proved that ELM with random hidden nodes has universal approximation capability, and the parameters of hidden nodes of ELM can be randomly assigned and remain unchanged, then learning an ELM is equivalently transformed into finding a least square solution
On the basis of ELM, Liang et al. 4 developed an online version of ELM to meet the needs of online learning. The learning process of OSELM includes two phases: an initialization phase and an online learning phase.
In the initialization phase, the standard ELM algorithm is used to initialize the OSELM. Given an initial training data set
In the online learning phase, the RLS algorithm is employed to update the output weights constantly. Upon the arrival of another chunk of data
Especially, when the k-th data chunk
M-OSELM algorithm
As described above, the OSELM algorithm adopts least square as learning criterion. Nevertheless, the least square method is easily affected by outliers and prone to generating overfitting models, thus making the performances of OSELM degenerate seriously in the presence of outliers. M-estimation is one of the most popular methods to solve the outliers problems. In order to deal with sequential data with outliers, an M-estimation based OSELM algorithm (M-OSELM) has been proposed in our previous work.
10
The objective function of M-OSELM is written as
Compared with the original OSELM, M-OSELM uses an M-estimation loss function to replace the traditional least square learning criterion, so as to enhance its robustness to outliers. It is easy to know that outliers in streams may produce large perturbation errors and affect the accuracy of the obtained learning model. Different from the least square method which minimizes the quadratic sum of all errors uniformly, the M-estimation technique uses some moderate loss functions to mitigate the large perturbation when the errors exceed a reasonable range, and hence reduces the influence of outliers on the learning model.
GRAF-OSELM algorithm
For nonstationary systems, the variation trends of the systems often evolve over time. Accordingly, the new samples are more helpful to depict the real-time status of the nonstationary systems and the out-of-date samples may be useless or even become noises to harm the learning model. However, during OSELM's online learning process, it just constantly adds new learning samples into the learner and learns them sequentially, but does nothing to the old samples, which greatly limits its tracking performance in nonstationary environments.
To better express the timeliness of the dynamic data and further improve the prediction accuracy of OSELM for nonstationary systems, an improved OSELM algorithm with generalized regularization and adaptive forgetting factor (GRAF-OSELM) has been proposed in our previous work.
12
The objective function of GRAF-OSELM is written as
Different from the OSELM which treats all the training samples equally, the GRAF-OSELM assigns a different forgetting factor as weight to each sample according to the changes of the residual error in the online learning process, so as to reflect the different contributions of the new samples and the old samples to the learning model, thus improving its dynamic tracking ability to nonstationary systems. Moreover, different from other forgetting factor based OSELM algorithms22,23 which commonly use a traditional exponential forgetting regularization, the GRAF-OSELM adopts a new generalized regularization approach to make the algorithm have a constant regularization effect and a persistent stability in all the online learning stages.
Proposed RA-OSELM
The learning model of RA-OSELM
As described in the Preliminaries section, in our previous works we have presented an M-OSELM algorithm
10
and a GRAF-OSELM algorithm
12
to handle the outliers problem and the nonstationary system modeling problem, respectively. However, in practical data streams applications, the streaming data are often of nonstationary nature and contain outliers, and a comprehensive online learning algorithm with both outliers resistance ability and dynamic tracking ability is urgently needed but still vacant. With this in mind, this paper proposes a novel robust adaptive OSELM algorithm (RA-OSELM) by integrating the merits of our previous works,10,12 so as to address the above two problems simultaneously. The objective function of RA-OSELM is written as
It can be seen that when the forgetting factor
Recursive solution of the RA-OSELM model
Next, we will solve the above RA-OSELM model and try to derive a recursive formula of the output weights for online learning. Differentiating equation (13) with respect to
Using the Sherman-Morrison-Woodbury formula,
31
the inverse matrices of
Let
Define
Also, equation (24) can be transformed and rewritten as
Combine equations (14) to (16), we can finally obtain the recursive formula for updating
Note that in equation (28), the calculation of
Online outlier detection for RA-OSELM
In this study, we adopt a modified Huber function as the M-estimation function of RA-OSELM. This Huber function is a piecewise function described below
According to the statistical analysis and central limit theorem, the estimation error of the learning model without outliers can be simply assumed to be normally distributed with mean zero and variance
The standard deviation
Adaptive forgetting scheme for RA-OSELM
Previous studies have proven that the forgetting factor method is effective for nonstationary system modeling and prediction.22–24 However, in practical applications the variation rates of the nonstationary systems are often irregular. In this situation, the preset fixed forgetting factor may not guarantee the global adaptability to the dynamic changes of nonstationary systems, and a variable forgetting factor (VFF) method is generally a more attractive choice. Up to now, several VFF-based OSELM algorithms25–30 have been developed and successfully applied to practical applications, but these methods still have some shortcomings such as complex parameter setting and poor practicability. In this section, a new variable forgetting factor method based on error gradient descent is derived to accommodate the proposed RA-OSELM, by which the forgetting factor can be automatically tuned without need for meticulous parameter setting.
With the strategy of adaptive forgetting, the forgetting factor in RA-OSELM is updated adaptively according to the changes of the online prediction errors so as to adapt itself to nonstationary environment. With this in mind, the goal is to compute the forgetting factor
Let
By moving and merging similar items, the
Perform matrix inverse calculation to the last term of equation (36) with the Sherman-Morrison formula,
31
and we further get
Combining with equation (25), we have
The
Substituting equation (39) into equation (38), we obtain
Provided that we obtain the gradient
Adopting the instantaneous estimation
Differences among RA-OSELM, M-OSELM and GRAF-OSELM
In this study, we consult our previous works10,12 and absorb some good ideas from them. But actually, compared with these two previous works, many improvements and innovations have been made in this paper to accommodate the new and more complex scenarios of this study. The main differences among the proposed RA-OSELM algorithm and our previous M-OSELM algorithm,
10
GRAF-OSELM algorithm
12
can be summarized as follows.
The research purposes of each paper are different. Guo et al.
10
aim to provide a robust online learning algorithm for predicting chaotic time series with outliers, and Guo et al.
12
aim to provide a stable and adaptive online learning algorithm for time-varying system prediction, while this paper aims to provide a comprehensive online learning algorithm with both robustness and adaptive tracking capability for predicting nonstationary data streams with outliers. In brief, in this paper we try to provide a complete solution by integrating the merits of our previous works,10,12 so as to address the outliers problem and the dynamic tracking problem simultaneously. The objective functions of each algorithm are different. Compared with M-OSELM (equation (11)), an extra forgetting factor is added in RA-OSELM (equation (13)) so as to better depict the dynamic behaviors of the nonstationary data streams; compared with GRAF-OSELM (equation (12)), an M-estimation function instead of the traditional least square function is used in RA-OSELM (equation (13)) so as to enhance its robustness to outliers. The obtained recursive updating formulas of each algorithm are different. For details see equations (23) to (26) and (28) in this paper, equation (26) in Guo et al.,
10
and equation (60) in Guo et al.
12
Comparing these equations we can deduce that if the forgetting factor in RA-OSELM is equal to one, its corresponding recursive updating formula will degenerate to equation (26), and if the adopted M-estimation function in RA-OSELM is the ordinary least squares function, its corresponding recursive updating formula will degenerate to equation (60). This further certifies that the proposed RA-OSELM algorithm is a unified online learning framework of our previous two algorithms. The outlier detection methods of RA-OSELM and M-OSELM are similar but not the same. For example, in M-OSELM a common sliding window strategy is used to estimate the error standard deviation (see equation (29) in Guo et al.
10
), while in RA-OSELM we adopt forgetting factor as a smoothing coefficient and estimate the error standard deviation in a recursive fashion (see equation (31) in this paper), so as to suit the change characteristics of the nonstationary environment. The forgetting factor adjustment schemes of RA-OSELM and GRAF-OSELM are similar but not the same. In section 3.4 of this paper, the item
In conclusion, though this paper absorbs some good ideas from Guo et al.,10,12 it is not just a simple combination of our previous works, and many new issues have been reconsidered and many improvements have been made. As a matter of fact, the main thought of this paper is try to provide a complete solution under a unified learning framework, so as to simultaneously address the above-mentioned outliers problem and dynamic tracking problem, which have only been considered separately in our previous works.10,12
Experiments
In this section, a simulative nonstationary data set, a parameter varying chaotic time series data set and a real stock price data set with dynamic behaviors are used to evaluate the performances of the proposed RA-OSELM algorithm. For comparison, the R-OSELM, 13 M-OSELM, 10 GRF-OSELM, 24 GRAF-OSELM 12 and DFF-OSELM 30 are employed as contrast algorithms and their experimental results are also given.
Data set description
The first data set is an artificial data set containing 2200 samples. For each sample, the input is composed of four random variables that are normally distributed with mean zero and standard deviation 0.1, and the output is the inner product of the input vector and the corresponding coefficient vector. To simulate the dynamic feature of the nonstationary data streams, the coefficient vector for each sample evolves continuously as time goes on.
36
To be specific, the coefficient vector is firstly initiated as
The second data set is a Logistic time series with time-varying characteristics. The Logistic time series has been widely used in dynamic system modeling due to its chaotic and dynamic nature, and here a parameter varying Logistic time series described in equation (43) is taken as an example to validate the tracking performance of the proposed RA-OSELM for dynamic system
37
As seen from equation (43), the system parameter
The third data set is a real stock price data streams from the financial field. Because the factors that affect stock price often change over time, the stock price usually exhibits some kind of nonstationary behaviors, and which is very appropriate to evaluate the tracking performance of the proposed method. The used stock price data set is downloaded from the https://www.pmel.noaa.gov/tao/drupal/disdel/, and the first 700 data are chosen for experiment. For this data set, the time delay and embedding dimension for phase space reconstruction are selected as 1 and 15, respectively. 21
Experimental setup
In our experiments, the R-OSELM, M-OSELM, GRF-OSELM, GRAF-OSELM, DFF-OSELM and RA-OSELM are conducted to model and predict the above three nonstationary data streams in an online mode. To be more specific, the data are scheduled to enter the learners one-by-one in the form of a stream after the initialization phase, and then the learners make online predictions for each input continuously and execute an online update of themselves at the same time. For comparison, the given experimental results of each case are the mean of 30 independent experiments and the root mean squared error (RMSE) is adopted as the evaluation criteria. For all the six OSELM-based algorithms, the same sigmoidal additive activation function is used; the hidden nodes number n is set as 50 and 100, respectively, for experiments, and the corresponding initial training sample number is taken as 200 uniformly.
In addition, other parameter settings for each algorithm in each case are as follows. In the first experiment, for R-OSELM,
Experimental results and analysis
Comparison of tracking capability
We first verify the tracking performances of the six online algorithms for tracking the dynamic changes of the nonstationary data streams. Tables 1 to 3 give the mean and SD (standard deviations) of the online prediction RMSE over 30 independent experiments of the six algorithms for conducting different prediction steps on the above three nonstationary data sets. From Table 1, it can be seen that the prediction errors of R-OSELM and M-OSELM increase gradually with prediction steps, and they are much larger than those of the other algorithms in the same situations. In contrast, the other four forgetting factor-based algorithms, GRF-OSELM, GRAF-OSELM, DFF-OSELM, RA-OSELM, can timely eliminate the influences of the outdated samples and better track the real-time status of the nonstationary data streams, so they obtain much smaller prediction RMSE than R-OSELM and M-OSELM. Furthermore, we can see that the proposed RA-OSELM equipped with variable forgetting scheme can always achieve best or nearly best results among the contrast algorithms and exhibit good tracking performances. The experimental results in Table 2 are similar to those in Table 1, and in this simulation, the proposed RA-OSELM algorithm and our previous GRAF-OSELM algorithm achieve the best prediction results with equal accuracy level. It is worth noting that the DFF-OSELM does not behave well in this experiment, and this may be because the used parameters of this algorithm recommended by Soares and Araújo 30 are not appropriate for this data set. In the last experiment, the corresponding experimental results in Table 3 also show that the RA-OSELM and GRAF-OSELM achieve better prediction performances than the counterparts and demonstrate their superiority in tracking in real application. In short, the proposed RA-OSELM algorithm has the same good tracking capability as our previous GRAF-OSELM algorithm and they can generally provide more accurate prediction results than other representative algorithms for predicting nonstationary data streams.
Mean and SD of the online prediction RMSE of the six algorithms on simulative nonstationary data streams.
The boldface values display the best results in each case.
Mean and SD of the online prediction RMSE of the six algorithms on time-varying time series.
The boldface values display the best results in each case.
Mean and SD of the online prediction RMSE of the six algorithms on stock price.
The boldface values display the best results in each case.
Comparison of robustness
To further verify and compare the robustness of the six algorithms in the presence of outliers, some outliers are added into the above three data sets, and another group of contrast tests are performed again on each contaminated data set, respectively. For the first and the second data sets, four outliers are added to replace the original samples at point 300, 800, 1300 and 1800; for the third data set, the outliers are added in the same way at point 100, 200, 300 and 400. Figures 1 to 3 present the typical graphs of the online prediction error between the real value and the predicted value of the six contrast algorithms on the three nonstationary data sets with outliers. As shown in the figures, the R-OSELM, GRF-OSELM, GRAF-OSELM and DFF-OSELM are very sensitive to outliers, and once the outliers occur in the streaming data, these algorithms will conduct an incorrect update calculation using the abnormal outliers and consequently produce biased online prediction models, so we can see the prediction errors on the subsequent prediction steps increase significantly or fluctuate dramatically. In contrast, by taking the M-estimation technique, the proposed RA-OSELM algorithm can effectively get away from the negative effects of the outliers in the online update process and generate a more accurate fitting model, so we can see the prediction errors of the RA-OSELM are very stable (Figure 1(f) and Figure 2(f)) or fluctuate in a normal range (Figure 3(f)) during the whole online prediction process except for the occurrence points of the outliers. From the figures we can also see that although the M-OSELM algorithm demonstrates the similar robustness to outliers as the proposed RA-OSELM algorithm, its online prediction errors are much larger than those of the RA-OSELM due to lack of tracking capability.

The typical graphs of the online prediction error on simulative nonstationary data streams with outliers. (a) R-OSELM; (b) GRF-OSELM; (c) GRAF-OSELM; (d) DFF-OSELM; (e) M-OSELM; (f) RA–OSELM.

The typical graphs of the online prediction error on time-varying time series with outliers. (a) R-OSELM; (b) GRF-OSELM; (c) GRAF-OSELM; (d) DFF-OSELM; (e) M-OSELM; (f) RA–OSELM.

The typical graphs of the online prediction error on stock price with outliers. (a) R-OSELM; (b) GRF-OSELM; (c) GRAF-OSELM; (d) DFF-OSELM; (e) M-OSELM; (f) RA–OSELM.
Combining the above two aspects together, we can conclude that the proposed RA-OSELM has good tracking capability with strong robustness and it can achieve more accurate and stable prediction results than its counterparts for predicting nonstationary data streams with outliers.
Comparison of learning efficiency
To compare the learning efficiency of the six OSELM-based algorithms, the online modeling and prediction time of these algorithms on each data sets are recorded, and the results are shown in Tables 4 to 6, respectively. The experimental results reported here are obtained in the case that the hidden nodes number is set as 100 for all the six algorithms. From the results, we can see first that for each single algorithm, the prediction time increases linearly with the prediction steps. In addition, by comparison of all the algorithms under the same condition, the prediction time of R-OSELM, M-OSELM and DFF-OSELM are basically the same and the shortest, followed by GRF-OSELM, followed by GRAF-OSELM, and the prediction time of RA-OSELM is the longest. Though the proposed RA-OSELM is the least efficient among the six algorithms for the reason that additional calculations are required for adjusting the forgetting factor recursively and detecting the potential outliers continually in the online prediction process, its adaptive tracking ability and anti-outlier ability have been greatly improved, so it is worth it.
Online prediction time (seconds) of the six algorithms on simulative nonstationary data streams.
Online prediction time (seconds) of the six algorithms on time-varying time series.
Online prediction time (seconds) of the six algorithms on stock price.
Conclusions
In this paper, a novel robust adaptive OSELM algorithm (RA-OSELM) based on M-estimation and variable forgetting factor is presented for predicting nonstationary data streams with outliers. By employing a more robust M-estimation loss function instead of the traditional least square learning criterion, the RA-OSELM’s online updating schemes are less affected by the large disturbances stemming from the outliers, thus improving the robustness of the algorithm in the presence of outliers. Besides, with the help of the variable forgetting factor strategy, the RA-OSELM is capable of tracking the dynamic changes of the nonstationary data streams automatically and reducing the adverse impacts of the outdated data timely, so it tends to produce satisfying tracking results in nonstationary environments. The performances of the proposed RA-OSELM algorithm have been evaluated and compared with other five representative algorithms by synthetic and real data sets. The experimental results indicate that the RA-OSELM has better adaptive tracking capability with stronger robustness than its counterparts, and which is applicable for the online modeling and prediction of nonstationary data streams with outliers.
Footnotes
Declaration of conflicting interests
The author(s) declared no potential conflicts of interest with respect to the research, authorship, and/or publication of this article.
Funding
The author(s) disclosed receipt of the following financial support for the research, authorship, and/or publication of this article: This research was supported by the National Natural Science Foundation of China (grant numbers 61603326, 61379064, 61273106); the research fund of Jiangsu Provincial Key Constructive Laboratory for Big Data of Psychology and Cognitive Science (grant number 72591962004 G).
