TY - JOUR
T1 - Meta Expert Learning and Efficient Pruning for Evolving Data Streams
AU - Azarafrooz, Mahdi
AU - Daneshmand, Mahmoud
N1 - Publisher Copyright:
© 2015 IEEE.
PY - 2015/8/1
Y1 - 2015/8/1
N2 - Researchers have proposed several ensemble methods for the data stream environments including online bagging and boosting. These studies show that bagging methods perform better than boosting methods although the opposite is known to be true in the batch setting environments. The reason behind the weaker performance of boosting methods in the streaming environments is not clear. We have taken advantage of the algorithmic procedure of meta expert learnings for the sake of our study. The meta expert learning differs from the classic expert learning methods in that each expert starts to predict from a different point in the history. Moreover, maintaining a collection of base learners follows an algorithmic procedure. The focus of this paper is on studying the pruning function for maintaining the appropriate set of experts rather than proposing a competitive algorithm for selecting the experts. It shows how a well-structured pruning method leads to a better prediction accuracy without necessary higher memory consumption. Next, it is shown how pruning the set of base learners in the meta expert learning (in order to avoid memory exhaustion) affects the prediction accuracy for different types of drifts. In the case of time-locality drifts, the prediction accuracy is highly tied to the mathematical structure of the pruning algorithms. This observation may explain the main reason behind the weak performance of previously studied boosting methods in the streaming environments. It shows that the boosting algorithms should be designed with respect to the suitable notion of the regret metrics.
AB - Researchers have proposed several ensemble methods for the data stream environments including online bagging and boosting. These studies show that bagging methods perform better than boosting methods although the opposite is known to be true in the batch setting environments. The reason behind the weaker performance of boosting methods in the streaming environments is not clear. We have taken advantage of the algorithmic procedure of meta expert learnings for the sake of our study. The meta expert learning differs from the classic expert learning methods in that each expert starts to predict from a different point in the history. Moreover, maintaining a collection of base learners follows an algorithmic procedure. The focus of this paper is on studying the pruning function for maintaining the appropriate set of experts rather than proposing a competitive algorithm for selecting the experts. It shows how a well-structured pruning method leads to a better prediction accuracy without necessary higher memory consumption. Next, it is shown how pruning the set of base learners in the meta expert learning (in order to avoid memory exhaustion) affects the prediction accuracy for different types of drifts. In the case of time-locality drifts, the prediction accuracy is highly tied to the mathematical structure of the pruning algorithms. This observation may explain the main reason behind the weak performance of previously studied boosting methods in the streaming environments. It shows that the boosting algorithms should be designed with respect to the suitable notion of the regret metrics.
UR - http://www.scopus.com/inward/record.url?scp=84938831009&partnerID=8YFLogxK
UR - http://www.scopus.com/inward/citedby.url?scp=84938831009&partnerID=8YFLogxK
U2 - 10.1109/JIOT.2015.2420689
DO - 10.1109/JIOT.2015.2420689
M3 - Article
AN - SCOPUS:84938831009
VL - 2
SP - 268
EP - 273
JO - IEEE Internet of Things Journal
JF - IEEE Internet of Things Journal
IS - 4
ER -