Combining labeled and unlabeled data with co-training
COLT' 98 Proceedings of the eleventh annual conference on Computational learning theory
Pattern Classification (2nd Edition)
Pattern Classification (2nd Edition)
ICTAI '04 Proceedings of the 16th IEEE International Conference on Tools with Artificial Intelligence
Semi-supervised regression with co-training
IJCAI'05 Proceedings of the 19th international joint conference on Artificial intelligence
Vector quantization: a weighted version for time-series forecasting
Future Generation Computer Systems
Using co-training and self-training in semi-supervised multiple classifier systems
SSPR'06/SPR'06 Proceedings of the 2006 joint IAPR international conference on Structural, Syntactic, and Statistical Pattern Recognition
A comparison between neural-network forecasting techniques-case study: river flow forecasting
IEEE Transactions on Neural Networks
MPEG-4 and H.263 video traces for network performance evaluation
IEEE Network: The Magazine of Global Internetworking
MCS '09 Proceedings of the 8th International Workshop on Multiple Classifier Systems
Hi-index | 0.00 |
In this paper we consider the problem of missing data in time series analysis. We propose a semi-supervised co-training method to handle the problem of missing data. We transform the time series data to set of labeled and unlabeled data. Different predictors are used to predict the unlabelled data and the most confident labeled patterns are used to retrain the predictors further to and enhance the overall prediction accuracy. By labeling the unknown patterns the missing data is compensated for. Experiments were conducted on different time series data and with varying percentage of missing data using a uniform distribution. We used KNN base predictors and Fuzzy Inductive Reasoning (FIR) base predictors and compared their performance using different confidence measures. Results reveal the effectiveness of the co-training method to compensate for the missing values and to improve prediction. The FIR model together with the "similarity" confidence measures obtained in most cases the best results in our study.