Multiple strategies detection in ontology mapping
WWW '05 Special interest tracks and posters of the 14th international conference on World Wide Web
Mining relational databases with multi-view learning
MRDM '05 Proceedings of the 4th international workshop on Multi-relational mining
Interactive wrapper generation with minimal user effort
Proceedings of the 15th international conference on World Wide Web
Mining relational data through correlation-based multiple view validation
Proceedings of the 12th ACM SIGKDD international conference on Knowledge discovery and data mining
Improving supervised learning performance by using fuzzy clustering method to select training data
Journal of Intelligent & Fuzzy Systems: Applications in Engineering and Technology - Fuzzy theory and technology with applications
Multirelational classification: a multiple view approach
Knowledge and Information Systems
Learning from Skewed Class Multi-relational Databases
Fundamenta Informaticae - Progress on Multi-Relational Data Mining
Multi-view Semi-supervised Learning: An Approach to Obtain Different Views from Text Datasets
Proceedings of the 2005 conference on Advances in Logic Based Intelligent Systems: Selected Papers of LAPTEC 2005
Active learning with multiple views
Journal of Artificial Intelligence Research
Active learning with strong and weak views: a case study on wrapper induction
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
Deploying information agents on the web
IJCAI'03 Proceedings of the 18th international joint conference on Artificial intelligence
Actively Learning Ontology Matching via User Interaction
ISWC '09 Proceedings of the 8th International Semantic Web Conference
A Multi-view Approach for Relation Extraction
WISM '09 Proceedings of the International Conference on Web Information Systems and Mining
Active learning strategies: a case study for detection of emotions in speech
ICDM'07 Proceedings of the 7th industrial conference on Advances in data mining: theoretical aspects and applications
Initial training data selection for active learning
Proceedings of the 5th International Conference on Ubiquitous Information Management and Communication
Multi-class ensemble-based active learning
ECML'06 Proceedings of the 17th European conference on Machine Learning
iASA: learning to annotate the semantic web
Journal on Data Semantics IV
Learning from Skewed Class Multi-relational Databases
Fundamenta Informaticae - Progress on Multi-Relational Data Mining
Hi-index | 0.00 |
Labeling training data for machine learning algorithms is tedious, time consuming, and error prone. Consequently, it is of utmost importance to minimize the amount of labeled data that is required to learn a target concept. In the work presented here, I focus on reducing the need for labeled data in multi-view learning tasks. The key characteristic of multi-view learning tasks is that the target concept can be independently learned within different views (i.e., disjoint sets of features that are sufficient to learn the concept of interest). For instance, robot navigation is a 2-view learning task because a robot can learn to avoid obstacles based on either sonar or vision sensors. In my dissertation, I make three main contributions. First, I introduce Co-Testing, which is an active learning algorithm that exploits multiple views. Co-Testing is based on the idea of learning from mistakes. More precisely, it queries examples on which the views predict a different label: if two views disagree, one of them is guaranteed to make a mistake. In a variety of real-world domains, from information extraction to text classification and discourse tree parsing, Co-Testing outperforms existing active learners. Second, I show that existing multi-view learners can perform unreliably if the views are incompatible or correlated. To cope with this problem, I introduce a robust multi-view learner, Co-EMT, which interleaves semi-supervised and active multi-view learning. My empirical results show that Co-EMT outperforms existing multi-view learners on a wide variety of learning tasks. Third, I introduce a view validation algorithm that predicts whether or not two views are adequate for solving a new, unseen learning task. View validation uses information acquired while solving several exemplar learning tasks to train a classifier that discriminates between tasks for which the views are adequate and inadequate for multi-view learning. My experiments on wrapper induction and text classification show that view validation requires a modest amount of training data to make high accuracy predictions.