Mistake bounds and logarithmic linear-threshold learning algorithms
Mistake bounds and logarithmic linear-threshold learning algorithms
COLT '90 Proceedings of the third annual workshop on Computational learning theory
Elements of information theory
Elements of information theory
The weighted majority algorithm
Information and Computation
Journal of the ACM (JACM)
Using and combining predictors that specialize
STOC '97 Proceedings of the twenty-ninth annual ACM symposium on Theory of computing
A decision-theoretic generalization of on-line learning and an application to boosting
Journal of Computer and System Sciences - Special issue: 26th annual ACM symposium on the theory of computing & STOC'94, May 23–25, 1994, and second annual Europe an conference on computational learning theory (EuroCOLT'95), March 13–15, 1995
General convergence results for linear discriminant updates
COLT '97 Proceedings of the tenth annual conference on Computational learning theory
A game of prediction with expert advice
Journal of Computer and System Sciences - Special issue on the eighth annual workshop on computational learning theory, July 5–8, 1995
Context-sensitive learning methods for text categorization
ACM Transactions on Information Systems (TOIS)
The robustness of the p-norm algorithms
COLT '99 Proceedings of the twelfth annual conference on Computational learning theory
Large Margin Classification Using the Perceptron Algorithm
Machine Learning - The Eleventh Annual Conference on computational Learning Theory
Analysis of two gradient-based algorithms for on-line regression
Journal of Computer and System Sciences
Linear hinge loss and average margin
Proceedings of the 1998 conference on Advances in neural information processing systems II
Relative Loss Bounds for Multidimensional Regression Problems
Machine Learning
Machine Learning
General Convergence Results for Linear Discriminant Updates
Machine Learning
A Second-Order Perceptron Algorithm
COLT '02 Proceedings of the 15th Annual Conference on Computational Learning Theory
Internal Regret in On-Line Portfolio Selection
Machine Learning
Improved second-order bounds for prediction with expert advice
Machine Learning
Regret Minimization Under Partial Monitoring
Mathematics of Operations Research
Worst-Case Analysis of Selective Sampling for Linear Classification
The Journal of Machine Learning Research
The communication complexity of uncoupled nash equilibrium procedures
Proceedings of the thirty-ninth annual ACM symposium on Theory of computing
Consistency of discrete Bayesian learning
Theoretical Computer Science
ICANNGA '07 Proceedings of the 8th international conference on Adaptive and Natural Computing Algorithms, Part I
On the convergence of regret minimization dynamics in concave games
Proceedings of the forty-first annual ACM symposium on Theory of computing
Classification of peptide mass fingerprint data by novel no-regret boosting method
Computers in Biology and Medicine
Regret Minimization and Job Scheduling
SOFSEM '10 Proceedings of the 36th Conference on Current Trends in Theory and Practice of Computer Science
An overview of AI research in Italy
Artificial intelligence
Recognition tasks are imitation games
ICAPR'05 Proceedings of the Third international conference on Advances in Pattern Recognition - Volume Part I
The missing consistency theorem for bayesian learning: stochastic model selection
ALT'06 Proceedings of the 17th international conference on Algorithmic Learning Theory
From external to internal regret
COLT'05 Proceedings of the 18th annual conference on Learning Theory
Sample complexity of risk-averse bandit-arm selection
IJCAI'13 Proceedings of the Twenty-Third international joint conference on Artificial Intelligence
Hi-index | 0.00 |
In this paper we show that several known algorithms for sequential prediction problems (including Weighted Majority and the quasi-additive family of Grove, Littlestone, and Schuurmans), for playing iterated games (including Freund and Schapire's Hedge and MW, as well as the Λ-strategies of Hart and Mas-Colell), and for boosting (including AdaBoost) are special cases of a general decision strategy based on the notion of potential. By analyzing this strategy we derive known performance bounds, as well as new bounds, as simple corollaries of a single general theorem. Besides offering a new and unified view on a large family of algorithms, we establish a connection between potential-based analysis in learning and their counterparts independently developed in game theory. By exploiting this connection, we show that certain learning problems are instances of more general game-theoretic problems. In particular, we describe a notion of generalized regret and show its applications in learning theory.