A generalization of the Mixing Past Posteriors algorithm for an uncountable set of experts

Автор: Zukhba R.D., Zukhba A.V.

Журнал: Труды Московского физико-технического института @trudy-mipt

Рубрика: Математика

Статья в выпуске: 4 (68) т.17, 2025 года.

Бесплатный доступ

A paper [2, Herbster, Warmuth, 1998] examines an on-line learning problem of aggregating the predictions of a finite set of experts with the goal of predicting almost as well as the best sequence of such experts. The regret in this problem is 𝑅𝑇 = Θ(𝑘 ln𝑁), that is, it gets arbitrary large with a large number of experts 𝑁. This paper considers a similar setting, except with an uncountable set of experts 𝒩 = [𝑐, 𝑑]. We propose a generalization of the Mixing Past Posteriors algorithm for an uncountable set of experts. We provide an upper bound on the regret of this generalization and a lower bound on regret of any algorithm that solves this problem. We also propose a generalization of a standard approach of estimating the regret of algorithms similar to MPP by taking into account additional information regarding the losses of individual experts.

Еще

On-line learning, prediction with experts’ advice, supervised learning, regret, compound expert, upper regret bound, lower regret bound

Короткий адрес: https://sciup.org/142247123

IDR: 142247123   |   УДК: 519.6