Regret Analysis of Stochastic and Nonstochastic Multi-Armed Bandit Problems (Foundations and Trends(r) in Machine Learning)

Bubeck, Sébastien; Nicolò, Cesa-Bianchi

ISBN 10: 1601986262 ISBN 13: 9781601986269
Verlag: Now Publishers, 2012
Neu Softcover

Verkäufer Ria Christie Collections, Uxbridge, Vereinigtes Königreich Verkäuferbewertung 5 von 5 Sternen 5 Sterne, Erfahren Sie mehr über Verkäufer-Bewertungen

AbeBooks-Verkäufer seit 25. März 2015

Dieses Buch ist nicht mehr verfügbar. AbeBooks führt Millionen von Büchern. Bitte geben Sie unten Suchbegriffe ein, um ähnliche Exemplare zu finden.

Beschreibung

Beschreibung:

In. Bestandsnummer des Verkäufers ria9781601986269_new

Diesen Artikel melden

Inhaltsangabe:

A multi-armed bandit problem - or, simply, a bandit problem - is a sequential allocation problem defined by a set of actions. At each time step, a unit resource is allocated to an action and some observable payoff is obtained. The goal is to maximize the total payoff obtained in a sequence of allocations. The name bandit refers to the colloquial term for a slot machine (a "one-armed bandit" in American slang). In a casino, a sequential allocation problem is obtained when the player is facing many slot machines at once (a "multi-armed bandit"), and must repeatedly choose where to insert the next coin. Multi-armed bandit problems are the most basic examples of sequential decision problems with an exploration-exploitation trade-off. This is the balance between staying with the option that gave highest payoffs in the past and exploring new options that might give higher payoffs in the future. Although the study of bandit problems dates back to the 1930s, exploration-exploitation trade-offs arise in several modern applications, such as ad placement, website optimization, and packet routing. Mathematically, a multi-armed bandit is defined by the payoff process associated with each option. In this book, the focus is on two extreme cases in which the analysis of regret is particularly simple and elegant: independent and identically distributed payoffs and adversarial payoffs. Besides the basic setting of finitely many actions, it also analyzes some of the most important variants and extensions, such as the contextual bandit model. This monograph is an ideal reference for students and researchers with an interest in bandit problems.

Reseña del editor: A multi-armed bandit problem - or, simply, a bandit problem - is a sequential allocation problem defined by a set of actions. At each time step, a unit resource is allocated to an action and some observable payoff is obtained. The goal is to maximize the total payoff obtained in a sequence of allocations. The name bandit refers to the colloquial term for a slot machine (a "one-armed bandit" in American slang). In a casino, a sequential allocation problem is obtained when the player is facing many slot machines at once (a "multi-armed bandit"), and must repeatedly choose where to insert the next coin. Multi-armed bandit problems are the most basic examples of sequential decision problems with an exploration-exploitation trade-off. This is the balance between staying with the option that gave highest payoffs in the past and exploring new options that might give higher payoffs in the future. Although the study of bandit problems dates back to the 1930s, exploration-exploitation trade-offs arise in several modern applications, such as ad placement, website optimization, and packet routing. Mathematically, a multi-armed bandit is defined by the payoff process associated with each option. In this book, the focus is on two extreme cases in which the analysis of regret is particularly simple and elegant: independent and identically distributed payoffs and adversarial payoffs. Besides the basic setting of finitely many actions, it also analyzes some of the most important variants and extensions, such as the contextual bandit model. This monograph is an ideal reference for students and researchers with an interest in bandit problems.

„Über diesen Titel“ kann sich auf eine andere Ausgabe dieses Titels beziehen.

Bibliografische Details

Titel: Regret Analysis of Stochastic and ...
Verlag: Now Publishers
Erscheinungsdatum: 2012
Einband: Softcover
Zustand: New

ZVAB ist ein Internet-Marktplatz für neue, gebrauchte, antiquarische und vergriffene Bücher. Bei uns finden Sie Tausende professioneller Buchhändler weltweit und Millionen Bücher. Einkaufen beim ZVAB ist einfach und zu 100% sicher — Suchen Sie nach Ihrem Buch, erwerben Sie es über unsere sichere Kaufabwicklung und erhalten Sie Ihr Buch direkt vom Händler.

Millionen neuer und gebrauchter Bücher bei tausenden Anbietern

Antiquarische Bücher

Antiquarische Bücher

Von seltenen Erstausgaben bis hin zu begehrten signierten Ausgaben – beim ZVAB finden Sie eine große Anzahl seltener, wertvoller Bücher und Sammlerstücke.

ZVAB Startseite

Erstausgaben

Erstausgaben

Erstausgaben sind besondere Bücher, die den ersten Abdruck des Textes in seiner ursprünglichen Form darstellen. Hier finden sie Erstausgaben von damals bis heute.

Erstausgaben

Gebrauchte Bücher

Gebrauchte Bücher

Ob Bestseller oder Klassiker, das ZVAB bietet Ihnen eine breite Auswahl an gebrauchten Büchern: Stöbern Sie in unseren Rubriken und entdecken Sie ein Buch-Schnäppchen.

Gebrauchte Bücher

Mehr Bücher entdecken