DHL Abholort 4.69 Hermes Kurierdienst 4.99 DHL-Kurier 3.99 Hermes-Stelle 4.49 GLS-Kurierdienst 3.99

Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems

Sprache EnglischEnglisch
Buch Broschur
Buch Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems Sebastian Bubeck
Libristo-Code: 04834934
Verlag now publishers Inc, Dezember 2012
A multi-armed bandit problem - or, simply, a bandit problem - is a sequential allocation problem def... Vollständige Beschreibung
? points 267 b
106.34 inkl. MwSt.
Externes Lager Wir versenden in 15-20 Tagen

30 Tage für die Rückgabe der Ware


Das könnte Sie auch interessieren


TOP
Reckless Lauren Roberts / Broschur
common.buy 11.39
TOP
Happiness 1 Shuzo Oshimi / Broschur
common.buy 13.11
TOP
The Course of Love Alain de Botton / Broschur
common.buy 11.59
TOP
Science of Cooking Stuart Farrimond / Hardcover
common.buy 27.84
TOP
Dead Boy Detectives Omnibus Toby Litt / Hardcover
common.buy 89.59
Japanese Sake Bible Takashi Eguchi / Broschur
common.buy 17.95
Dog Behaviour, Evolution, and Cognition Adam Miklosi / Broschur
common.buy 75.76
Damn Delicious Meal Prep Chungah Rhee / Hardcover
common.buy 28.54
Albert Camus: A Life Olivier Todd / Broschur
common.buy 24.41
Design Principles for Photography Jeremy Webb / Broschur
common.buy 48.62
Art of SEO Stephan Spencer / Broschur
common.buy 62.34
Battles of Tolkien David Day / Broschur
common.buy 15.83
How to Be Better at Almost Everything Pat Flynn / Hardcover
common.buy 25.82

A multi-armed bandit problem - or, simply, a bandit problem - is a sequential allocation problem defined by a set of actions. At each time step, a unit resource is allocated to an action and some observable payoff is obtained. The goal is to maximize the total payoff obtained in a sequence of allocations. The name bandit refers to the colloquial term for a slot machine (a "one-armed bandit" in American slang). In a casino, a sequential allocation problem is obtained when the player is facing many slot machines at once (a "multi-armed bandit"), and must repeatedly choose where to insert the next coin. Multi-armed bandit problems are the most basic examples of sequential decision problems with an exploration-exploitation trade-off. This is the balance between staying with the option that gave highest payoffs in the past and exploring new options that might give higher payoffs in the future. Although the study of bandit problems dates back to the 1930s, exploration-exploitation trade-offs arise in several modern applications, such as ad placement, website optimization, and packet routing. Mathematically, a multi-armed bandit is defined by the payoff process associated with each option. In this book, the focus is on two extreme cases in which the analysis of regret is particularly simple and elegant: independent and identically distributed payoffs and adversarial payoffs. Besides the basic setting of finitely many actions, it also analyzes some of the most important variants and extensions, such as the contextual bandit model. This monograph is an ideal reference for students and researchers with an interest in bandit problems.

Informationen zum Buch

Vollständiger Name Regret Analysis of Stochastic and Nonstochastic Multi-armed Bandit Problems
Sprache Englisch
Einband Buch - Broschur
Datum der Veröffentlichung 2012
Anzahl der Seiten 138
EAN 9781601986269
ISBN 1601986262
Libristo-Code 04834934
Gewicht 208
Abmessungen 234 x 159 x 8
Verschenken Sie dieses Buch noch heute
Es ist ganz einfach
1 Legen Sie das Buch in Ihren Warenkorb und wählen Sie den Versand als Geschenk 2 Wir schicken Ihnen umgehend einen Gutschein 3 Das Buch wird an die Adresse des beschenkten Empfängers geliefert

Anmeldung

Melden Sie sich bei Ihrem Konto an. Sie haben noch kein Libristo-Konto? Erstellen Sie es jetzt!

 
obligatorisch
obligatorisch

Sie haben kein Konto? Nutzen Sie die Vorteile eines Libristo-Kontos!

Mit einem Libristo-Konto haben Sie alles unter Kontrolle.

Erstellen Sie ein Libristo-Konto