Decision-making in the face of uncertainty is a significant challenge in machine learning, and the multi-armed bandit model is a commonly used framework to address it. This comprehensive and rigorous introduction to the multi-armed bandit problem examines all the major settings, including stochastic, adversarial, and Bayesian frameworks. A focus on both mathematical intuition and carefully worked proofs makes this an excellent reference for established researchers and a helpful resource for graduate students in computer science, engineering, statistics, applied mathematics and economics. Linear bandits receive special attention as one of the most useful models in applications, while other chapters are dedicated to combinatorial bandits, ranking, non-stationary problems, Thompson sampling and pure exploration. The book ends with a peek into the world beyond bandits with an introduction to partial monitoring and learning in Markov decision processes.
Die Inhaltsangabe kann sich auf eine andere Ausgabe dieses Titels beziehen.
Tor Lattimore is a research scientist at DeepMind. His research is focused on decision making in the face of uncertainty, including bandit algorithms and reinforcement learning. Before joining DeepMind he was an assistant professor at Indiana University and a postdoctoral fellow at the University of Alberta.
Csaba Szepesvári is a Professor in the Department of Computing Science at the University of Alberta and a Principal Investigator of the Alberta Machine Intelligence Institute. He also leads the 'Foundations' team at DeepMind. He has co-authored a book on nonlinear approximate adaptive controllers and authored a book on reinforcement learning, in addition to publishing over 200 journal and conference papers. He is an action editor of the Journal of Machine Learning Research.
„Über diesen Titel“ kann sich auf eine andere Ausgabe dieses Titels beziehen.
Gratis für den Versand innerhalb von/der USA
Versandziele, Kosten & DauerEUR 7,65 für den Versand innerhalb von/der USA
Versandziele, Kosten & DauerAnbieter: BooksRun, Philadelphia, PA, USA
Hardcover. Zustand: Very Good. 1. It's a well-cared-for item that has seen limited use. The item may show minor signs of wear. All the text is legible, with all pages included. It may have slight markings and/or highlighting. Bestandsnummer des Verkäufers 1108486827-8-1
Anzahl: 1 verfügbar
Anbieter: Goodwill Books, Hillsboro, OR, USA
Zustand: good. Signs of wear and consistent use. Bestandsnummer des Verkäufers 3IIT7G0063MN_ns
Anzahl: 1 verfügbar
Anbieter: Best Price, Torrance, CA, USA
Zustand: New. SUPER FAST SHIPPING. Bestandsnummer des Verkäufers 9781108486828
Anzahl: 2 verfügbar
Anbieter: GreatBookPrices, Columbia, MD, USA
Zustand: New. Bestandsnummer des Verkäufers 40407250-n
Anzahl: Mehr als 20 verfügbar
Anbieter: Lucky's Textbooks, Dallas, TX, USA
Zustand: New. Bestandsnummer des Verkäufers ABLIING23Mar2317530285308
Anzahl: Mehr als 20 verfügbar
Anbieter: BargainBookStores, Grand Rapids, MI, USA
Hardback or Cased Book. Zustand: New. Bandit Algorithms 2.3. Book. Bestandsnummer des Verkäufers BBS-9781108486828
Anzahl: 5 verfügbar
Anbieter: California Books, Miami, FL, USA
Zustand: New. Bestandsnummer des Verkäufers I-9781108486828
Anzahl: Mehr als 20 verfügbar
Anbieter: GreatBookPrices, Columbia, MD, USA
Zustand: As New. Unread book in perfect condition. Bestandsnummer des Verkäufers 40407250
Anzahl: Mehr als 20 verfügbar
Anbieter: Grand Eagle Retail, Mason, OH, USA
Hardcover. Zustand: new. Hardcover. Decision-making in the face of uncertainty is a significant challenge in machine learning, and the multi-armed bandit model is a commonly used framework to address it. This comprehensive and rigorous introduction to the multi-armed bandit problem examines all the major settings, including stochastic, adversarial, and Bayesian frameworks. A focus on both mathematical intuition and carefully worked proofs makes this an excellent reference for established researchers and a helpful resource for graduate students in computer science, engineering, statistics, applied mathematics and economics. Linear bandits receive special attention as one of the most useful models in applications, while other chapters are dedicated to combinatorial bandits, ranking, non-stationary problems, Thompson sampling and pure exploration. The book ends with a peek into the world beyond bandits with an introduction to partial monitoring and learning in Markov decision processes. Decision-making in the face of uncertainty is a challenge in machine learning, and the multi-armed bandit model is a common framework to address it. This comprehensive introduction is an excellent reference for established researchers and a resource for graduate students interested in exploring stochastic, adversarial and Bayesian frameworks. Shipping may be from multiple locations in the US or from the UK, depending on stock availability. Bestandsnummer des Verkäufers 9781108486828
Anzahl: 1 verfügbar
Anbieter: Ria Christie Collections, Uxbridge, Vereinigtes Königreich
Zustand: New. In. Bestandsnummer des Verkäufers ria9781108486828_new
Anzahl: Mehr als 20 verfügbar