Dekorationsartikel gehören nicht zum Leistungsumfang.
Sprache:
Englisch
72,85 €*
Versandkostenfrei per Post / DHL
Lieferzeit 1-2 Wochen
Kategorien:
Beschreibung
A description of perturbation-based methods developed in machine learning to augment novel optimization methods with strong statistical guarantees.
In nearly all machine learning, decisions must be made given current knowledge. Surprisingly, making what is believed to be the best decision is not always the best strategy, even when learning in a supervised learning setting. An emerging body of work on learning under different rules applies perturbations to decision and learning procedures. These methods provide simple and highly efficient learning rules with improved theoretical guarantees. This book describes perturbation-based methods developed in machine learning to augment novel optimization methods with strong statistical guarantees, offering readers a state-of-the-art overview.
Chapters address recent modeling ideas that have arisen within the perturbations framework, including Perturb & MAP, herding, and the use of neural networks to map generic noise to distribution over highly structured data. They describe new learning procedures for perturbation models, including an improved EM algorithm and a learning algorithm that aims to match moments of model samples to moments of data. They discuss understanding the relation of perturbation models to their traditional counterparts, with one chapter showing that the perturbations viewpoint can lead to new algorithms in the traditional setting. And they consider perturbation-based regularization in neural networks, offering a more complete understanding of dropout and studying perturbations in the context of deep neural networks.
In nearly all machine learning, decisions must be made given current knowledge. Surprisingly, making what is believed to be the best decision is not always the best strategy, even when learning in a supervised learning setting. An emerging body of work on learning under different rules applies perturbations to decision and learning procedures. These methods provide simple and highly efficient learning rules with improved theoretical guarantees. This book describes perturbation-based methods developed in machine learning to augment novel optimization methods with strong statistical guarantees, offering readers a state-of-the-art overview.
Chapters address recent modeling ideas that have arisen within the perturbations framework, including Perturb & MAP, herding, and the use of neural networks to map generic noise to distribution over highly structured data. They describe new learning procedures for perturbation models, including an improved EM algorithm and a learning algorithm that aims to match moments of model samples to moments of data. They discuss understanding the relation of perturbation models to their traditional counterparts, with one chapter showing that the perturbations viewpoint can lead to new algorithms in the traditional setting. And they consider perturbation-based regularization in neural networks, offering a more complete understanding of dropout and studying perturbations in the context of deep neural networks.
A description of perturbation-based methods developed in machine learning to augment novel optimization methods with strong statistical guarantees.
In nearly all machine learning, decisions must be made given current knowledge. Surprisingly, making what is believed to be the best decision is not always the best strategy, even when learning in a supervised learning setting. An emerging body of work on learning under different rules applies perturbations to decision and learning procedures. These methods provide simple and highly efficient learning rules with improved theoretical guarantees. This book describes perturbation-based methods developed in machine learning to augment novel optimization methods with strong statistical guarantees, offering readers a state-of-the-art overview.
Chapters address recent modeling ideas that have arisen within the perturbations framework, including Perturb & MAP, herding, and the use of neural networks to map generic noise to distribution over highly structured data. They describe new learning procedures for perturbation models, including an improved EM algorithm and a learning algorithm that aims to match moments of model samples to moments of data. They discuss understanding the relation of perturbation models to their traditional counterparts, with one chapter showing that the perturbations viewpoint can lead to new algorithms in the traditional setting. And they consider perturbation-based regularization in neural networks, offering a more complete understanding of dropout and studying perturbations in the context of deep neural networks.
In nearly all machine learning, decisions must be made given current knowledge. Surprisingly, making what is believed to be the best decision is not always the best strategy, even when learning in a supervised learning setting. An emerging body of work on learning under different rules applies perturbations to decision and learning procedures. These methods provide simple and highly efficient learning rules with improved theoretical guarantees. This book describes perturbation-based methods developed in machine learning to augment novel optimization methods with strong statistical guarantees, offering readers a state-of-the-art overview.
Chapters address recent modeling ideas that have arisen within the perturbations framework, including Perturb & MAP, herding, and the use of neural networks to map generic noise to distribution over highly structured data. They describe new learning procedures for perturbation models, including an improved EM algorithm and a learning algorithm that aims to match moments of model samples to moments of data. They discuss understanding the relation of perturbation models to their traditional counterparts, with one chapter showing that the perturbations viewpoint can lead to new algorithms in the traditional setting. And they consider perturbation-based regularization in neural networks, offering a more complete understanding of dropout and studying perturbations in the context of deep neural networks.
Über den Autor
Tamir Hazan is Assistant Professor at Technion, Israel Institute of Technology.
George Papandreou is a Research Scientist for Google, Inc.
Daniel Tarlow is a Researcher at Microsoft Research Cambridge, UK.
George Papandreou is a Research Scientist for Google, Inc.
Daniel Tarlow is a Researcher at Microsoft Research Cambridge, UK.
Details
Erscheinungsjahr: | 2023 |
---|---|
Genre: | Importe, Informatik |
Rubrik: | Naturwissenschaften & Technik |
Medium: | Taschenbuch |
Inhalt: | Einband - flex.(Paperback) |
ISBN-13: | 9780262549943 |
ISBN-10: | 0262549948 |
Sprache: | Englisch |
Einband: | Kartoniert / Broschiert |
Redaktion: |
Hazan, Tamir
Papandreou, George Tarlow, Daniel |
Hersteller: | MIT Press |
Verantwortliche Person für die EU: | Libri GmbH, Europaallee 1, D-36244 Bad Hersfeld, gpsr@libri.de |
Maße: | 254 x 203 x 22 mm |
Von/Mit: | Tamir Hazan (u. a.) |
Erscheinungsdatum: | 05.12.2023 |
Gewicht: | 0,88 kg |
Über den Autor
Tamir Hazan is Assistant Professor at Technion, Israel Institute of Technology.
George Papandreou is a Research Scientist for Google, Inc.
Daniel Tarlow is a Researcher at Microsoft Research Cambridge, UK.
George Papandreou is a Research Scientist for Google, Inc.
Daniel Tarlow is a Researcher at Microsoft Research Cambridge, UK.
Details
Erscheinungsjahr: | 2023 |
---|---|
Genre: | Importe, Informatik |
Rubrik: | Naturwissenschaften & Technik |
Medium: | Taschenbuch |
Inhalt: | Einband - flex.(Paperback) |
ISBN-13: | 9780262549943 |
ISBN-10: | 0262549948 |
Sprache: | Englisch |
Einband: | Kartoniert / Broschiert |
Redaktion: |
Hazan, Tamir
Papandreou, George Tarlow, Daniel |
Hersteller: | MIT Press |
Verantwortliche Person für die EU: | Libri GmbH, Europaallee 1, D-36244 Bad Hersfeld, gpsr@libri.de |
Maße: | 254 x 203 x 22 mm |
Von/Mit: | Tamir Hazan (u. a.) |
Erscheinungsdatum: | 05.12.2023 |
Gewicht: | 0,88 kg |
Sicherheitshinweis