Chong / Zak / Lu | An Introduction to Optimization | Buch | 978-1-119-87763-9 | sack.de

Buch, Englisch, 672 Seiten, Format (B × H): 183 mm x 260 mm, Gewicht: 1437 g

Chong / Zak / Lu

An Introduction to Optimization

With Applications to Machine Learning

Buch, Englisch, 672 Seiten, Format (B × H): 183 mm x 260 mm, Gewicht: 1437 g

ISBN: 978-1-119-87763-9
Verlag: John Wiley & Sons Inc


An Introduction to Optimization
Accessible introductory textbook on optimization theory and methods, with an emphasis on engineering design, featuring MATLAB® exercises and worked examples
Fully updated to reflect modern developments in the field, the Fifth Edition of An Introduction to Optimization fills the need for an accessible, yet rigorous, introduction to optimization theory and methods, featuring innovative coverage and a straightforward approach. The book begins with a review of basic definitions and notations while also providing the related fundamental background of linear algebra, geometry, and calculus.
With this foundation, the authors explore the essential topics of unconstrained optimization problems, linear programming problems, and nonlinear constrained optimization. In addition, the book includes an introduction to artificial neural networks, convex optimization, multi-objective optimization, and applications of optimization in machine learning.
Numerous diagrams and figures found throughout the book complement the written presentation of key concepts, and each chapter is followed by MATLAB® exercises and practice problems that reinforce the discussed theory and algorithms.
The Fifth Edition features a new chapter on Lagrangian (nonlinear) duality, expanded coverage on matrix games, projected gradient algorithms, machine learning, and numerous new exercises at the end of each chapter.
An Introduction to Optimization includes information on: - The mathematical definitions, notations, and relations from linear algebra, geometry, and calculus used in optimization
- Optimization algorithms, covering one-dimensional search, randomized search, and gradient, Newton, conjugate direction, and quasi-Newton methods
- Linear programming methods, covering the simplex algorithm, interior point methods, and duality - Nonlinear constrained optimization, covering theory and algorithms, convex optimization, and Lagrangian duality
- Applications of optimization in machine learning, including neural network training, classification, stochastic gradient descent, linear regression, logistic regression, support vector machines, and clustering.

An Introduction to Optimization is an ideal textbook for a one- or two-semester senior undergraduate or beginning graduate course in optimization theory and methods. The text is also of value for researchers and professionals in mathematics, operations research, electrical engineering, economics, statistics, and business.
Chong / Zak / Lu An Introduction to Optimization jetzt bestellen!

Weitere Infos & Material


Preface xv

About the Companion Website xviii

Part I Mathematical Review 1

1 Methods of Proof and Some Notation 3

1.1 Methods of Proof 3

1.2 Notation 5

Exercises 5

2 Vector Spaces and Matrices 7

2.1 Vector and Matrix 7

2.2 Rank of a Matrix 11

2.3 Linear Equations 16

2.4 Inner Products and Norms 18

Exercises 20

3 Transformations 23

3.1 Linear Transformations 23

3.2 Eigenvalues and Eigenvectors 24

3.3 Orthogonal Projections 26

3.4 Quadratic Forms 27

3.5 Matrix Norms 32

Exercises 35

4 Concepts from Geometry 39

4.1 Line Segments 39

4.2 Hyperplanes and Linear Varieties 39

4.3 Convex Sets 41

4.4 Neighborhoods 43

4.5 Polytopes and Polyhedra 44

Exercises 45

5 Elements of Calculus 47

5.1 Sequences and Limits 47

5.2 Differentiability 52

5.3 The Derivative Matrix 54

5.4 Differentiation Rules 57

5.5 Level Sets and Gradients 58

5.6 Taylor Series 61

Exercises 65

Part II Unconstrained Optimization 67

6 Basics of Set-Constrained and Unconstrained Optimization 69

6.1 Introduction 69

6.2 Conditions for Local Minimizers 70

Exercises 78

7 One-Dimensional Search Methods 87

7.1 Introduction 87

7.2 Golden Section Search 87

7.3 Fibonacci Method 91

7.4 Bisection Method 97

7.5 Newton’s Method 98

7.6 Secant Method 101

7.7 Bracketing 103

7.8 Line Search in Multidimensional Optimization 103

Exercises 105

8 Gradient Methods 109

8.1 Introduction 109

8.2 Steepest Descent Method 110

8.3 Analysis of Gradient Methods 117

Exercises 126

9 Newton’s Method 133

9.1 Introduction 133

9.2 Analysis of Newton’s Method 135

9.3 Levenberg–Marquardt Modification 138

9.4 Newton’s Method for Nonlinear Least Squares 139

Exercises 142

10 Conjugate Direction Methods 145

10.1 Introduction 145

10.2 Conjugate Direction Algorithm 146

10.2.1 Basic Conjugate Direction Algorithm 146

10.3 Conjugate Gradient Algorithm 151

10.4 Conjugate Gradient Algorithm for Nonquadratic Problems 154

Exercises 156

11 Quasi-Newton Methods 159

11.1 Introduction 159

11.2 Approximating the Inverse Hessian 160

11.3 Rank One Correction Formula 162

11.4 DFP Algorithm 166

11.5 BFGS Algorithm 170

Exercises 173

12 Solving Linear Equations 179

12.1 Least-Squares Analysis 179

12.2 Recursive Least-Squares Algorithm 187

12.3 Solution to a Linear Equation with Minimum Norm 190

12.4 Kaczmarz’s Algorithm 191

12.5 Solving Linear Equations in General 194

Exercises 201

13 Unconstrained Optimization and Neural Networks 209

13.1 Introduction 209

13.2 Single-Neuron Training 211

13.3 Backpropagation Algorithm 213

Exercises 222

14 Global Search Algorithms 225

14.1 Introduction 225

14.2 Nelder–Mead Simplex Algorithm 225

14.3 Simulated Annealing 229

14.3.1 Randomized Search 229

14.3.2 Simulated Annealing Algorithm 229

14.4 Particle Swarm Optimization 231

14.4.1 Basic PSO Algorithm 232

14.4.2 Variations 233

14.5 Genetic Algorithms 233

14.5.1 Basic Description 233

14.5.1.1 Chromosomes and Representation Schemes 234

14.5.1.2 Selection and Evolution 234

14.5.2 Analysis of Genetic Algorithms 238

14.5.3 Real-Number Genetic Algorithms 243

Exercises 244

Part III Linear Programming 247

15 Introduction to Linear Programming 249

15.1 Brief History of Linear Programming 249

15.2 Simple Examples of Linear Programs 250

15.3 Two-Dimensional Linear Programs 256

15.4 Convex Polyhedra and Linear Programming 258

15.5 Standard Form Linear Programs 260

15.6 Basic Solutions 264

15.7 Properties of Basic Solutions 267

15.8 Geometric View of Linear Programs 269

Exercises 273

16 Simplex Method 277

16.1 Solving Linear Equations Using Row Operations 277

16.2 The Canonical Augmented Matrix 283

16.3 Updating the Augmented Matrix 284

16.4 The Simplex Algorithm 285

16.5 Matrix Form of the Simplex Method 291

16.6 Two-Phase Simplex Method 294

16.7 Revised Simplex Method 297

Exercises 301

17 Duality 309

17.1 Dual Linear Programs 309

17.2 Properties of Dual Problems 316

17.3 Matrix Games 321

Exercises 324

18 Nonsimplex Methods 331

18.1 Introduction 331

18.2 Khachiyan’s Method 332

18.3 Affine Scaling Method 334

18.3.1 Basic Algorithm 334

18.3.2 Two-Phase Method 337

18.4 Karmarkar’s Method 339

18.4.1 Basic Ideas 339

18.4.2 Karmarkar’s Canonical Form 339

18.4.3 Karmarkar’s Restricted Problem 341

18.4.4 From General Form to Karmarkar’s Canonical Form 342

18.4.5 The Algorithm 345

Exercises 349

19 Integer Linear Programming 351

19.1 Introduction 351

19.2 Unimodular Matrices 351

19.3 The Gomory Cutting-Plane Method 358

Exercises 366

Part IV Nonlinear Constrained Optimization 369

20 Problems with Equality Constraints 371

20.1 Introduction 371

20.2 Problem Formulation 373

20.3 Tangent and Normal Spaces 374

20.4 Lagrange Condition 379

20.5 Second-Order Conditions 387

20.6 Minimizing Quadratics Subject to Linear Constraints 390

Exercises 394

21 Problems with Inequality Constraints 399

21.1 Karush–Kuhn–Tucker Condition 399

21.2 Second-Order Conditions 406

Exercises 410

22 Convex Optimization Problems 417


Edwin K. P. Chong, PhD, is Professor and Head of Electrical and Computer Engineering and Professor of Mathematics at Colorado State University. He is a Fellow of the IEEE and AAAS, and was Senior Editor of the IEEE Transactions on Automatic Control.
Wu-Sheng Lu, PhD, is Professor Emeritus of Electrical and Computer Engineering at the University of Victoria, Canada. He is a Fellow of the IEEE and former Associate Editor of the IEEE Transactions on Circuits and??Systems.
Stanislaw H. Zak, PhD, is Professor in the School of Electrical and Computer Engineering at Purdue University. He is former Associate Editor of Dynamics and Control and the IEEE Transactions on Neural Networks.


Ihre Fragen, Wünsche oder Anmerkungen
Vorname*
Nachname*
Ihre E-Mail-Adresse*
Kundennr.
Ihre Nachricht*
Lediglich mit * gekennzeichnete Felder sind Pflichtfelder.
Wenn Sie die im Kontaktformular eingegebenen Daten durch Klick auf den nachfolgenden Button übersenden, erklären Sie sich damit einverstanden, dass wir Ihr Angaben für die Beantwortung Ihrer Anfrage verwenden. Selbstverständlich werden Ihre Daten vertraulich behandelt und nicht an Dritte weitergegeben. Sie können der Verwendung Ihrer Daten jederzeit widersprechen. Das Datenhandling bei Sack Fachmedien erklären wir Ihnen in unserer Datenschutzerklärung.