Notice bibliographique

  • Notice

Type(s) de contenu et mode(s) de consultation : Texte : électronique

Auteur(s) : Nash, John C. (1947-....)  Voir les notices liées en tant qu'auteur

Titre(s) : Nonlinear parameter optimization using R tools [Texte électronique] / John C. Nash, Telfer School of Management, University of Ottawa

Publication : . - Chichester, West Sussex, United Kingdom : John Wiley & Sons, Inc., 2014

Description matérielle : 1 online resource

Note(s) : Includes bibliographical references and index
. - Print version record and CIP data provided by publisher.
The aim of this book is to provide an appreciation of the R tools available for optimization problems. Most users of R are not specialists in computation and the workings of the specialized tools are a black box. This can lead to mis-application, therefore users need help in making appropriate choices. This book looks at the principal tools available for users of the R statistical computing system for function minimization, optimization, and nonlinear parameter determination, featuring numerous examples throughout

Sujet(s) : R  Voir les notices liées en tant que sujet
Optimisation mathématique  Voir les notices liées en tant que sujet
Théories non linéaires  Voir les notices liées en tant que sujet

Numéros : ISBN 9781118884003

Notice n° :  FRBNF44655520 (notice reprise d'un réservoir extérieur)

Table des matières : Cover; Title Page; Copyright; Contents; Preface; Chapter 1 Optimization problem tasks and how they arise; 1.1 The general optimization problem; 1.2 Why the general problem is generally uninteresting; 1.3 (Non- )Linearity; 1.4 Objective function properties; 1.4.1 Sums of squares; 1.4.2 Minimax approximation; 1.4.3 Problems with multiple minima; 1.4.4 Objectives that can only be imprecisely computed; 1.5 Constraint types; 1.6 Solving sets of equations; 1.7 Conditions for optimality; 1.8 Other classifications; References; Chapter 2 Optimization algorithms-an overview.
2.1 Methods that use the gradient2.2 Newton-like methods; 2.3 The promise of Newton's method; 2.4 Caution: convergence versus termination; 2.5 Difficulties with Newton's method; 2.6 Least squares: Gauss-Newton methods; 2.7 Quasi-Newton or variable metric method; 2.8 Conjugate gradient and related methods; 2.9 Other gradient methods; 2.10 Derivative-free methods; 2.10.1 Numerical approximation of gradients; 2.10.2 Approximate and descend; 2.10.3 Heuristic search; 2.11 Stochastic methods; 2.12 Constraint-based methods-mathematical programming; References.
Chapter 3 Software structure and interfaces3.1 Perspective; 3.2 Issues of choice; 3.3 Software issues; 3.4 Specifying the objective and constraints to the optimizer; 3.5 Communicating exogenous data to problem definition functions; 3.5.1 Use of "global'' data and variables; 3.6 Masked (temporarily fixed) optimization parameters; 3.7 Dealing with inadmissible results; 3.8 Providing derivatives for functions; 3.9 Derivative approximations when there are constraints; 3.10 Scaling of parameters and function; 3.11 Normal ending of computations; 3.12 Termination tests-abnormal ending.
3.13 Output to monitor progress of calculations3.14 Output of the optimization results; 3.15 Controls for the optimizer; 3.16 Default control settings; 3.17 Measuring performance; 3.18 The optimization interface; References; Chapter 4 One-parameter root-finding problems; 4.1 Roots; 4.2 Equations in one variable; 4.3 Some examples; 4.3.1 Exponentially speaking; 4.3.2 A normal concern; 4.3.3 Little Polly Nomial; 4.3.4 A hypothequial question; 4.4 Approaches to solving 1D root-finding problems; 4.5 What can go wrong?; 4.6 Being a smart user of root-finding programs.
4.7 Conclusions and extensionsReferences; Chapter 5 One-parameter minimization problems; 5.1 The optimize() function; 5.2 Using a root-finder; 5.3 But where is the minimum?; 5.4 Ideas for 1D minimizers; 5.5 The line-search subproblem; References; Chapter 6 Nonlinear least squares; 6.1 nls() from package stats; 6.1.1 A simple example; 6.1.2 Regression versus least squares; 6.2 A more difficult case; 6.3 The structure of the nls() solution; 6.4 Concerns with nls(); 6.4.1 Small residuals; 6.4.2 Robustness-"singular gradient'' woes; 6.4.3 Bounds with nls().

Localiser ce document(1 Exemplaire)

Document numérique : 

1 partie d'exemplaire regroupée

support : document électronique dématérialisé