Muutke küpsiste eelistusi

Linear Dimensionality Reduction [Pehme köide]

  • Formaat: Paperback / softback, 130 pages, kõrgus x laius: 235x155 mm, 3 Illustrations, color; 21 Illustrations, black and white; X, 130 p. 24 illus., 3 illus. in color., 1 Paperback / softback
  • Sari: Lecture Notes in Statistics 228
  • Ilmumisaeg: 26-Sep-2025
  • Kirjastus: Springer International Publishing AG
  • ISBN-10: 3031957849
  • ISBN-13: 9783031957840
Teised raamatud teemal:
  • Pehme köide
  • Hind: 76,49 €*
  • * hind on lõplik, st. muud allahindlused enam ei rakendu
  • Tavahind: 89,99 €
  • Säästad 15%
  • See raamat ei ole veel ilmunud. Raamatu kohalejõudmiseks kulub orienteeruvalt 2-4 nädalat peale raamatu väljaandmist.
  • Kogus:
  • Lisa ostukorvi
  • Tasuta tarne
  • Tellimisaeg 2-4 nädalat
  • Lisa soovinimekirja
  • Formaat: Paperback / softback, 130 pages, kõrgus x laius: 235x155 mm, 3 Illustrations, color; 21 Illustrations, black and white; X, 130 p. 24 illus., 3 illus. in color., 1 Paperback / softback
  • Sari: Lecture Notes in Statistics 228
  • Ilmumisaeg: 26-Sep-2025
  • Kirjastus: Springer International Publishing AG
  • ISBN-10: 3031957849
  • ISBN-13: 9783031957840
Teised raamatud teemal:

This book provides an overview of some classical linear methods in Multivariate Data Analysis. This is an old domain, well established since the 1960s, and refreshed timely as a key step in statistical learning. It can be presented as part of statistical learning, or as dimensionality reduction with a geometric flavor. Both approaches are tightly linked: it is easier to learn patterns from data in low-dimensional spaces than in high-dimensional ones. It is shown how a diversity of methods and tools boil down to a single core method, PCA with SVD, so that the efforts to optimize codes for analyzing massive data sets like distributed memory and task-based programming, or to improve the efficiency of algorithms like Randomized SVD, can focus on this shared core method, and benefit all methods.

This book is aimed at graduate students and researchers working on massive data who have encountered the usefulness of linear dimensionality reduction and are looking for a recipe to implement it. It has been written according to the view that the best guarantee of a proper understanding and use of a method is to study in detail the calculations involved in implementing it. With an emphasis on the numerical processing of massive data, it covers the main methods of dimensionality reduction, from linear algebra foundations to implementing the calculations. The basic requisite elements of linear and multilinear algebra, statistics and random algorithms are presented in the appendix.

-
1. Introduction.-
2. Principal Component Analysis (PCA).-
3.
Complements on PCA.-
4. PCA with Metrics on Rows and Columns.-
5.
Correspondence Analysis.-
6. PCA with Instrumental Variables.-
7. Canonical
Correlation Analysis.-
8. Multiple Canonical Correlation Analysis.-
9.
Multidimensional Scaling.
Alain Franc is a senior researcher at INRAE (National Research Institute for Agriculture, Food and the Environment) and INRIA (National Institute for Research in Digital Science and Technology). He works on dimension reduction and statistical modelling with applications to the discovery of patterns in biodiversity. His focus is on the development of methods for handling massive data sets, which is a challenge for high-performance computing.