Atjaunināt sīkdatņu piekrišanu

Mathematical Foundations of Large Language Models [Hardback]

  • Formāts: Hardback, height x width: 235x155 mm, 25 Illustrations, black and white
  • Izdošanas datums: 19-Jun-2026
  • Izdevniecība: Springer Nature
  • ISBN-10: 9819204658
  • ISBN-13: 9789819204656
  • Hardback
  • Cena: 116,69 €*
  • * ši ir gala cena, t.i., netiek piemērotas nekādas papildus atlaides
  • Standarta cena: 155,59 €
  • Ietaupiet 25%
  • Grāmata tiks piegādāta 3-6 nedēļas pēc tās publicēšanas.
  • Daudzums:
  • Ielikt grozā
  • Piegādes laiks - 2-4 nedēļas
  • Pievienot vēlmju sarakstam
  • Formāts: Hardback, height x width: 235x155 mm, 25 Illustrations, black and white
  • Izdošanas datums: 19-Jun-2026
  • Izdevniecība: Springer Nature
  • ISBN-10: 9819204658
  • ISBN-13: 9789819204656
This book emerged from a simple observation: while Large Language Models (LLMs) have become ubiquitous, their mathematical foundations remain opaque to many practitioners, students, and enthusiasts. This book bridges that gap by presenting LLMs from first principles not as a black box, but as an elegant mathematical construction. The journey begins with the simplest language models bigrams and n-grams and progressively builds toward the transformer architecture that powers modern LLMs. The emphasize is on clarity over completeness, intuition over implementation details, and mathematical rigor over hand-waving explanations. This book is self-contained and only assumes familiarity with undergraduate-level linear algebra, probability, and calculus. Wherever possible, the abstract concepts are connected to concrete examples, often using minimal two word vocabularies to illuminate general principles. The hope is that this primer serves as both an introduction for newcomers and a reference for practitioners seeking deeper understanding of the mathematical machinery underlying todays most influential AI systems.
Preamble.- ForegroundingLargeLanguageModels.- Methodspre-DatingLLM.
Tokens and the Sentence Matrix.- Transformer and the Attention Matrix.- The
Multi-Head Attention Matrix.- Computational Complexity of Attention.-
Conditional Probability in LLM.- Softmax.- Training Large Language Models.
Prof. Belal Ehsan Baaquie holds a B.S. in Physics from Caltech and a Ph.D. in Theoretical Physics from Cornell University, USA. His main research interest is in the study and application of mathematical methods from quantum field theory. His research includes the theoretical foundations of quantum  mechanics and  quantum computers. He has applied the mathematical formalism of field theory to finance and been a major contributor to the emerging field of quantum finance. His current focus is on developing the formalism of quantum finance and applying it to option pricing, corporate coupon bonds, and the theory of interest rates, as well as the study of equity, foreign exchange, and commodities. He is also applying methodologies from statistical mechanics and quantum field theory to the study of microeconomics and macroeconomics.