Muutke küpsiste eelistusi

Databases on Modern Hardware: How to Stop Underutilization and Love Multicores [Pehme köide]

  • Formaat: Paperback / softback, 113 pages, kõrgus x laius: 235x191 mm, kaal: 235 g
  • Sari: Synthesis Lectures on Data Management
  • Ilmumisaeg: 14-Aug-2017
  • Kirjastus: Morgan & Claypool Publishers
  • ISBN-10: 1681731533
  • ISBN-13: 9781681731537
Teised raamatud teemal:
  • Formaat: Paperback / softback, 113 pages, kõrgus x laius: 235x191 mm, kaal: 235 g
  • Sari: Synthesis Lectures on Data Management
  • Ilmumisaeg: 14-Aug-2017
  • Kirjastus: Morgan & Claypool Publishers
  • ISBN-10: 1681731533
  • ISBN-13: 9781681731537
Teised raamatud teemal:
Data management systems enable various influential applications from high-performance online services (e.g., social networks like Twitter and Facebook or financial markets) to big data analytics (e.g., scientific exploration, sensor networks, business intelligence). As a result, data management systems have been one of the main drivers for innovations in the database and computer architecture communities for several decades. Recent hardware trends require software to take advantage of the abundant parallelism existing in modern and future hardware. The traditional design of the data management systems, however, faces inherent scalability problems due to its tightly coupled components. In addition, it cannot exploit the full capability of the aggressive micro-architectural features of modern processors. As a result, today's most commonly used server types remain largely underutilized leading to a huge waste of hardware resources and energy.

In this book, we shed light on the challenges present while running DBMS on modern multicore hardware. We divide the material into two dimensions of scalability: implicit/vertical and explicit/horizontal.

The first part of the book focuses on the vertical dimension: it describes the instruction- and data-level parallelism opportunities in a core coming from the hardware and software side. In addition, it examines the sources of under-utilization in a modern processor and presents insights and hardware/software techniques to better exploit the microarchitectural resources of a processor by improving cache locality at the right level of the memory hierarchy.

The second part focuses on the horizontal dimension, i.e., scalability bottlenecks of database applications at the level of multicore and multisocket multicore architectures. It first presents a systematic way of eliminating such bottlenecks in online transaction processing workloads, which is based on minimizing unbounded communication, and shows several techniques that minimize bottlenecks in major components of database management systems. Then, it demonstrates the data and work sharing opportunities for analytical workloads, and reviews advanced scheduling mechanisms that are aware of nonuniform memory accesses and alleviate bandwidth saturation.
1 Introduction
1(6)
1.1 Implicit/Vertical Dimension
1(2)
1.2 Explicit/Horizontal Dimension
3(1)
1.3 Structure of the Book
4(3)
PART I Implicit/Vertical Scalability
7(36)
2 Exploiting Resources of a Processor Core
9(18)
2.1 Instruction and Data Parallelism
9(5)
2.2 Multithreading
14(3)
2.3 Horizontal Parallelism
17(10)
2.3.1 Horizontal parallelism in advanced database scenarios
19(6)
2.3.2 Conclusions
25(2)
3 Minimizing Memory Stalls
27(16)
3.1 Workload Characterization for Typical Data Management Workloads
27(3)
3.2 Roadmap for this
Chapter
30(1)
3.3 Prefetching
31(3)
3.3.1 Techniques that are Common in Modern Hardware
31(1)
3.3.2 Temporal Streaming
32(1)
3.3.3 Software-guided Prefetching
33(1)
3.4 Being Cache-conscious while Writing Software
34(5)
3.4.1 Code Optimizations
35(1)
3.4.2 Data Layouts
36(1)
3.4.3 Changing Execution Models
37(2)
3.5 Exploiting Common Instructions
39(1)
3.6 Conclusions
40(3)
PART II Explicit/Horizontal Scalability
43(32)
4 Scaling-up OLTP
45(12)
4.1 Focus on Unscalable Components
48(7)
4.1.1 Locking
48(3)
4.1.2 Latching
51(1)
4.1.3 Logging
52(2)
4.1.4 Synchronization
54(1)
4.2 Non-uniform Communication
55(1)
4.3 Conclusions
56(1)
5 Scaling-up OLAP Workloads
57(18)
5.1 Sharing Across Concurrent Queries
58(5)
5.1.1 Reactive Sharing
59(1)
5.1.2 Proactive Sharing
60(2)
5.1.3 Systems with Sharing Techniques
62(1)
5.2 NUMA-awareness
63(10)
5.2.1 Analytical Operators
65(3)
5.2.2 Task Scheduling
68(3)
5.2.3 Coordinated Data Placement and Task Scheduling
71(2)
5.3 Conclusions
73(2)
PART III Conclusions
75(10)
6 Outlook
77(6)
6.1 Dark Silicon and Hardware Specialization
77(1)
6.2 Non-volatile RAM and Durability
78(1)
6.3 Hardware Transactional Memory
79(1)
6.4 Task Scheduling for Mixed Workloads and Energy Efficiency
80(3)
7 Summary
83(2)
Bibliography 85(14)
Authors' Biographies 99