0% found this document useful (0 votes)
34 views22 pages

Scalable Storage Configuration For The Physics Database Services

This document summarizes a presentation about scalable storage configuration for physics database services. It discusses challenges in storage configuration, best practices like RAID 1 for high availability and RAID 0 for performance, and Oracle's Automatic Storage Management (ASM) which provides striping and mirroring. It provides examples of ASM configurations and answers frequently asked questions about datafiles, tablespaces, and storage sizing based on performance measurements. Benchmark data is also presented to help size new projects and upgrades.

Uploaded by

Dev Raj
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
34 views22 pages

Scalable Storage Configuration For The Physics Database Services

This document summarizes a presentation about scalable storage configuration for physics database services. It discusses challenges in storage configuration, best practices like RAID 1 for high availability and RAID 0 for performance, and Oracle's Automatic Storage Management (ASM) which provides striping and mirroring. It provides examples of ASM configurations and answers frequently asked questions about datafiles, tablespaces, and storage sizing based on performance measurements. Benchmark data is also presented to help size new projects and upgrades.

Uploaded by

Dev Raj
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPT, PDF, TXT or read online on Scribd
You are on page 1/ 22

Scalable Storage Configuration for the

Physics Database Services

Luca Canali, CERN IT


LCG Database Deployment and
Persistency Workshop
October, 2005
Outline
• In this talk I will discuss
– Storage configuration for scalable database services
• Main challenges
• Best practices
– An implementation of scalable storage
• Impacts on DB logical to physical mapping
• Performance and resource allocations
– How we can help you to size new database projects or
to scale up existing applications
• Performance testing
• Benchmark data

Database Workshop, October 2005 L.Canali, CERN 2


Oracle Scalable Architecture
Goal: A database infrastructure that provides the required system
resources to the end-users and applications.
How: A modular architecture that can scale up to a large number of
components

Database Workshop, October 2005 L.Canali, CERN 3


RAID 1: HA Storage
• Mirroring
– 2-Way mirroring (RAID 1) protects against single point
of failures
– Can be used to redistribute I/O load (performance)

Database Workshop, October 2005 L.Canali, CERN 4


RAID 0: Scalable Performances
• RAID 0 (Striping) automatically redistributes files across
multiple disks.
• Performance and scalability are increased
• Error resiliency is decreased

Unstriped Disks Striped Disks

Database Workshop, October 2005 L.Canali, CERN 5


Mechanical and Geometrical constraints
• The external part of the disk provides
– More throughput
– Less latency

Database Workshop, October 2005 L.Canali, CERN 6


S.A.M.E Strategy
• Goal: optimize storage I/O utilization
• S.A.M.E. (Stripe And Mirror Everything) Strategy
– Built on the concepts of RAID 1 + 0
– Proposed by J. Loaiza (Oracle) in 1999
– Replaces “old recipes”: manual balancing across
volumes
• Need a Software or Hardware Volume Manager
– ASM is Oracle’s solution with 10g “S.A.M.E. out of the
box”
– Other solutions available from different vendors
require configuration

Database Workshop, October 2005 L.Canali, CERN 7


Storage Configuration Guidelines
• Use all available disk drives
• Place frequently used data at outer half of disk
– Fastest transfer rate
– Minimize seek time
• Stripe data at 1MB extents
– Distribute the workload across disks
– Eliminate hot spots
– Optimum sequential bandwidth gained with1MB I/O
• Stripe redo logs across multiple drives
– Maximize write throughput for small writes
– Smaller stripe size (128KB) and/or dedicated disks
• Use cache on the controller
– ‘Write-back’ cache
– Battery-backed cache

Database Workshop, October 2005 L.Canali, CERN 8


Oracle’s ASM Main Features
• Mirror protection:
– 2-way and 3-way mirroring available.
– Mirror on a per-file basis
– Can mirror across storage arrays
• Data striping across the volume:
– 1MB and 128KB stripes available
• Supports clustering and single instance
• Dynamic data distribution
– A solution to avoid ‘hot spots’
– On-line add/drop disk with minimal data relocation
– Automatic database file management
• Database File System with performance of RAW I/O

Database Workshop, October 2005 L.Canali, CERN 9


ASM’s Configuration – Examples
• ASM is a volume manager, its output are disk groups (DG) that Oracle
databases can mount to allocate their files

DATA-DG RECOVERY-DG

Config 1: Disk groups created Config 2: Disk groups created


with dedicated disks by ‘horizontal’ slicing

Database Workshop, October 2005 L.Canali, CERN 10


Proposed Storage Configuration
• Proposed storage configuration:
– High availability - Allows backups to disk
– High performance - Allows clusterware mirroring (10.2)
– DBs have dedicated resources

DB N.1 DB N.2
Oracle RAC
Nodes

Data DG-1 Recovery DG-2 Data DG-2 Recovery DG-1 Disk Groups
(ASM)

Storage
Arrays

Database Workshop, October 2005 L.Canali, CERN 11


FAQ 1: Datafiles
• Do I need to worry on the number and names of
the datafiles allocated for each tablespace?
• “Traditional” storage allocation across multiple volumes:
– Requires a careful allocation of multiple datafiles across logical
volumes and/or filesystems
– Datafile-to-filesystem and filesystem-to-physical storage mappings
have to be frequently tuned
• S.A.M.E. storage, such as Oracle ASM, provides balanced I/O access
across disks
– There is NO NEED, for performance reasons, to allocate multiple
datafiles per tablespace.
– 10g new feature “bigfile tablespace” allows for tablespaces with a
single datafile that can grow up to 32 TB (db_block_size=8k)

Database Workshop, October 2005 L.Canali, CERN 12


FAQ 2: Data and Index Tablespaces
• Do I need dedicated tablespaces for indexes and
tables?
• Separation of indexes and tables has often been advised to:
– Distribute I/O
– Reduce fragmentation
– Allow separate backup of tables and indexes
• S.A.M.E. storage, such as Oracle ASM, provides balanced I/O access
across disks
– No performance gains are expected by using dedicated
tablespaces for INDEXes and TABLEs.
• Additional Notes:
– Tablespaces fragmentation has little impact when using locally managed
TBSs and automatic segment space management (9i and 10g)
– Very large database can profit from using multiple tablespaces for admin
purposes and logical separation of objects

Database Workshop, October 2005 L.Canali, CERN 13


FAQ 3: Sizing Storage
• Storage sizing for database should not take size as the
first requirement
– Bandwidth and performance metrics are bound to the number of
disk spindles
– Magnetic HD technology has improved the GByte/$ ratio
– The rest of HD technology has not seen much improvements in
the last 5 years (since 15K rpms HDs)
• Sizing for storage requirements should
– Be based on stress test measurements
– Past performance measurements on comparable systems
– New projects can leverage benchmark data.
• Extra HD space is not wasted
– Can be used to strengthen the B&R policy with Disk Backups

Database Workshop, October 2005 L.Canali, CERN 14


IO Benchmark Measurements
• Benchmark data can be used for
– Sizing of new projects and upgrades
– Performance baseline, testing for new hardware
• The following metrics have been measured:
– Sequential throughput (full scans)
– Random access (indexed access)
– I/O per second (indexed access)
– Metrics are measured as a function of workload
• Other test details
– Benchmark tool: Oracle’s ORION
– Infortrend Storage array: 16 SATA x 400 GB disks, 1
controller and 1 GB cache

Database Workshop, October 2005 L.Canali, CERN 15


IO Benchmark Data
Sequential Workload Performance
(Read Only Workload)
Throughput:
+ 25%
160

140

120
Throughput (MBps)

100
12 LUNs, noraid
80
16 LUNs, no raid,
60 external partition

40

20

0
1 2 4 6 8 10 12 14 16 18 20 22 24 26 28 30 32
Load (N# Outstanding asynch I/Os)

Database Workshop, October 2005 L.Canali, CERN 16


IO Benchmark Data
Latency for Small Random I/O Latency:
(Read Only Workload)
- 35%
160

140

120
Latency (ms)

100

80

60
12 LUNs, noraid
40
16 LUNs, no raid,
external partition
20

0
1 2 3 4 8 12 16 20 24 28 32 36 40 44 48 52 56 60 64 68 72 76 80

Load (N# Outstanding asynch I/Os)

Database Workshop, October 2005 L.Canali, CERN 17


IO Benchmark Data
Small Random I/O Performance I/O /sec:
(Read Only Workload)
+60%
1000
900
800
700
I/O per second

600
500
400 12 LUNs, noraid
300 16 LUNs, no raid,
external partition
200
100
0
1 4 16 28 40 52 64 76
Load (N# Outstanding asynch I/Os)

Database Workshop, October 2005 L.Canali, CERN 18


Conclusions
• The Database Services for Physics can provide
– Scalable Database services
• Scalable on CPU and Memory resources
• Scalable on Storage resources
– Sizing for new projects or upgrades
• Stress/Performance testing
• Integration and Validation Testing
• Benchmark data for capacity planning

Database Workshop, October 2005 L.Canali, CERN 19


Additional Benchmark Data
Sequential Workload Performance
(Read Only Workload)

140

120
Throughput (MBps)

100

80
7 LUNs, raid0

60 14 LUNs, no raid

14 LUNs, no raid, outer


40 partition used
14 LUN, no raid, inner
partition used
20

0
1 2 4 6 8 10 12 14 16 18 20 22 24 26 28 30 32
Load (N# Outstanding asynch I/Os)

Database Workshop, October 2005 L.Canali, CERN 20


Additional Benchmark Data
Latency for Small Random I/O
(Read Only Workload)

140

120

100
Latency (ms)

80

60 7 LUNs, raid0

14 LUNs, no raid
40
14 LUNs, no raid, outer
20 partition used
14 LUNS, no raid, inner
partition used
0
1 2 3 4 8 12 16 20 24 28 32 36 40 44 48 52 56 60 64 68 72 76 80

Load (N# Outstanding asynch I/Os)

Database Workshop, October 2005 L.Canali, CERN 21


Additional Benchmark Data
Small Random I/O Performance
(Read Only Workload)

900

800

700

600
I/O per second

500
7 LUNs, raid0
400
14 LUNs, no raid
300
14 LUNs, noraid, outer
200 partition used
14 LUNS, noraid, inner
100 partition used

0
1 4 16 28 40 52 64 76
Load (N# Outstanding asynch I/Os)

Database Workshop, October 2005 L.Canali, CERN 22

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy