Download Extreme Performance - Exadata

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Entity–attribute–value model wikipedia , lookup

SQL wikipedia , lookup

Concurrency control wikipedia , lookup

Open Database Connectivity wikipedia , lookup

Extensible Storage Engine wikipedia , lookup

Microsoft Jet Database Engine wikipedia , lookup

PL/SQL wikipedia , lookup

Database wikipedia , lookup

Microsoft SQL Server wikipedia , lookup

Relational model wikipedia , lookup

Functional Database Model wikipedia , lookup

Database model wikipedia , lookup

Clusterpoint wikipedia , lookup

Oracle Database wikipedia , lookup

Transcript
<Insert Picture Here>
Extreme Performance…
Thomas Kyte
http://asktom.oracle.com/
The Beginning...
•
•
•
•
Data Model with Structure
Data Independent of Code
Set-oriented
1977 the work begins
GPS
1978
GPS
1978
First RDBMS: Version 2
June 1979
• FIRST Commercial SQL RDBMS
• Impressive First SQL
• Joins, Subqueries
• Outer Joins, Connect By
• A Simple Server
• No transactions, ‘Limited’ Reliability
• Portability from the Start
• Written in Fortran
• But multi-platform – PDP11, Dec VAX
IBM PC – 1981
IBM model
number
5150,
introduced
on August
12, 1981.
Internet (as we know it) –
1983
The first TCP/IPbased wide-area
network was
operational by
January 1, 1983
when all hosts on
the ARPANET were
switched over from
the older NCP
protocols.
Portability: Version 3
March 1983
• New Implementation Designed for Portability
• Written in ‘C’
• Single Source
• Architectural Changes
• Transactions, multi-versioning, no read consistency
• AI/BI files
• Oracle Corporation – name established
Cooperative Server: Version 5
April 1985
• My First Oracle Experience
• 1st Client/Server
• Cooperative Server
• Distributed Processing
• Parallel Server
• Portability
• V5 was first to go beyond 640K memory on PCs
• Single-user for Macintosh o/s
• SQL_TRACE
• select trace('sql',1),1 from dual;
Transaction Processing: Version
July 1988
6
• New Architecture
•
•
•
•
Performance (first SMP)
Availability
TPO
PL/SQL
• V6 Lays Architectural Groundwork for the Future
• This was a rewrite of the entire database fundamentally
World Wide Web – 1990’ish
The World Wide Web
was created in 1989 by
British scientist Tim
Berners-Lee, working at
the European
Organization for
Nuclear Research
(CERN) in Geneva,
Switzerland, and
released in 1992.
Oracle7.3
February 1996
• Partitioned Views
• Bitmapped Indexes
• Asynchronous read ahead for
table scans
• Standby Database
• Deferred transaction recovery
on instance startup
• Updatable Join View
• SQLDBA no longer shipped.
• Index rebuilds
• DBV introduced
• Context Option
• PL/SQL - UTL_FILE
 Spatial Data Option
 Tablespaces changes - Coalesce,
Temporary Permanent,
 Trigger compilation, debug
 Unlimited extents on STORAGE
clause.
 Some init.ora parameters
modifiable - TIMED_STATISTICS
 HASH Joins, Antijoins
 Histograms
 Oracle Trace
 Advanced Replication Object
Groups
Data Warehouses Growing Rapidly
Tripling In Size Every Two Years
Size of the Largest Data Warehouses
Terabytes of Data
1000
800
600
Actual
Projected
400
200
1998 2000 2002 2004 2006 2008 2010 2012
Source: Winter TopTen Survey, Winter Corporation, Waltham MA, 2008.
<Insert Picture Here>
Extreme Performance
The Performance Challenge
Storage Data Bandwidth Bottleneck
• Current warehouse deployments often have bottlenecks limiting the
movement of data from disks to servers
• Storage Array internal bottlenecks on processors and Fibre Channel Loops
• Limited Fibre Channel host bus adapters in servers
• Under configured and complex SANs
• Pipes between disks and servers are 10x to 100x too slow for data size
Data Warehouses Start Slowdown at 1TB
10 Hours
Table
Scan
Time
5 Hours
Typical
NAS
Typical
Mid-Range
Array
Typical
High-End
Array
1 Hour
Table Size
1TB
10 TB
100TB
Solutions To Data Bandwidth Bottleneck
• Add more pipes – Massively parallel architecture
• Make the pipes wider – 5X faster than conventional storage
• Ship less data through the pipes – Process data in storage
What We Announced at OpenWorld
Extreme Performance | Unlimited Scalability | Enterprise Ready
HP Oracle Database Machine:
The next step in DW Hardware Solutions
Custom
Reference
Configurations
• Complete Flexibility
• Any OS, any platform
• Easy fit into a
company’s IT
standards
• Documented bestpractice
configurations for
data warehousing
Optimized
Warehouse
HP Oracle
Database
Machine
• Scalable systems
pre-installed and preconfigured: ready to
run out-of-the-box
• Highest performance
• Pre-installed and preconfigured
• Sold by Oracle
Products Announced
HP Exadata Storage Server Hardware
• Paired with Oracle Exadata Storage Server
Software
• Delivers database intelligence in storage tier
• Supported for Oracle Enterprise and Red
Hat servers running Oracle Database 11g
Enterprise Edition*
HP Oracle Database Machine
• Simplicity of appliance seamlessly
integrated with the database
• Eliminates all bottlenecks preventing high
performance data scans
• Includes Exadata Storage Server
Dramatic performance improvement for data warehouses
* Linux 5.1 releases with appropriate Infiniband drivers,
Oracle Database 11g Enterprise Edition vers. 11.1.0.7
First
HP Oracle Database Machine
The HP Oracle Database Machine
Sales / Support Model
• System Delivery
• Hardware Service
• System Sales
• System Support
Oracle Technology Sales Manager is the single point for sales
Oracle is single point for Support
Second
HP Exadata Storage Server Hardware
• 2 Intel processors, 8 cores
• 12 disk drives, up to 12 TB raw storage
• 2 Infiniband connections
• Oracle Enterprise Linux OS
Oracle Exadata Storage Server Software
Reduces Data Going through the Pipes
• Intelligent storage server
• Unique ‘smart scan’ technology
• Returns query result set
• Not disk blocks
Traditional Scan Processing

SELECT
customer_name
FROM calls
WHERE amount >
200;

Table
Extents
Identified

I/Os Issued
• With traditional storage, all
database intelligence
resides in the database
hosts
• Very large percentage of

data returned from storage
DB Host reduces
terabyte of data to 1000 is discarded by database
customer names that
servers
are returned to client
• Discarded data consumes
valuable resources, and
impacts the performance of
other workloads


Rows Returned
I/Os Executed:
1 terabyte of data
returned to hosts
Exadata Smart Scan Processing

SELECT
customer_name
FROM calls
WHERE amount >
200;

Rows Returned

Smart Scan
Constructed And
Sent To Cells

Consolidated
Result Set
Built From All
Cells

Smart Scan
identifies rows and
columns within
terabyte table that
match request

2MB of data
returned to server
• Only the relevant columns
• customer_name
and required rows
• where amount>200
are are returned to hosts
• CPU consumed by predicate
evaluation is offloaded
• Moving scan processing off the
database host frees host CPU
cycles and eliminates massive
amounts of unproductive
messaging
• Returns the needle, not the entire
hay stack
Additional Smart Scan Functionality
• Join filtering
• Star join filtering is performed within Exadata storage cells
• Dimension table predicates are transformed into filters that are
applied to scan of fact table
• Backups
• I/O for incremental backups is much more efficient since only
changed blocks are returned
• Create Tablespace (file creation)
• Formatting of tablespace extents eliminates the I/O associated with
the creation and writing of tablespace blocks
Smart Scan Transparency
• Smart scans are transparent to the application
• No application or SQL changes required
• Returned data is fully consistent and transactional
• If a cell dies during a smart scan, the uncompleted portions
of the smart scan are transparently routed to another cell
• Smart Scans correctly handle complex cases including
•
•
•
•
•
•
•
Uncommitted data and locked rows
Chained rows
Compressed tables
National Language Processing
Date arithmetic
Regular expression searches
Partitioned tables
High Throughput, Reduced Overhead, No Complex Tuning
Massively Parallel Storage Grid
• Exadata Storage servers are organized into a
massively parallel storage grid
• Scalable
• Scales to hundreds of storage servers
• Data automatically distributed across storage
servers by ASM
• Transparently redistributed when storage
servers are added or removed
• Data bandwidth scales linearly with capacity
16 GB/sec
8 GB/sec
…
• Available
4 GB/sec
• Data is mirrored across storage servers
• Failure of disk or storage server transparently
tolerated
• Simple
Exadata bandwidth scales
• Works transparently - no application changes
linearly with capacity
Exadata Performance Scales
10 Hour
• Exadata delivers brawny
hardware for use by Oracle’s
brainy software
Table Scan Time
Typical
Warehouse
• Performance scales with size
5 Hour
• Result
• More business insight
• Better decisions
• Improved competitiveness
1 Hour
Exadata
1TB
10 TB
100TB
Table Size
Exadata Co-Existence and Migration
• Databases can be concurrently
deployed on Exadata and
traditional storage
• Tablespaces can exist on Exadata
storage, traditional storage, or a
combination of the two, and is
transparent to database applications
• SQL offload processing requires all
pieces of a tablespace reside on
Exadata
• Online migration if currently using
ASM and ASM redundancy
• Migration can be done using
RMAN or Data Guard
Database
Server
Exadata
Non-Exadata
Online Migration
M-Tel Exadata Speedup
10X to 72X Performance Improvement
28x
Average
Tablespace Creation
Speedup
Index Creation
Handset to Customer Mapping Report
CRM Customer Discount Report
CRM Service Order Report
Warehouse Inventory Report
CDR Full Table Scan
0
10
20
30
40
50
60
70
80
Plamen Zyumbyulev
Head of Database Administration
M-Tel
“Every query was faster on Exadata compared to our
current systems. The smallest performance improvement
was 10x and the biggest one was an incredible 72x.”
Grant Salmon
Chief Executive Officer
LGR Telecommunications
“Call Data Record queries that used to run for over 30
minutes now complete in under 1 minute.
That's extreme performance.”
Giant Eagle Exadata Speedup
3X to 48X Performance Improvement
Merchandising Level 1 Detail:
Period Ago
Merchandising Level 1 Detail:
Current - 52 weeks
Supply Chain Vendor - Year - Item
Movement
Merchandising Level 1 Detail by
Week
Materialized Views Rebuild
Date to Date Movement
Comparison - 53 weeks
16x
Prompt04 Clone for ACL audit
Average
Sales and Customer Counts
Speedup
Gift Card Activations
Recall Query
-
5.0
10.0
15.0
20.0
25.0
30.0
35.0
40.0
45.0
50.0
Walt Litzenberger
Director Enterprise Database Systems
The CME Group
“Oracle Exadata outperforms anything we’ve
tested to date by 10 to 15 times.
This product flat-out screams.”
Where Can You Try It?
• North America Enterprise Technology Centers (ETC)
• Atlanta, GA
• Two HP Oracle Database Machines, each with 8 database server
nodes and 14 HP Exadata Storage Server cells holding 12 drives
of 1 TB in size each
• Capacity of about 46 TB of uncompressed data / rack
• Reston, VA
• Three HP Oracle Database Machines, each with 8 database
server nodes and 14 HP Exadata Storage Server cells holding 12
drives of 300 GB in size each
• Capacity of about 14 TB of uncompressed data / rack
A queue has already formed!
Exadata Benefits
• Extreme Performance
• 10X and more speedup for data warehousing
• Database Aware Storage
• Smart Scans
• Massively Parallel Architecture
• Dynamically Scalable to hundreds of cells
• Linear Scaling of Data Bandwidth
• Transaction/Job level Quality of Service
• Mission Critical Availability and Protection
• Disaster recovery, backup, point-in-time recovery, data validation,
encryption
Resources
• Oracle.com:
http://www.oracle.com/exadata
• Oracle Exadata Technology Portal on OTN:
http://www.oracle.com/technology/products/bi/db/exadata
• Oracle Exadata white papers:
http://www.oracle.com/technology/products/bi/db/exadata/pdf/exad
ata-technical-whitepaper.pdf
http://www.oracle.com/technology/products/bi/db/exadata/pdf/migr
ation-to-exadata-whitepaper.pdf