advertisement
Volume 3: Annex 2 - Data manipulation and management issues Deliverable 1
The hot backup without transaction load was performed with 35 IBM Magstar 3590 drives and was wall-clock measured at 985 GB/hour and 6% total system overhead.
TOTALS
Database
Elapsed Time
Sustained Xfer
Hot Backup
1.0265 TB Oracle7 database
64 minutes
1.3 TB/hour
Wall clock Xfer
985 GB/hour
Total System Overhead
6% (94% of the system was still available for other processes)
•
Hot Backup with Transaction Load at 901 GB/hour with 35 drives
The hot backup with transaction load was performed with 35 IBM Magstar 3590 drives and was wall-clock measured at 901 GB/hour and 21% total system overhead.
Scripts were used to generate a load of 75 transactions per second (4500 tpm).
TOTALS
Database
Elapsed Time
Sustained Xfer
Wall clock Xfer
Total System Overhead
Hot Backup with Transaction Load
1.0265 TB Oracle7 database
70 minutes
1.1 TB/hour
901 GB/hour
21% @ 4500 updates per minute load (79% of the system was still available for other processes)
B.3
Interpreting the Results
Demonstrating the ability to back up a true terabyte database in approximately one hour is an important landmark for company’s with VLDBs looking for a capable backup solution.
Why is one hour so important? Taking a VLDB off-line directly cuts into a 24x7 company’s bottom line. In a recent Information Week article, one of the largest banks in the country estimated the bank would lose close to $50 million for every 24 hours, or $2.08 million per hour, their system is down.1
Backup of terabyte databases requires certain features and functionality to take advantage of their sheer size:
•
Hot (Online) backup capability: As applications increasingly demand 24x7 uptime, "hot" online database backup will take precedence over "cold" offline database backup.
•
Availability: Database availability is key. The backup system should have minimal impact on availability, so that as the database grows the backup system will not interfere with its’ availability to the end-users.
•
Ability to Scale: the system must be designed so that it will not collapse under the weight of its own growth in 18 to 24 months. 10% growth of terabyte database will use considerably more resources than 10% growth of a 10 gigabyte database.
•
Support for Heterogeneous Environments: today’s terabyte UNIX sites consist of a wide variety of platforms, operating systems and tape libraries. The backup software should not limit your choices.
page 118 (120)
1999 EURESCOM Participants in Project P817-PF
advertisement
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Related manuals
advertisement
Table of contents
- 15 Part 1 Transaction Processing Monitors
- 15 1 Introduction
- 15 2 Concepts of Transactions
- 15 2.1 ACID Properties
- 15 2.2 Two Phase Commit Protocol
- 16 3 Concepts of TP Monitors
- 16 3.1 Why should you use a TP Monitor?
- 18 3.2 Standards and Architecture
- 20 3.3 Transaction management
- 21 3.4 Process management
- 21 3.4.1 Server classes
- 21 3.4.2 Reduced server resources
- 22 3.4.3 Dynamic load balancing
- 22 3.5 Robustness
- 23 3.6 Scalability
- 23 3.6.1 Shared process resources
- 23 3.6.2 Flexible hardware requirements
- 23 3.7 Performance
- 24 3.8 Security
- 24 3.9 Transaction profiles
- 25 3.10 Administration
- 25 3.11 Costs
- 26 3.12 3-tier architecture framework
- 26 3.13 When not to use a TP Monitor
- 27 4 Commercial TP Monitors
- 27 4.1 BEA Systems Inc.’s Tuxedo
- 27 4.1.1 Summary
- 28 4.1.2 History
- 29 4.1.3 Architecture
- 30 4.1.4 Web Integration
- 31 4.1.5 When to use
- 31 4.1.6 Future plans
- 32 4.1.7 Pricing
- 32 4.2 IBM’s TXSeries (Transarc’s Encina)
- 32 4.2.1 Summary
- 33 4.2.2 History
- 33 4.2.3 Architecture
- 35 4.2.4 Web Integration
- 35 4.2.5 When to use
- 36 4.2.6 Future plans
- 36 4.2.7 Pricing
- 36 4.3 IBM’s CICS
- 36 4.3.1 Summary
- 37 4.3.2 History
- 37 4.3.3 Architecture
- 39 4.3.4 Web integration
- 40 4.3.5 When to use
- 40 4.3.6 Future plans
- 41 4.3.7 Pricing
- 41 4.4 Microsoft Transaction Server MTS
- 41 4.4.1 Summary
- 41 4.4.2 History
- 42 4.4.3 Architecture
- 43 4.4.4 Web Integration
- 43 4.4.5 When to use
- 43 4.4.6 Future plans
- 43 4.4.7 Pricing
- 44 4.5 NCR TOP END
- 44 4.5.1 Summary
- 44 4.5.2 History
- 45 4.5.3 Architecture
- 46 4.5.4 Web Integration
- 47 4.5.5 When to use
- 47 4.5.6 Future plans
- 48 4.5.7 Pricing
- 48 4.6 Itautec’s Grip
- 48 4.6.1 Summary
- 48 4.6.2 History
- 49 4.6.3 Architecture
- 50 4.6.4 Web Integration
- 50 4.6.5 When to use
- 50 4.6.6 Future plans
- 51 4.6.7 Pricing
- 51 5 Analysis and recommendations
- 51 5.1 Analysis
- 51 5.2 Recommendations
- 52 References
- 53 Part 2 Retrieval and Manipulation
- 53 1 Introduction
- 53 1.1 General architecture of distributed Databases
- 53 1.1.1 Components of a distributed DBMS
- 55 1.1.2 Distributed versus Centralised databases
- 55 1.2 General architecture of federated Databases
- 56 1.2.1 Constructing Federated Databases
- 58 1.2.2 Implementing federated database systems
- 60 1.2.3 Data Warehouse Used To Implement Federated System
- 61 1.2.4 Query Processing in Federated Databases
- 61 1.2.5 Conclusion: Federated Databases
- 62 2 Organisation of distributed data
- 62 2.1 Schema integration in Federated Databases
- 63 2.2 Data Placement in Distributed Databases
- 64 2.2.1 Data Fragmentation
- 64 2.2.2 Criteria for the distribution of fragments
- 65 3 Parallel processing of retrieval
- 65 3.1 Query Processing
- 65 3.2 Query optimisation
- 66 4 Parallel processing of transactions
- 66 4.1 Characteristics of transaction management
- 66 4.2 Distributed Transaction
- 67 5 Commercial products
- 67 5.1 Tandem
- 67 5.1.1 Designed for scalability
- 67 5.1.2 High degree of manageability
- 67 5.1.3 Automatic process migration and load balancing
- 67 5.1.4 High level of application and system availability
- 68 5.2 Oracle
- 68 5.2.1 Oracle
- 69 5.2.2 A Family of Products with Oracle
- 74 5.3 Informix
- 74 5.3.1 Informix Dynamic Server
- 74 5.3.2 Basic Database Server Architecture
- 76 5.3.3 Informix Dynamic Server Features
- 78 5.3.4 Supported Interfaces and Client Products
- 80 5.4 IBM
- 80 5.4.1 DB2 Universal Database
- 83 5.4.2 IBM’s Object-Relational Vision and Strategy
- 85 5.4.3 IBM’s Business Intelligence Software Strategy
- 87 5.5 Sybase
- 87 5.5.1 Technology Overview: Sybase Computing Platform
- 90 Customer-Centric Development
- 91 5.5.3 Java for Logic in the Database
- 93 5.6 Microsoft
- 93 5.6.1 Overview
- 95 5.6.2 Microsoft Cluster Server
- 97 5.7 NCR Teradata
- 97 5.7.1 Data Warehousing with NCR Teradata
- 98 5.7.2 Teradata Architecture
- 99 5.7.3 Application Programming Interfaces
- 99 5.7.4 Language Preprocessors
- 100 5.7.5 Data Utilities
- 100 5.7.6 Database Administration Tools
- 100 5.7.7 Internet Access to Teradata
- 100 5.7.8 NCR's Commitment to Open Standards
- 101 5.7.9 Teradata at work
- 101 6 Analysis and recommendations
- 102 References
- 105 Part 3 Backup and Recovery
- 105 1 Introduction
- 105 2 Security aspects
- 107 3 Backup and Recovery Strategies
- 109 3.1 Recovery
- 110 3.2 Strategies
- 110 3.2.1 Requirements
- 111 3.2.2 Characteristics
- 111 4 Overview of commercial products
- 112 4.1 Tools
- 112 4.1.1 PC-oriented backup packages
- 113 4.1.2 UNIX packages
- 114 4.2 Databases
- 114 4.2.1 IBM DB
- 115 4.2.2 Informix
- 116 4.2.3 Microsoft SQL Server
- 116 4.2.4 Oracle
- 117 4.2.5 Oracle
- 119 4.2.6 Sybase SQL Server
- 119 5 Analysis and recommendations
- 120 References
- 121 Appendix A: Backup and Restore Investigation of Terabyte-scale Databases
- 121 A.1 Introduction
- 121 A.2 Requirements
- 121 A.3 Accurate benchmarking
- 122 A.4 The benchmark environment
- 123 A.5 Results
- 123 A.5.1 Executive summary
- 125 A.5.2 Detailed results
- 127 A.6 Interpreting the results
- 127 A.7 Summary
- 129 Appendix B: True Terabyte Database Backup Demonstration
- 129 B.1 Executive Summary
- 130 B.1.1 Definitions
- 130 B.2 Detailed Results
- 130 B.2.1 Demonstration Environment
- 131 B.2.2 Results
- 132 B.3 Interpreting the Results
- 133 B.4 Summary