Download Computing Resources Management Unit- II

Survey
yes no Was this document useful for you?
   Thank you for your participation!

* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project

Document related concepts

Database model wikipedia , lookup

Transcript
Computing Resources Management (CRM) Unit- III
Storage Management
“Storage management” is a process used to optimize the use of storage devices and to
protect the integrity of data for any media on which it resides.”
Optimizing the use of storage devices translates into making sure the maximum amount
of usable data is written to and read from these units at an acceptable rate of response.
Optimizing these resources also means ensuring that there is an adequate amount of
storage space available while guarding (protecting) against having expensive excess amounts.
This notion of optimal use ties in to two of the main areas of storage management: capacity and
performance.
Protecting the integrity of data means that the data will always be accessible to those
authorized to it and that it will not be changed unless the authorized owner specifically intends
for it to be changed.
Data integrity also implies that, should the data inadvertently become inaccessible or destroyed,
reliable backup copies will enable its complete recovery. These explanations of data integrity tie
into the other two main areas of storage management: reliability and recoverability.
1. Desired Traits (Qualities) of a Storage Management Process Owner
The individual who is assigned as the owner of the storage management process should have
a broad basis of knowledge in several areas related to disk space resources.
These areas include applications, backup systems, hardware configurations, and database
systems. Table lists, in priority order, these and other desirable characteristics to look for when
selecting this individual.
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
Table: Prioritized Characteristics of a Storage Management Process Owner
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
2. Four Keys To Storage Management
1. Storage Management Capacity
2. Storage Management Performance
3. Storage Management Reliability
4. Storage Management Recoverability
1. Storage Management Capacity
Storage management capacity consists of providing
1. sufficient data storage to authorized users at a reasonable cost.
2. Storage capacity is often thought of as large quantities of disk farms accessible to servers
or mainframes.
3. Data storage capacity includes main memory and magnetic disk storage for mainframe
processors, midrange computers, workstations, servers, and desktop computers in all their
various flavors.
4. Data storage capacity also includes alternative storage devices such as optical disks,
magnetic drums, open reel magnetic tape, magnetic tape cartridges and cassettes, digital
audio tape, and digital linear tape.
2. Storage Management Performance
There are a variety of considerations that come into play when configuring infrastructure storage
for optimal performance. The following list shows some of the most common of these.
1. Size and type of processor main memory
2. Number and size of buffers
3. Size of swap space
4. Number and type of channels
5. Device controller configuration
6. Logical volume groups
7. Amount of disk array cache memory
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
8. Storage area networks (SANs)
9. Network-attached storage (NAS)
SAN is a configuration enhancement that places a high-speed fiber-optic switch between
servers and disk arrays. The two primary advantages are speed and flexibility.
NAS is similar in concept to SAN except that the switch in NAS is replaced by a
network. This enables data to be shared between storage devices and processors across a
network.
3. Storage Management Reliability
Manufacturers began designing into their disk storage systems redundant components such as
backup power systems, dual channel ports to disk controllers, and dual pathing between drives
and controllers. These improvements significantly increased the reliability of disk and tape
storage equipment, in some instances more than doubling the usual one year mean time between
failures (MTBF) for a disk drive. Redundant array of inexpensive disks (RAID)
Performance was improved by disk striping, which writes data across multiple drives,
increasing data paths and transfer rates and allowing simultaneous reading and writing to
multiple disks. This implementation of RAID is referred to as level 0.
Mirroring at RAID level 1 means that all data is duplicated on separate drives so that if
one drive malfunctions, its mirrored drive maintains uninterrupted operation of all read and write
transactions.
RAID level 3 stripes the data for performance reasons similar to RAID level 0 and, for
high reliability, assigns a dedicated parity drive on which parity bits are written to recover and
rebuild data in the event of a data drive malfunction. There are usually two to four data drives
supported by a single parity drive.
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
RAID level 5 is similar to level 3 except that parity bits are shared on non associated data
drives. For example, for three data drives labeled A, B, and C, the parity bits for data striped
across drives B and C would reside on drive A; the parity bits for data striped across drives A
and C would reside on drive B; the parity bits for data striped across drives A and B would reside
on drive C.
4. Storage Management Recoverability
There are several methods available for recovering data that has been altered, deleted, damaged,
or otherwise made inaccessible. Determining the correct recovery technique depends on the
manner in which the data was backed up.
Table “A” lists four common types of data backups.
The first three are referred to as physical backups because operating system software or
specialized program products copy the data as it physically resides on the disk without regard to
database structures or logical organization—it is purely a physical backup. The fourth is called a
logical backup because database management software reads—or backs up—logical parts of the
database, such as tables, schemas, data dictionaries, or indexes; the software then writes the
output to binary files.
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
Physical offline backups require shut down of all online systems, applications, and databases
residing on the volume being backed up prior to starting the backup process.
Incremental backups also require systems and databases to be shut down, but for much shorter
periods of time. Since only the data that has changed since the last backup is what is copied,
incremental backups can usually be completed within a few hours if done on a nightly basis.
A physical online backup is a powerful backup technique that offers two very valuable and
distinct benefits:
■ Databases can remain open to users during the backup process.
■ Recovery can be accomplished back to the last transaction processed.
Full recovery is accomplished by restoring the last full backup and the incremental backups
taken since the last full backup and then doing a forward recovery utilizing the archive and log
tapes.
Logical backups are less complicated and more time consuming to perform. There are three
advantages to performing logical backups in concert with physical backups:
1. Exports can be made online, enabling 24/7 applications and databases to remain operational
during the copying process.
2. Small portions of a database can be exported and imported, efficiently enabling maintenance
to be performed on only the data required.
3. Exported data can be imported into databases or schemas at a higher version level than the
original database, allowing for testing at new software levels.
Another approach to safeguarding data becoming more prevalent today is the disk-to-disk
backup. As the size of critical databases continues to grow, and as allowable backup windows
continue to shrink, there are a number of advantages to this approach that help to justify its
obvious costs. Among these advantages are:
■ Significant reduction in backup and recovery time. Copying directly to disk is orders of
magnitude faster than copying to tape. This benefit also applies to online backups, which, while
allowing databases to be open and accessible during backup processing, still incur a performance
hit that is noticeably reduced by this method.
■ The stored copy can be used for other purposes.
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
These can include testing or report generation which, if done with the original data, could
impact database performance.
 Data replication can be employed.
This is a specialized version of a disk backup in which the data is mirrored on one of two drives
in real time. At short specified intervals, such as every 30 minutes, the data is copied to an offsite
recovery center
■ Help to cost justify tape backups.
Copying the second stored disk files to tape can be scheduled at any time, provided it ends prior
to the beginning of the next disk backup. It may even reduce investment in tape equipment,
which can offset the costs of additional disks.
Several other backup considerations need to be kept in mind when designing such a process.
They are as follows:
1. Backup window
2. Restore times
3. Expiration dates
4. Retention (maintenance) periods
5. Recycle (recover) periods
6. Generation data groups
7. Offsite retrieval (recovery) times
8. Tape density (compactness)
9. Tape format (arrangement)
10. Tape packaging (wrapping / covering)
11. Automation techniques.
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
Assessing
an
Infrastructure’s
Process: (Missing Topic)
Storage
Management
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
3. Measuring and Streamlining the Storage Management Process
We can measure and streamline the storage management process with the help of the
assessment worksheet shown in previous figures. We can measure the effectiveness of a storage
management process with service metrics such as outages caused by either lack of disk space or
disk hardware problems and poor response due to fragmentation or lack of reorganizations.
Process metrics—such as elapsed time for backups, errors during backups, and time to restore—
help us gauge the efficiency of this process. And we can streamline the storage management
process by automating actions such as the initiation of backups, restoring of files, or the
changing of sizes or the number of buffers or cache storage.
Network Management
Network management
“A process to maximize the reliability and utilization of network components in order to
optimize network availability and responsiveness.”
1. Key decisions about network management
Before an effective, high-level network management process can be designed, six key
decisions need to be made that influence the strategy, direction, and cost of the process.
The list below consists of the questions that these decisions answer. Once answered,
these decisions will define the scope of responsibilities.
1. What will be managed by this process?
2. Who will manage it?
3. How much authority will be given?
4. What types of tools and support will be provided?
5. To what extent will other processes be integrated with this process?
6. What levels of service and quality will be expected?
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
1. What will be managed by this process?
Ans: The Network
One of the problems that infrastructures experience in this regard is being too unclear as to who
is responsible for which aspects of the network.
This is specially true in large, complex environments where world-wide networks may vary in
terms of topology, platform, protocols, security, and suppliers.
2. Who will manage it?
Initially, this decision should determine which department within the infrastructure will be
assigned the responsibility for heading (leading) up the design, implementation, and ongoing
management of this process. Within that department will be assigned the person who will have
overall responsibility for the network management process. In essence, this decision deals with
determining who the process owner will be. The ideal candidate to serve as owner of the network
management process will have strong people skills, knowledge of network resources, and a sense of
urgency. Table “B” offers a more comprehensive list, in priority order, of the traits desired in such an
individual. The people skills involve working with developers about application profiles, transaction mix
and arrival patterns, and planned increases in workloads, and working effectively with users about
desktop requirements, connectivity, and security.
A network management process owner should be knowledgeable about network
operating systems, utility programs, support software, and key hardware components, such as
routers switchers, hubs, and repeaters. One of the most valuable traits for the owner of this
process is a sense of urgency in tactically responding to and resolving a variety of network
problems. Other desirable traits include knowledge of infrastructure software and hardware and
the ability to analyze metrics.
Table B: Prioritized characteristics of a network management process owner
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
3. How much authority will this person be given?
Executive support plays a key role here from several perspectives. First, managers must
be willing to surrender some of their authority over the network to the individual now
responsible for overseeing its management. Second, managers must be willing to support their
person in situations of conflict that will surely arise. Conflicts involving network management
can originate from several sources.
One source of conflict can come from enforcing network security policies such as if,
when, and how data over the network will be encrypted or under what instances a single sign-on
scheme will be employed for network and application logons. Another common source of
network conflict involves user expectations about the infrastructure backing up desktop data
instead of the more practical method of saving it to the server. The degree of authority a network
management process owner has over network suppliers can also instigate conflict, particularly if
managers have not communicated clearly to supplier management about the delegation that is in
effect.
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
4. What types of tools and support will be provided?
Decisions about the types of tools and the amount of vendor support that will be provided
directly influence the costs of managing networks.
A variety of network tools are available to monitor and manage a network. Some are
incorporated in hardware components, but most are software-based.
ICMP Ping
When you receive a 'device down' alert, the first condition that you might want to assess is if the
device is reachable. From the device snapshot page in OpManager, do an instant ping and check
for response. You can troubleshoot further using the other network monitoring tools if a ping to
the device fails, or if the response time is very high.
Traceroute
When you troubleshoot a device down alert using Ping and the device fails to respond, you can
Traceroute to determine if the device is not reachable because of a failure in the path. Trace the
route from OpManager to the destination device, check the number of hops to the monitored
device and spot the exact point of delay or outage. Again, this serves as a first level
troubleshooting and based on the response, you can switch to other monitoring tools to drill
down a fault.
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
Switch Port Mapper
This tool helps you see the port-wise connectivity of devices to the network switches which is
necessary to help troubleshoot high traffic issues. The Switch Port Mapper is a network
monitoring tool that gives you the MAC address, IP Address and DNS names of the devices
connected to the switch.
5. To what extent will other processes be integrated with this process?
A well-designed network management process will have strong relationships to six other
systems management processes. These processes are availability, performance and tuning,
change management, problem management, capacity planning, and security. The extent to which
network management integrates with these other processes by sharing tools, databases,
procedures, or cross-trained personnel is a decision that will influence the effectiveness of the
process.
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
For example, capacity planning and network management could both share a tool that
simulates network workloads to make resource forecasts more accurate. Similarly, a tool that
controls network access may have database access controls, which the security process could
use.
6. What levels of service and quality will be expected?
The levels of service and quality that network groups negotiate with their customers directly
affect the cost of hardware, software, training, and support. This is a key decision that should be
thoroughly understood and committed to by the groups responsible for budgeting its costs, and
delivering on its agreements.
Suppliers of network components, such as routers, switches, hubs, and repeaters, should also be
part of these negotiations, since their equipment has a direct impact on availability and
performance.
Long distance carriers are another group to include in the development of service level
agreements. Many carriers will not stipulate a guaranteed percentage of uptime in their service
contracts because of situations beyond their control such as natural disasters or man-made
accidents such as construction mishaps. In these instances, one can specify conditions in the
service contract for which carriers will assume responsibility such as spare parts, time to repair,
and on-call response times.
Service level agreements (SLAs) for the network should be revisited whenever major upgrades
to hardware or bandwidth are put in place. The SLAs should also be adjusted whenever
significant workloads are added to the network to account for increases line traffic, contention on
resources, or extended hours of service.
4. Assessing an infrastructure’s network management process
The worksheet in Figure C is used without weighting factors, meaning that all 10 categories
are weighted evenly for the assessment of a network management process. Sample ratings are
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
inserted to illustrate the use of the worksheet. In this case, the network management process
scored a total of 23 points for an overall non weighted assessment score of 58%, as compared to
the second sample worksheet, which compiled a weighted assessment score of 61%.
Figure C
Measuring and streamlining the network management process
A network management process can be measured and streamlined with the help of the
assessment worksheet shown in Figure D. The effectiveness of a network management
process can be measured with service metrics, such as network availability, network response
Department of Computer Science & Engineering, PRMIT&R, Badnera.
Computing Resources Management (CRM) Unit- III
times, and elapsed time to log on. The efficiency of this process can be measured with
process metrics, such as outages caused by network design, maintenance, carriers testing,
nonstandard devices, lack of training, and negligence. A network management process can be
streamlined by automating actions such as the notification of network analysts when nodes
go offline or when other network triggers are activated.
Figure D
Department of Computer Science & Engineering, PRMIT&R, Badnera.