Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Computing Resources Management (CRM) Unit- III Storage Management “Storage management” is a process used to optimize the use of storage devices and to protect the integrity of data for any media on which it resides.” Optimizing the use of storage devices translates into making sure the maximum amount of usable data is written to and read from these units at an acceptable rate of response. Optimizing these resources also means ensuring that there is an adequate amount of storage space available while guarding (protecting) against having expensive excess amounts. This notion of optimal use ties in to two of the main areas of storage management: capacity and performance. Protecting the integrity of data means that the data will always be accessible to those authorized to it and that it will not be changed unless the authorized owner specifically intends for it to be changed. Data integrity also implies that, should the data inadvertently become inaccessible or destroyed, reliable backup copies will enable its complete recovery. These explanations of data integrity tie into the other two main areas of storage management: reliability and recoverability. 1. Desired Traits (Qualities) of a Storage Management Process Owner The individual who is assigned as the owner of the storage management process should have a broad basis of knowledge in several areas related to disk space resources. These areas include applications, backup systems, hardware configurations, and database systems. Table lists, in priority order, these and other desirable characteristics to look for when selecting this individual. Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III Table: Prioritized Characteristics of a Storage Management Process Owner Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III 2. Four Keys To Storage Management 1. Storage Management Capacity 2. Storage Management Performance 3. Storage Management Reliability 4. Storage Management Recoverability 1. Storage Management Capacity Storage management capacity consists of providing 1. sufficient data storage to authorized users at a reasonable cost. 2. Storage capacity is often thought of as large quantities of disk farms accessible to servers or mainframes. 3. Data storage capacity includes main memory and magnetic disk storage for mainframe processors, midrange computers, workstations, servers, and desktop computers in all their various flavors. 4. Data storage capacity also includes alternative storage devices such as optical disks, magnetic drums, open reel magnetic tape, magnetic tape cartridges and cassettes, digital audio tape, and digital linear tape. 2. Storage Management Performance There are a variety of considerations that come into play when configuring infrastructure storage for optimal performance. The following list shows some of the most common of these. 1. Size and type of processor main memory 2. Number and size of buffers 3. Size of swap space 4. Number and type of channels 5. Device controller configuration 6. Logical volume groups 7. Amount of disk array cache memory Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III 8. Storage area networks (SANs) 9. Network-attached storage (NAS) SAN is a configuration enhancement that places a high-speed fiber-optic switch between servers and disk arrays. The two primary advantages are speed and flexibility. NAS is similar in concept to SAN except that the switch in NAS is replaced by a network. This enables data to be shared between storage devices and processors across a network. 3. Storage Management Reliability Manufacturers began designing into their disk storage systems redundant components such as backup power systems, dual channel ports to disk controllers, and dual pathing between drives and controllers. These improvements significantly increased the reliability of disk and tape storage equipment, in some instances more than doubling the usual one year mean time between failures (MTBF) for a disk drive. Redundant array of inexpensive disks (RAID) Performance was improved by disk striping, which writes data across multiple drives, increasing data paths and transfer rates and allowing simultaneous reading and writing to multiple disks. This implementation of RAID is referred to as level 0. Mirroring at RAID level 1 means that all data is duplicated on separate drives so that if one drive malfunctions, its mirrored drive maintains uninterrupted operation of all read and write transactions. RAID level 3 stripes the data for performance reasons similar to RAID level 0 and, for high reliability, assigns a dedicated parity drive on which parity bits are written to recover and rebuild data in the event of a data drive malfunction. There are usually two to four data drives supported by a single parity drive. Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III RAID level 5 is similar to level 3 except that parity bits are shared on non associated data drives. For example, for three data drives labeled A, B, and C, the parity bits for data striped across drives B and C would reside on drive A; the parity bits for data striped across drives A and C would reside on drive B; the parity bits for data striped across drives A and B would reside on drive C. 4. Storage Management Recoverability There are several methods available for recovering data that has been altered, deleted, damaged, or otherwise made inaccessible. Determining the correct recovery technique depends on the manner in which the data was backed up. Table “A” lists four common types of data backups. The first three are referred to as physical backups because operating system software or specialized program products copy the data as it physically resides on the disk without regard to database structures or logical organization—it is purely a physical backup. The fourth is called a logical backup because database management software reads—or backs up—logical parts of the database, such as tables, schemas, data dictionaries, or indexes; the software then writes the output to binary files. Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III Physical offline backups require shut down of all online systems, applications, and databases residing on the volume being backed up prior to starting the backup process. Incremental backups also require systems and databases to be shut down, but for much shorter periods of time. Since only the data that has changed since the last backup is what is copied, incremental backups can usually be completed within a few hours if done on a nightly basis. A physical online backup is a powerful backup technique that offers two very valuable and distinct benefits: ■ Databases can remain open to users during the backup process. ■ Recovery can be accomplished back to the last transaction processed. Full recovery is accomplished by restoring the last full backup and the incremental backups taken since the last full backup and then doing a forward recovery utilizing the archive and log tapes. Logical backups are less complicated and more time consuming to perform. There are three advantages to performing logical backups in concert with physical backups: 1. Exports can be made online, enabling 24/7 applications and databases to remain operational during the copying process. 2. Small portions of a database can be exported and imported, efficiently enabling maintenance to be performed on only the data required. 3. Exported data can be imported into databases or schemas at a higher version level than the original database, allowing for testing at new software levels. Another approach to safeguarding data becoming more prevalent today is the disk-to-disk backup. As the size of critical databases continues to grow, and as allowable backup windows continue to shrink, there are a number of advantages to this approach that help to justify its obvious costs. Among these advantages are: ■ Significant reduction in backup and recovery time. Copying directly to disk is orders of magnitude faster than copying to tape. This benefit also applies to online backups, which, while allowing databases to be open and accessible during backup processing, still incur a performance hit that is noticeably reduced by this method. ■ The stored copy can be used for other purposes. Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III These can include testing or report generation which, if done with the original data, could impact database performance. Data replication can be employed. This is a specialized version of a disk backup in which the data is mirrored on one of two drives in real time. At short specified intervals, such as every 30 minutes, the data is copied to an offsite recovery center ■ Help to cost justify tape backups. Copying the second stored disk files to tape can be scheduled at any time, provided it ends prior to the beginning of the next disk backup. It may even reduce investment in tape equipment, which can offset the costs of additional disks. Several other backup considerations need to be kept in mind when designing such a process. They are as follows: 1. Backup window 2. Restore times 3. Expiration dates 4. Retention (maintenance) periods 5. Recycle (recover) periods 6. Generation data groups 7. Offsite retrieval (recovery) times 8. Tape density (compactness) 9. Tape format (arrangement) 10. Tape packaging (wrapping / covering) 11. Automation techniques. Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III Assessing an Infrastructure’s Process: (Missing Topic) Storage Management Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III 3. Measuring and Streamlining the Storage Management Process We can measure and streamline the storage management process with the help of the assessment worksheet shown in previous figures. We can measure the effectiveness of a storage management process with service metrics such as outages caused by either lack of disk space or disk hardware problems and poor response due to fragmentation or lack of reorganizations. Process metrics—such as elapsed time for backups, errors during backups, and time to restore— help us gauge the efficiency of this process. And we can streamline the storage management process by automating actions such as the initiation of backups, restoring of files, or the changing of sizes or the number of buffers or cache storage. Network Management Network management “A process to maximize the reliability and utilization of network components in order to optimize network availability and responsiveness.” 1. Key decisions about network management Before an effective, high-level network management process can be designed, six key decisions need to be made that influence the strategy, direction, and cost of the process. The list below consists of the questions that these decisions answer. Once answered, these decisions will define the scope of responsibilities. 1. What will be managed by this process? 2. Who will manage it? 3. How much authority will be given? 4. What types of tools and support will be provided? 5. To what extent will other processes be integrated with this process? 6. What levels of service and quality will be expected? Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III 1. What will be managed by this process? Ans: The Network One of the problems that infrastructures experience in this regard is being too unclear as to who is responsible for which aspects of the network. This is specially true in large, complex environments where world-wide networks may vary in terms of topology, platform, protocols, security, and suppliers. 2. Who will manage it? Initially, this decision should determine which department within the infrastructure will be assigned the responsibility for heading (leading) up the design, implementation, and ongoing management of this process. Within that department will be assigned the person who will have overall responsibility for the network management process. In essence, this decision deals with determining who the process owner will be. The ideal candidate to serve as owner of the network management process will have strong people skills, knowledge of network resources, and a sense of urgency. Table “B” offers a more comprehensive list, in priority order, of the traits desired in such an individual. The people skills involve working with developers about application profiles, transaction mix and arrival patterns, and planned increases in workloads, and working effectively with users about desktop requirements, connectivity, and security. A network management process owner should be knowledgeable about network operating systems, utility programs, support software, and key hardware components, such as routers switchers, hubs, and repeaters. One of the most valuable traits for the owner of this process is a sense of urgency in tactically responding to and resolving a variety of network problems. Other desirable traits include knowledge of infrastructure software and hardware and the ability to analyze metrics. Table B: Prioritized characteristics of a network management process owner Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III 3. How much authority will this person be given? Executive support plays a key role here from several perspectives. First, managers must be willing to surrender some of their authority over the network to the individual now responsible for overseeing its management. Second, managers must be willing to support their person in situations of conflict that will surely arise. Conflicts involving network management can originate from several sources. One source of conflict can come from enforcing network security policies such as if, when, and how data over the network will be encrypted or under what instances a single sign-on scheme will be employed for network and application logons. Another common source of network conflict involves user expectations about the infrastructure backing up desktop data instead of the more practical method of saving it to the server. The degree of authority a network management process owner has over network suppliers can also instigate conflict, particularly if managers have not communicated clearly to supplier management about the delegation that is in effect. Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III 4. What types of tools and support will be provided? Decisions about the types of tools and the amount of vendor support that will be provided directly influence the costs of managing networks. A variety of network tools are available to monitor and manage a network. Some are incorporated in hardware components, but most are software-based. ICMP Ping When you receive a 'device down' alert, the first condition that you might want to assess is if the device is reachable. From the device snapshot page in OpManager, do an instant ping and check for response. You can troubleshoot further using the other network monitoring tools if a ping to the device fails, or if the response time is very high. Traceroute When you troubleshoot a device down alert using Ping and the device fails to respond, you can Traceroute to determine if the device is not reachable because of a failure in the path. Trace the route from OpManager to the destination device, check the number of hops to the monitored device and spot the exact point of delay or outage. Again, this serves as a first level troubleshooting and based on the response, you can switch to other monitoring tools to drill down a fault. Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III Switch Port Mapper This tool helps you see the port-wise connectivity of devices to the network switches which is necessary to help troubleshoot high traffic issues. The Switch Port Mapper is a network monitoring tool that gives you the MAC address, IP Address and DNS names of the devices connected to the switch. 5. To what extent will other processes be integrated with this process? A well-designed network management process will have strong relationships to six other systems management processes. These processes are availability, performance and tuning, change management, problem management, capacity planning, and security. The extent to which network management integrates with these other processes by sharing tools, databases, procedures, or cross-trained personnel is a decision that will influence the effectiveness of the process. Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III For example, capacity planning and network management could both share a tool that simulates network workloads to make resource forecasts more accurate. Similarly, a tool that controls network access may have database access controls, which the security process could use. 6. What levels of service and quality will be expected? The levels of service and quality that network groups negotiate with their customers directly affect the cost of hardware, software, training, and support. This is a key decision that should be thoroughly understood and committed to by the groups responsible for budgeting its costs, and delivering on its agreements. Suppliers of network components, such as routers, switches, hubs, and repeaters, should also be part of these negotiations, since their equipment has a direct impact on availability and performance. Long distance carriers are another group to include in the development of service level agreements. Many carriers will not stipulate a guaranteed percentage of uptime in their service contracts because of situations beyond their control such as natural disasters or man-made accidents such as construction mishaps. In these instances, one can specify conditions in the service contract for which carriers will assume responsibility such as spare parts, time to repair, and on-call response times. Service level agreements (SLAs) for the network should be revisited whenever major upgrades to hardware or bandwidth are put in place. The SLAs should also be adjusted whenever significant workloads are added to the network to account for increases line traffic, contention on resources, or extended hours of service. 4. Assessing an infrastructure’s network management process The worksheet in Figure C is used without weighting factors, meaning that all 10 categories are weighted evenly for the assessment of a network management process. Sample ratings are Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III inserted to illustrate the use of the worksheet. In this case, the network management process scored a total of 23 points for an overall non weighted assessment score of 58%, as compared to the second sample worksheet, which compiled a weighted assessment score of 61%. Figure C Measuring and streamlining the network management process A network management process can be measured and streamlined with the help of the assessment worksheet shown in Figure D. The effectiveness of a network management process can be measured with service metrics, such as network availability, network response Department of Computer Science & Engineering, PRMIT&R, Badnera. Computing Resources Management (CRM) Unit- III times, and elapsed time to log on. The efficiency of this process can be measured with process metrics, such as outages caused by network design, maintenance, carriers testing, nonstandard devices, lack of training, and negligence. A network management process can be streamlined by automating actions such as the notification of network analysts when nodes go offline or when other network triggers are activated. Figure D Department of Computer Science & Engineering, PRMIT&R, Badnera.