Survey
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
* Your assessment is very important for improving the workof artificial intelligence, which forms the content of this project
Cracking of wireless networks wikipedia , lookup
Network tap wikipedia , lookup
IEEE 802.1aq wikipedia , lookup
Zero-configuration networking wikipedia , lookup
Airborne Networking wikipedia , lookup
Nonblocking minimal spanning switch wikipedia , lookup
List of wireless community networks by region wikipedia , lookup
VALIDATED HARDWARE CONFIGURATION Partner Name Deployment Guide Content Document History ...................................................................................................................... 3 Introduction ................................................................................................................................ 4 Notes ......................................................................................................................................... 5 Switches .................................................................................................................................... 5 Rack Physical Cabling schema............................................................................................... 6 Logical networking schema .................................................................................................... 7 Switch configuration ............................................................................................................... 8 Server configuration ................................................................................................................... 8 RAID configuration ................................................................................................................. 8 BIOS configuration ................................................................................................................. 8 Controller and Storage Nodes ............................................................................................. 9 Compute and Infrastructure Nodes ..................................................................................... 9 Storage node boot order ..................................................................................................... 9 Infrastructure node ..................................................................................................................... 9 Installation .............................................................................................................................. 9 Configuration .........................................................................................................................10 Fuel Installation .....................................................................................................................10 OpenStack Cluster Configuration ..............................................................................................12 Network configuration ............................................................................................................12 Configure VLANS...............................................................................................................12 Configure “Public” section: .................................................................................................13 Configure “Neutron L2 configuration”: ................................................................................13 Network interfaces layout ......................................................................................................13 Controllers .........................................................................................................................13 Computes ..........................................................................................................................14 Storages ............................................................................................................................15 Other settings ........................................................................................................................16 Verifying network setup .........................................................................................................16 Disk layout .............................................................................................................................16 Controllers .........................................................................................................................17 Computes .......................................................................................................................17 Storage nodes ................................................................................................................17 Deploy ...................................................................................................................................18 Post Deployment Verification .................................................................................................18 Appendix ...................................................................................................................................19 Cabling tables........................................................................................................................19 <10G Switch>-1..............................................................................................................20 <10G Switch>-2..............................................................................................................22 Switch “show running-config” output ......................................................................................24 Document History Version Revision Date Description 0.1 DD-MM-YYYY Initial Version Introduction The <Vendor>-Mirantis OpenStack Reference Architecture describes a modular, scalable rack configuration of <Vendor> compute, storage and network hardware, validated with Mirantis OpenStack, and engineered to <detail specific rationale, compatible workload types, target cluster size and other specifics>. This companion document details steps required to deploy such a cluster and start it running. The guide assumes that users are familiar with basic OpenStack operations, use of the Linux command line, with enterprise networking, virtualization and with the <Vendor> switch CLI. Basic knowledge of Fuel operations is also required. Notes Please note that IP addresses, VLAN IDs and network interface names are provided for reference only. Feel free to replace them with your own preferred values. Switches The rack contains three (3) <Vendor> Networking switches: one <1G Switch>- and pair of <10G Switches>, interconnected by Multi-chassis Link Aggregation Protocol (MC-LAG). Physical cabling and logical networking schemas can be found below. Cabling tables can be found in the Appendix. Rack Physical Cabling schema Logical networking schema Switch configuration Please refer to the Appendix for complete switch configuration listings. Server configuration Please refer to <Link to IPMI user guide> to learn more about how to use IPMI to configure RAID and BIOS. RAID configuration <Provide an actual disks configuration using the table below as an example> Node Disks RAID type Notes Fuel 2x1.2 TB SATA RAID1 OS Controller 2x 400 Gb SSD None OS: sda MongoDB: sdb Compute 1x 400 Gb SSD None OS: sda Storage 4x 200 Gb SSD None Ceph journal: sda-sdd 2x 1.2 TB SAS RAID1 OS: sde 20x 1.2 TB SAS RAID0 Ceph OSD: sdf-sdy (each disk configured separately as RAID0 to use controller’s cache) BIOS configuration The guide assumes that all BIOS settings are initially set to factory defaults. If they are not, please reset BIOS to default settings before proceeding. Common Settings, shown below, should be applied to all nodes. Additional settings for Compute, Storage, or other nodes, also shown below, should also be applied, as appropriate. Common settings Display Name <Sample Name> Attribute <Sample Attribute> Settings <Sample value> Controller and Storage Nodes Display Name <Sample Name> Attribute <Sample Attribute> Settings <Sample value> Compute and Infrastructure Nodes Display Name <Sample Name> Attribute <Sample Attribute> Settings <Sample value> Storage node boot order <Specify any changes in boot device priority if necessary. Typically, Storage nodes require precise boot configuration.> Infrastructure node Installation The Infrastructure node serves as a hypervisor host for the Fuel Master VM and other VM nodes involved in the testing process. To install and configure it: 1. Download the Ubuntu server 14.04 ISO from http://www.ubuntu.com/download/server to your local computer. 2. Use IPMI to mount the Ubuntu installation ISO on the Infrastructure node as a virtual CD-ROM. Please refer to the <Link to IPMI user guide> for more details. 3. Install Ubuntu with the following roles selected: ● SSH server ● Virtualization server. Leave other settings with default values. Networking will be configured in the next steps. Configuration 1. Log into the Infrastructure node and configure the resolver: # echo "nameserver 8.8.8.8" > /etc/resolv.conf 2. Configure bridges at /etc/network/interfaces. Use the following configuration as an example. Please pay attention to interface names and ip addresses, changing them if needed: auto em1 iface em1 inet manual auto em2 iface em2 inet manual auto br-ext iface br-ext inet static bridge_ports em2 address 172.16.224.2 netmask 255.255.255.128 gateway 172.16.224.1 bridge_stp off auto br-pxe iface br-pxe inet manual bridge_ports em1 bridge_stp off 3. Reboot the server and make sure the bridges are up and the gateway is accessible: # brctl show # ping 172.16.224.1 Fuel Installation 1. Download MOS: # cd ~ # wget http://9f2b43d3ab92f886c3f0e8d43ffad23ec549234584e5c62a6e24.r60.cf1.rackcdn.com/MirantisOpenStack-7.0.iso 2. Connect to the foundation node with virt-manager and create a new VM with the following parameters: Name: Fuel Networking: 2x NIC, each connected to corresponding bridge: br-ext and br-pxe Memory: 8Gb CPU: 2x core Disk: 100GB Another way to create this VM is with the following command: # virt-install --os-variant=ubuntutrusty --ram 8192 --vcpus=2 --network bridge=brext,model=virtio --network bridge=br-pxe,model=virtio --name fuel1 --disk path=/var/lib/libvirt/images/fuel1.qcow2,cache=none,size=100 -c ~/MirantisOpenStack7.0.iso --graphics vnc --autostart --noautoconsole 3. Use virt-manager or VNC console to connect to the VM and proceed with the installation: a) Wait until Fuel-menu appears, go to the “Network setup” tab, select eth0 and set parameters as follows: b) Go to the “Quit setup” tab and select “Save and continue.” Wait until installation completes. 4. Open your browser and check if the Fuel web UI is accessible at http://172.16.224.3:8000/ OpenStack Cluster Configuration 1. Log into each hardware server’s IPMI and turn the power on. 2. Wait until all servers are discovered by Fuel and shown in its Web UI. 3. Go to to the Fuel web UI at http://172.16.224.3:8000/ and create a new environment with the following parameters (change them if needed): Parameter Name and release Compute Networking setup Storage backends Additional services Value cloud1, Kilo on Ubuntu 14.04 Set radiobox “KVM” Set radiobox “Neutron with VLAN segmentation” Set radiobox “Yes, use Ceph” Set checkbox “Install Ceilometer (OpenStack Telemetry)” Use <Vendor> <SERVER MODEL 1> servers to add 3 nodes with the following roles: Controller Telemetry - Mongo DB Use <Vendor> <SERVER MODEL 1> servers to add 3 nodes with the following role: Compute Use <Vendor> <SERVER MODEL 2> servers to add 3 nodes with the following role: Storage – Ceph OSD Network configuration Configure VLANS Go to the “Networks” tab in the Fuel Web UI and select “Use VLAN tagging” checkbox for all the networks. VLAN IDs should be set according to the following table: Network Public Storage VLAN ID 160 180 Management 140 Configure “Public” section: Note: Make sure that “Gateway” field is filled in with your external router’s IP address from 172.16.224.0/25 subnet (VLAN 160) and that this provides internet access for the whole subnet. Field IP range start IP range end CIDR Gateway Floating IP range start Floating IP range end Value 172.16.224.4 172.16.224.14 172.16.224.0/25 172.16.224.1 172.16.224.15 172.16.224.126 Configure “Neutron L2 configuration”: Field VLAN ID range start VLAN ID range end Value 200 1000 Network interfaces layout Go to the “Nodes” tab, and configure the network interface layout for each server according to its role. Nodes with the same role may be configured together by selecting all servers in a group, instead of a single server. Use “Bond interfaces” to create bonds with the following parameters: Parameter Xmit Hash Policy Mode Value layer2 802.3ad(LACP) Controllers Interface bond0 Slave interfaces eth0 Assigned networks Public eth1 eth2 Storage Management Private Admin(PXE) Computes Interface bond0 Slave interfaces eth0 eth1 eth2 Assigned networks Public Storage Management Private Admin(PXE) Storages Storage traffic can have a strong performance impact on other networks within the same physical link, so we will separate it out. Interface bond0 bond1 Slave interfaces eth0 eth2 Assigned networks Public Management Private eth1 Storage eth3 eth4 Admin(PXE) Other settings Go to the “Settings tab” -> “Neutron advanced configuration” and select the “Neutron DVR” checkbox. Verifying network setup Return to the “Networks” tab and press “Verify Network.” Please note that full network check is not possible on that stage because bonded interfaces are dropped from the verification list. Make sure other network checks pass successfully. If not, double check your external router and network settings. Disk layout <This configuration depends on actual disks available on each server. Change the description below if needed> Go to the “Nodes” tab and configure disk layout for each node. Controllers MongoDB should be located on a separate physical disk for best performance. Disk sda sdb Roles Base system, Logs, Mysql Database MongoDB Computes Leave with the defaults. Storage nodes <This configuration based on 20 disks for Ceph, 4 SSD for Ceph Journal, and 2 disks in RAID1 for OS. Change it in case of different disk layout.> Use first 4 disks for the Ceph Journals, 5th disk for the Base System and leave all the rest for Ceph. Disk sda-sdd sde sdf through sdy Roles Ceph journal Base system Ceph Deploy Go to the “Dashboard” tab and click “Deploy.” Normally, the deployment process takes a few hours. Post Deployment Verification Go to the “Health Check” tab and click “Run Tests.” Wait for a while and make sure all the tests (except for the credentials tests, which fail because of default username and password) are green (passed). If the tests are fine, then everything is installed, working properly and ready for use. Appendix Cabling tables <1G Switch> Interface Name Gi 0/0 Gi 0/1 Gi 0/2 Gi 0/3 Gi 0/4 Gi 0/5 Gi 0/6 Gi 0/7 Gi 0/8 Gi 0/9 Gi 0/10 Gi 0/11 Gi 0/12 Gi 0/13 Gi 0/14 Gi 0/15 Gi 0/16 Gi 0/17 Gi 0/18 Connector 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper Destination Device Name OS Controller 1 OS Controller 2 OS Controller 3 Compute 1 Compute 2 Compute 3 Storage 1 Storage 2 Storage 3 Infrastructure node OS Controller 1 OS Controller 2 OS Controller 3 Port IPMI IPMI IPMI IPMI IPMI IPMI IPMI IPMI IPMI IPMI 1G1 1G1 1G1 VLANs Untagged 100 100 100 100 100 100 100 100 100 100 120 120 120 Tagged LAG # - Mode - - - Gi 0/19 Gi 0/20 Gi 0/21 Gi 0/22 Gi 0/23 Gi 0/24 Gi 0/25 Gi 0/26 Gi 0/27 Gi 0/28 Gi 0/29 Gi 0/30 Gi 0/31 Gi 0/32 Gi 0/33 Gi 0/34 Gi 0/35 Gi 0/36 Gi 0/37 Gi 0/38 Gi 0/39 Gi 0/40 Gi 0/41 Gi 0/42 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper 1G Copper Compute 1 Compute 2 Compute 3 Storage 1 Storage 2 Storage 3 1G1 1G1 1G1 1G1 1G1 1G1 120 120 120 120 120 120 1G Copper 1G1 Gi 0/43 1G Copper Gi 0/44 1G Copper Infrastructure node Infrastructure node <10G Switch>-1 Gi 0/45 1G Copper <10G Switch>-2 Gi 0/46 Gi 0/47 Stack 0/48 Stack 0/49 Te 0/50 1G Copper 1G Copper Uplink OOB Uplink Public 10G Fiber SR 10G Fiber SR <10G Switch>-1 Te 0/47 <10G Switch>-2 Te 1/47 Te 0/51 120 - - 1G2 160 - - Managem ent Managem ent 100 - - 100 - - 100 160 1 LACP 1 LACP 100,120, 160 100,120, 160 Management 0/0 - <10G Switch>-1 Interface Name Connector Device Name Port VLANs Untagged Tagged LAG # Mode Te 0/0 10G 10G1 160 140,180,200-1000 1 LACP 10G1 160 140,180,200-1000 2 LACP 10G1 160 140,180,200-1000 3 LACP 10G 10G 10G OS Controller 1 OS Controller 2 OS Controller 3 Compute 1 Compute 2 Compute 3 Te 0/1 10G Te 0/2 10G Te 0/3 Te 0/4 Te 0/5 Te 0/6 Te 0/7 Te 0/8 Te 0/9 Te 0/10 Te 0/11 Te 0/12 Te 0/13 Te 0/14 Te 0/15 Te 0/16 Te 0/17 Te 0/18 Te 0/19 Te 0/20 Te 0/21 Te 0/22 Te 0/23 Te 0/24 Te 0/25 Te 0/26 Te 0/27 Te 0/28 Te 0/29 Te 0/30 Te 0/31 Te 0/32 Te 0/33 Te 0/34 Te 0/35 Te 0/36 Te 0/37 Te 0/38 Te 0/39 Te 0/40 Te 0/41 10G1 10G1 10G1 160 160 160 140,180,200-1000 140,180,200-1000 140,180,200-1000 4 5 6 - LACP LACP LACP - 10G 10G 10G Storage 1 Storage 2 Storage 3 10G1 10G1 10G1 140,160 140,160 140,160 10G 10G 10G Storage 1 Storage 2 Storage 3 10G3 10G3 10G3 180,200-1000 180,200-1000 180,200-1000 7 8 9 10 11 12 - LACP LACP LACP LACP LACP LACP - - - Te 0/42 Te 0/43 Te 0/44 Te 0/45 Te 0/46 Te 0/47 10G fortyGig 0/48 40G fortyGig 0/52 40G fortyGig 0/56 fortyGig 0/60 Management 0/0 1G Copper LACP <AGGR EGATI ON PROTO COL> <AGGR EGATI ON PROTO COL> - <1G Switch> <10G Switch>-2 Te 0/50 fortyGig 1/48 - <10G Switch>-2 fortyGig 1/52 - <1G Switch> 100,120,160 13 - Gi 0/46 <10G Switch>-2 Interface Name Connector Te 0/0 10G Te 0/1 10G Te 0/2 10G Te 0/3 Te 0/4 Te 0/5 Te 0/6 Te 0/7 Te 0/8 Te 0/9 Te 0/10 Te 0/11 Te 0/12 Te 0/13 Te 0/14 10G 10G 10G Destination Device Name OS Controller 1 OS Controller 2 OS Controller 3 Compute 1 Compute 2 Compute 3 Port VLANs Untagged Tagged LAG # Mode 10G3 160 140,190,200-1000 1 LACP 10G3 160 140,190,200-1000 2 LACP 10G3 160 140,190,200-1000 3 LACP 10G3 10G3 10G3 160 160 160 140,190,200-1000 140,190,200-1000 140,190,200-1000 4 5 6 - LACP LACP LACP - - Te 0/15 Te 0/16 Te 0/17 Te 0/18 Te 0/19 Te 0/20 Te 0/21 Te 0/22 Te 0/23 Te 0/24 Te 0/25 Te 0/26 Te 0/27 Te 0/28 Te 0/29 Te 0/30 Te 0/31 Te 0/32 Te 0/33 Te 0/34 Te 0/35 Te 0/36 Te 0/37 Te 0/38 Te 0/39 Te 0/40 Te 0/41 Te 0/42 Te 0/43 Te 0/44 Te 0/45 Te 0/46 Te 0/47 10G 10G 10G Storage 1 Storage 2 Storage 3 10G2 10G2 10G2 140,160 140,160 140,160 10G 10G 10G Storage 1 Storage 2 Storage 3 10G4 10G4 10G4 180,200-1000 180,200-1000 180,200-1000 10G fortyGig 0/48 40G fortyGig 0/52 40G fortyGig 0/56 fortyGig 0/60 Management 0/0 1G Copper LACP LACP LACP LACP LACP LACP - 13 LACP <AGGR EGATI ON PROTO COL> <AGGR EGATI ON PROTO COL> - <1G Switch> <10G Switch>-1 Te 0/51 fortyGig 1/48 - <10G Switch>-1 fortyGig 1/52 - <1G Switch> Gi 0/47 100,120,160 7 8 9 10 11 12 - - Switch “show running-config” output <Provide a complete configuration of all three switches here>