Sei sulla pagina 1di 38

VMware vSphere 5 on Cisco UCS and EMC VNXe

Technical Integration Howto

Kenny Garreau @kennega 10/28/2012

VMware vSphere 5 on Cisco UCS and EMC VNXe

Table of Contents
Introduction ............................................................................................................................................ 2 Intended Audience .............................................................................................................................. 2 Special Thanks ..................................................................................................................................... 2 Vendor Products...................................................................................................................................... 3 VMware vSphere 5 .............................................................................................................................. 3 Cisco UCS B-series Servers ................................................................................................................... 3 EMC VNXe Unified Storage .................................................................................................................. 3 Implementation....................................................................................................................................... 4 Architecture ........................................................................................................................................ 4 Challenges ........................................................................................................................................... 5 Best Practices ...................................................................................................................................... 7 Cisco UCS Configuration .......................................................................................................................... 7 Create Pools and Policies ..................................................................................................................... 7 Create vNIC Templates ........................................................................................................................ 9 Create Service Profile Template ......................................................................................................... 10 Deploy and Modify Service Profiles .................................................................................................... 17 EMC VNXe Configuration ....................................................................................................................... 18 Add Generic iSCSI Storage.................................................................................................................. 18 vSphere Installation............................................................................................................................... 24 ESXi Network Configuration ............................................................................................................... 28 ESXi iSCSI Configuration ..................................................................................................................... 30 EMC VNXe Configuration Redux ............................................................................................................ 32 Add VMware Storage......................................................................................................................... 32 Conclusion............................................................................................................................................. 36 References ............................................................................................................................................ 37

VMware vSphere 5 on Cisco UCS and EMC VNXe

Introduction
Being in the technical solutions and integration space will always present challenges, and the more technologies involved in that integration, the greater the challenge. From a technological perspective, the protocols and products in this whitepaper arent the most advanced in the market, but I found some unique issues that arose during the work. Many of the troubleshooting steps I followed that were outlined on the Cisco, EMC and VMware forums didnt apply, or ended up being red herrings to the true problem. I spent a lot of time sifting through reference documentation from each vendor, and unfortunately the trusty Google search for EMC VNXe UCS yielded less fruitful information than I had hoped to find. That is why I decided to pen this whitepaper.

Intended Audience
This document is meant to provide specific technical guidance with regards to deploying VMware vSphere 5 in an environment backed by Cisco UCS B-series blade servers and EMCs VNXe storage offering. These are a collection of best practices from Cisco, VMware and EMC, as well as my professional recommendations and experiences from several of the deployments I have been charged with. The knowledge level assumed is the ability to deploy all three technologies in at least an SMB environment. Standard setup tasks such as rack and stack, array/chassis cabling and basic setup and initialization procedures are outside the scope of this document.

Special Thanks
Before I get started, putting this together would not have been possible without the help of a couple good friends. Jason Goldrick (@J_Goldrick) for all your help with the VNXe minutiae. Anderson Nichols (@badrockjones) the best technical writer I know!

VMware vSphere 5 on Cisco UCS and EMC VNXe

Vendor Products
Covered in this whitepaper is the integration of the following technologies: VMware vSphere 5 Cisco UCS B-series Servers EMC VNXe Unified Storage

VMware vSphere 5
The deployment here utilized VMware vSphere 5.0, with ESXi 5.0 U1 hosts deployed from Ciscos vendor-specific ISO. This is available from VMwares Downloads site, under the vSphere 5.0 download page, under Drivers and Tools. vCenter 5.0 U1b was installed and configured as the management platform for the underlying ESXi hypervisor hosts.

Cisco UCS B-series Servers


Im a huge advocate of Ciscos UCS blade servers. They are flexible, scalable and highly customizable. However, Cisco assumes a certain level of knowledge when using all of its products, and UCS is no different. The platform is extremely powerful, but you will need to know what you are configuring and more importantly, why youre configuring it in said fashion. The level of control you have over even the smallest parts or policies within the system, Service Profile mobility, as well as a low entry cost into 10 Gigabit networking makes it the best server platform on the market. Components used in this whitepaper are Ciscos 5108 UCS blade chassis, 6248UP Fabric Interconnects, 2208XP Fabric Extenders, B200 M3 blade servers and the Cisco VIC1280 mezzanine card.

EMC VNXe Unified Storage


For SMB customers looking to deploy an entry-level unified storage platform, EMC offers the VNXe platform in two models, the VNXe 3150 and the VNXe 3300. Configuration options will vary based on the customers storage, performance and connectivity requirements. Packing advanced storage features such as Flash drives, multi-protocol support and 10GbE connectivity, it is a fantastic entry-level platform for the SMB market. This installation included 10GbE iSCSI modules for block storage.

VMware vSphere 5 on Cisco UCS and EMC VNXe

Implementation
Architecture
Now that weve covered the key components involved, lets discuss how all of this is going to fit together. Keep in mind that most SMB customers who have purchased this type of solution arent going to have an existing 10GbE switching infrastructure. This will be their first exposure to 10Gb, and this may even be their first time using a dedicated storage or virtualization platform! There are lots of moving parts, so lets cover the basics. We have four 10GbE SFP+ Twinax cables going from each 2208 Fabric Extender (FEX) to its corresponding 6248UP Fabric Interconnect (FI). The customer does NOT have an existing 10GbE switching infrastructure, so in this scenario we have opted to utilize UCS Appliance Ports to connect the VNXes 10GbE iSCSI modules. Cisco UCS bundles will typically include four SFP -10GB-SRs, which are 10GbE optical SFPs youll need for connectivity between those 10GbE modules in the VNXe and the 6248 FIs1. Two appliance ports will be configured per FI, so you should cable module ports A-0 and B-1 to FI-A, then module ports A-1 and B-0 to FI-B. This customer had a single-switch core, and many SMBs will have a similar setup, or perhaps a stack of Cisco Catalyst 3750 switches. Connectivity choices upstream can vary wildly, but a quick Google search for Cisco UCS Networking Best Practices will yield plenty of results 2. For this install, the customer has two two-port 1GbE LACP bundles from each FI into a single switch, so we acquired four GLC-T GBICs and placed two into each FI and configured two LACP port-channels within UCS Manager. In order to enable Service Profile mobility, each blade server has a VIC1280 mezzanine card installed with no local drives. We will boot each Service Profile (ESXi host) from an iSCSI LUN presented via the VNXe, and shared iSCSI datastores for VMware virtual machine storage will be presented to each host and formatted with VMFS-5. The following diagram shows the Visio layout of how weve connected the components:

You may be able to unplug the 10GbE SFPs in the VNXes modules a nd use 10GbE SFP+ Twinax cabling, but I have not tested this. EMC VNX requires Active Twinax cabling, which may also be a requirement for the VNXe. Hit me up on Twitter (@kennega) if youve tested this! 2 See also the References section of this document.

VMware vSphere 5 on Cisco UCS and EMC VNXe

Challenges
Throughout the course of testing and validating this solution, I ran into quite a few issues. As of now they seem quite simple and obvious, but during the integration and deployment, things arent always so clear. The biggest challenge with this V+C+E integration is the lack of Storage Groups within the VNXe. In a VNX, LUNs are assigned to a storage group, and that storage group is in turn masked to a host. Therefore, masking a boot LUN ID of 0 to each host is uniform throughout the environment. The VNXe however, has no concept of Storage Groups. LUNs are assigned at the iSCSI Server level within the system, at which point you assign Virtual Disk or Datastore access to a Host. Its a slightly different way of thinking than with a mid-range or enterprise storage array. While this impacts how you end up deploying and organizing your storage, it also has an effect on your UCS Service Profiles and Service Profile Templates.

VMware vSphere 5 on Cisco UCS and EMC VNXe In a VNX-backed UCS deployment on FC or iSCSI, the array storage group decides how to mask LUN IDs. Therefore each Service Profile (host) could be assigned a 5GB boot LUN within the storage group. Service Profiles can always stay bound to, and in compliance with, a Service Profile Template within UCS, since LUN 0 will always be the boot LUN. This is not so with the VNXe. Since the iSCSI Server will assign LUN IDs in ascending order, the Service Profiles will have to be unbound from the Service Profile Template in order to modify the iSCSI Boot Parameters. Theres also an order of operations to be followed that I will outline shortly. Its certainly not efficient by any means, but its something youll need to take into consideration for properly automated VNXe/vSphere integration. Another thing to keep in mind is that, as an integrator, when you leave the customer site, youll have to be aware that if and when the customer purchases a new blade or set of blades to install into the chassis, the above considerations should be outlined to them, or scoped into any future professional services engagement. The Service Profile Templates will get them 95% of the way there with the VNXe, but that last 5% will come in the form of unbinding the new Service Profiles and editing the iSCSI Boot Parameters under the Boot Order tab of the Service Profile. Finally, be sure you size your pools to at least the capacity of the UCS chassis youre deploying. You could create your MAC/IQN/Initiator IP pools with a size of 4, but why not size it to the capacity of the chassis? This will make future expansion a plug and play operation until you fill the capacity of that chassis.

VMware vSphere 5 on Cisco UCS and EMC VNXe

Best Practices
The following best practices are followed as part of this implementation: Dedicated storage VLANs for iSCSI fabrics; one for A side, one for B side. Differing subnets in each storage VLAN. In this deployment, we utilize 172.16.10.0/24 and 172.16.11.0/24 LUN masking at the array for boot LUNs and vSphere datastores Do not use link aggregation for any iSCSI interfaces iSCSI vmkernel port binding within vSphere iSCSI vmkernel port groups are contained within a dedicated vSwitch Balanced storage presentation from iSCSI Server A and iSCSI Server B All boot LUNs presented from one iSCSI Server Boot LUNs should always utilize the lowest tier of storage available iSCSI Overlay vNICs do not utilize a MAC address (inherited from parent vNIC) Do not enable Fabric Failover for iSCSI-bound vNICs All iSCSI connections are set as access ports to their respective iSCSI VLANs Jumbo frames are enabled end-to-end Utilize Round Robin PSP within vSphere for storage multipathing

Cisco UCS Configuration


From this point forward, I assume that all racking, stacking, cabling, uplinks, hardware/software initialization and storage pool creation have been completed. At this stage, the environment is ready to have Service Profile Templates and Services Profiles configured, boot LUNs to be created and assigned to initiators, and for ESXi to be installed on the boot LUN.

Create Pools and Policies


Create the following pools within UCS Manager: Management IP, iSCSI Initiator IP, IQN, MAC, Server, UUID. For MAC pools, I recommend creating several MAC pools, and in my example here I create four MAC pools. One pool each for iSCSI vNICs on FI-A, iSCSI vNICs on FI-B, SC/VM vNICs on FI-A and SC/VM vNICs on FI-B. This is extremely useful in troubleshooting connectivity issues! Under the LAN tab, expand Policies Appliances Network Control Policies, right-click and select Create Network Control Policy. Name the policy something descriptive, in my example Ive called it iSCSI_CDP. Choose to Enable CDP, and leave all other settings at their default. Click OK:

VMware vSphere 5 on Cisco UCS and EMC VNXe

Create your Appliance Interfaces, and be sure to set them as Access ports to the appropriate iSCSI VLAN. For ease, Ive created a VLAN label iSCSI and assigned VLAN 500 on FI-A and VLAN 501 on FI-B. Remember, youll need to create your iSCSI VLANs in two places: under Appliances Fabric A/B VLANs, and under LAN LAN Cloud Fabric A/B VLANs. While creating the Appliance Interfaces, be sure to assign the Network Control Policy you just created:

VMware vSphere 5 on Cisco UCS and EMC VNXe

Under the LAN tab, under Policies root Network Control Policies, right-click and Create Network Control Policy. Wait... Didnt we do this above? Sort of. We created a Network Control Policy before for our Appliance Ports. This policy were creating will be applied to our vNIC Templates were about to create. Set the same options as before however: Enable CDP, leave all others at their defaults. Give it a descriptive name and click OK.

Create vNIC Templates


Next well be creating vNIC Templates. For an SMB customer, I generally recommend the following vNIC layout: One vNIC per Fabric for Service Console and vMotion traffic One vNIC per Fabric for Virtual Machine traffic One vNIC per Fabric for iSCSI traffic (do not enable Fabric Failover for these vNIC Templates!)

Create your six vNIC Templates, assign the appropriate VLANs, and be sure you set MTU 9000 for your iSCSI vNIC Templates. Unfortunately I didnt get a screen grab of my vNIC Templates as a n example, but I did get a screen cap of the vNICs within the Service Profile Templates as they were built from the vNIC Connectivity Templates:

Youll see above that this iSCSI vNIC within the Service Profile Template is bound to my iSCSI_A LAN template, is owned by Fabric A WITHOUT Failover, has an MTU of 9000, Adapter Policy set to VMware and Network Control Policy set to my CDP-enabled NCP I created earlier. 9

VMware vSphere 5 on Cisco UCS and EMC VNXe At this stage, you can create individual Policies in the Servers tab, or you can elect to Create Service Profile Template and have the template prompt and create these policies for you. Ill walk these steps as if creating them from the Create Service Profile Template Wizard, but approach however youd like.

Create Service Profile Template


Create a Service Profile Template. Name it, select Initial or Updating, select a UUID pool you created earlier, and click Next.

Select Create Local Disk Configuration Policy, give it a name. Ive called mine NoLocalStorage - select No Local Storage from the drop-down, uncheck Protect Configuration and click OK. Choose the policy you just created from the drop-down. Under How would you like to configure SAN connectivity? select No vHBAs. Click Next.

10

VMware vSphere 5 on Cisco UCS and EMC VNXe

On the Networking screen, skip any configuration of Dynamic vNIC Connection Policies. Select Expert for configuring LAN connectivity, and now create all of your vNICs for the Service Profile Template. Select Add, Name your vNIC to be attached to the Service Profile Template, and check Use LAN Connectivity Template select your vNIC Template created earlier, set your Adapter Policy to VMware.

Create a vNIC for each vNIC template you created earlier. Your screen should look similar to this:

11

VMware vSphere 5 on Cisco UCS and EMC VNXe

Expand iSCSI vNICs and select Add:

Give the iSCSI vNIC a name, select the vNIC created just now to Overlay iSCSI traffic, and select the appropriate VLAN. Click Create iSCSI Adapter Policy and create a policy with a Connection Timeout of 30s and Busy Retry of 30s:

12

VMware vSphere 5 on Cisco UCS and EMC VNXe

Click OK, click OK, and click Next. Under the vNIC/vHBA Placement screen, I recommend moving the iSCSI vNICs to last in the placement order. If you leave the iSCSI vNICs as first in the placement order, they will be assigned as vmnic0 and vmnic1 in the vmkernel. This in itself is not a bad thing until you attempt to load ESXi onto the LUN you assign to the Service Profile. At the end of the install process, ESXi will attempt to re-initialize vmnic0 as DHCP and the Service Profile will lose IP connectivity to the iSCSI target. Youll wind up with an error a s follows:

13

VMware vSphere 5 on Cisco UCS and EMC VNXe So be safe and change the placement order:

On the Server Boot Order screen, create a Boot Policy to use. Boot off the CD-ROM first, and then add an iSCSI Boot option. Be sure you use the same iSCSI Overlay vNIC name you created earlier. Click OK.

14

VMware vSphere 5 on Cisco UCS and EMC VNXe Select the boot policy from the drop-down. Select the iSCSI option, and click Set iSCSI Boot Parameters. Select a pool for Initiator Name assignment, and set Initiator Addressing to Pool. Now select to Add an iSCSI Target Interface. At this point youll need to head over to the VNXe quickly to copy/paste in your target IQN. Under Settings iSCSI Server Settings you can select your iSCSI Server and select Details, and on this screen youll be able to copy the IQN.

Paste the IQN of the VNXe iSCSI Server into the Name field, set the IPv4 address to the iSCSI Server Aside IP address. The default LUN ID of 0 is acceptable, this will change for the Service Profiles anyway when we un-bind them from the Template after deploying. Note: I only deploy one iSCSI vNIC at this time. The reason for this is that UCS currently does not have the ability to deploy multiple IP pools for iSCSI vNICs. It would be great if this feature was enabled as wed be able to simply create two iSCSI vNICs, bind distinct IPs to each initiator on different subnets, and VMware would be crafty enough to recognize both of them and assign vmkernel ports to them. Our goal is to automate what we can though, so that is why I do not create a second iSCSI vNIC. Click OK, and your screen should look like this:

Click OK, then click Next.

15

VMware vSphere 5 on Cisco UCS and EMC VNXe Create a Maintenance Policy that requires User Ack before reboot and click Next:

Select a Server Assignment Pool to use, your desired power state at association, and any Pool Qualifications or Firmware Management Policies youd like to apply, if any. Click Next.

16

VMware vSphere 5 on Cisco UCS and EMC VNXe Set any operational policies you wish to set, and click Finish. Your Service Profile Template should be created, and at this point you can deploy Service Profiles from this template.

Deploy and Modify Service Profiles


At this point you should deploy X number of Service Profiles, where X is the number of blades within the chassis that youre deploying. Its easiest to deploy all at once, then you can unbind each individually to apply specific iSCSI Boot Parameters to each blade. When deploying, IPs, MACs, IQNs, etc. will all be assigned and visible within each Service Profile. Deploy your Service Profiles, and then select the first in the list. Right-click and select Unbind from the Template. Select the Boot Order tab and select the iSCSI boot at the bottom. Click Set iSCSI Boot Parameters. Youll see that you can now modify all of the settings, including Pools and IQNs, but you should leave these as-is. Remember, automate! Select the Static Target at the bottom and select Modify. Youll be able to change/update the boot LUN ID that the iSCSI Server on the VNXe will present. When I am doing a net-new installation for a customer, I utilize the first several LUNs created on the VNXe iSCSI Server as my boot LUNs. You can see in the following screenshots the difference between Service Profile 1s boot LUN config and Service Profile 2s boot LUN config: (note the LUN ID column, and difference in initiator IPs and IQNs)

17

VMware vSphere 5 on Cisco UCS and EMC VNXe

After you modify these parameters and click OK and Save Settings, youll be prompted with a User Acknowledgment to reboot the blade to apply the new parameters. Click Yes, and under Pending Activities, acknowledge the reboot. You can do these one at a time for each Service Profile, or edit and reboot them all in bulk. Dont close UCS Manager quite yet; well need some information for the VNXe.

EMC VNXe Configuration


Add Generic iSCSI Storage
Well need to add Generic iSCSI Storage at first so that we can provide boot LUNs to the ESXi h osts. Creating VMware iSCSI Storage within the VNXe attempts to create VMFS volumes automatically, and thats not what were looking to do initially. At the Dashboard screen after logging into the VNXe, select Create storage for generic iSCSI. The first screen asks you to name the generic storage. Give it a descriptive name; here Ive named it ESX13:

Id recommend something a little more descriptive, maybe ESXnBootLUN, where n is the Service Profile number. I did a bad job here of clearly delineating between the Host ID and Storage ID.

18

VMware vSphere 5 on Cisco UCS and EMC VNXe

Click Next. Choose a Pool from which to provision the LUN, preferably the pool with the lowest performance. Set your LUN size to 5GB, and since you can have two iSCSI Servers, make sure youre provisioning from iSCSI Server A:

19

VMware vSphere 5 on Cisco UCS and EMC VNXe Click Next. Select to Not Configure Protection Storage:

Click Next. Under Configure Host Access, select Create Host:

20

VMware vSphere 5 on Cisco UCS and EMC VNXe Specify a name for the host:

Click Next. Under the Operating System dropdown, select VMware ESX:

21

VMware vSphere 5 on Cisco UCS and EMC VNXe Click Next. Select IP address, and enter the IP address of the iSCSI initiator that corresponds with this host. Youll find that under the Service Profiles iSCSI Boot Parameters that we modified earlier.

Click Next. Grab the IQN off the iSCSI Boot Parameters property sheet as well, and input that on this screen.

22

VMware vSphere 5 on Cisco UCS and EMC VNXe Click Next. Verify the Summary information and click Finish. Youll return to this page:

Select Virtual Disk from the Access drop-down. Click Next. Verify the summary information, and click Finish.

23

VMware vSphere 5 on Cisco UCS and EMC VNXe You will repeat this process for each Service Profile youre deploying. Once youve finished creating these generic iSCSI LUNs, pop back over to UCS Manager and its time to begin the process of installing from the Cisco-customized ESXi 5.0 ISO.

vSphere Installation
Log back into the UCS Manager interface, and select your first Service Profile. Reboot the Service Profile since weve added our LUN masking configuration on the VNXe for our boot LUNs. During the boot-up process, if you have disabled the Quiet Boot option in Service Profile BIOS Policy4, you should see the boot LUN appear in the Cisco VIC iSCSI Boot Driver as follows:

If you do NOT see the above, do not proceed with mounting the Cisco ISO to the Virtual Media drive. Double-check the configuration items preceding this step. The command-line for nxos within the FIs is extremely useful. Verify that the MAC address for the VNXes iSCSI Server IP is assigned to the proper VLAN, and that you are seeing the proper iSCSI vNIC MAC addresses on that VLAN as well:

Id highly recommend disabling Quiet Boot in the BIOS policies f or your Service Profiles for the initial part of the implementation. Its extremely useful to be able to look at the POST process and be sure your VIC is initializing and seeing the boot LUN properly.

24

VMware vSphere 5 on Cisco UCS and EMC VNXe

This is actually a bad example as only one port is active currently (Eth1/24), which doesnt have an IP assigned to it within the VNXe yet. But a show mac address-table vlan 500 will show the MAC membership for your iSCSI VLAN on FI-A. If you are seeing MAC addresses on the opposite FIs that youre expecting, you may need to edit your iSCSI Server configuration within the VNXe, or do some cable swaps between FIs. If all is well, mount your Cisco custom ISO in the Virtual Media tab:

25

VMware vSphere 5 on Cisco UCS and EMC VNXe After booting from the ISO, youll land at the ESXi Installer welcome screen:

Run through the install prompts, and when the installer scans for disks, you should see the following:

26

VMware vSphere 5 on Cisco UCS and EMC VNXe After walking through the rest of the prompts, youll again be prompted to install to your iSCSI boot device. Hit F11 to confirm and run the install.

After the install finishes, the virtual media ISO will be ejected and the Service Profile will reboot. Youll need to perform this procedure for each Service Profile. If you dont see your boot LUN at first, dont panic. Sometimes its as simple as a Service Profile Reset and on the next boot it will discover the target and login successfully.5

CLI tools are invaluable for troubleshooting. From the adapter prompt, the iscsi_ping and iscsi_get_config commands display a wealth of information. However, they are only available when booting or when the server has entered the BIOS screen.

27

VMware vSphere 5 on Cisco UCS and EMC VNXe

ESXi Network Configuration


You should now complete basic setup of the ESXi host. From the DCUI, configure IP addressing and continue basic configuration via the vSphere Client. The applicable configuration items for iSCSI vmkernel port binding, network configuration and iSCSI multipathing are represented here via the GUI, but you can also complete these tasks via the command line. In the vSphere Client, when you connect to the ESXi host and look at the Configuration tab under Networking, you should see two vSwitches created for you; vSwitch0 will contain your Management Network vmkernel port. The ESXi install will also have created an iScsiBootvSwitch with an iScsiBootPG vmkernel port. The following graphic shows this, as well as the fact that Ive created a second Port Group and vmkernel port attached to iScsiBootvSwitch, as well as vmnic5, my second iSCSI vNIC.

A pre-requisite for iSCSI vmkernel port binding is to bind each vmkernel port in the iSCSI vSwitch to its own vmnic by overriding the vSwitch Failover order. For the iScsiBootPG port group, edit the NIC Teaming properties and select to override the switch failover order:

28

VMware vSphere 5 on Cisco UCS and EMC VNXe

Commit this change and repeat the same process for the other port group, except youll reverse the adapter usage:

At this point youll be able to configure iSCSI vmkernel port binding.6

Im not quite sure why these vmnics show a speed of 20000. This usually happens when Fabric Failover is enabled, as the NIC technically sees two 10Gb connections. I can assure you FF is not enabled for the vNICs backing these vmnics however.

29

VMware vSphere 5 on Cisco UCS and EMC VNXe

ESXi iSCSI Configuration


To configure iSCSI vmkernel port binding, youll need to edit the vmhba properties for the iSCSI adapter. From the vSphere Client, go to Configuration Storage Adapters, right-click the iSCSI Software Initiator, and select Properties. Select the Network Configuration tab:

Click Add, and select a vmkernel port to use. Both of the vmkernel ports you created before should show as compliant; if they do not, go back and double-check your bindings for the port groups.

30

VMware vSphere 5 on Cisco UCS and EMC VNXe Click OK to bind one vmkernel port, and then click Add and add the other vmkernel port. Your Network Configuration screen should show a summary as follows:

You can choose to add discovery targets now, but the VNXes vSphere integration will actually do all of this for you.

31

VMware vSphere 5 on Cisco UCS and EMC VNXe

EMC VNXe Configuration Redux


Our Service Profile hosts have already been added to the VNXe, but theyve been added as generic initiators. To take advantage of the VNXes vSphere integration features, its necessary to delete the Generic Host and re-add it as a VMware Host. After that well need to re-add the generic boot LUN to the host, as well as configure its shared storage. This should be done while the Service Profile is online. If you remove the host and try to re-add it while the server is offline, the operation will fail as this process requires an active Management Network connection to the ESXi host.

Add VMware Storage


From the VNXe Unisphere window, select Hosts and highlight your first Service Profile. Select Delete to delete the Host from the VNXe. This does not delete any storage associated with the Host; it simply removes the Host record from the array:

Navigate to Hosts VMware and select Find ESX Hosts.

32

VMware vSphere 5 on Cisco UCS and EMC VNXe

Click Next, and enter the IP address of the Service Profiles Management Network vmkernel port. Then click Find.

33

VMware vSphere 5 on Cisco UCS and EMC VNXe Youll need to then enter credentials for the host 7, and click OK.

At this point the VNXe should find and recognize the host as an ESXi server.

If the password is ever changed (or the account deleted if you use a non-root account) youll need to be sure to update the credentials for the host(s) affected.

34

VMware vSphere 5 on Cisco UCS and EMC VNXe Click Next, and click Finish on the Summary screen. You should now see a host listed in the Virtualization Hosts list:

The host has now been re-added to the VNXe, but well need to make sure we re-assign its boot LUN to the host. Navigate to Storage Generic iSCSI Storage and select the boot LUN for the host you just readded. Click Details, and then select Host Access.

35

VMware vSphere 5 on Cisco UCS and EMC VNXe Here youll see a list of available hosts that you can present the LUN to. Notice the difference between the newly-created VMware host and the pre-existing Generic hosts. The Generic hosts have only registered a single initiator IP, whereas the VMware host has registered three. Two will be your iSCSI vmkernel ports, and one will be the Management Network vmkernel port. 8 Select the Access drop-down for the host youll be adding the disk to, and select Virtual Disk. Click Apply Changes. That is the only modification required for re-adding the boot LUN. Remember, there are no Storage Groups the LUNs are assigned per iSCSI Server, so the LUN ID will never change.

At this point, youll be able to add shared datastores to your ESXi hosts by selecting Create Storage for VMware from the Dashboard. This is a well-covered topic in EMCs documentation, as well as on several blogs and YouTube, so for the sake of brevity Ill leave those instructions to their docs.9

Conclusion
I hope this document is of assistance to anyone looking to virtualize with VMware vSphere on Cisco Unified Compute Servers backed by EMCs VNXe on iSCSI. If you have any feedback or questions, or just want to give a shout out that this was helpful, I can be found on Twitter @kennega, or via my website: http://www.dudewheresmycloud.com/

8 9

This is for reference only. The host will never try to use the Management Network to log in via iSCSI. Depending on the VNXe code revision, the datastores created by the VNXe vSphere engine may create the datastores as VMFS-3 instead of VMFS-5. The newest code rev from EMC is supposed to address this.

36

VMware vSphere 5 on Cisco UCS and EMC VNXe

References
EMC VNXe High Availability https://community.emc.com/docs/DOC-12551 Multipathing Configuration for Software iSCSI Using Port Binding http://www.vmware.com/files/pdf/techpaper/vmware-multipathing-configuration-software-iSCSI-portbinding.pdf Cisco UCS Manager GUI Configuration Guide http://www.cisco.com/en/US/docs/unified_computing/ucs/sw/gui/config/guide/2.0/b_UCSM_GUI_Conf iguration_Guide_2_0_chapter_011101.html#concept_D7BF302366F24CF5A602B0E0BD18787C Cisco UCS Networking Best Practices http://bradhedlund.com/2010/06/22/cisco-ucs-networking-best-practices/

37

Potrebbero piacerti anche