6.0.1
Table Of Contents
- vSphere Storage
- Contents
- About vSphere Storage
- Updated Information
- Introduction to Storage
- Overview of Using ESXi with a SAN
- Using ESXi with Fibre Channel SAN
- Configuring Fibre Channel Storage
- Configuring Fibre Channel over Ethernet
- Booting ESXi from Fibre Channel SAN
- Booting ESXi with Software FCoE
- Best Practices for Fibre Channel Storage
- Using ESXi with iSCSI SAN
- Configuring iSCSI Adapters and Storage
- ESXi iSCSI SAN Requirements
- ESXi iSCSI SAN Restrictions
- Setting LUN Allocations for iSCSI
- Network Configuration and Authentication
- Set Up Independent Hardware iSCSI Adapters
- About Dependent Hardware iSCSI Adapters
- Dependent Hardware iSCSI Considerations
- Configure Dependent Hardware iSCSI Adapters
- About the Software iSCSI Adapter
- Modify General Properties for iSCSI Adapters
- Setting Up iSCSI Network
- Using Jumbo Frames with iSCSI
- Configuring Discovery Addresses for iSCSI Adapters
- Configuring CHAP Parameters for iSCSI Adapters
- Configuring Advanced Parameters for iSCSI
- iSCSI Session Management
- Booting from iSCSI SAN
- Best Practices for iSCSI Storage
- Managing Storage Devices
- Storage Device Characteristics
- Understanding Storage Device Naming
- Storage Refresh and Rescan Operations
- Identifying Device Connectivity Problems
- Edit Configuration File Parameters
- Enable or Disable the Locator LED on Storage Devices
- Working with Flash Devices
- About VMware vSphere Flash Read Cache
- Working with Datastores
- Understanding VMFS Datastores
- Understanding Network File System Datastores
- Creating Datastores
- Managing Duplicate VMFS Datastores
- Upgrading VMFS Datastores
- Increasing VMFS Datastore Capacity
- Administrative Operations for Datastores
- Set Up Dynamic Disk Mirroring
- Collecting Diagnostic Information for ESXi Hosts on a Storage Device
- Checking Metadata Consistency with VOMA
- Configuring VMFS Pointer Block Cache
- Understanding Multipathing and Failover
- Raw Device Mapping
- Working with Virtual Volumes
- Virtual Machine Storage Policies
- Upgrading Legacy Storage Profiles
- Understanding Virtual Machine Storage Policies
- Working with Virtual Machine Storage Policies
- Creating and Managing VM Storage Policies
- Storage Policies and Virtual Machines
- Default Storage Policies
- Assign Storage Policies to Virtual Machines
- Change Storage Policy Assignment for Virtual Machine Files and Disks
- Monitor Storage Compliance for Virtual Machines
- Check Compliance for a VM Storage Policy
- Find Compatible Storage Resource for Noncompliant Virtual Machine
- Reapply Virtual Machine Storage Policy
- Filtering Virtual Machine I/O
- VMkernel and Storage
- Storage Hardware Acceleration
- Hardware Acceleration Benefits
- Hardware Acceleration Requirements
- Hardware Acceleration Support Status
- Hardware Acceleration for Block Storage Devices
- Hardware Acceleration on NAS Devices
- Hardware Acceleration Considerations
- Storage Thick and Thin Provisioning
- Using Storage Providers
- Using vmkfstools
- vmkfstools Command Syntax
- vmkfstools Options
- -v Suboption
- File System Options
- Virtual Disk Options
- Supported Disk Formats
- Creating a Virtual Disk
- Example for Creating a Virtual Disk
- Initializing a Virtual Disk
- Inflating a Thin Virtual Disk
- Removing Zeroed Blocks
- Converting a Zeroedthick Virtual Disk to an Eagerzeroedthick Disk
- Deleting a Virtual Disk
- Renaming a Virtual Disk
- Cloning or Converting a Virtual Disk or RDM
- Example for Cloning or Converting a Virtual Disk
- Migrate Virtual Machines Between DifferentVMware Products
- Extending a Virtual Disk
- Upgrading Virtual Disks
- Creating a Virtual Compatibility Mode Raw Device Mapping
- Example for Creating a Virtual Compatibility Mode RDM
- Creating a Physical Compatibility Mode Raw Device Mapping
- Listing Attributes of an RDM
- Displaying Virtual Disk Geometry
- Checking and Repairing Virtual Disks
- Checking Disk Chain for Consistency
- Storage Device Options
- Index
If there are no running virtual machines on the VMFS datastore, after you change the ID of the LUN,
you must use rescan to reset the ID on your host. For information on using rescan, see “Storage Refresh
and Rescan Operations,” on page 124.
n
If you need to change the default iSCSI name of your iSCSI adapter, make sure the name you enter is
worldwide unique and properly formaed. To avoid storage access problems, never assign the same
iSCSI name to dierent adapters, even on dierent hosts.
Optimizing iSCSI SAN Storage Performance
Several factors contribute to optimizing a typical SAN environment.
If the network environment is properly congured, the iSCSI components provide adequate throughput and
low enough latency for iSCSI initiators and targets. If the network is congested and links, switches or routers
are saturated, iSCSI performance suers and might not be adequate for ESXi environments.
Storage System Performance
Storage system performance is one of the major factors contributing to the performance of the entire iSCSI
environment.
If issues occur with storage system performance, consult your storage system vendor’s documentation for
any relevant information.
When you assign LUNs, remember that you can access each shared LUN through a number of hosts, and
that a number of virtual machines can run on each host. One LUN used by the ESXi host can service I/O
from many dierent applications running on dierent operating systems. Because of this diverse workload,
the RAID group that contains the ESXi LUNs should not include LUNs that other hosts use that are not
running ESXi for I/O intensive applications.
Enable read caching and write caching.
Load balancing is the process of spreading server I/O requests across all available SPs and their associated
host server paths. The goal is to optimize performance in terms of throughput (I/O per second, megabytes
per second, or response times).
SAN storage systems require continual redesign and tuning to ensure that I/O is load balanced across all
storage system paths. To meet this requirement, distribute the paths to the LUNs among all the SPs to
provide optimal load balancing. Close monitoring indicates when it is necessary to manually rebalance the
LUN distribution.
Tuning statically balanced storage systems is a maer of monitoring the specic performance statistics (such
as I/O operations per second, blocks per second, and response time) and distributing the LUN workload to
spread the workload across all the SPs.
Server Performance with iSCSI
You must consider several factors to ensure optimal server performance.
Each server application must have access to its designated storage with the following conditions:
n
High I/O rate (number of I/O operations per second)
n
High throughput (megabytes per second)
n
Minimal latency (response times)
vSphere Storage
116 VMware, Inc.