CLI Guide
Table Of Contents
- Dell EMC Storage Systems CLI Guide for PowerStore and UnityXT metro node feature
- Introduction
- Using the VPlexcli
- Commands
- advadm dismantle
- alias
- array claim
- array forget
- array re-discover
- array used-by
- back-end degraded list
- back-end degraded recover
- batch-migrate cancel
- batch-migrate check-plan
- batch-migrate clean
- batch-migrate commit
- batch-migrate create-plan
- batch-migrate pause
- batch-migrate remove
- batch-migrate resume
- batch-migrate start
- batch-migrate summary
- capture begin
- capture end
- capture pause
- capture replay
- capture resume
- cd
- cluster add
- cluster configdump
- cluster expel
- cluster forget
- cluster show-remote-devices
- cluster shutdown
- cluster status
- cluster summary
- cluster unexpel
- collect-diagnostics
- configuration get-product-type
- configuration join-clusters
- configuration metadata-backup
- configuration remote-clusters add-addresses
- configuration remote-clusters clear-addresses
- configuration show-meta-volume-candidates
- configuration upgrade-meta-slot-count
- connect
- connectivity director
- connectivity list all
- connectivity list directors
- connectivity list initiators
- connectivity list storage-volumes
- connectivity show
- connectivity validate-be
- connectivity validate-local-com
- connectivity validate-wan-com
- consistency-group add-virtual-volumes
- consistency-group choose-winner
- consistency-group convert-to-local
- consistency-group create
- consistency-group destroy
- consistency-group list-eligible-virtual-volumes
- consistency-group remove-virtual-volumes
- consistency-group resolve-conflicting-detach
- consistency-group resume-at-loser
- consistency-group set-detach-rule no-automatic-winner
- consistency-group set-detach-rule winner
- consistency-group summary
- date
- describe
- device attach-mirror
- device collapse
- device detach-mirror
- device mirror-isolation auto-unisolation disable
- device mirror-isolation auto-unisolation enable
- device mirror-isolation disable
- device mirror-isolation enable
- device mirror-isolation show
- device resume-link-down
- device resume-link-up
- device resurrect-dead-storage-volumes
- director commission
- director decommission
- director fc-port-stats
- director firmware show-banks
- director forget
- director passwd
- director ping
- director shutdown
- director tracepath
- director uptime
- dirs
- disconnect
- dm migration cancel
- dm migration clean
- dm migration commit
- dm migration pause
- dm migration remove
- dm migration resume
- dm migration start
- drill-down
- ds dd convert-to-local
- ds dd create
- ds dd declare-winner
- ds dd destroy
- ds dd remove-all-rules
- ds dd set-log
- ds rule destroy
- ds rule island-containing
- ds rule-set copy
- ds rule-set create
- ds rule-set destroy
- ds rule-set what-if
- ds summary
- esrs status
- esrs un-register
- exec
- exit
- export initiator-port discovery
- export initiator-port register
- export initiator-port register-host
- export initiator-port show-logins
- export initiator-port unregister
- export port summary
- export storage-view addinitiatorport
- export storage-view addport
- export storage-view addvirtualvolume
- export storage-view checkconfig
- export storage-view create
- export storage-view destroy
- export storage-view find
- export storage-view find-unmapped-volumes
- export storage-view map
- export storage-view removeinitiatorport
- export storage-view removeport
- export storage-view removevirtualvolume
- export storage-view show-powerpath-interfaces
- export storage-view summary
- export target-port renamewwns
- extent create
- extent destroy
- extent summary
- find
- front-end-performance-stats start
- front-end-performance-stats status
- front-end-performance-stats stop
- getsysinfo
- health-check
- help
- history
- local-device create
- local-device destroy
- local-device summary
- log filter create
- log filter destroy
- log filter list
- log source create
- log source destroy
- log source list
- logging-volume add-mirror
- logging-volume create
- logging-volume detach-mirror
- logging-volume destroy
- logical-unit forget
- ls
- manifest upgrade
- manifest version
- meta-volume attach-mirror
- meta-volume backup
- meta-volume create
- meta-volume destroy
- meta-volume detach-mirror
- meta-volume move
- meta-volume verify-on-disk-consistency
- monitor add-console-sink
- monitor add-file-sink
- monitor collect
- monitor create
- monitor destroy
- monitor get-stats
- monitor remove-sink
- monitor stat-list
- ndu pre-check
- ndu pre-config-upgrade
- ndu recover
- ndu start
- ndu status
- plugin addurl
- plugin listurl
- plugin register
- popd
- pushd
- rebuild set-transfer-size
- rebuild show-transfer-size
- rebuild status
- report capacity-arrays
- report capacity-clusters
- report capacity-hosts
- rm
- schedule add
- schedule list
- schedule modify
- schedule remove
- scheduleSYR add
- scheduleSYR list
- scheduleSYR remove
- script
- sessions
- set
- set topology
- show-use-hierarchy
- sms dump
- source
- storage-tool dismantle
- storage-tool compose
- storage-volume auto-unbanish-interval
- storage-volume claim
- storage-volume claimingwizard
- storage-volume find-array
- storage-volume forget
- storage-volume list-banished
- storage-volume list-thin-capable
- storage-volume resurrect
- storage-volume summary
- storage-volume unbanish
- storage-volume unclaim
- storage-volume used-by
- syrcollect
- tree
- unalias
- validate-system-configuration
- version
- virtual-volume create
- virtual-volume destroy
- virtual-volume expand
- virtual-volume list-thin
- virtual-volume re-initialize
- virtual-volume set-thin-enabled
- virtual-volume summary
- wait
- webserver
- Index
When the inter-cluster link is restored, the clusters learn that I/O has proceeded independently.
I/O continues at both clusters until the administrator picks a winning cluster whose data image will be used as the source to
synchronize the data images.
Use this command to pick the winning cluster. For the distributed volumes in the consistency group:
● I/O at the losing cluster is suspended (there is an impending data change)
● The administrator stops applications running at the losing cluster.
● Any dirty cache data at the losing cluster is discarded
● The legs of distributed volumes rebuild, using the legs at the winning cluster as the rebuild source.
When the applications at the losing cluster are shut down, use the consistency-group resume-after-data-loss-
failure command to allow the system to service I/O at that cluster again.
Example
Select cluster-1 as the winning cluster for consistency group “TestCG” from the TestCG context:
VPlexcli:/clusters/cluster-1/consistency-groups/TestCG> resolve-conflicting-detach
This will cause I/O to suspend at clusters in conflict with cluster cluster-1, allowing
you to stop applications at those clusters. Continue? (Yes/No) yes
Select cluster-1 as the winning cluster for consistency group “TestCG” from the root context:
VPlexcli:/> consistency-group resolve-conflicting-detach --cluster cluster-1 --
consistency-group /clusters/cluster-1/consistency-groups/TestCG
This will cause I/O to suspend at clusters in conflict with cluster cluster-1, allowing
you to stop applications at those clusters. Continue? (Yes/No) Yes
In the following example, I/O has resumed at both clusters during an inter-cluster link outage. When the inter-cluster link is
restored, the two clusters will come back into contact and learn that they have each detached the other and carried on I/O.
● The ls command shows the operational-status as ok, requires-resolve-conflicting-detach at both clusters.
● The resolve-conflicting-detach command selects cluster-1 as the winner.
Cluster-2 will have its view of the data discarded.
I/O is suspended on cluster-2.
● The ls command displays the change in operational status.
○ At cluster-1, I/O continues, and the status is ok.
○ At cluster-2, the view of data has changed and so I/O is suspended pending the consistency-group resume-at-
loser command.
VPlexcli:/clusters/cluster-1/consistency-groups/cg1> ls
Attributes:
Name Value
-------------------- -----------------------------------------
active-clusters [cluster-1, cluster-2]
cache-mode synchronous
detach-rule no-automatic-winner
operational-status [(cluster-1,{ summary:: ok, details:: [requires-resolve-
conflicting-detach] }),
(cluster-2,{ summary:: ok, details:: [requires-resolve-conflicting-detach] })]
passive-clusters []
read-only false
recoverpoint-enabled false
storage-at-clusters [cluster-1, cluster-2]
virtual-volumes [dd1_vol, dd2_vol]
visibility [cluster-1, cluster-2]
Contexts:
advanced recoverpoint
VPlexcli:/clusters/cluster-1/consistency-groups/cg1> resolve-conflicting-detach -c
cluster-1
This will cause I/O to suspend at clusters in conflict with cluster cluster-1, allowing
you to stop applications at those clusters. Continue? (Yes/No) Yes
VPlexcli:/clusters/cluster-1/consistency-groups/cg1> ls
Attributes:
Commands
107