Command Line Interface Reference for File

Command Line Interface Reference for File
EMC® VNX™ Series
Release 8.1
VNX™ Command Line Interface Reference for File
P/N 300-014-338 REV 02
EMC Corporation
Corporate Headquarters:
Hopkinton, MA 01748-9103
1-508-435-1000
www.EMC.com
Copyright © 1998-2013 EMC Corporation. All rights reserved. Published in the USA.
Published February 2014
EMC believes the information in this publication is accurate of its publication date. The information is subject
to change without notice.
THE INFORMATION IN THIS PUBLICATION IS PROVIDED AS IS. EMC CORPORATION MAKES NO
REPRESENTATIONS OR WARRANTIES OF ANY KIND WITH RESPECT TO THE INFORMATION IN THIS
PUBLICATION, AND SPECIFICALLY DISCLAIMS IMPLIED WARRANTIES OF MERCHANTABILITY OR
FITNESS FOR A PARTICULAR PURPOSE.
Use, copying, and distribution of any EMC software described in this publication requires an applicable
software license.
EMC2, EMC, and the EMC logo are registered trademarks or trademarks of EMC Corporation in the United
States and other countries. All other trademarks used herein are the property of their respective owners.
For the most up-to-date regulatory document for your product line, go to the technical documentation and
advisories section on the EMC online support website.
2
Command Line Interface Reference for File 8.1
Contents
Preface.............................................................................................................................. 7
Chapter 1
Overview
VNX System.......................................................................................
Main components ......................................................................
Control Station ..................................................................................
Accessing the Control Station ..................................................
Accessing the command line interface ...................................
Role-based access ..............................................................................
Command set conventions ..............................................................
Synopsis ......................................................................................
Command prefixes ....................................................................
General notes .....................................................................................
Chapter 2
12
12
13
13
13
15
16
16
17
18
The cel and cs Commands
cel_fs ................................................................................................... 20
clariion_mgmt ................................................................................... 25
cs_standby.......................................................................................... 32
Chapter 3
The fs Commands
fs_ckpt................................................................................................. 36
fs_dedupe........................................................................................... 49
fs_dhsm .............................................................................................. 68
fs_group.............................................................................................. 97
fs_rdf ................................................................................................. 101
fs_timefinder.................................................................................... 108
Command Line Interface Reference for File 8.1
3
Contents
Chapter 4
The get and set Commands
get_attributes................................................................................... 120
set_attributes ................................................................................... 126
Chapter 5
The nas Commands
nas_acl ..............................................................................................
nas_automountmap........................................................................
nas_ca_certificate ............................................................................
nas_cel ..............................................................................................
nas_checkup ....................................................................................
nas_ckpt_schedule..........................................................................
nas_connecthome............................................................................
nas_config ........................................................................................
nas_copy...........................................................................................
nas_cs................................................................................................
nas_dbtable......................................................................................
nas_devicegroup .............................................................................
nas_disk............................................................................................
nas_diskmark ..................................................................................
nas_emailuser..................................................................................
nas_environment ............................................................................
nas_event .........................................................................................
nas_fs ................................................................................................
nas_fsck ............................................................................................
nas_halt ............................................................................................
nas_inventory..................................................................................
nas_license .......................................................................................
nas_logviewer .................................................................................
nas_message ....................................................................................
nas_mview .......................................................................................
nas_pool ...........................................................................................
nas_quotas .......................................................................................
nas_rdf..............................................................................................
nas_replicate ....................................................................................
nas_rp ...............................................................................................
nas_server ........................................................................................
nas_slice ...........................................................................................
nas_stats ...........................................................................................
nas_storage ......................................................................................
nas_task ............................................................................................
nas_version ......................................................................................
nas_volume......................................................................................
4
Command Line Interface Reference for File 8.1
133
136
138
141
154
159
166
174
178
183
190
194
198
206
210
213
221
228
266
270
275
279
281
284
285
294
311
328
349
368
376
391
395
402
418
422
424
Contents
Chapter 6
The server Commands
server_archive.................................................................................. 433
server_arp......................................................................................... 447
server_certificate ............................................................................. 449
server_cdms ..................................................................................... 455
server_cepp ...................................................................................... 461
server_checkup................................................................................ 464
server_cifs......................................................................................... 481
server_cifssupport........................................................................... 509
server_cpu ........................................................................................ 523
server_date ....................................................................................... 525
server_dbms..................................................................................... 530
server_devconfig ............................................................................. 537
server_df ........................................................................................... 545
server_dns ........................................................................................ 547
server_export ................................................................................... 549
server_file ......................................................................................... 563
server_fileresolve ............................................................................ 564
server_ftp.......................................................................................... 568
server_http ....................................................................................... 584
server_ifconfig ................................................................................. 590
server_ip ........................................................................................... 597
server_kerberos ............................................................................... 602
server_ldap....................................................................................... 609
server_log ......................................................................................... 621
server_mount ................................................................................... 628
server_mountpoint.......................................................................... 637
server_mpfs...................................................................................... 639
server_mt.......................................................................................... 642
server_name ..................................................................................... 644
server_netstat................................................................................... 645
server_nfs ......................................................................................... 649
server_nis.......................................................................................... 658
server_nsdomains ........................................................................... 662
server_param ................................................................................... 665
server_pax ........................................................................................ 673
server_ping....................................................................................... 677
server_ping6..................................................................................... 679
server_rip.......................................................................................... 681
server_route ..................................................................................... 682
server_security................................................................................. 685
server_setup ..................................................................................... 688
server_snmpd .................................................................................. 691
Command Line Interface Reference for File 8.1
5
Contents
server_ssh ........................................................................................
server_standby ................................................................................
server_stats ......................................................................................
server_sysconfig..............................................................................
server_sysstat ..................................................................................
server_tftp........................................................................................
server_umount ................................................................................
server_uptime .................................................................................
server_user.......................................................................................
server_usermapper.........................................................................
server_version .................................................................................
server_viruschk...............................................................................
server_vtlu .......................................................................................
Appendix A
695
707
710
758
767
769
772
774
775
778
782
783
787
Using Command Scripts for VNX
Scripting guidelines........................................................................ 798
Appendix B
GNU General Public License
GNU General Public License ........................................................ 806
Index .............................................................................................................................. 813
6
Command Line Interface Reference for File 8.1
Preface
As part of an effort to improve and enhance the performance and capabilities
of its product line, EMC periodically releases revisions of its hardware and
software. Therefore, some functions described in this document may not be
supported by all revisions of the software or hardware currently in use. For
the most up-to-date information on product features, refer to your product
release notes.
If a product does not function properly or does not function as described in
this document, please contact your EMC representative.
Audience
This guide is part of the VNX documentation set, and is intended for
use by administrators and users of the EMC VNX system.
Related
documentation
Other VNX system publications are available on the EMC Online
Support website. To search for technical documentation, go to
http://Support.EMC.com. After logging in to the website, click
the VNX Support by Product page to locate information for the
specific feature required.
Conventions used in
this guide
EMC uses the following conventions for notes, cautions, warnings,
and danger notices.
Note: A note presents information that is important, but not hazard-related.
!
CAUTION
A caution contains information essential to avoid data loss or
damage to the system or equipment. The caution may apply to
hardware or software.
Command Line Interface Reference for File 8.1
7
Preface
WARNING
A warning contains information essential to avoid a hazard that can
cause severe personal injury, death, or substantial property damage
if you ignore the warning.
DANGER
A danger notice contains information essential to avoid a hazard
that will cause severe personal injury, death, or substantial property
damage if you ignore the message.
Typographical conventions
EMC uses the following type style conventions in this guide:
Bold
• User actions (what the user clicks, presses, or selects)
• Interface elements (button names, dialog box names)
• Names of keys, commands, programs, scripts, applications,
utilities, processes, notifications, system calls, services,
applications, and utilities in text
italic
• Book titles
• New terms in text
• Emphasis in text
Courier
•
•
•
•
•
•
Prompts
System output
Filenames
Pathnames
URLs
Syntax when shown in command line or other examples
Courier, bold • User entry
• Options in command-line syntax
8
Courier italic
• Arguments in examples of command-line syntax
• Variables in examples of screen or file output
• Variables in pathnames
<>
Angle brackets for parameter values (variables) supplied by user.
[]
Square brackets for optional values.
|
Vertical bar symbol for alternate selections. The bar means or.
...
Ellipsis for nonessential information omitted from the example.
Command Line Interface Reference for File 8.1
Preface
Where to get help
EMC support, product, and licensing information can be obtained as
follows.
Product information — For documentation, release notes, software
updates, or for information about EMC products, licensing, and
service, go to the EMC Online Support website (registration required)
at
http://Support.EMC.com
Technical support — For technical support and service requests, go
to EMC Customer Service on the EMC Online Support website. After
logging in, locate the applicable Support by Product page, and select
either Live Chat or Create a service request. To open a service
request through EMC Online Support, you must have a valid support
agreement. Contact your EMC Sales Representative for details about
obtaining a valid support agreement or with any questions about
your account.
Your comments
Your suggestions will help us continue to improve the accuracy,
organization, and overall quality of the user publications. Please send
your opinion of this guide to:
techpubcomments@EMC.com
Command Line Interface Reference for File 8.1
9
Preface
10
Command Line Interface Reference for File 8.1
1
Overview
This chapter provides a brief description of the commands that can be
used to configure and manage the VNX. Topics included are:
◆
◆
◆
◆
◆
VNX System........................................................................................
Control Station....................................................................................
Role-based access ...............................................................................
Command set conventions ...............................................................
General notes ......................................................................................
Overview
12
13
15
16
18
11
Overview
VNX System
The EMC® VNX™ is a dedicated file server solution that is easily
integrated into existing networks by using standard network
interface protocols.
Main components
12
The VNX hardware platform is the VNX cabinet that works with a
system to produce shared network-attached storage. The VNX
cabinet:
◆
Contains up to eight Data Movers available as dedicated file
servers that provide access to file systems resident on the system
to a requesting client on the network.
◆
Includes a Control Station that manages all Data Mover functions
both locally and remotely.
Command Line Interface Reference for File 8.1
Overview
Control Station
The Control Station provides utilities for managing, configuring, and
monitoring of the Data Movers in the VNX.
As the system administrator, you may type commands through the
Control Station to perform tasks that include the following:
Accessing the
Control Station
◆
Managing and Configuring the database and Data Movers
◆
Monitoring statistics of the VNX cabinet components
You may use either local or remote access to the Control Station.
Note: To access locally a connection to serial port have to be established.
Local
Local access to the command line interface is available directly at the
Control Station console.
Remote
Remote access to the command line interface by using a secure,
encrypted login application allows the use of the VNX command set.
Accessing the
command line
interface
A description of how to gain local or remote access to the command
line interface for the VNX follows.
Note: For a local connection, connect a client to the Control Station serial
port.
Local
For local access to the command line interface, at the prompt, log in
with your administrative username and password.
Establish the connection to the Control Station with the following
settings:
Table 1
Control Station serial port connection settings (1 of 2)
Setting
Value
Bits per second
19200
Data biits
8
Parity
None
Control Station
13
Overview
Table 1
Remote
Control Station serial port connection settings (2 of 2)
Setting
Value
Stop bits
1
Flow control
None
Emulation
Auto Detect
Telnet terminal ID
ANSI
For remote access to the command line interface:
1. Use a secure, encrypted, remote login application capable of SSH.
Type the IP address of the Control Station.
2. Log in with your administrative username and password.
14
Command Line Interface Reference for File 8.1
Overview
Role-based access
The administrative user account you use to access the command line
interface is associated with specific privileges, also referred to as
roles. A role defines the privileges (operations) a user can perform on
a particular VNX object. The ability to select a predefined role or
define a custom role that gives a user certain privileges is supported
for users who access VNX through the CLI, EMC Unisphere™, and
the XML API.
The Security Configuration Guide for VNX provides detailed
information about how role-based access is used to determine the
commands a particular user can execute. You create and manage user
accounts and roles in Unisphere by using Settings > User
Management.
Role-based access
15
Overview
Command set conventions
This manual uses commonly known command set conventions for
the VNX for file man pages. Each man page presents the command
name at the top of the man page followed by a brief overview of what
the command does. The synopsis contains the actual command
usage. The description contains a more detailed breakdown of the
features of the command, and the options describe what each switch
or option does specifically.
The ’See Also’ section refers to the technical modules that support the
feature, in addition to any other commands that interact with the
command. The examples are at the end of the command.
The naming convention for the Data Mover variable in the command
line interface is <movername> (default = server_2 to server_9).
The commands are prefixed, then appear in alphabetical order.
Synopsis
The synopsis is usage of each command. The synopsis appears in
courier typeface, with variables such as movername, filename, and
device name enclosed by angle brackets, with the command name
appearing in bold. The switches and other options also appear in
bold and, in most cases, are prefixed by a minus sign:
server_umount {<movername>|ALL}[-perm|-temp]{-all|<fs_name>|<pathname>}
Required entries
A switch or variable enclosed with curly brackets, or not enclosed at
all, indicates a required entry:
{<movername>|ALL}
Optional entries
A switch or variable enclosed with square brackets indicates an
optional entry:
[-perm|-temp]
Formatting
The variable name enclosed by angle brackets indicates the name of a
specified object:
{<movername>|ALL}
Options
An option is prefixed with a minus (-) sign: -perm
If the option is spelled out, for example, -perm, in the command
syntax, you may use just the first letter: -p
16
Command Line Interface Reference for File 8.1
Overview
Options and names are case-sensitive. If an uppercase letter is
specified in the syntax, a lowercase letter is not accepted.
The vertical bar symbol ( | ) represents or, meaning an alternate
selection:
{-all|<fs_name>|<pathname>}
Command prefixes
Commands are prefixed depending on what they are administering.
For example, commands prefixed with:
◆
cel_ execute to the remotely linked VNX system.
◆
cs_ execute to the Control Station.
◆
fs_ execute to the specified file system.
◆
nas_ execute directly to the Control Station database.
◆
server_ require a movername entry and execute directly to a Data
Mover.
Command set conventions
17
Overview
General notes
◆
If a command is interrupted by using Ctrl-C, then the following
messages or traces on the console are expected:
• nas_cmd: system execution failed.
• nas_cmd: PANIC: caught signal #11 (Segmentation fault) -Giving up.
◆
18
Use VNX CLI for file to add IPv6 addresses to the NFS export
host list. Enclose the IPv6 address in { } or square brackets in the
CLI. The IPv6 addresses added to the NFS export list by using the
CLI are displayed as read-only fields in the Unisphere software.
Command Line Interface Reference for File 8.1
2
The cel and cs
Commands
This chapter lists the VNX Command Set provided for managing,
configuring, and monitoring of Data Movers. The commands are
prefixed with cel or cs, and appear alphabetically. The prefix assigned
to a command indicates what the command is administering. The cel
prefix administers to the remote VNX, and cs administers to the local
Control Station. The command line syntax (Synopsis), a description
of the options, and an example of usage are provided for each
command. Commands included are:
◆
◆
◆
cel_fs..................................................................................................... 20
clariion_mgmt..................................................................................... 25
cs_standby........................................................................................... 32
The cel and cs Commands
19
The cel and cs Commands
cel_fs
Manages file systems residing on a remotely linked VNX in an RDF
FarCopy or NearCopy environment..
SYNOPSIS
cel_fs <cel_name>
-list
| -info [-size]{-all|{<fs_name>|id=<fs_id>},...}
| -Extract {-all|{<fs_name>|id=<fs_id>},...}
| -Import {{<fs_name>,...}|-file <filename>}
DESCRIPTION
From the remotely linked VNX, cel_fs imports and enables the
definition of a remote file system known as the R2 onto the local
system known as the R1. The <cel_name> is the default hostname of
the remote Control Station active during installation, or when
performing a nas_rdf -init to set up the remote data facility (RDF)
environment. To find the <cel_name>, go to nas_cel and execute the
-list option. The local VNX, by default, has an ID of zero.
cel_fs displays a listing of the file systems residing on the remote
VNX including their attributes.
This command is used in FarCopy and NearCopy environments.
OPTIONS
-list
Lists all of the file systems residing on the remote VNX as specified
by its <cel_name>.
Note: The ID of the object is an integer and is assigned automatically. The
name of a file system may be truncated if it is too long for the display. To
display the full name, use the -info option with a file system ID.
-info [-size] {-all|{<fs_name>|id=<fs_id>},...}
Displays the attributes and size of the file system residing on the
<cel_name>. The <fs_name> can be a file system name or a group
name and can be up to 16 characters.
-Extract {-all|{<fs_name>|id=<fs_id>},...}
Extracts all file system attributes that can be redirected for import
into the VNX. The second <fs_name> specifies a file system where
the attributes can be imported.
20
Command Line Interface Reference for File 8.1
The cel and cs Commands
-Import {{<fs_name>,...}|-file <filename>}
Imports file system attributes obtained from the extraction of a file
system definition either from the specified <fs_name> or <filename>.
SEE ALSO
EXAMPLE #1
id
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
inuse type acl
n
1
0
y
1
0
y
1
0
y
1
0
y
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
y
1
0
n
5
0
y
1
0
Using SRDF/S with VNX for Disaster Recovery, nas_cel, nas_fs,
nas_rdf, and fs_group.
To display a listing of the file systems residing on the remote VNX
cs30, type:
$ cel_fs cs30 -list
volume
name
66
root_fs_1
68
root_fs_2
70
root_fs_3
72
root_fs_4
74
root_fs_5
76
root_fs_6
78
root_fs_7
80
root_fs_8
82
root_fs_9
84
root_fs_10
86
root_fs_11
88
root_fs_12
90
root_fs_13
92
root_fs_14
94
root_fs_15
96
root_fs_common
145
root_fs_ufslog
156
ufs1
server
1
2
3
4
4,3,2,1
1
Where:
Value
id
inuse
type
acl
volume
name
server
Definition
ID of a file system (assigned automatically).
Whether a file system has been registered into the mount table.
What type of file system. nas_fs provides information.
Access control value for a file system. nas_fs provides information.
ID of the volume on which a file system resides.
Name assigned to a file system.
Server associated with a file system.
EXAMPLE #2
To display the attributes and the size for ufs1 on cs30, type:
$ cel_fs cs30 -info -size ufs1
id
name
acl
in_use
type
=
=
=
=
=
18
ufs1
0
True
uxfs
cel_fs
21
The cel and cs Commands
volume
= mtv1
pool
=
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
size
= total = 22660 avail = 22659 used = 0 ( 0% ) (sizes in MB)
volume: total = 23012 (sizes in MB)
stor_devs =
002804000190-0034,002804000190-0035,002804000190-0036,002804000190-0037
disks
= d3,d4,d5,d6
disk=d3
stor_dev=002804000190-0034
addr=c0t3l8-15-0
server=server_2
disk=d4
stor_dev=002804000190-0035
addr=c0t3l9-15-0
server=server_2
disk=d5
stor_dev=002804000190-0036
addr=c0t3l10-15-0
server=server_2
disk=d6
stor_dev=002804000190-0037
addr=c0t3l11-15-0
server=server_2
Where:
Value
id
name
acl
inuse
type
volume
pool
rw_servers
ro_servers
rw_vdms
ro_vdms
size
volume
stor_devs
disks
Definition
ID of a file system (assigned automatically).
Name of the imported file system.
Access control value for a file system. nas_fs provides information.
Whether a file system has been registered into the mount table.
What type of file system. nas_fs provides information.
Name of the volume on which a file system resides.
Storage pool for the file system.
Servers with read-write access.
Servers with read-only access.
VDM servers with read-write access to a file system.
VDM servers with read-only access to a file system.
Total size, available size, and amount used in MB.
Volume size that is used.
ID of the system device and the serial number.
The disks on which a file system resides. - dd indicates a remote disk.
EXAMPLE #3
To extract a file system definition for ufs1 from cs30, type:
$ cel_fs cs30 -Extract ufs1
0:bofs:18
18:ufs1:0:y:1:164:1::::0:::
164:mtv1:0:y:3:2:18:156:
156:stv1:0:y:2:1:164:64:3,4,5,6:
6:d6:0:y:4:1:156:6:
6:d6:0:y:5753:002804000190:1,2,3,4:6:0037:5:
5:d5:0:y:4:1:156:5:
5:d5:0:y:5753:002804000190:1,2,3,4:5:0036:5:
4:d4:0:y:4:1:156:4:
4:d4:0:y:5753:002804000190:1,2,3,4:4:0035:5:
22
Command Line Interface Reference for File 8.1
The cel and cs Commands
3:d3:0:y:4:1:156:3:
3:d3:0:y:5753:002804000190:1,2,3,4:3:0034:5:
0:eof:18
Where:
EXAMPLE #4
The output is the extracted file system definition.
To extract a file system definition for ufs1 from cs30 into the
extract-file, type:
$ cel_fs cs30 -Extract ufs1 > extract-file
EXAMPLE #5
To import a remote file system definition for ufs1 from cs30 onto the
current VNX, type:
$ cel_fs cs30 -Import ufs2
id
= 18
name
= ufs2
acl
= 0
in_use
= False
type
= uxfs
volume
= mtv2
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = 19@1 Fri Apr 23 15:25:59 EDT 2004
remainder = 0 MB (0%)
stor_devs =
002804000218-0038,002804000218-0039,002804000218-003A,002804000218-003B
disks
= dd7,dd8,dd9,dd10
EXAMPLE #2 provides an output description of all values except
those which are unique to the -Import option.
Where:
Value
backup_of
remainder
Definition
System ID of the remote file system.
Number of MB copied from the file system on the remote VNX.
EXAMPLE #6
To import a remote file system definition from the file, extract-file,
from cs30 into the current VNX, type:
$ cel_fs cs30 -Import -file extract-file
id
name
acl
in_use
type
volume
=
=
=
=
=
=
19
ufs1
0
False
uxfs
mmtv1
cel_fs
23
The cel and cs Commands
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = 18@1 Fri Apr 23 15:26:25 EDT 2004
remainder = 0 MB (0%)
stor_devs =
002804000218-0034,002804000218-0035,002804000218-0036,002804000218-0037
disks
= dd3,dd4,dd5,dd6
EXAMPLE #2 and EXAMPLE #5 provide a description of command
output.
24
Command Line Interface Reference for File 8.1
The cel and cs Commands
clariion_mgmt
SPA and SPB public/private IP address management.
SYNOPSIS
clariion_mgmt {
| -start [-spa_ip <Public IP for SPA>][-spb_ip <Public IP for SPB>]
[-use_proxy_nd] [-retry]
| -modify [-spa_ip <Public IP for SPA>][-spb_ip <Public IP for SPB>]
[-retry]
| -recover [-o]
| -stop [-use_proxy_nd][-retry]
| -info [-use_proxy_nd]
}
DESCRIPTION
OPTIONS
Sets up routing information so that the SPs can be accessed from the
public network by means of the Control Station.
-start
Changes IP address of both SPA and SPB from private to public IP
address.
-spa_ip
Indicates the parameter is used to provide SPA public IP address.
-spb_ip
Indicates the parameter is used to provide SPB public IP address.
-use_proxy_nd
Uses proxy Neighbor Discovery (ND) protocol as an option changing
IP addresses of SPA and SPB.
-retry
Indicates the parameter is used to bypass health checks.
-modify
Changes IP address of SPA and SPB at a time. Modifies existing
public IP address to other public IP address.
-recover
Rebuilds and restores the damaged configuration files.
-o
Indicates the parameter is used to bypass any prompting of
questions.
-stop
Changes public IP address to default private IP address of both SPA
and SPB.
clariion_mgmt
25
The cel and cs Commands
-info
Displays all information of proxy ARP services. It includes public IP
address of SPA and SPB, status of services, proxy ARP
implementation type and booting enable, and so on.
TROUBLESHOOTING
If a system is not properly configured, here are some ways to get the
state back. If the configuration files become damaged, use the
-recover option. This will rebuild and restore the configuration files
(if possible). Run the -info command and confirm that the system is
now OK:
◆
If the system is halfway setup, where one IP address is on the
public network and one is on the internal network due to a failure
in the initial setups, the ARP settings will be rolled back during a
-start or -modify operation. To continue, run the operation again
with the -retry flag. This will retry the task and not attempt to run
the health check before resuming.
◆
If it fails to successfully complete, there may be underlying
problems with the Storage Processors. Verify that you can
communicate with them through naviseccli. Because the
operation is not fully complete, and the security case has not been
updated, the credentials must be manually specified to naviseccli
to use the commands.
◆
If unable to ping the storage processors because ARP settings are
rolled back, the parameters -ip_already_set can be supplied to a
partial command to restore ARP access.
Example:
clariion_mgmt -start -spa_ip <IP address> -ip_already_set
◆
If unable to change SPA and SPB to public because of initialize
security, following command steps may resolve the issue.
Example:
cd /nas/tools/
tar zxvf tftpboot.tar.gz
cd tftpboot//setupbackend
./setup_clariion_security <spa_ip> <spb_ip> <username>
<password> -initialize
EXAMPLE #1
To use the -start option, type:
$ /nas/sbin/clariion_mgmt -start -spa_ip 10.6.2.80
-spb_ip 10.6.2.81
26
Command Line Interface Reference for File 8.1
The cel and cs Commands
Sets SPA and SPB IP address from private to public:
Checking if running as root...yes
Checking if model is supported...yes
Checking for integrated system...yes
Checking if interface eth3 is configured...yes
Checking if interface eth3:1 is configured...no
Checking if interface eth3:2 is configured...no
Checking if IP (10.6.2.80) is available...yes
Checking if IP (10.6.2.81) is available...yes
Checking if SP (128.221.252.200) is up...yes
Checking if SP (128.221.253.201) is up...yes
Checking if a gateway is setup for eth3...yes
Step 1/12 [28185788431]: Setting up Proxy ARP for SPA on Control Station
Adding host specific route for SPA
Adding rules to allow outbound traffic from SPA
Adding ARP entry for SPA
Updating /etc/hosts entry for SPA
Step 2/12 [28185788417]: Changing SPA IP address.
Changing SPA IP from 128.221.252.200 to 10.6.2.80 (subnetmask 255.255.255.0,
gateway 10.6.2.1)
Step 3/12 [28185788421]: Waiting for SPA to restart.
Waiting for SPA to go down.....done (18 secs)
Step 4/12 [28185788425]: Waiting for ping response from SPA.
Waiting for 10.6.2.80 to respond.....done (15 secs)
Step 5/12 [28185788427]: Waiting for CLARiiON software to start on SPA.
Waiting for CLARiiON software to start on SPA...done (37 secs)
Waiting until SPB sees 10.6.2.80 in the domain...done (3 secs)
Step 6/12 [28185788429]: Updating NAS database with SPA IP address.
Updating SYMAPI database with new CLARiiON IP addresses...done (21 secs)
Step 7/12 [28185788432]: Setting up Proxy ARP for SPB on Control Station
Adding host specific route for SPB
Adding rules to allow outbound traffic from SPB
Adding ARP entry for SPB
Updating /etc/hosts entry for SPB
Step 8/12 [28185788418]: Changing SPB IP address.
Changing SPB IP from 128.221.253.201 to 10.6.2.81 (subnetmask 255.255.255.0,
gateway 10.6.2.1)
Step 9/12 [28185788422]: Waiting for SPB to restart.
Waiting for SPB to go down.....done (18 secs)
Step 10/12 [28185788426]: Waiting for ping response from SPB.
Waiting for 10.6.2.81 to respond......done (20 secs)
Step 11/12 [28185788428]: Waiting for CLARiiON software to start on SPB.
Waiting for CLARiiON software to start on SPB...done (25 secs)
Waiting until SPA sees 10.6.2.81 in the domain....done (28 secs)
Step 12/12 [28185788430]: Updating NAS database with SPB IP address.
Updating SYMAPI database with new CLARiiON IP addresses...done (18 secs)
FINISH: Operation took a total time of 4 minutes 3 seconds to complete.
EXAMPLE #2
To modify already existing public IP address of SPB, type:
clariion_mgmt
27
The cel and cs Commands
$ /nas/sbin/clariion_mgmt -modify -spb_ip 10.6.2.84
Note: SPA and SPB must be set to public IP to use this command.
Output:
Checking if running as root...yes
Checking if model is supported...yes
Checking for integrated system...yes
Checking if interface eth3 is configured...yes
Checking if interface eth3:1 is configured...no
Checking if interface eth3:2 is configured...no
Checking if IP (10.6.2.84) is available...yes
Checking if SP (10.6.2.80) is up...yes
Checking if SP (10.6.2.81) is up...yes
Checking if a gateway is setup for eth3...yes
Step 1/7 [28185788432]: Setting up Proxy ARP for SPB on Control Station
Adding host specific route for SPB
Adding rules to allow outbound traffic from SPB
Adding ARP entry for SPB
Updating /etc/hosts entry for SPB
Step 2/7 [28185788418]: Changing SPB IP address.
Changing SPB IP from 10.6.2.81 to 10.6.2.84 (subnetmask 255.255.255.0, gateway
10.6.2.1)
Step 3/7 [28185788422]: Waiting for SPB to restart.
Waiting for SPB to go down.....done (19 secs)
Step 4/7 [28185788426]: Waiting for ping response from SPB.
Waiting for 10.6.2.84 to respond.....done (14 secs)
Step 5/7 [28185788428]: Waiting for CLARiiON software to start on SPB.
Waiting for CLARiiON software to start on SPB...done (37 secs)
Waiting until SPA sees 10.6.2.84 in the domain...done (1 secs)
Step 6/7 [28185788430]: Updating NAS database with SPB IP address.
Updating SYMAPI database with new CLARiiON IP addresses...done (17 secs)
Step 7/7 [28185788436]: Removing old Proxy ARP setup for SPB on Control Station
Removing host specific route for SPB
Removing rules that allow outbound traffic from SPB
Removing ARP entry for SPB
FINISH: Operation took a total time of 1 minute 49 seconds to complete.
EXAMPLE #3
To modify already existing public IP address of SPA and SPB, type:
$ /nas/sbin/clariion_mgmt -modify -spa_ip 10.6.2.79
-spb_ip 10.6.2.81 { -o }
Output:
Checking
Checking
Checking
Checking
Checking
Checking
28
if running as root...yes
if model is supported...yes
for integrated system...yes
if interface eth3 is configured...yes
if interface eth3:1 is configured...no
if interface eth3:2 is configured...no
Command Line Interface Reference for File 8.1
The cel and cs Commands
Checking if IP (10.6.2.84) is available...yes
Checking if SP (10.6.2.80) is up...yes
Checking if SP (10.6.2.81) is up...yes
Checking if a gateway is setup for eth3...yes
Step 1/7 [28185788432]: Setting up Proxy ARP for SPB on Control Station
Adding host specific route for SPB
Adding rules to allow outbound traffic from SPB
Adding ARP entry for SPB
Updating /etc/hosts entry for SPB
Step 2/7 [28185788418]: Changing SPB IP address.
Changing SPB IP from 10.6.2.81 to 10.6.2.84 (subnetmask 255.255.255.0, gateway
10.6.2.1)
Step 3/7 [28185788422]: Waiting for SPB to restart.
Waiting for SPB to go down.....done (19 secs)
Step 4/7 [28185788426]: Waiting for ping response from SPB.
Waiting for 10.6.2.84 to respond.....done (14 secs)
Step 5/7 [28185788428]: Waiting for CLARiiON software to start on SPB.
Waiting for CLARiiON software to start on SPB...done (37 secs)
Waiting until SPA sees 10.6.2.84 in the domain...done (1 secs)
Step 6/7 [28185788430]: Updating NAS database with SPB IP address.
Updating SYMAPI database with new CLARiiON IP addresses...done (17 secs)
Step 7/7 [28185788436]: Removing old Proxy ARP setup for SPB on Control Station
Removing host specific route for SPB
Removing rules that allow outbound traffic from SPB
Removing ARP entry for SPB
FINISH: Operation took a total time of 1 minute 49 seconds to complete.
EXAMPLE #4
To recover configuration files related to proxy ARP configuration,
type:
$ /nas/sbin/clariion_mgmt -recover
Validating primary configuration file
Restored configuration files
$ /nas/sbin/clariion_mgmt -recover -o
Note: Output is the same as the above command, but with no prompting of
questions.
Validating primary configuration file
Restored configuration files
EXAMPLE #5
To use the -stop option, type:
$ /nas/sbin/clariion_mgmt -stop
Sets default private IP address of SPA (128.221.252.200) and SPB
(128.221.253.201).
Checking if running as root...yes
clariion_mgmt
29
The cel and cs Commands
Checking if model is supported...yes
Checking for integrated system...yes
Checking if interface eth3 is configured...yes
Checking if SP (10.6.2.80) is up...yes
Checking if SP (10.6.2.84) is up...yes
Step 1/12 [28185788417]: Changing SPA IP address.
Changing SPA IP from 10.6.2.80 to 128.221.252.200 (subnetmask 255.255.255.0,
gateway 128.221.252.104)
Step 2/12 [28185788421]: Waiting for SPA to restart.
Waiting for SPA to go down.....done (18 secs)
Step 3/12 [28185788425]: Waiting for ping response from SPA.
Waiting for 128.221.252.200 to respond.....done (15 secs)
Step 4/12 [28185788427]: Waiting for CLARiiON software to start on SPA.
Waiting for CLARiiON software to start on SPA...done (37 secs)
Waiting until SPB sees 128.221.252.200 in the domain...done (5 secs)
Step 5/12 [28185788429]: Updating NAS database with SPA IP address.
Adding rules to allow outbound traffic from SPB
Updating SYMAPI database with new CLARiiON IP addresses...done (21 secs)
Step 6/12 [28185788433]: Removing Proxy ARP for SPA on Control Station
Removing host specific route for SPA
Removing rules that allow outbound traffic from SPA
Removing ARP entry for SPA
Updating /etc/hosts entry for SPA
Step 7/12 [28185788418]: Changing SPB IP address.
Changing SPB IP from 10.6.2.84 to 128.221.253.201 (subnetmask 255.255.255.0,
gateway 128.221.253.104)
Step 8/12 [28185788422]: Waiting for SPB to restart.
Waiting for SPB to go down.....done (20 secs)
Step 9/12 [28185788426]: Waiting for ping response from SPB.
Waiting for 128.221.253.201 to respond.....done (14 secs)
Step 10/12 [28185788428]: Waiting for CLARiiON software to start on SPB.
Waiting for CLARiiON software to start on SPB...done (24 secs)
Waiting until SPA sees 128.221.253.201 in the domain...done (22 secs)
Step 11/12 [28185788430]: Updating NAS database with SPB IP address.
Updating SYMAPI database with new CLARiiON IP addresses...done (18 secs)
Step 12/12 [28185788434]: Removing Proxy ARP for SPB on Control Station
Removing host specific route for SPB
Removing rules that allow outbound traffic from SPB
Removing ARP entry for SPB
Updating /etc/hosts entry for SPB
FINISH: Operation took a total time of 4 minutes 12 seconds to complete.
EXAMPLE #6
To display the default information of proxy ARP services, type:
$ /nas/sbin/clariion_mgmt -info
Displays default information of proxy ARP services.
Note: The command will not display anything if there is nothing configured.
30
Command Line Interface Reference for File 8.1
The cel and cs Commands
Public IP address for SPA:
Public IP address for SPB:
Start on boot
:
Current implementation
:
Status
:
10.6.2.80
10.6.2.81
yes
Proxy-ARP
Started
clariion_mgmt
31
The cel and cs Commands
cs_standby
Initiates a takeover and failover of a Control Station on a VNX with
dual Control Stations.
SYNOPSIS
cs_standby
{-takeover|-failover}
DESCRIPTION
The cs_standby command initiates a Control Station takeover and
failover. When a Control Station is activated, the name of the primary
Control Station is displayed.
This command must be executed from the /nas/sbin or
/nasmcd/sbin directory. su to root to execute this command.
Note: EMC SRDF® is not supported on the secondary Control Station.
OPTIONS
-takeover
Executed from the standby Control Station, initiates a reboot of the
primary Control Station, then changes the state of the standby to that
of the primary. The original primary Control Station now becomes
the standby Control Station. The -takeover option can be used to
failback Control Station 0 to the role of primary Control Station after a
failover, or to set Control Station 1 to the role of primary Control
Station on demand.
!
CAUTION
When executing a takeover or failover, Data Movers performing
functions such as RDF, EMC TimeFinder®/FS, file system extends,
or quotas may be interrupted.
!
CAUTION
If a primary Control Station fails over to a standby Control Station,
the remote replication service continues to run, but replication
management capabilities are no longer available.
Note: After executing a takeover or failover, a few minutes may be needed to
stop Linux and other services active on the Control Station.
32
Command Line Interface Reference for File 8.1
The cel and cs Commands
-failover
Executed from the primary Control Station, initiates a reboot of the
primary Control Station, then activates the standby to take over the
role of the primary Control Station. The -failover option can be used
to complete a failback by forcing a failover from Control Station 1
back to Control Station 0 after Control Station 0 had failed over, or to
set Control Station 1 to the role of primary Control Station on
demand.
To display the primary Control Station, type:
$ nas/sbin/getreason
EXAMPLE #1
To change the state of the standby Control Station to primary, cd to
the /nasmcd/sbin directory of the standby Control Station, then
type:
#./cs_standby -takeover
Taking over as Primary Control Station............done
If the takeover command is executed on the primary Control Station,
the following error message appears:
The -takeover option is only valid on a standby Control
Station
EXAMPLE #2
To initiate a failover from the primary Control Station to the standby
Control Station, cd to the /nas/sbin directory of the primary Control
Station, then type:
#./cs_standby -failover
The system will reboot, do you wish to continue [yes or no]: y
Failing over from Primary Control Station
cs_standby
33
The cel and cs Commands
34
Command Line Interface Reference for File 8.1
3
The fs Commands
This chapter lists the VNX Command Set provided for managing,
configuring, and monitoring the specified file system. The commands
are prefixed with fs and appear alphabetically. The command line
syntax (Synopsis), a description of the options, and examples of
usage are provided for each command. Commands included are:
◆
◆
◆
◆
◆
◆
fs_ckpt.................................................................................................. 36
fs_dedupe ............................................................................................ 49
fs_dhsm ............................................................................................... 68
fs_group............................................................................................... 97
fs_rdf .................................................................................................. 101
fs_timefinder..................................................................................... 108
The fs Commands
35
The fs Commands
fs_ckpt
Manages checkpoints using the EMC SnapSure™ functionality.
SYNOPSIS
fs_ckpt {<fs_name>|id=<fs_id>}
-list [-all]
| [-name <name>] -Create [-readonly {y|n}][<volume_name>][-option <options>]
| [-name <name>] -Create [-readonly {y|n}][size=<integer>[T|G|M|%]]
[pool=<pool>][storage=<system_name>][-option <options>]
| -refresh [-option <options>]
| [-name <name>] -Restore [-Force][-option <options>]
| -modify [%full=<value>][maxsavsize=<integer>[T|G|M]]
DESCRIPTION
What is a checkpoint
file system?
OPTIONS
The fs_ckpt command creates a checkpoint of a Production File
System (PFS), lists associated checkpoints, refreshes a checkpoint to
the current time, and restores a PFS back to a specific point in time
using a checkpoint. Checkpoints are deleted using nas_fs.
A PFS is made up of blocks. When a block within a PFS is modified, a
copy containing the original contents of that block is saved to a
metavolume called the SavVol. Subsequent changes made to the same
block in the PFS are not copied into the SavVol. The original blocks
from the PFS (in the SavVol) and the unchanged PFS blocks (that
remain in the PFS) are read according to a bitmap and blockmap data
tracking structure. These blocks combine to provide a complete
point-in-time file system image which is called a checkpoint.
-list
Displays all of the associated checkpoints for the specified file
system. The -all option displays system-generated Replication
checkpoints in addition to checkpoints created by the user.
[-name <name>] -Create
Creates, mounts, and optionally assigns a name to the checkpoint of
the PFS. The checkpoint must be unmounted prior to unmounting
the PFS. Names assigned to a checkpoint cannot be all numeric. If a
name is not chosen, one is assigned by default.
[-readonly {y|n}]
Specifies whether a checkpoint is read only or not. y (default) sets
the checkpoint as read only; n sets the checkpoint as writeable.
36
Command Line Interface Reference for File 8.1
The fs Commands
[<volume_name>]
Specifies an unused metavolume for the checkpoint.
Note: A volume can be specified for only the first checkpoint of a PFS
since all of the subsequent checkpoints share the same SavVol. The
minimum size required for a SavVol is 64 MB. The volume size is 10 GB
unless the PFS is less than 10 GB, then the volume is the same size as the
file system.
[-option <options>]
Specifies the following comma-separated options:
%full=<value>
Specifies a value as the percentage threshold permitted for the
SavVol. When that value is reached, a warning is sent to the
server_log and the syslog files. The Control Station acknowledges
the event and automatically extends the checkpoint. The SavVol is
automatically extended by 10 GB if its default %full value is
reached. If the %full value is set to zero, the option is disabled.
maxsavsize=<integer>[T|G|M]
Limits the final size to which the SavVol can be automatically
extended when the high watermark value specified in %full has
been reached. Automatic extension of the SavVol stops when the
size of the SavVol reaches the value specified in maxsavsize. The
range for maxsavsize is 64 MB to 16 TB.
automount=no
Stops the checkpoint from being automatically mounted.
[-name <name>] -Create
Creates, mounts, and optionally assigns a name to the checkpoint of
the PFS. The checkpoint must be unmounted prior to unmounting
the PFS. Names assigned to a checkpoint cannot be all numeric. If a
name is not chosen, one is assigned by default.
[-readonly {y|n}]
Specifies whether a checkpoint is read only or not. y (default) sets
the checkpoint as read only; n sets the checkpoint as writeable.
[size=<integer>[T|G|M|%]]
Specifies a size for the checkpoint file system. Type an integer
between 1 and 1024, specify T for terabytes, G for gigabytes
(default), or M for megabytes, or type an integer representing the
percentage of a file system’s size, followed by the percent sign.
fs_ckpt
37
The fs Commands
[pool=<pool>]
Specifies the storage pool to be used for the checkpoint. Storage
pools can either be user-defined or system-defined. The nas_pool
-list command displays a listing of available pool types.
[storage=<system_name>]
Specifies the attached system for the checkpoint SavVol to reside.
[-option <options>]
Specifies the following comma-separated options:
%full=<value>
Specifies a value as the percentage threshold permitted for the
SavVol. When that value is reached, a warning is sent to the
server_log and the syslog files. The Control Station
acknowledges the event and automatically extends the
checkpoint. The SavVol is automatically extended by 10 GB if
its default %full value is reached. If the %full value is set to
zero, the option is disabled. The default for <value> is 90 and
it can be within the range of 10 to 99.
automount=no
Stops the checkpoint from being automatically mounted.
-refresh
Initiates an immediate update of a checkpoint, thereby allowing the
SavVol space to be reused. Refreshing a checkpoint does not add to
the number of checkpoints of the PFS.
[-option <options>] %full=<value>
Specifies a value as the percentage threshold permitted for the
metavolume. When that value is reached, a warning is sent to the
server_log and the syslog files. The Control Station acknowledges
the event and automatically extends the checkpoint. The SavVol is
automatically extended by 10 GB if its default %full value is
reached. If the %full value is set to zero, the option is disabled.
The default for <value> is 90.
-modify
Modifies one or all of the following options:
Note: The -modify action works only on the PFS and not on the checkpoint.
[%full=<value>]
Modifies the value of the percentage threshold permitted for the
metavolume.
38
Command Line Interface Reference for File 8.1
The fs Commands
[maxsavsize=<integer>[T|G|M]]
Modifies the final size to which the SavVol can be automatically
extended, when the size specified in %full is reached.
[-name <name>] -Restore
Restores the PFS from the specified checkpoint and optionally assigns
a name to the automatically created checkpoint. If a name is not
chosen, one is assigned by default.
Note: As part of the restore, a new checkpoint is automatically created to
capture the latest point-in-time image of the PFS. This is for protection in the
event that the restored image is discarded.
[-Force]
The -Force option must be used when restoring a production file
system with File-Level Retention enabled.
!
CAUTION
Forcing a restore of a production file system with File-Level
Retention enabled from a checkpoint will delete or overwrite
files that were written after this checkpoint was created or
refreshed.
[-option <options>]
Specifies the following comma-separated options:
%full=<value>
Specifies a value as the percentage threshold permitted for the
SavVol. When that value is reached, a warning is sent to the
server_log and the syslog files. The Control Station acknowledges
the event and automatically extends the checkpoint. The SavVol is
automatically extended by 10 GB if its default %full value is
reached. If the %full value is set to zero, the option is disabled.
The <value> can be an integer between 10 and 75 (default).
automount=no
Stops the checkpoint from being automatically mounted.
SEE ALSO
SYSTEM OUTPUT
Using VNX Snapsure, nas_fs, and nas_pool.
The number associated with the storage device is dependent on the
attached system. VNX for block displays a prefix of APM or CK
before a set of integers, for example, APM00033900124-0019. For
example, EMC Symmetrix® systems display as 002804000190-003C.
The outputs displayed in the examples use a VNX for block.
fs_ckpt
39
The fs Commands
EXAMPLE #1
To display the checkpoint for the file system fs4, type:
$ fs_ckpt fs4 -list
id ckpt_name creation_time inuse fullmark total_savvol_used ckpt_usage_on_savvol
1406 fs4_ckpt1 05/26/2008-16:22:19-EDT y 90% 51% 0%
id wckpt_name inuse fullmark total_savvol_used base ckpt_usage_on_savvol
EXAMPLE #2
To display all checkpoints including internal checkpoints for the file
system fs4, type:
$ fs_ckpt fs4 -list -all
id ckpt_name creation_time inuse fullmark total_savvol_used ckpt_usage_on_savvol
1401 root_rep_ckpt_1398_21625_1 05/26/2008-16:11:10-EDT y 90% 51% 0%
1402 root_rep_ckpt_1398_21625_2 05/26/2008-16:11:22-EDT y 90% 51% 0%
1406 fs4_ckpt1 05/26/2008-16:22:19-EDT y 90% 51% 0%
id wckpt_name inuse fullmark total_savvol_used base ckpt_usage_on_savvol
EXAMPLE #3
To create a checkpoint of ufs1, on the volume, ssmtv1, type:
$ fs_ckpt ufs1 -Create ssmtv1
operation in progress (not interruptible)...id
= 22
name
= ufs1
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= mtv1
pool
=
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
ckpts
= ufs1_ckpt1
stor_devs = APM00043807043-0010,APM00043807043-0014
disks
= d7,d9
disk=d7
stor_dev=APM00043807043-0010 addr=c0t1l0
disk=d7
stor_dev=APM00043807043-0010 addr=c16t1l0
disk=d9
stor_dev=APM00043807043-0014 addr=c0t1l4
disk=d9
stor_dev=APM00043807043-0014 addr=c16t1l4
id
=
name
=
acl
=
in_use
=
type
=
worm
=
volume
=
pool
=
member_of =
rw_servers=
ro_servers=
rw_vdms
=
40
24
ufs1_ckpt1
0
True
ckpt
off
vp132
server_2
Command Line Interface Reference for File 8.1
server=server_2
server=server_2
server=server_2
server=server_2
The fs Commands
ro_vdms
=
checkpt_of=
used
=
full(mark)=
stor_devs =
disks
=
disk=d12
disk=d12
disk=d15
disk=d15
ufs1 Wed Oct 13 18:01:04 EDT 2004
0%
90%
APM00043807043-0011,APM00043807043-0017
d12,d15
stor_dev=APM00043807043-0011 addr=c16t1l1
stor_dev=APM00043807043-0011 addr=c0t1l1
stor_dev=APM00043807043-0017 addr=c16t1l7
stor_dev=APM00043807043-0017 addr=c0t1l7
server=server_2
server=server_2
server=server_2
server=server_2
Where:
Value
id
name
acl
in_use
type
worm
volume
pool
member_of
rw_servers
ro_servers
rw_vdms
ro_vdms
ckpts
checkpt_of
used
full(mark)
stor_devs
disks
Definition
Automatically assigned ID of a file system or the checkpoint.
Name assigned to the file system or the checkpoint.
Access control value for a file system. nas_acl provides information.
If a file system is registered into the mount table of a Data Mover.
Type of file system. -list provides a description of the types.
Whether the feature is enabled.
Volume on which a file system resides.
Storage pool for the file system.
Group to which the file system belongs.
Servers with read-write access to a file system.
Servers with read-only access to a file system.
VDM servers with read-write access to a file system.
VDM servers with read-only access to a file system.
Associated checkpoints for the file system.
Name of the PFS related to the existing checkpoints.
Percentage of SavVol space used by the checkpoints of the PFS.
SavVol usage point which, when reached, sends a warning message to the system log, and auto-extends
the SavVol as system space permits.
System devices associated with a file system.
Disks on which the metavolume resides.
EXAMPLE #4
To create a checkpoint of ufs1 named ufs1_ckpt2 with a size of 2 GB
by using the clar_r5_performance pool, with the specified system,
with the %full set to 95, type:
$ fs_ckpt ufs1 -name ufs1_ckpt2 -Create size=2G
pool=clar_r5_performance storage=APM00043807043 -option
%full=95
operation in progress (not interruptible)...id
= 27
name
= ufs1
acl
= 0
in_use
= True
type
= uxfs
fs_ckpt
41
The fs Commands
worm
=
volume
=
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
ckpts
=
stor_devs =
disks
=
disk=d7
disk=d7
disk=d9
disk=d9
off
mtv1
id
=
name
=
acl
=
in_use
=
type
=
worm
=
volume
=
pool
=
member_of =
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
checkpt_of=
used
=
full(mark)=
stor_devs =
disks
=
disk=d12
disk=d12
disk=d15
disk=d15
30
ufs1_ckpt2
0
True
ckpt
off
vp145
server_2
ufs1_ckpt1,ufs1_ckpt2
APM00043807043-0010,APM00043807043-0014
d7,d9
stor_dev=APM00043807043-0010 addr=c0t1l0
stor_dev=APM00043807043-0010 addr=c16t1l0
stor_dev=APM00043807043-0014 addr=c0t1l4
stor_dev=APM00043807043-0014 addr=c16t1l4
server=server_2
server=server_2
server=server_2
server=server_2
server_2
ufs1 Wed Nov 10 14:00:20 EST 2004
0%
95%
APM00043807043-0011,APM00043807043-0017
d12,d15
stor_dev=APM00043807043-0011 addr=c16t1l1
stor_dev=APM00043807043-0011 addr=c0t1l1
stor_dev=APM00043807043-0017 addr=c16t1l7
stor_dev=APM00043807043-0017 addr=c0t1l7
server=server_2
server=server_2
server=server_2
server=server_2
EXAMPLE #3 provides a description of command output.
EXAMPLE #5
To create a checkpoint of ufs2 named ufs2_ckpt1 with a size of 2 GB
by using the clar_mapped_pool VNX mapped pool, with the
specified system, with the %full set to 95, type:
$ fs_ckpt ufs2 -name ufs2_ckpt1 -Create size=2G
pool=clar_mapped_pool storage=APM00043807043 -option
%full=95
operation in progress (not interruptible)...id
= 435
name
= ufs2
acl
= 0
in_use
= True
type
= uxfs
42
Command Line Interface Reference for File 8.1
The fs Commands
worm
= off
volume
= v731
pool
= clar_mapped_pool
member_of = root_avm_fs_group_50
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
fast_clone_level = 1
deduplication
= Off
thin_storage
= False
tiering_policy = N/A/Optimize Pool
compressed= False
mirrored = False
ckpts
= ufs2_ckpt1
stor_devs =
FNM00103400314-0036,FNM00103400314-0037,FNM00103400314-0038,FNM00103400314-00
39
disks
= d60,d61,d62,d63
disk=d60
stor_dev=FNM00103400314-0036 addr=c0t1l0
server=server_2
disk=d60
stor_dev=FNM00103400314-0036 addr=c16t1l0
server=server_2
disk=d61
stor_dev=FNM00103400314-0037 addr=c0t1l1
server=server_2
disk=d61
stor_dev=FNM00103400314-0037 addr=c16t1l1
server=server_2
disk=d62
stor_dev=FNM00103400314-0038 addr=c0t1l2
server=server_2
disk=d62
stor_dev=FNM00103400314-0038 addr=c16t1l2
server=server_2
disk=d63
stor_dev=FNM00103400314-0039 addr=c0t1l3
server=server_2
disk=d63
stor_dev=FNM00103400314-0039 addr=c16t1l3
server=server_2
id
= 438
name
= ufs2_ckpt1
acl
= 0
in_use
= True
type
= ckpt
worm
= off
volume
= vp735
pool
= clar_mapped_pool
member_of =
rw_servers=
ro_servers= server_2
rw_vdms
=
ro_vdms
=
checkpt_of= ufs2 Fri Jan 4 01:43:20 EST 2013
deduplication
= Off
thin_storage
= False
tiering_policy = N/A/Optimize Pool
compressed= False
mirrored = False
used
= 13%
full(mark)= 95%
fs_ckpt
43
The fs Commands
stor_devs =
FNM00103400314-0036,FNM00103400314-0037,FNM00103400314-0038,FNM00103400314-00
39
disks
= d60,d61,d62,d63
disk=d60
stor_dev=FNM00103400314-0036 addr=c0t1l0
server=server_2
disk=d60
stor_dev=FNM00103400314-0036 addr=c16t1l0
server=server_2
disk=d61
stor_dev=FNM00103400314-0037 addr=c0t1l1
server=server_2
disk=d61
stor_dev=FNM00103400314-0037 addr=c16t1l1
server=server_2
disk=d62
stor_dev=FNM00103400314-0038 addr=c0t1l2
server=server_2
disk=d62
stor_dev=FNM00103400314-0038 addr=c16t1l2
server=server_2
disk=d63
stor_dev=FNM00103400314-0039 addr=c0t1l3
server=server_2
disk=d63
stor_dev=FNM00103400314-0039 addr=c16t1l3
server=server_2
Where:
Value
Definition
thin_storage
Indicates whether the VNX for block system uses thin provisioning. Values are: True, False, Mixed.
tiering_policy Indicates the tiering policy is in effect. If the initial tier and the tiering policy are the same, the values are:
Auto-Tier, Highest Available Tier, Lowest Available Tier. If the initial tier and the tiering policy are not the
same, the values are: Auto-Tier/No Data Movement, Highest Available Tier/No Data Movement, Lowest
Available Tier/No Data Movement.
compressed
Indicates whether data is compressed. Values are True, False, Mixed (indicates some of the LUNs, but
not all, are compressed).
mirrored
Indicates whether the disk is mirrored.
EXAMPLE #6
To create a writeable checkpoint of baseline checkpoint ufs1_ckpt1,
type:
$ fs_ckpt ufs1_ckpt1 -Create -readonly n
operation in progress (not interruptible)...id
= 45
name
= ufs1_ckpt1
acl
= 0
in_use
= False
type
= ckpt
worm
= off
volume
= vp145
pool
= clar_r5_performance
member_of =
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
checkpt_of= ufs1 Tue Nov 6 14:56:43 EST 2007
ckpts
= ufs1_ckpt1_writeable1
used
= 38%
full(mark)= 90%
stor_devs =
APM00042000814-0029,APM00042000814-0024,APM00042000814-0021,APM000420
00814-001C
disks
= d34,d17,d30,d13
44
Command Line Interface Reference for File 8.1
The fs Commands
id
= 46
name
= ufs1_ckpt1_writeable1
acl
= 0
in_use
= True
type
= wckpt
worm
= off
volume
= vp145
pool
= clar_r5_performance
member_of =
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
checkpt_of= ufs1
baseline_ckpt = ufs1_ckpt1 Tue Nov 6 14:56:43 EST 2007
used
= 38%
full(mark)= 90%
stor_devs =
APM00042000814-0029,APM00042000814-0024,APM00042000814-0021,APM000420
00814-001C
disks
= d34,d17,d30,d13
disk=d34
stor_dev=APM00042000814-0029 addr=c16t2l9
server=server_2
disk=d34
stor_dev=APM00042000814-0029 addr=c32t2l9
server=server_2
disk=d34
stor_dev=APM00042000814-0029 addr=c0t2l9
server=server_2
disk=d34
stor_dev=APM00042000814-0029 addr=c48t2l9
server=server_2
disk=d17
stor_dev=APM00042000814-0024 addr=c0t2l4
server=server_2
disk=d17
stor_dev=APM00042000814-0024 addr=c48t2l4
server=server_2
disk=d17
stor_dev=APM00042000814-0024 addr=c16t2l4
server=server_2
disk=d17
stor_dev=APM00042000814-0024 addr=c32t2l4
server=server_2
disk=d30
stor_dev=APM00042000814-0021 addr=c16t2l1
server=server_2
disk=d30
stor_dev=APM00042000814-0021 addr=c32t2l1
server=server_2
disk=d30
stor_dev=APM00042000814-0021 addr=c0t2l1
server=server_2
disk=d30
stor_dev=APM00042000814-0021 addr=c48t2l1
server=server_2
disk=d13
stor_dev=APM00042000814-001C addr=c0t1l12
server=server_2
disk=d13
stor_dev=APM00042000814-001C addr=c48t1l12
server=server_2
disk=d13
stor_dev=APM00042000814-001C addr=c16t1l12
server=server_2
disk=d13
stor_dev=APM00042000814-001C addr=c32t1l12
server=server_2
Where:
Value
Definition
baseline_ckpt Name of the read-only checkpoint from which the writeable checkpoint is created.
EXAMPLE #3 provides a description of command output.
EXAMPLE #7
id
29
30
ckpt_name
ufs1_ckpt1
ufs1_ckpt2
To list checkpoints for ufs1, type:
$ fs_ckpt ufs1 -list
creation_time
11/04/2004-14:54:06-EST
11/10/2004-14:00:20-EST
inuse full(mark) used
n
95%
0%
y
95%
0%
fs_ckpt
45
The fs Commands
Where:
Value
id
ckpt_name
creation_time
inuse
full(mark)
used
Definition
Automatically assigned ID of a file system or checkpoint.
Name assigned to the checkpoint.
Date and time the checkpoint was created.
If a checkpoint is registered into the mount table of a Data Mover.
SavVol-usage point which, when reached, sends a warning message to the system log, and
auto-extends the SavVol as system space permits.
Percentage of SavVol space used by checkpoints of the PFS.
EXAMPLE #8
To refresh ufs1_ckpt2 by using the %full at 85, type:
$ fs_ckpt ufs1_ckpt2 -refresh -option
operation in progress (not interruptible)...id
= 30
name
= ufs1_ckpt2
acl
= 0
in_use
= True
type
= ckpt
worm
= off
volume
= vp145
pool
=
member_of =
rw_servers=
ro_servers= server_2
rw_vdms
=
ro_vdms
=
checkpt_of= ufs1 Wed Nov 10 14:02:59 EST 2004
used
= 0%
full(mark)= 85%
stor_devs = APM00043807043-0011,APM00043807043-0017
disks
= d12,d15
disk=d12
stor_dev=APM00043807043-0011 addr=c16t1l1
disk=d12
stor_dev=APM00043807043-0011 addr=c0t1l1
disk=d15
stor_dev=APM00043807043-0017 addr=c16t1l7
disk=d15
stor_dev=APM00043807043-0017 addr=c0t1l7
%full=85
server=server_2
server=server_2
server=server_2
server=server_2
EXAMPLE #3 provides a description of command output.
EXAMPLE #9
Using root command, to restore ufs1_ckpt2 and capture the latest
point-in-time image of the PFS on ufs1_ckpt3, type:
$ /nas/sbin/rootfs_ckpt ufs1_ckpt2 -name ufs1_ckpt3
-Restore
operation in progress (not interruptible)...id
= 30
name
= ufs1_ckpt2
acl
= 0
in_use
= True
type
= ckpt
worm
= off
volume
= vp145
46
Command Line Interface Reference for File 8.1
The fs Commands
pool
=
member_of =
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
checkpt_of=
used
=
full(mark)=
stor_devs =
disks
=
disk=d12
disk=d12
disk=d15
disk=d15
server_2
ufs1 Wed Nov 10 14:02:59 EST 2004
0%
90%
APM00043807043-0011,APM00043807043-0017
d12,d15
stor_dev=APM00043807043-0011 addr=c16t1l1
stor_dev=APM00043807043-0011 addr=c0t1l1
stor_dev=APM00043807043-0017 addr=c16t1l7
stor_dev=APM00043807043-0017 addr=c0t1l7
server=server_2
server=server_2
server=server_2
server=server_2
EXAMPLE #3 provides a description of command output.
EXAMPLE #10
To modify the %full value of the SavVol associated with the file
system ufs1 and set it to 95, type:
$ fs_ckpt ufs1 -modify %full=95
operation in progress (not interruptible)...id
= 33
name
= ufs1
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= vp145
pool
=
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,virtual_provision=no
ckpts
= wipckpt
stor_devs = APM00062400708-0014,APM00062400708-0016
disks
= d26,d27
disk=d26
stor_dev=APM00062400708-0014 addr=c0t1l4
disk=d26
stor_dev=APM00062400708-0014 addr=c16t1l4
disk=d27
stor_dev=APM00062400708-0016 addr=c0t1l6
disk=d27
stor_dev=APM00062400708-0016 addr=c16t1l6
EXAMPLE #11
operation
name
acl
in_use
type
server=server_2
server=server_2
server=server_2
server=server_2
To modify the maxsavsize value of the SavVol associated with the file
system ufs1 and set it to 65 GB, type:
$ fs_ckpt ufs1 -modify maxsavsize=65G
in progress (not interruptible)...id
= 33
= ufs1
= 0
= True
= uxfs
fs_ckpt
47
The fs Commands
worm
= off
volume
= vp145
pool
=
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,virtual_provision=no
ckpts
= wipckpt
stor_devs = APM00062400708-0014,APM00062400708-0016
disks = d26,d27
disk=d26 stor_dev=APM00062400708-0014 addr=c0t1l4 server=server_2
disk=d26 stor_dev=APM00062400708-0014 addr=c16t1l4 server=server_2
disk=d27 stor_dev=APM00062400708-0016 addr=c0t1l6 server=server_2
disk=d27 stor_dev=APM00062400708-0016 addr=c16t1l6 server=server_2
DIAGNOSTICS
48
The fs_ckpt command returns one of the following return codes:
◆
0 — Command completed successfully
◆
1 — Usage error
◆
2 — Invalid object error
◆
3 — Unable to acquire lock
◆
4 — Permission error
◆
5 — Communication error
◆
6 — Transaction error
◆
7 — Dart error
◆
8 — Backend error
Command Line Interface Reference for File 8.1
The fs Commands
fs_dedupe
Manages file system deduplication state.
SYNOPSIS
fs_dedupe {
-list
| -info {-all|<fs_name>|id=<fs_id>}
| -modify {<fs_name>|id=<fs_id>} [-state {off|suspended|on}]
[-minimum_scan_interval <days>][-minimum_size <KB>]
[-maximum_size <MB>][-access_time <days>][-modification_time <days>]
[-case_sensitive {yes|no}][-pathname_exclude_list <path_list>]
[-file_ext_exclude_list <ext_list>]
[-duplicate_detection_method {sha1|byte|off}]
[-savvol_threshold <percent>][-backup_data_threshold <percent>]
[-cifs_compression_enabled {yes|no}][-compression_method {fast|deep}]
| -clear {<fs_name>|id=<fs_id>}[-minimum_scan_interval][-minimum_size]
[-maximum_size][-access_time][-modification_time][-case_sensitive]
[-pathname_exclude_list][-file_ext_exclude_list]
[-duplicate_detection_method][-savvol_threshold][-backup_data_threshold]
[-cifs_compression_enabled][-compression_method]
| -default {
-info {<mover_name>|-all}
| -set {<mover_name>|-all}[-minimum_scan_interval <days>]
[-minimum_size<KB>][-maximum_size <MB>][-access_time
<days>][-modification_time <days>][-case_sensitive
{yes|no}][-file_ext_exclude_list <ext_list>]
[-duplicate_detection_method {sha1|byte|off}][-savvol_threshold
<percent>][-cpu_usage_low_watermark <percent> ]
[-cpu_usage_high_watermark <percent>][-backup_data_threshold
<percent>] [-cifs_compression_enabled {yes|no}]
| -clear {<mover_name>|-all}
[-minimum_scan_interval][-minimum_size][-maximum_size][-access_time]
[-modification_time][-case_sensitive][-file_ext_exclude_list]
[-duplicate_detection_method][-savvol_threshold]
[-cpu_usage_low_watermark][-cpu_usage_high_watermark]
[-backup_data_threshold <percent>][-cifs_compression_enabled]
}
}
DESCRIPTION
fs_dedupe allows the VNX administrator to enable, suspend, and
undo all deduplication processing on a file system or a Data Mover.
The Data Mover settings are the global settings that can be used for
both the Data Mover and the file system. If a user sets a value for a
specific file system, then that value overrides the Data Mover global
value. If a user clears a value set for a specific file system, then that
value is reset to the Data Mover global value.
fs_dedupe
49
The fs Commands
OPTIONS
-list
Lists all deduplication-enabled file systems on the VNX.
-info {-all|<fs_name>|id=<fs_id>}
Lists the existing file systems and provides information on the state
of deduplication processing.
-all
Lists all file systems and provides detailed information on the
state of deduplication processing.
<fs_name>
Lists the file system information for the specified file system
name.
id=<fs_id>
Lists the file system information for the specified identifier.
The file system state and status information displayed includes:
◆
If the state is off and the status is not reduplicating:
• ID
• Name
• Deduplication state
◆
If the state is off and the status is reduplicating:
•
•
•
•
◆
If the state of the file system is on or suspended, and the status is
Idle or Scanning:
•
•
•
•
•
•
•
•
•
•
•
50
ID
Name
Deduplication state
Progress information (the percentage of files scanned)
ID
Name
Reduplication state
Status
The percentage of files scanned
Last system scan time
Number of files scanned
Number of files deduplicated
The percentage of files deduplicated
File system capacity
Logical data size
Command Line Interface Reference for File 8.1
The fs Commands
• Percentage of file system usage
• Space saved (in MB and percent)
-modify {<fs_name>|id=<fs_id>} [-state {off|
suspended|on}]
Modifies the deduplication state of the file system for each specified
file system identifier or file system name. The state can be set to off,
on, or suspended.
[-minimum_scan_interval <days>]
Defines the minimum number of days between completing one
scan of a file system and before scanning the same file system
again. The values range from 1 to 365 and the default value is 7
days.
[-minimum_size <KB>]
Defines the file size in KB that limits deduplication. File sizes
equal to this value or smaller will not be deduplicated. Setting
this value to zero disables it. This value should not be set lower
than 24 KB. The values range from 0 to 1000 and the default value
is 24 KB.
[-maximum_size <MB>]
Defines the file size in MB of the largest file to be processed for
deduplication. Files larger than this size in MB will not be
deduplicated. Setting this value to zero disables it. The values
range from 0 to 8388608 and the default value is 8388608 MB.
[-access_time <days>]
Defines the minimum required file age in days based on read
access time. Files that have been read within the specified number
of days will not be deduplicated. This setting does not apply to
files with FLR locked state. Setting this value to zero disables it.
The values range from 0 to 365 and the default value is 15 days.
[-modification_time <days>]
Defines the minimum required file age in days based on
modification time. Files updated within the specified number of
days will not be deduplicated. Setting this value to zero disables
it. The values range from 0 to 365 and the default value is 15 days.
[-case_sensitive {yes|no}]
Defines whether case-sensitive (for NPS environments) or
case-insensitive (for CIFS environments) string comparisons will
fs_dedupe
51
The fs Commands
be used during scans. By default, case-insensitive comparisons
will be done to be consistent for CIFS environments. The default
value is zero (false).
[-pathname_exclude_list <path_list>]
This is a file system setting only (no global setting). It is empty by
default.
Defines a semicolon-delimited list of relative pathnames, in
UTF-8 format, to be excluded from deduplication. Any directory
below a specified pathname will be excluded from deduplication.
You can specify a maximum of 10 pathnames and each one can be
up to 1024 bytes. The default value is ' ' (empty).
[-file_ext_exclude_list <ext_list>]
Specifies a colon-delimited list of filename extensions to be
excluded from deduplication. Each extension must include the
leading dot. The default value is ' ' (empty).
[-duplicate_detection_method {sha1|byte|off}]
0 (off) — This means that duplicate data detection is disabled.
With this setting, every deduplicated file is considered unique
and the only space savings made are accomplished with
compression.
1 (sha1) — The SHA-1 hash is used to detect duplicate data. It
is faster than a byte comparison. This is the default method.
2 (byte) — This will use a byte-by-byte comparison to detect
duplicate data. This adds considerable overhead especially for
large files.
[-savvol_threshold <percent>]
Represents the percentage of the configured save volume
(SavVol) auto extension threshold that can be used during
deduplication. When the specified amount of SavVol is used,
deduplication stops on this file system. By default, this value is 90
percent and the SavVol auto extension is also 90 percent; this
option will apply when the SavVol is 81 percent full (90 * 90).
Setting this value to zero disables it. The values range from 0 to
100.
52
Command Line Interface Reference for File 8.1
The fs Commands
[-backup_data_threshold <percent>]
Indicates the full percentage that a deduplicated file has to be
below in order to trigger space-reduced backups for NDMP. For
example, when set to 90, any deduplicated file whose physical
size (compressed file plus changed blocks) is greater than 90
percent of the logical size of the file will have the entire file data
backed up without attempting to back it up in a space-reduced
format. Setting this value to zero disables it. The values range
from 0 to 200 and the default value is 90 percent.
[-cifs_compression_enabled {yes|no}]
This option controls whether CIFS compression is allowed. When
the default is yes, enable CIFS compression is allowed. When set
to yes and the deduplication state of the file system is either on or
suspended, then CIFS compression is enabled. If the
deduplication state is either off or in the process of being turned
off, then CIFS compression is not allowed, regardless of whether
this option is set to yes.
[-compression_method {fast|deep}]
Indicates whether the compression algorithm is set to fast
(default setting) or deep. This option is valid for VNX systems
that use version 7.1 and later. You can set this value for file
systems only. You cannot set it as a Data Mover global value.
The fast option is the default compression algorithm that achieves
the original compression ratios and performance.
The deep option is the compression algorithm that achieves space
savings up to 30% greater than the fast method. For example, if a
file is 50% compressible then the deep algorithm can compress
the same file up to 65%. However, the compression and
decompression time when using this deep option is longer than
when using the fast option. You obtain more storage space at the
cost of slower access. Selecting this deep compression method
applies only to new files that are subsequently compressed, and
not to existing compressed files.
When using VNX Replicator™, VNX systems that use version 7.0
and earlier cannot read the deep compression format and will
return an I/O error if a read operation is attempted. Select the
deep compression format only if downstream replication sessions
are using compatible software or are scheduled to be upgraded
soon.
fs_dedupe
53
The fs Commands
-clear {<fs_name>|id=<fs_id>}
Sets the file system setting back to the Data Mover setting, which is
the default setting.
[-minimum_scan_interval]
Defines the minimum number of days between completing one
scan of a file system and before scanning the same file system
again. The values range from 1 to 365 and the default value is 7
days.
[-minimum_size]
Defines the file size in KB that limits deduplication. File sizes
equal to this value or smaller will not be deduplicated. File sizes
greater than this value will be candidates for deduplication.
Setting this value to zero disables it. This value should not be set
lower than 24 KB. The values range from 0 to 1000 and the default
value is 24 KB.
[-maximum_size]
Specifies the file size in MB of the largest file to be processed for
deduplication. Files larger than this size in MB will not be
deduplicated. Setting this value to zero disables it. The values
range from 0 to 8388608 and the default value is 8388608 MB.
[-access_time]
Defines the minimum required file age in days based on read
access time. Files that have been read within the specified number
of days will not be deduplicated. This setting does not apply to
files with FLR locked state. Setting this value to zero disables it.
The values range from 0 to 365 and the default value is 15 days.
[-modification_time]
Defines the minimum required file age in days based on
modification time. Files updated within the specified number of
days will not be deduplicated. Setting this value to zero disables
it. The values range from 0 to 365 and the default value is 15 days.
[-case_sensitive]
Defines whether case-sensitive (for NPS environments) or
case-insensitive (for CIFS environments) string comparisons will
be used during scans. By default, case-insensitive comparisons
will be done to be consistent for CIFS environments. The default
value is zero (false).
[-pathname_exclude_list]
This is a file system setting only (no global setting). It is empty by
default.
54
Command Line Interface Reference for File 8.1
The fs Commands
Specifies a semicolon-delimited list of relative pathnames, in
UTF-8 format, to be excluded from deduplication. Any directory
below a specified pathname will be excluded from deduplication.
You can specify a maximum of 10 pathnames and each one can be
up to 1024 bytes. The default value is ' ' (empty).
[-file_ext_exclude_list]
Specifies a colon-delimited list of filename extensions to be
excluded from deduplication. Each extension must include the
leading dot. The default value is ' ' (empty).
[-duplicate_detection_method {sha1|byte|off}]
0 (off) — This means that duplicate data detection is disabled.
With this setting, every deduplicated file is considered unique
and the only space savings made are accomplished with
compression.
1 (sha1) — The SHA-1 hash is used to detect duplicate data. It
is faster than a byte comparison. This is the default method.
2 (byte) — This will use a byte-by-byte comparison to detect
duplicate data. This adds considerable overhead especially for
large files.
[-savvol_threshold]
Represents the percentage of the configured save volume
(SavVol) auto extension threshold that can be used during
deduplication. After the specified amount of SavVol is used,
deduplication stops on this file system. By default, this value is 90
percent and the SavVol auto extension is also 90 percent; this
option will apply when the SavVol is 81 percent full (90 percent of
90 percent). Setting this value to zero disables it. The values range
from 0 to 100.
WARNING
If you set the SavVol threshold option to 0 to disable it, be aware
that the SavVol may grow up to the size of the compressed version of
the data, consuming disk space that cannot be reclaimed unless you
delete all checkpoints.
[-backup_data_threshold]
Indicates the full percentage that a deduplicated file has to be
below in order to trigger space-reduced backups for NDMP. For
example, when set to 90, any deduplicated file whose physical
size (compressed file plus changed blocks) is greater than 90
fs_dedupe
55
The fs Commands
percent of the logical size of the file will have the entire file data
backed up without attempting to back it up in a space-reduced
format. Setting this value to zero disables it. The values range
from 0 to 200 and the default value is 90 percent.
[-cifs_compression_enabled]
This option controls whether CIFS compression is allowed. When
the default is yes, enable CIFS compression. When set to yes and
the deduplication state of the file system is either on or
suspended, then CIFS compression is allowed. If the
deduplication state is either off or in the process of being turned
off, then CIFS compression is not allowed, regardless of whether
this option is set to yes.
[-compression_method]
This is a file system setting only (no global setting). Identifies the
compression algorithm: fast (default) or deep.
| -default {-info {<mover_name>|-all}| -set
{<mover_name>|-all}
Manages the Data Mover settings. The -set option determines the
Data Mover settings.
[-minimum_scan_interval <days>]
Defines the minimum number of days between completing one
scan of a file system and before scanning the same file system
again. The values range from 1 to 365 and the default value is 7
days.
[-minimum_size <KB>]
Defines the file size in KB that limits deduplication. File sizes
equal to this value or smaller will not be deduplicated. File sizes
greater than this value will be candidates for deduplication.
Setting this value to zero disables it. This value should not be set
lower than 24 KB. The values range from 0 to 1000 and the default
value is 24 KB.
[-maximum_size <MB>]
Defines the file size in MB of the largest file to be processed for
deduplication. Files larger than this size in MB will not be
deduplicated. Setting this value to zero disables it. The values
range from 0 to 8388608 and the default value is 8388608 MB.
[-access_time <days>]
Specifies the minimum required file age in days based on read
access time. Files that have been read within the specified number
56
Command Line Interface Reference for File 8.1
The fs Commands
of days will not be deduplicated. This setting does not apply to
files with FLR locked state. Setting this value to zero disables it.
The values range from 0 to 365 and the default value is 15 days.
[-modification_time <days>]
Specifies the minimum required file age in days based on
modification time. Files updated within the specified number of
days will not be deduplicated. Setting this value to zero disables
it. The values range from 0 to 365 and the default value is 15 days.
[-case_sensitive {yes|no}]
Defines whether case-sensitive (for NPS environments) or
case-insensitive (for CIFS environments) string comparisons will
be used during scans. By default, case-insensitive comparisons
will be done to be consistent for CIFS environments. The default
value is zero (false).
[-file_ext_exclude_list <ext_list>]
Specifies a colon-delimited list of filename extensions to be
excluded from deduplication. Each extension must include the
leading dot. The default value is ' ' (empty).
[-duplicate_detection_method {sha1|byte|off}]
0 (off) — This means that duplicate data detection is disabled.
With this setting, every deduplicated file is considered unique
and the only space savings made are accomplished with
compression.
1 (sha1) — The SHA-1 hash is used to detect duplicate data. It
is faster than a byte comparison. This is the default method.
2 (byte) — This will use a byte-by-byte comparison to detect
duplicate data. This adds considerable overhead especially for
large files.
[-savvol_threshold <percent>]
Represents the percentage of the configured save volume
(SavVol) auto extension threshold that can be used during
deduplication. After the specified amount of SavVol is used,
deduplication stops on this file system. By default, this value is 90
percent and the SavVol auto extension is also 90 percent; this
option will apply when the SavVol is 81 percent full (90 * 90).
Setting this value to zero disables it. The values range from 0 to
100.
fs_dedupe
57
The fs Commands
WARNING
If you set the SavVol threshold option to 0 to disable it, be aware
that the SavVol may grow up to the size of the compressed version of
the data, consuming disk space that cannot be reclaimed unless you
delete all checkpoints.
[-cpu_usage_low_watermark <percent>]
Defines the average percent of CPU usage that can be used
during the deduplication process at which full throttle mode is
re-entered. The values range from 0 to 100 and the default value is
40 percent. This is a global setting only.
[-cpu_usage_high_watermark <percent>]
Defines the average percent of CPU usage that can be used
during the deduplication process which should trigger a slow
throttle mode. The system starts in full throttle mode. The values
range from 0 to 100 and the default value is 75 percent. This is a
global setting only.
[-backup_data_threshold <percent>]
Defines the full percentage that a deduplicated file has to be
below in order to trigger space-reduced backups for NDMP. For
example, when set to 90, any deduplicated file whose physical
size (compressed file plus changed blocks) is greater than 90
percent of the logical size of the file will have the entire file data
backed up without attempting to back it up in a space-reduced
format. Setting this value to zero disables it. The values range
from 0 to 200 and the default value is 90 percent.
[-cifs_compression_enabled {yes|no}]
This option controls whether CIFS compression is allowed. When
the default is yes, enable CIFS compression. When set to yes and
the deduplication state of the file system is either on or
suspended, then CIFS compression is allowed. If the
deduplication state is either off or in the process of being turned
off, then CIFS compression is not allowed, regardless of whether
this option is set to yes.
| -clear {<mover_name>|-all}
The -clear option sets the global setting back to the default value.
58
Command Line Interface Reference for File 8.1
The fs Commands
[-minimum_scan_interval]
Defines the minimum number of days between completing one
scan of a file system and before scanning the same file system
again. The values range from 1 to 365 and the default value is 7
days.
[-minimum_size]
Defines the file size in KB that limits deduplication. File sizes
equal to this value or smaller will not be deduplicated. File sizes
greater than this value will be candidates for deduplication.
Setting this value to zero disables it. This value should not be set
lower than 24 KB. The values range from 0 to 1000 and the default
value is 24 KB.
[-maximum_size]
Defines the file size in MB of the largest file to be processed for
deduplication. Files larger than this size in MB will not be
deduplicated. Setting this value to zero disables it. The values
range from 0 to 8388608 and the default value is 8388608 MB.
[-access_time]
Defines the minimum required file age in days based on read
access time. Files that have been read within the specified number
of days will not be deduplicated. This setting does not apply to
files with FLR locked state. Setting this value to zero disables it.
The values range from 0 to 365 and the default value is 15 days.
[-modification_time]
Defines the minimum required file age in days based on
modification time. Files updated within the specified number of
days will not be deduplicated. Setting this value to zero disables
it. The values range from 0 to 365 and the default value is 15 days.
[-case_sensitive]
Defines whether case-sensitive (for NPS environments) or
case-insensitive (for CIFS environments) string comparisons will
be used during scans. By default, case-insensitive comparisons
will be done to be consistent for CIFS environments. The default
value is zero (false).
[-file_ext_exclude_list]
Specifies a colon-delimited list of filename extensions to be
excluded from deduplication. Each extension must include the
leading dot. The default value is ' ' (empty).
fs_dedupe
59
The fs Commands
[-duplicate_detection_method]
0 (off) — This means that duplicate data detection is disabled.
With this setting, every deduplicated file is considered unique
and the only space savings made are accomplished with
compression.
1 (sha1) — The SHA-1 hash is used to detect duplicate data. It
is faster than a byte comparison. This is the default method.
2 (byte) — This will use a byte-by-byte comparison to detect
duplicate data. This adds considerable overhead especially for
large files.
[-savvol_threshold]
Represents the percentage of the configured save volume
(SavVol) auto extension threshold that can be used during
deduplication. After the specified amount of SavVol is used,
deduplication stops on this file system. By default, this value is 90
percent and the SavVol auto extension is also 90 percent; this
option will apply when the SavVol is 81 percent full (90 * 90).
Setting this value to zero disables it. The values range from 0 to
100.
[-cpu_usage_low_watermark]
Specifies the average percent of CPU usage that can be used
during the deduplication process at which full throttle mode is
re-entered. The values range from 0 to 100 and the default value is
25 percent. This is a global setting only.
[-cpu_usage_high_watermark]
Specifies the average percent of CPU usage that can be used
during the deduplication process which should trigger a slow
throttle mode. The system starts in full throttle mode. The values
range from 0 to 100 and the default value is 75 percent. This is a
global setting only.
[-backup_data_threshold <percent>]
Specifies the full percentage that a deduplicated file has to be
below in order to trigger space-reduced backups for NDMP. For
example, when set to 90, any deduplicated file whose physical
size (compressed file plus changed blocks) is greater than 90
percent of the logical size of the file will have the entire file data
backed up without attempting to back it up in a space-reduced
format. Setting this value to zero disables it. The values range
from 0 to 200 and the default value is 90 percent.
60
Command Line Interface Reference for File 8.1
The fs Commands
[-cifs_compression_enabled]
This option controls whether CIFS compression is allowed. When
the default is yes, enable CIFS compression. When set to yes and
the deduplication state of the file system is either on or
suspended, then CIFS compression is allowed. If the
deduplication state is either off or in the process of being turned
off, then CIFS compression is not allowed, regardless of whether
this option is set to yes.
SEE ALSO:
EXAMPLE #1
nas_fs
To list the file systems and their deduplication states, type:
$ fs_dedupe -list
id
name
state
status time_of_last_scan original_
data_size
141 ranap1_
Suspended Idle
Wed Nov 12
5 MB
replica
09:04:45 EST 2008
104 ds850gb_ On
Idle
Fri Nov 21
875459 MB
replica1
10:31:15 EST 2008
495 cworm
On
Idle
Thu Nov 20
3 MB
09:14:09 EST 2008
33 chrisfs1 On
Idle
Sat Nov 22
1100 MB
10:04:33 EST 2008
usage space_saved
0%
0 MB (0%)
84%
341590 MB(39%)
0%
0 MB (0%)
18%
424 MB (38%)
Where:
Value
id
name
state
status
time_of_last_scan
Definition
File system identifier.
Name of the file system.
Deduplication state of the file system. The file data is transferred to the storage which performs
the deduplication and compression on the data. The states are:
• On — Deduplication on the file system is enabled.
• Suspended — Deduplication on the file system is suspended. Deduplication does not
perform any new space reduction but existing space-reduced files remain the same.
• Off — Deduplication on the file system is disabled. Deduplication does not perform any new
space reduction and the data is now reduplicated, which is the process used to restore a file
that was deduplicated to its original condition.
Current state of the deduplication-enabled file system. The progress statuses are:
• Idle — Deduplication process is currently idle.
• Scanning — File system is being scanned for deduplication. It displays the percentage of
scanned files in the file system.
• Reduplicating — File system files are being reduplicated from the deduplicated files. It
displays the percentage of reduplicated files.
Time when the file system was last scanned.
fs_dedupe
61
The fs Commands
Value
original_data_size
usage
space_saved
EXAMPLE #2
Definition
Original size of the file system before deduplication.
The current space usage of the file system.
The file system space saved after deduplication.
To list the file systems and provide detailed reports on the state of the
deduplication processing, type:
$ fs_dedupe -info -all
Id
Name
Deduplication
File system parameters:
Case Sensitive
Duplicate Detection Method
Access Time
Modification Time
Minimum Size
Maximum Size
File Extension Exclude List
Minimum Scan Interval
Savevol Threshold
Backup Data Threshold
Cifs Compression Enabled
Pathname Exclude List
Compression Method
= 53
= svr2fs1
= Off
=
=
=
=
=
=
=
=
=
=
=
=
=
no
sha1
15
15
24 KB
8388608 MB
7
90
90
yes
fast
Id
= 2040
Name
= server_2_fsltest2
Deduplication
= Suspended
As of the last file system scan (Mon Aug 17 11:33:38 EDT 2009):
Files scanned
= 4
Files deduped
= 3 (75% of total files)
File system capacity
= 2016 MB
Original data size
= 6 MB (0% of current file system capacity)
Space saved
= 0 MB (0% of original data size)
File system parameters:
Case Sensitive
= no
Duplicate Detection Method = sha1
Access Time
= 15
Modification Time
= 15
Minimum Size
= 24 KB
Maximum Size
= 8388608 MB
File Extension Exclude List =
Minimum Scan Interval
= 7
Savevol Threshold
= 90
Backup Data Threshold
= 90
Cifs Compression Enabled
= yes
Pathname Exclude List
=
Compression Method
= fast
62
Command Line Interface Reference for File 8.1
The fs Commands
Id
Name
Deduplication
File system parameters:
Case Sensitive
Duplicate Detection Method
Access Time
Modification Time
Minimum Size
Maximum Size
File Extension Exclude List
Minimum Scan Interval
Savevol Threshold
Backup Data Threshold
Cifs Compression Enabled
Pathname Exclude List
Compression Method
= 506
= demofs
= Off
=
=
=
=
=
=
=
=
=
=
=
=
=
no
sha1
15
15
24 KB
8388608 MB
7
90
90
yes
fast
Id
= 2113
Name
= testrdefs
Deduplication
= Suspended
As of the last file system scan (Thu Aug 13 14:22:31 EDT 2009):
Files scanned
= 1
Files deduped
= 0 (0% of total files)
File system capacity
= 1008 MB
Original data size
= 0 MB (0% of current file system capacity)
Space saved
= 0 MB (0% of original data size)
File system parameters:
Case Sensitive
= no
Duplicate Detection Method = sha1
Access Time
= 15
Modification Time
= 15
Minimum Size
= 24 KB
Maximum Size
= 8388608 MB
File Extension Exclude List =
Minimum Scan Interval
= 7
Savevol Threshold
= 90
Backup Data Threshold
= 90
Cifs Compression Enabled
= yes
Pathname Exclude List
=
Compression Method
= deep
Id
Name
Deduplication
File system parameters:
Case Sensitive
Duplicate Detection Method
Access Time
Modification Time
Minimum Size
= 2093
= kfs_ckpt1
= Off
=
=
=
=
=
no
sha1
15
15
24 KB
fs_dedupe
63
The fs Commands
Maximum Size
File Extension Exclude List
Minimum Scan Interval
Savevol Threshold
Backup Data Threshold
Cifs Compression Enabled
Pathname Exclude List
Compression Method
=
=
=
=
=
=
=
=
8388608 MB
7
90
90
yes
fast
Id
= 2095
Name
= ranap-test3
Deduplication
= On
Status
= Idle
As of the last file system scan (Tue Aug 11 17:37:58 EDT 2009):
Files scanned
= 30
Files deduped
= 2 (7% of total files)
File system capacity
= 5041 MB
Original data size
= 1109 MB (22% of current file system capacity)
Space saved
= 0 MB (0% of original data size)
File system parameters:
Case Sensitive
= no
Duplicate Detection Method = sha1
Access Time
= 15
Modification Time
= 15
Minimum Size
= 24 KB
Maximum Size
= 8388608 MB
File Extension Exclude List =
Minimum Scan Interval
= 7
Savevol Threshold
= 90
Backup Data Threshold
= 90
Cifs Compression Enabled
= yes
Pathname Exclude List
=
Compression Method
= deep
Where:
Value
Deduplication
Status
Name
Id
Files scanned
Files deduped
Original data size
Filesystem capacity
Space saved
Case Sensitive
Duplicate Detection Method
Access Time
64
Definition
Current deduplication state of the file system.
Progress status of the files being scanned.
Name of the file system.
File system identifier.
Number of files scanned.
Number of files in the file system that has been deduplicated.
Proportion of space in use with respect to the file system capacity.
Current space usage of the file system.
Proportion of space saved with respect to the original data size.
Method of string comparison: case-sensitive or case-insensitive.
Method of duplication detection: 0, sha-1, or byte-by-byte.
Minimum required file age in days based on read access time.
Command Line Interface Reference for File 8.1
The fs Commands
Modification Time
Minimum required file age in days based on modification time.
Minimum Size
Minimum file size to be processed for deduplication.
Maximum Size
Maximum file size to be processed for deduplication.
File Extension Exclude List Lists filename extensions to be excluded from the deduplication.
Minimum Scan Interval
Minimum number of days between completing one scan of a file system and
before scanning the same file system again.
SavVol Threshold
Percentage of SavVol space that can be used during deduplication.
Backup Data Threshold
Percentage below which a deduplicated file has to be in order to trigger
space-reduced NDMP backups.
Cifs Compression Enabled
Controls whether CIFS permission is enabled.
Pathname Exclude List
Lists relative pathnames to be excluded from the deduplication.
Compression Method
Compression algorithm used: fast or deep.
Note: If reduplication fails, then the state transitions to the suspended state
and a CCMD message will be sent to the server’s event log. If reduplication
succeeds, then it remains in the off state.
EXAMPLE #3
To list the file systems for a given file system name, type:
$ fs_dedupe -info server3_fs3
Id
= 98
Name
= server3_fs3
Deduplication
= On
Status
= Idle
As of the last file system scan on Tue Sep 23 13:28:01 EDT 2008:
Files deduped
= 30 (100%)
File system capacity
= 413590 MB
Original data size
= 117 MB (0% of current file system capacity)
Space saved
= 106 MB (90% of original data size)
File system parameters:
Case Sensitive
= yes
Duplicate Detection Method
= sha1
Access Time
= 30
Modification Time
= 30
Minimum Size
= 20
Maximum Size
= 200
File Extension Exclude List
= .jpg:.db:.pst
Minimum Scan Interval
= 1
SavVol Threshold
= 90
Backup Data Threshold
= 90
Pathname Exclude List
= root;etc
Compression Method
= fast
EXAMPLE #2 provides a description of command output.
EXAMPLE #4
To list the duplication properties of a given Data Mover, type:
$ fs_dedupe -default -info server_2
fs_dedupe
65
The fs Commands
Server parameters:
Case Sensitive
Duplicate Detection Method
Access Time = 30
Modification Time
Minimum Size
Maximum Size
File Extension Exclude List
Minimum Scan Interval
SavVol Threshold
Backup Data Threshold
CPU % Usage Low Water Mark
CPU % Usage High Water Mark
Cifs Compression Enabled
= yes
= sha1
=
=
=
=
=
=
=
=
=
=
30
20
200
.jpg:.db:.pst
1
90
90
25
90
yes
Where:
Value
Deduplication
Status
Name
Id
Files scanned
Files deduped
Original data size
File system capacity
Space saved
Case Sensitive
Duplicate Detection
Method
Access Time
Modification Time
Minimum Size
Maximum Size
File Extension
Exclude List
Minimum Scan Interval
Definition
Current deduplication state of the file system.
Progress status of the files being scanned.
Name of the file system.
File system identifier.
Number of files scanned.
Number of files in the file system that has been deduplicated.
Proportion of space in use with respect to the file system capacity.
Current space usage of the file system.
Proportion of space saved with respect to the original data size.
Method of string comparison: case-sensitive or case-insensitive.
Method of duplication detection: 0, sha-1, or byte-by-byte.
Minimum required file age in days based on read access time.
Minimum required file age in days based on modification time.
Minimum file size to be processed for deduplication.
Maximum file size to be processed for deduplication.
Lists filename extensions to be excluded from the deduplication.
Minimum number of days between completing one scan of a file system and before scanning
the same file system again.
SavVol Threshold
Percentage of SavVol space that can be used during deduplication.
Backup Data Threshold Percentage below which a deduplicated file has to be in order to trigger space-reduced
NDMP backups.
CPU % Usage Low Water Average percentage of CPU usage which should trigger full throttle mode.
Mark
CPU % Usage High
Average percentage of CPU usage which should trigger slow throttle mode.
Water Mark
66
Command Line Interface Reference for File 8.1
The fs Commands
EXAMPLE #5
To modify the file system, type:
$ fs_dedupe -modify testrdefs -state on
Done
EXAMPLE #6
To modify the file system settings to the user-specified values, type:
$ fs_dedupe -modify testrdefs -maximum_size 100
-file_extension_exclude_list .jpg:.db:.pst
Done
EXAMPLE #7
To modify specific Data Mover settings, type:
$ fs_dedupe -default -set server_2 -maximum_size 100
-minimum_size 20 -duplicate_detection_method sha1
server_2: Done
EXAMPLE #8
To reset the file system settings to the default settings (which are the
Data Mover settings), type:
$ fs_dedupe -clear testrdefs -maximum_size -minimum_size
-duplicate_detection_method
Done
EXAMPLE #9
To reset specific Data Mover settings to the default settings, type:
$ fs_dedupe -default -clear server_2 -maximum_size
-minimum_size -duplicate_detection_method
server_2: Done
EXAMPLE #10
To reset all options for a specific Data Mover to the default settings,
type:
$ fs_dedupe -default -clear server_2
server_2: Done
EXAMPLE #11
To reset all options on all Data Movers to the default settings, type:
$ fs_dedupe -default -clear -all
server_2: Done
server_3: Done
fs_dedupe
67
The fs Commands
fs_dhsm
Manages the VNX FileMover file system connections.
SYNOPSIS
fs_dhsm
-list
| -info [<fs_name>|id=<fs_id>]
| -modify {<fs_name>|id=<fs_id>}[-state enabled]
[-popup_timeout <sec>][-backup {offline|passthrough}]
[-log {on|off}][-max_log_size <mb>][-offline_attr {on|off}]
[-read_policy_override {none|full|passthrough|partial}]}
| -modify {<fs_name>|id=<fs_id>}[-state disabled]
| -connection {<fs_name>|id=<fs_id>}
-list
| -info [<cid>]
| -create -type {nfsv3|nfsv2} -secondary <nfs_server>:/<path>
[-read_policy_override {full|passthrough|partial|none}]
[-useRootCred {true|false}][-proto {UDP|TCP}][-nfsPort <port>]
[-mntPort <port>][-mntVer {3|2|1}][-localPort <port>]
|-create -type cifs -admin [<fqdn>\]<admin_name>
-secondary \\<fqdn>\<share>[\<path>]
-local_server <host_name> [-wins <address>][-password <password>]
[-read_policy_override {full|passthrough|partial|none}]
|-create -type http -secondary http://<host><url_path>
[-read_policy_override {full|passthrough|partial|none}]
[-httpPort <port>][-localPort <port>]
[-user <username> [-password <password>]]
[-timeout <seconds>][-cgi {y|n}]
|-create -type https -secondary https://<host><url_path>
[-read_policy_override {full|passthrough|partial|none}]
[-httpsPort <port>][-localPort <port>]
[-user <username> [-password <password>]]
[-timeout <seconds>][-cgi {y|n}]
| -delete {-all|<cid>[,<cid>...]} [-recall_policy {check|no|yes}]
| -modify {-all|<cid>[,<cid>...]} [-state {enabled|disabled|recallonly}]
[-read_policy_override {full|passthrough|partial|none}]
[{[-nfs_server <address>] [-localPort <port>]
[-proto {TCP|UDP}] [-useRootCred {true|false}]}
|{[-cifs_server <fqdn>][-local_server <host_name>]
[-password <password>][-admin [<fqdn>\]<admin_name>]
[-wins <address>]}
|{[-http_server <host>][-httpPort <port>][-httpsPort <port>]
[-localPort <port>][-user <username>]
[-password <password>][-timeout <seconds>]}
68
Command Line Interface Reference for File 8.1
The fs Commands
DESCRIPTION
OPTIONS
The fs_dhsm command modifies the properties on file systems
enabled for VNX FileMover. The fs_dhsm command creates, deletes,
and modifies NFS, CIFS, and HTTP connections to remote hosts, lists
VNX FileMover file systems, and provides information on the
connections.
-list
Lists all file systems enabled with the VNX FileMover.
-info [<fs_name>|id=<fs_id>]
Displays information for the specified VNX FileMover file systems.
-modify {<fs_name>|id=<fs_id>}
Sets VNX FileMover parameters for the specified file system.
Note: When specifying the -modify option on a disabled file system, the state
is automatically changed to enabled. When specifying the -state disabled
option, it is not possible to specify any other parameter to modify.
[-state enabled]
Enables VNX FileMover operations on the specified file system.
The file system must be enabled to accept other options.
[-state disabled]
Disables VNX FileMover operations on the specified file system.
New FileMover attributes cannot be specified as part of a disable
command, nor can be specified for a file system that is in the
disabled state. The attributes persist. If the file system is enabled
after a disable command, then the attributes prior to the disable
command take effect.
[-popup_timeout <sec>]
Specifies the Windows popup timeout value in seconds. If a CIFS
I/O request cannot be processed within the specified time, then a
popup notification of the delay is sent to the CIFS client. The
default for <sec> is 0 (zero) which disables Windows popups.
Note: It may take up to 10 seconds before the popup is displayed.
[-backup {offline|passthrough}]
Specifies the nature of CIFS network backups. The offline option
backs up the stub file only. The passthrough (default) option
backs up all of the file data by using passthrough read.
fs_dhsm
69
The fs Commands
[-log {on|off}]
Enables or disables VNX FileMover logging. The default log
filename is dhsm.log; it resides in the /.etc directory on the
FileMover-enabled file system.
[-max_log_size <mb>]
Specifies the maximum size of the log file. The current log file, in
addition to four old log files, is saved. The minimum log file size
is 10 MB.
[-offline_attr {on|off}]
Specifies whether the Data Mover should set the CIFS offline file
attributes on the stub files. The default is on.
!
CAUTION
It is recommended that you do not disable the CIFS offline
attributes.
[-read_policy_override {none|full|passthrough|
partial}]
Specifies the migration method option used by the VNX, in the
connection level or file system level, to override the migration
method specified in the stub file. none (default) specifies no
override, full recalls the whole file to the VNX on a read request
before the data is returned, passthrough retrieves data without
recalling the data to the VNX, and partial recalls only the blocks
required to satisfy the client read request.
Note: The full migration may take several minutes or hours if the file is
very large.
-connection {<fs_name>|id=<fs_id>} -list
Lists all connections for the specified file system.
-connection {<fs_name>|id=<fs_id>} -info [<cid>]
Displays details on all connections for the specified file system. If the
<cid> is specified, only information for that connection is displayed.
Note: A connection ID is automatically created when a connection is
established. The connection ID is displayed using the -list and is referred to
as the <cid> in other commands.
70
Command Line Interface Reference for File 8.1
The fs Commands
NFS CONNECTIONS
-connection {<fs_name>|id=<fs_id>} -create -type
{nfsv3|nfsv2} -secondary <nfs_server>:/<path>
Creates a connection using the NFS protocol between the specified
file system and the secondary file system. The secondary file system
stores migrated data. The -type option specifies the NFS version that
the Data Mover should use when connecting to the secondary server.
Note: VNX FileMover does not currently support NFSv4 protocol.
The -secondary option specifies the location of the remote file system.
Note: Although an IP address can be specified for an <nfs_server>, EMC
strongly suggests using the hostname of the server, which allows you to take
advantage of Domain Name System (DNS) failover capability.
[-read_policy_override {full|passthrough|partial
|none}]
Specifies the migration method for data recall in response to
client read requests. full migrates the whole file before it returns
the requested blocks. passthrough leaves the stub file, but
retrieves the requested data from the secondary file system.
partial migrates only the blocks required to satisfy the client read
request. none (default) defaults to the read method option
specified in the stub file.
Note: The full migration may take several minutes or hours if the file is
very large.
[-useRootCred {true|false}]
Specifies the user credentials that the Data Mover uses when
requesting data from the secondary VNX. When set to true, the
Data Mover requests data as the root user (UID 0). When set to
false (default), the Data Mover requests data as the owner of the
file as specified in the stub file.
Note: If the -useRootCred option is set to true, the secondary storage
NFS server must grant the Data Mover root privilege for NFS traffic.
[-proto {TCP|UDP}]
Specifies the protocol for the Data Movers to use for
communication to the secondary <nfs_server>. TCP is the
default.
fs_dhsm
71
The fs Commands
[-nfsPort <port>]
Specifies an NFS port on the secondary <nfs_server>. A default
port is discovered automatically.
[-mntPort <port>]
Specifies a mount port on the secondary <nfs_server>. A default
mount port is discovered automatically.
Note: The -nfsPort and -mntPort options are used for secondary servers
that do not have the Portmapper running. The admin starts the nfsd and
mountd daemons on specific ports to avoid hackers.
[-mntVer {1|2|3}]
Specifies the mount version for the NFS connection. If the -type is
nfsv3, then the -mntVer must be 3. If the -type is nfsv2, then 1 or 2
can be specified. The default for nfsv2 is 2.
[-localPort <port>]
Overrides the default port that the Data Mover uses during
connection to be compatible with firewalls. The default for UDP is
1020. By default, TCP uses a random port over 1024 to make the
connection.
-connection {<fs_name>|id=<fs_id>} -modify {-all|
<cid>[,<cid>...]}
Changes parameters on an existing NFS VNX FileMover connection.
Either all connections can be removed or just the specified <cid>
connection can be removed.
[-state {enabled|disabled|recallonly}]
Sets the state of VNX FileMover operations on the specified file
system. enabled (default) allows both the creation of stub files
and data migration through reads and writes. If the state is
disabled, neither stub files nor data migration is possible. Data
currently on the VNX can be read and written to in the disabled
state.
If the state is recallonly, the policy engine is not allowed to create
stub files, but the user is still able to trigger data migration using
a read or write request from the secondary file system to the VNX.
[-read_policy_override {full|passthrough|partial
|none}]
Specifies the migration method option used by the VNX, in the
connection level or file system level, to override the migration
method specified in the stub file. none (default) specifies no
72
Command Line Interface Reference for File 8.1
The fs Commands
override, full recalls the whole file to the VNX on read request
before the data is returned, passthrough retrieves data without
recalling the data to the VNX, and partial recalls only the blocks
required to satisfy the client read request.
Note: The full migration may take several minutes or hours if the file is
very large.
[-nfs_server <address>]
Specifies the name or IP address of the secondary NFS server.
Note: Although an IP address can be specified for the <nfs_server>, EMC
strongly suggests using the hostname of the server, which allows use of
the DNS failover capability.
[-localPort <port>]
Specifies a port to override the default port used by the Data
Mover during connection for compatibility with firewalls.
[-proto {TCP|UDP}]
Specifies the protocol for the Data Mover to use for NFS
communications to the secondary <nfs_server>. TCP is the
default.
[-useRootCred {true|false}]}
Specifies the user credentials that the Data Mover uses when
requesting data from the secondary VNX. When set to true,
the Data Mover requests data as the root user (UID 0). When
set to false (default), the Data Mover requests data as the
owner of the file as specified in the stub file.
Note: If the -useRootCred option is set to true, the secondary storage
NFS server must grant the Data Mover root privilege for NFS traffic.
-connection <fs_name> -delete {-all|<cid> [,<cid>...]}
Removes an existing NFS connection between the file system and the
secondary file system. Either all connections can be removed or just
the specified <cid> connection can be removed.
[-recall_policy {check|no|yes}]
Specifies the recall policy for any migrated file during the -delete.
check (default) scans the file system for stub files that depend on
the connection and fails on the first one. no deletes the connection
without checking for stub files that depend on the connection,
fs_dhsm
73
The fs Commands
and yes migrates the files back to the VNX before the connection
is removed. If no is specified and stub files exist, an I/O error
appears when the file is read because the connection no longer
exists.
CIFS CONNECTIONS
-connection {<fs_name>|id=<fs_id>} -create -type cifs
Creates a connection using the CIFS protocol between the specified
file system and a secondary file system. A connection ID is
automatically created when a connection is established. The
connection ID is seen using the -list and is referred to as the <cid> in
other commands.
-admin [<fqdn>\]<admin_name>
Specifies the <admin_name> used to make the CIFS connection.
If an optional <fqdn> is specified, it must be a fully qualified
domain name. The [<fqdn>\]<admin_name> entry must be
enclosed within quotes as shown in EXAMPLE #2. If the <fqdn>
is not specified, the -local_server domain is used.
-secondary \\<fqdn>\<share>[\<path>]
Specifies the CIFS server, the share, and path for the secondary
server for connection. The <fqdn>\<share>[\<path>] entry must
be enclosed within quotes. The domain must be fully qualified; an
IP address will not work.
-local_server <host_name>
Specifies the NetBIOS name or computer name of the local CIFS
server on the Data Mover.
[-wins <address>]
Specifies a WINS server to resolve names in a Windows domain.
[-password <password>]
Allows the user to specify the admin password. The password is
not recorded in the command log. If the -password option is
given but no password is specified, the user is prompted
interactively.
!
CAUTION
When specifying the password with this option, be aware it is
unmasked, and visible to other users. The command may also
be read from the log of the shell.
74
Command Line Interface Reference for File 8.1
The fs Commands
[-read_policy_override {full|passthrough|partial
|none}]
Specifies the migration method for data recall in response to
client read requests. full migrates the whole file before it returns
the requested blocks. passthrough leaves the stub file, but
retrieves the requested data from the secondary file system.
partial migrates only the blocks required to satisfy the client read
request. none (default) defaults to the read method option
specified in the stub file.
Note: The full migration may take several minutes or hours if the file is
very large.
-connection {<fs_name>|id=<fs_id>} -modify {-all|
<cid>[,<cid>...]}
Changes parameters on an existing NFS VNX FileMover connection.
[-state {enabled|disabled|recallonly}]
Sets the state of VNX FileMover operations on the specified file
system. enabled (default) allows both the creation of stub files
and data migration through reads and writes. If the state is
disabled, neither stub files nor data migration is possible. Data
currently on the VNX can be read and written to in the disabled
state.
If the state is recallonly, the policy engine is not allowed to create
stub files, but the user is still able to trigger data migration using
a read or write request from the secondary file system to the VNX.
[-read_policy_override {full|passthrough|partial
|none}]
Specifies the migration method option used by the VNX, in the
connection level or file system level, to override the migration
method specified in the stub file. none (default) specifies no
override, full recalls the whole file to the VNX on read request
before the data is returned, passthrough retrieves data without
recalling the data to the VNX, and partial recalls only the blocks
required to satisfy the client read request.
Note: The full migration may take several minutes or hours if the file is
very large.
fs_dhsm
75
The fs Commands
[-cifs_server <fqdn>]
Specifies the fully qualified domain name of the secondary CIFS
server.
[-local_server <host_name>]
Specifies the NetBIOS name or computer name of the local
CIFS server on the Data Mover.
[-password <password>]
Allows the user to specify the admin password. The password
is not recorded in the command log. If the -password option is
given but no password is specified, the user is prompted
interactively.
!
CAUTION
When specifying the password with this option, be aware it is
unmasked, and visible to other users. The command may also be read
from the log of the shell.
[-admin [<fqdn>\]<admin_name>]
Specifies the <admin_name> used to make the CIFS
connection. If an optional <fqdn> is specified, it must be a
fully qualified domain name. If the <fqdn> is not specified, the
-local_server domain is used.
[-wins <address>]}
Specifies a WINS server to resolve names in a Windows
domain.
-connection <fs_name> -delete {-all|<cid> [,<cid>...]}
Removes an existing CIFS connection between the file system and the
secondary file system.
[-recall_policy {check|no|yes}]
Specifies the recall policy for any migrated file during the -delete
option. check (default) scans the file system for stub files that
depend on the connection and fails on the first one. no deletes the
connection without checking for stub files that depend on the
connection, and yes migrates the files back to the VNX before the
connection is removed. If no is specified and stub files exist, an
I/O error appears when the file is read because the connection no
longer exists.
HTTP CONNECTIONS
76
-connection {<fs_name>|id=<fs_id>} -create -type http
-secondary http://<host><url_path>
Command Line Interface Reference for File 8.1
The fs Commands
Creates a connection using the HTTP protocol between the specified
primary file system and a secondary file system. There are two types
of HTTP connections: CGI and non-CGI. For CGI connections, the
value of the -secondary option specifies the hostname of the server
running the secondary storage HTTP server and the location of the
CGI application that provides access to a system. For non-CGI
connections, the value for the -secondary option specifies the
hostname and, optionally, a portion of the hierarchical namespace
published by the web server.
Note: Although an IP address can be specified for a <host>, EMC strongly
suggests using the hostname of the server, which allows the DNS failover
capability.
[-read_policy_override {full|passthrough|partial
|none}]
Specifies the migration method option used by the VNX, in the
connection level or file system level, to override the migration
method specified in the stub file. none (default) specifies no
override, full recalls the whole file to the VNX on read request
before the data is returned, passthrough retrieves data without
recalling the data to the VNX, and partial recalls only the blocks
required to satisfy the client read request.
Note: The full migration may take several minutes or hours if the file is
very large.
-httpPort <port>
Specifies the remote port number that the Data Mover delivers
the HTTP request to. If not specified, the Data Mover issues HTTP
requests to port 80 on the secondary storage HTTP server.
-localPort <port>
Specifies the local port number the Data Mover uses to issue
HTTP requests to the web server active on the secondary storage.
The <port> specified should be an integer number less than 1024.
If not specified, the Data Mover selects a port to issue the HTTP
requests.
fs_dhsm
77
The fs Commands
Note: The two end points of an HTTP connection are specified by the file
system name and the value specified for the -secondary option. If
multiple connections are created by using identical end points with
different attributes such as -cgi, -user, -password, -localPort, -httpPort,
the connection will fail.
[-user <username>]
Defines the username the HTTP client uses if digest
authentication is required by the secondary storage HTTP server.
[-password <password>]
Allows the user to specify the admin password. The password is
not recorded in the command log. If the -password option is
given but no password is specified, the user is prompted
interactively.
Use the -password option when digest authentication is required
by the secondary storage HTTP server.
[-timeout <seconds>
Specifies the timeout value in seconds. By default, the VNX HTTP
client waits 30 seconds for a reply from the HTTP server and then
retries the operation once.
[-cgi {y|n}
Specifies the HTTP connection type: CGI or non-CGI. By default,
FileMover assumes that the web server is using CGI connections
to access migrated file data by using a CGI application. For
non-CGI connections, set the -cgi option to n; FileMover then
assumes the web server has direct access to migrated file content
on secondary storage.
-connection {<fs_name>|id=<fs_id>} -modify {-all|
<cid>[,<cid>...]}
Changes parameters on an existing NFS VNX FileMover connection.
[-state {enabled|disabled|recallonly}]
Sets the state of VNX FileMover operations on the specified file
system. enabled (default) allows both the creation of stub files
and data migration through reads and writes. If the state is
disabled, neither stub files nor data migration is possible. Data
currently on the VNX can be read and written to in the disabled
state.
78
Command Line Interface Reference for File 8.1
The fs Commands
If the state is recallonly, the policy engine is not allowed to create
stub files, but the user is still able to trigger data migration by
using a read or write request from the secondary file system to the
VNX.
[-read_policy_override {full|passthrough|partial
|none}]
Specifies the migration method option used by the VNX, in the
connection level or file system level, to override the migration
method specified in the stub file. none (default) specifies no
override, full recalls the whole file to the VNX on read request
before the data is returned, passthrough retrieves data without
recalling the data to the VNX, and partial recalls only the blocks
required to satisfy the client read request.
Note: The full migration may take several minutes or hours if the file is
very large.
[-http_server <host>]
Specifies the hostname of the secondary storage HTTP server.
-httpPort <port>
Specifies the remote port number that the Data Mover delivers
the HTTP request to. If not specified, the Data Mover issues
HTTP requests to port 80 on the secondary storage HTTP
server.
-localPort <port>
Specifies the local port number the Data Mover uses to issue
HTTP requests to the web server active on the secondary
storage. The <port> specified should be an integer number
less than 1024. If not specified, the Data Mover selects a port to
issue the HTTP requests.
Note: If you attempt to create multiple HTTP connections by using
identical end points with different attributes such as -cgi, -user,
-password, -localPort, -httpPort, the connection will fail.
[-user <username>]
An optional attribute used to define the username the HTTP
client uses if digest authentication is required by the
secondary storage HTTP server.
fs_dhsm
79
The fs Commands
[-password <password>]
Allows the user to specify the admin password. The password
is not recorded in the command log. If the -password option is
given but no password is specified, the user is prompted
interactively.
[-timeout <sec>]
Specifies the timeout value in seconds. By default, VNX’s
HTTP client waits 30 seconds for a reply from the HTTP server
and then retries the operation once before commencing the
failover operation.
-connection <fs_name> -delete {-all |<cid> [,<cid>...]}
Removes an existing HTTP connection between the file system and
the secondary file system. Either all connections can be removed or
just the specified <cid> connection can be removed.
[-recall_policy {check|no|yes}]
Specifies the recall policy for any migrated file during the -delete
option. The check (default) argument scans the file system for
stub files that depend on the connection and fails on the first one.
no deletes the connection without checking for stub files that
depend on the connection, and yes migrates the files back to the
VNX before the connection is removed. If no is specified and stub
files exist, an I/O error appears when the file is read because the
connection no longer exists.
HTTPS CONNECTIONS
-connection {<fs_name>|id=<fs_id>} -create -type https
-secondary https://<host><url_path>
Creates a connection by using the HTTPS protocol between the
specified primary file system and a secondary file system. There are
two types of HTTPS connections: CGI and non-CGI. For CGI
connections, the value of the -secondary option specifies the
hostname of the server running the secondary storage HTTPS server
and the location of the CGI application that provides access to a
system. For non-CGI connections, the value for the -secondary option
specifies the hostname and, optionally, a portion of the hierarchical
namespace published by the web server.
Note: Although an IP address can be specified for a <host>, EMC strongly
suggests using the hostname of the server, which allows the DNS failover
capability.
80
Command Line Interface Reference for File 8.1
The fs Commands
[-read_policy_override {full|passthrough|partial
|none}]
Specifies the migration method option used by the VNX, in the
connection level or file system level, to override the migration
method specified in the stub file. none (default) specifies no
override, full recalls the whole file to the VNX on read request
before the data is returned, passthrough retrieves data without
recalling the data to the VNX, and partial recalls only the blocks
required to satisfy the client read request.
Note: The full migration may take several minutes or hours if the file is
very large.
[-httpsPort <port> ]
Specifies the remote port number that the Data Mover delivers
the HTTPS request to. If not specified, the Data Mover issues
HTTPS requests to port 443 on the secondary storage HTTPS
server.
[-localPort <port> ]
Specifies the local port number the Data Mover uses to issue
HTTPS requests to the web server active on the secondary
storage. The <port> specified should be an integer number less
than 1024. If not specified, the Data Mover selects a port to issue
the HTTPS requests.
Note: The two end points of an HTTPS connection are specified by the
file system name and the value specified for the -secondary option. If
multiple connections are created by using identical end points with
different attributes such as -cgi, -user, -password, -localPort, -httpsPort,
the connection will fail.
[-user <username>]
Defines the username the HTTPS client uses if digest
authentication is required by the secondary storage HTTPS
server.
[-password <password>]
Allows the user to specify the admin password. The password is
not recorded in the command log. If the -password option is
given but no password is specified, the user is prompted
interactively.
Use the -password option when digest authentication is required
by the secondary storage HTTPS server.
fs_dhsm
81
The fs Commands
[-timeout <seconds>
Specifies the timeout value in seconds. By default, the VNX
HTTPS client waits 30 seconds for a reply from the HTTPS server
and then retries the operation once.
[-cgi {y|n}
Specifies the HTTPS connection type: CGI or non-CGI. By default,
FileMover assumes that the web server is using CGI connections
to access migrated file data by using a CGI application. For
non-CGI connections, set the -cgi option to n; FileMover then
assumes the web server has direct access to migrated file content
on secondary storage.
-connection {<fs_name>|id=<fs_id>} -modify {-all|
<cid>[,<cid>...]}
Changes parameters on an existing NFS VNX FileMover connection.
[-state {enabled|disabled|recallonly}]
Sets the state of VNX FileMover operations on the specified file
system. enabled (default) allows both the creation of stub files
and data migration through reads and writes. If the state is
disabled, neither stub files nor data migration is possible. Data
currently on the VNX can be read and written to in the disabled
state.
If the state is recallonly, the policy engine is not allowed to create
stub files, but the user is still able to trigger data migration by
using a read or write request from the secondary file system to the
VNX.
[-read_policy_override {full|passthrough|partial
|none}]
Specifies the migration method option used by the VNX, in the
connection level or file system level, to override the migration
method specified in the stub file. none (default) specifies no
override, full recalls the whole file to the VNX on read request
before the data is returned, passthrough retrieves data without
recalling the data to the VNX, and partial recalls only the blocks
required to satisfy the client read request.
Note: The full migration may take several minutes or hours if the file is
very large.
[-http_server <host>]
Specifies the hostname of the secondary storage HTTPS server.
82
Command Line Interface Reference for File 8.1
The fs Commands
-httpsPort <port>
Specifies the remote port number that the Data Mover delivers
the HTTPS request to. If not specified, the Data Mover issues
HTTPS requests to port 443 on the secondary storage HTTPS
server.
Note: Although the -http_server option is used to modify the name
of the secondary storage HTTPS server, files that can be converted
into a stub by using an HTTPS connection can be brought back online
using only HTTPS and not using NFS, CIFS, or even HTTP.
-localPort <port>
Specifies the local port number the Data Mover uses to issue
HTTPS requests to the web server active on the secondary
storage. The <port> specified should be an integer number
less than 1024. If not specified, the Data Mover selects a port to
issue the HTTPS requests.
Note: If you attempt to create multiple HTTPS connections by using
identical end points with different attributes such as -cgi, -user,
-password, -localPort, -httpsPort, the connection will fail.
[-user <username>]
An optional attribute used to define the username the HTTPS
client uses if digest authentication is required by the
secondary storage HTTPS server.
[-password <password>]
Allows the user to specify the admin password. The password
is not recorded in the command log. If the -password option is
given but no password is specified, the user is prompted
interactively.
[-timeout <sec>]
Specifies the timeout value in seconds. By default, VNX’s
HTTPS client waits 30 seconds for a reply from the HTTPS
server and then retries the operation once before commencing
the failover operation.
-connection <fs_name> -delete {-all|<cid> [,<cid>...]}
Removes an existing HTTPS connection between the file system and
the secondary file system. Either all connections can be removed or
just the specified <cid> connection can be removed.
fs_dhsm
83
The fs Commands
[-recall_policy {check|no|yes}]
Specifies the recall policy for any migrated file during the -delete.
check (default) scans the file system for stub files that depend on
the connection and fails on the first one. no deletes the connection
without checking for stub files that depend on the connection,
and yes migrates the files back to the VNX before the connection
is removed. If no is specified and stub files exist, an I/O error
appears when the file is read because the connection no longer
exists.
SEE ALSO
Using VNX FileMover, server_cifs, server_http, and server_nfs.
EXAMPLE #1
To enable VNX FileMover on a file system, type:
$ fs_dhsm -modify ufs1 -state enabled
ufs1:
state
offline attr
popup timeout
backup
read policy override
log file
max log size
=
=
=
=
=
=
=
enabled
on
0
passthrough
none
on
10MB
Done
Where:
Value
state
offline attr
popup timeout
backup
read policy override
log file
max log size
EXAMPLE #2
Definition
Whether VNX FileMover is enabled or disabled on the file system.
Whether CIFS clients should be notified that a file is migrated.
Timeout value in seconds, before Windows popup notification is sent to the CIFS client.
Nature of CIFS network backups.
Migration method option used to override the read method specified in the stub file.
Whether FileMover logging is enabled or disabled.
Maximum size of the log file.
To create a CIFS connection for ufs1 to the secondary file system
\\winserver2.nasdocs.emc.com\dhsm1 with a specified
administrative account nasdocs.emc.com\Administrator and local
server dm102-cge0:
$ fs_dhsm -connection ufs1 -create -type cifs -admin
'nasdocs.emc.com\Administrator' -secondary
'\\winserver2.nasdocs.emc.com\dhsm1'
-local_server dm102-cge0
Enter Password:********
ufs1:
84
Command Line Interface Reference for File 8.1
The fs Commands
state
= enabled
offline attr
= on
popup timeout
= 0
backup
= passthrough
read policy override = none
log file
= on
max log size
= 10MB
cid
= 0
type
= CIFS
secondary
= \\winserver2.nasdocs.emc.com\dhsm1\
state
= enabled
read policy override = none
write policy
= full
local_server
= DM102-CGE0.NASDOCS.EMC.COM
admin
= nasdocs.emc.com\Administrator
wins
=
Done
Where:
Value
state
offline attr
popup timeout
backup
read policy override
log file
max log size
cid
type
secondary
state
read policy override
write policy
local_server
EXAMPLE #3
Definition
Whether VNX FileMover is enabled or disabled on the file system.
Whether CIFS clients should be notified that a file is migrated.
Timeout value, in seconds, before a popup notification is sent to CIFS client.
Nature of CIFS network backups.
Migration method option used to override the read method specified in the stub file.
Whether FileMover logging is enabled or disabled.
Maximum size of the log file.
Connection ID.
Type of file system. The -list option provides information for a description of the types.
Hostname or IP address of the remote file system.
Specifies whether VNX FileMover is enabled or disabled on the file system.
Migration method option used to override the read method specified in the stub file.
Write policy option used to recall data from secondary storage.
Name of the local CIFS server used to authenticate the CIFS connection.
To create a CIFS connection for ufs1 to the secondary file system
\\winserver2.nasdocs.emc.com\dhsm2 with a specified
administrative account nasdocs.emc.com\Administrator, local
server dm102-cge0, WINS server, and with the migration method set
to full, type:
$ fs_dhsm -connection ufs1 -create -type cifs -admin
'nasdocs.emc.com\Administrator' -secondary
'\\winserver2.nasdocs.emc.com\dhsm1'
-local_server dm102-cge0 -wins 172.24.102.25
-read_policy_override full
fs_dhsm
85
The fs Commands
Enter Password:********
ufs1:
state
= enabled
offline attr
= on
popup timeout
= 0
backup
= passthrough
read policy override = none
log file
= on
max log size
= 10MB
cid
= 0
type
= CIFS
secondary
= \\winserver2.nasdocs.emc.com\dhsm1\
state
= enabled
read policy override = full
write policy
= full
local_server
= DM102-CGE0.NASDOCS.EMC.COM
admin
= nasdocs.emc.com\Administrator
wins
= 172.24.102.25
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #4
To display connection information for ufs1, type:
$ fs_dhsm -connection ufs1 -info 1
ufs1:
state
offline attr
popup timeout
backup
read policy override
log file
max log size
=
=
=
=
=
=
=
enabled
on
0
passthrough
none
on
10MB
EXAMPLE #2 provides a description of command output.
EXAMPLE #5
To modify the read_policy_override setting for connection 0 for ufs1,
type:
$ fs_dhsm -connection ufs1 -modify 0
-read_policy_override passthrough
ufs1:
state
offline attr
popup timeout
backup
read policy override
log file
max log size
cid
type
secondary
86
=
=
=
=
=
=
=
=
enabled
on
0
passthrough
none
on
10MB
0
= CIFS
= \\winserver2.nasdocs.emc.com\dhsm1\
Command Line Interface Reference for File 8.1
The fs Commands
state
read policy override
write policy
local_server
admin
wins
=
=
=
=
=
=
enabled
pass
full
DM102-CGE0.NASDOCS.EMC.COM
nasdocs.emc.com\Administrator
172.24.102.25
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #6
To modify the VNX FileMover connection for ufs1, type:
$ fs_dhsm -connection ufs1 -modify 0 -nfs_server
172.24.102.115 -proto TCP
ufs1:
state
= enabled
offline attr
= on
popup timeout
= 10
backup
= offline
read policy override = full
log file
= on
max log size
= 25MB
cid
= 0
type
= NFSV3
secondary
= 172.24.102.115:/export/dhsm1
state
= enabled
read policy override = full
write policy
= full
options
= useRootCred=true proto=TCP
cid
= 1
type
= CIFS
secondary
= \\winserver2.nasdocs.emc.com\dhsm1\
state
= enabled
read policy override = none
write policy
= full
local_server
= DM102-CGE0.NASDOCS.EMC.COM
admin
= nasdocs.emc.com\Administrator
wins
= 172.24.102.25
cid
= 2
type
= HTTP
secondary
= http://172.24.102.115/export/dhsm1
state
= enabled
read policy override = none
write policy
= full
user
=
options
= cgi=n
Done
EXAMPLE #2 provides a description of command output.
fs_dhsm
87
The fs Commands
EXAMPLE #7
To create the NFSv3 connection for ufs1 to the secondary file system
172.24.102.115:/export/dhsm1 with the migration method set to full,
the -useRootCred set to true and the protocol set to UDP, type:
$ fs_dhsm -connection ufs1 -create -type nfsv3 -secondary
172.24.102.115:/export/dhsm1 -read_policy_override full
-useRootCred true
-proto UDP
ufs1:
state
= enabled
offline attr
= on
popup timeout
= 0
backup
= passthrough
read policy override = none
log file
= on
max log size
= 10MB
cid
= 0
type
= CIFS
secondary
= \\winserver2.nasdocs.emc.com\dhsm1\
state
= enabled
read policy override = pass
write policy
= full
local_server
= DM102-CGE0.NASDOCS.EMC.COM
admin
= nasdocs.emc.com\Administrator
wins
= 172.24.102.25
cid
= 1
type
= NFSV3
secondary
= 172.24.102.115:/export/dhsm1
state
= enabled
read policy override = full
write policy
= full
options
= useRootCred=true proto=UDP
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #8
To modify the VNX FileMover connection for ufs1, type:
$ fs_dhsm -connection ufs1 -modify 1 -proto TCP
ufs1:
state
offline attr
popup timeout
backup
read policy override
log file
max log size
cid
type
secondary
state
88
=
=
=
=
=
=
=
=
enabled
on
0
passthrough
none
on
10MB
0
= CIFS
= \\winserver2.nasdocs.emc.com\dhsm1\
= enabled
Command Line Interface Reference for File 8.1
The fs Commands
read policy override =
write policy
=
local_server
=
admin
=
wins
=
cid
= 1
type
=
secondary
=
state
=
read policy override =
write policy
=
options
=
Done
pass
full
DM102-CGE0.NASDOCS.EMC.COM
nasdocs.emc.com\Administrator
172.24.102.25
NFSV3
172.24.102.115:/export/dhsm1
enabled
full
full
useRootCred=true proto=TCP
EXAMPLE #2 provides a description of command output.
EXAMPLE #9
To display VNX FileMover connection information for ufs1, type:
$ fs_dhsm -info ufs1
ufs1:
state
offline attr
popup timeout
backup
read policy override
log file
max log size
=
=
=
=
=
=
=
enabled
on
0
passthrough
none
on
10MB
EXAMPLE #1 provides a description of command output.
id
29
29
29
EXAMPLE #10
To list VNX FileMover connections, type:
name
ufs1
ufs1
ufs1
$ fs_dhsm -connection ufs1 -list
cid
0
1
2
EXAMPLE #11
To modify the VNX FileMover connection for ufs1, type:
$ fs_dhsm -modify ufs1 -popup_timeout 10 -backup offline
-log on -max_log_size 25 -offline_attr on
-read_policy_override full
ufs1:
state
offline attr
popup timeout
backup
read policy override
log file
max log size
cid
=
=
=
=
=
=
=
=
enabled
on
10
offline
full
on
25MB
0
fs_dhsm
89
The fs Commands
type
=
secondary
=
state
=
read policy override =
write policy
=
local_server
=
admin
=
wins
=
cid
= 1
type
=
secondary
=
state
=
read policy override =
write policy
=
options
=
CIFS
\\winserver2.nasdocs.emc.com\dhsm1\
enabled
pass
full
DM102-CGE0.NASDOCS.EMC.COM
nasdocs.emc.com\Administrator
172.24.102.25
NFSV3
172.24.102.115:/export/dhsm1
enabled
full
full
useRootCred=true proto=TCP
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #12
To modify the state of the VNX FileMover connection 0 for ufs1, type:
$ fs_dhsm -connection ufs1 -modify 0 -state disabled
ufs1:
state
= enabled
offline attr
= on
popup timeout
= 10
backup
= offline
read policy override = full
log file
= on
max log size
= 25MB
cid
= 0
type
= CIFS
secondary
= \\winserver2.nasdocs.emc.com\dhsm1\
state
= disabled
read policy override = pass
write policy
= full
local_server
= DM102-CGE0.NASDOCS.EMC.COM
admin
= nasdocs.emc.com\Administrator
wins
= 172.24.102.25
cid
= 1
type
= NFSV3
secondary
= 172.24.102.115:/export/dhsm1
state
= enabled
read policy override = full
write policy
= full
options
= useRootCred=true proto=TCP
Done
EXAMPLE #2 provides a description of command output.
90
Command Line Interface Reference for File 8.1
The fs Commands
EXAMPLE #13
To modify the state of the VNX FileMover connection 1 for ufs1, type:
$ fs_dhsm -connection ufs1 -modify 1 -state recallonly
ufs1:
state
= enabled
offline attr
= on
popup timeout
= 10
backup
= offline
read policy override = full
log file
= on
max log size
= 25MB
cid
= 0
type
= CIFS
secondary
= \\winserver2.nasdocs.emc.com\dhsm1\
state
= enabled
read policy override = pass
write policy
= full
local_server
= DM102-CGE0.NASDOCS.EMC.COM
admin
= nasdocs.emc.com\Administrator
wins
= 172.24.102.25
cid
= 1
type
= NFSV3
secondary
= 172.24.102.115:/export/dhsm1
state
= recallonly
read policy override = full
write policy
= full
options
= useRootCred=true proto=TCP
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #14
To delete the VNX FileMover connections 1 and 2 for ufs1, and
specify the recall policy for any migrated files during the delete, type:
$ fs_dhsm -connection ufs1 -delete 0,1 -recall_policy
no
ufs1:
state
offline attr
popup timeout
backup
read policy override
log file
max log size
=
=
=
=
=
=
=
enabled
on
10
offline
full
on
25MB
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #15
To change the state of the VNX FileMover connection for ufs1 to
disabled, type:
fs_dhsm
91
The fs Commands
$ fs_dhsm -modify ufs1 -state disabled
ufs1:
state
offline attr
popup timeout
backup
read policy override
log file
max log size
=
=
=
=
=
=
=
disabled
on
10
offline
full
on
25MB
Done
EXAMPLE #1 provides a description of command output.
EXAMPLE #16
To create an HTTP connection for ufs1 to the secondary file system
/export/dhsm1 on the web server http://172.24.102.115 that has direct
access to the storage, type:
$ fs_dhsm -connection ufs1 -create -type http -secondary
http://172.24.102.115/export/dhsm1 -cgi n
ufs1:
state
= enabled
offline attr
= on
popup timeout
= 10
backup
= offline
read policy override = full
log file
= on
max log size
= 25MB
cid
= 2
type
= HTTP
secondary
= http://172.24.102.115/export/dhsm1
state
= enabled
read policy override = none
write policy
= full
user
=
options
= cgi=n
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #17
To create an HTTP connection for ufs1 to the secondary file system
using CGI connections to access migrated file data using a CGI
application, type:
$ fs_dhsm -connection ufs1 -create -type http -secondary
http://www.nasdocs.emc.com/cgi-bin/access.sh
ufs1:
state
offline attr
popup timeout
backup
92
=
=
=
=
enabled
on
0
passthrough
Command Line Interface Reference for File 8.1
The fs Commands
read policy override = none
log file
= on
max log size
= 10MB
cid
= 0
type
= HTTP
secondary
= http://www.nasdocs.emc.com/cgi-bin/access.sh
state
= enabled
read policy override = none
write policy
= full
user
=
options
=
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #18
To create an HTTPS connection for server2_fs1 on the web server
https://int16543 with read_policy_override set to full, type:
$ fs_dhsm -connection server2_fs1 -create -type https
-secondary https://int16543 -read_policy_override full
-cgi n
server2_fs1:
state
= enabled
offline attr
= on
popup timeout
= 0
backup
= passthrough
read policy override = passthrough
log file
= on
max log size
= 10MB
cid
= 0
type
= HTTPS
secondary
= https://int16543
state
= enabled
read policy override = full
write policy
= full
user
=
options
=
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #19
To create an HTTPS connection for ufs1 to the secondary file system
using CGI connections to access migrated file data using a CGI
application, type:
$ fs_dhsm -connection ufs1 -create -type https -secondary
https://www.nasdocs.emc.com/cgi-bin/access.sh
ufs1:
state = enabled
offline attr = on
popup timeout = 0
backup = passthrough
fs_dhsm
93
The fs Commands
read policy override = none
log file = on
max log size = 10MB
cid = 0
type = HTTPS
secondary = https://www.nasdocs.emc.com/cgi-bin/access.sh
state = enabled
read policy override = none
write policy = full
user =
options =
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #20
To create an HTTPS connection on httpsPort 443 for server2_ufs1 on
the web server https://int16543 with read_policy_override set to
passthrough, type:
$ fs_dhsm -connection server2_fs1 -create -type https
-secondary https://int16543 -read_policy_override
passthrough -httpsPort 443 -cgi n
server2_fs1:
state
= enabled
offline attr
= on
popup timeout
= 0
backup
= passthrough
read policy override = passthrough
log file
= on
max log size
= 10MB
cid
= 1
type
= HTTPS
secondary
= https://int16543
state
= enabled
read policy override = pass
write policy
= full
user
=
options
=
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #21
To create an HTTPS connection on localPort 80 for server2_ufs1 on
the web server https://int16543 with read_policy_override set to
passthrough, type:
$ fs_dhsm -connection server2_fs1 -create -type https
-secondary https://int16543 -read_policy_override
passthrough -localPort 80 -cgi n
server2_fs1:
state
offline attr
94
= enabled
= on
Command Line Interface Reference for File 8.1
The fs Commands
popup timeout
= 0
backup
= passthrough
read policy override = passthrough
log file
= on
max log size
= 10MB
cid
= 0
type
= HTTPS
secondary
= https://int16543
state
= enabled
read policy override = pass
write policy
= full
user
=
options
=
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #22
To create an HTTPS connection on httpsPort 443 for server2_ufs1 on
the web server https://int16543 with a specified user dhsm_user,
type:
$ fs_dhsm -connection server2_fs1 -create -type https
-secondary https://int16543 -read_policy_override full
-httpsPort 443 -user dhsm_user -password dhsm_user -cgi n
server2_fs1:
state
= enabled
offline attr
= on
popup timeout
= 0
backup
= passthrough
read policy override = passthrough
log file
= on
max log size
= 10MB
cid
= 1
type
= HTTPS
secondary
= https://int16543
state
= enabled
read policy override = full
write policy
= full
user
= dhsm_user
options
=
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #23
To modify the read_policy_override setting for connection 1 from
server2_fs1, type:
$ fs_dhsm -connection server2_fs1 -modify 1
-read_policy_override passthrough
server2_fs1:
state
= enabled
fs_dhsm
95
The fs Commands
offline attr
= on
popup timeout
= 0
backup
= passthrough
read policy override = passthrough
log file
= on
max log size
= 10MB
cid
= 1
type
= HTTPS
secondary
= https://int16543
state
= enabled
read policy override = pass
write policy
= full
user
= dhsm_user
options
=
Done
EXAMPLE #2 provides a description of command output.
EXAMPLE #24
To delete the VNX FileMover connection 0 for ufs1, type:
$ fs_dhsm -connection ufs1 -delete 0
ufs1:
state
offline attr
popup timeout
backup
read policy override
log file
max log size
=
=
=
=
=
=
=
enabled
on
0
passthrough
none
on
10MB
Done
EXAMPLE #1 provides a description of command output.
96
Command Line Interface Reference for File 8.1
The fs Commands
fs_group
Creates a file system group from the specified file systems or a single
file system.
SYNOPSIS
fs_group
-list
| -delete <fs_group_name>
| -info {<fs_group_name>|id=<fs_group_id>}
| [-name <name>] -create {<fs_name>,...}
| -xtend <fs_group_name> {<fs_name>,...}
| -shrink <fs_group_name> {<fs_name>,...}
DESCRIPTION
OPTIONS
The fs_group command combines file systems to be acted upon
simultaneously as a single group for TimeFinder/FS.
-list
Displays a listing of all file system groups.
Note: The ID of the object is an integer and is assigned automatically. The
name of a file system may be truncated if it is too long for the display. To
display the full name, use the -info option with a file system ID.
-delete <fs_group_name>
Deletes the file system group configuration. Individual file systems
are not deleted.
-info {<fs_group_name>|id=<fs_group_id>}
Displays information about a file system group, either by name or
group ID.
[-name <name>] -create {<fs_name>,...}
Creates a file system group from the specified file systems. If a name
is not specified, one is assigned by default.
-xtend <fs_group_name> {<fs_name>,...}
Adds the specified file systems or group to a file system group.
-shrink <fs_group_name> {<fs_name>,...}
Removes the specified file systems or group from a file system group.
Individual file systems are not deleted.
SEE ALSO
Managing Volumes and File Systems for VNX Manually and Using
TimeFinder/FS, NearCopy, and FarCopy on VNX for File, fs_timefinder,
and nas_fs.
fs_group
97
The fs Commands
SYSTEM OUTPUT
EXAMPLE #1
The number associated with the storage device is dependent on the
attached system. VNX for block displays a prefix of APM before a set
of integers, for example, APM00033900124-0019. Symmetrix systems
appear as 002804000190-003C.
To create a file system group named, ufsg1, and add ufs1, type:
$ fs_group -name ufsg1 -create ufs1
id
= 22
name
= ufsg1
acl
= 0
in_use
= False
type
= group
fs_set
= ufs1
pool
=
stor_devs =
000187940268-0006,000187940268-0007,000187940268-0008,000187940268-0009
disks
= d3,d4,d5,d6
Where:
Value
id
name
acl
in_use
type
fs_set
pool
stor_devs
disks
Definition
ID of the group that is automatically assigned.
Name assigned to the group.
Access control value for the group.
Whether a file system is used by a group.
Type of file system.
File systems that are part of the group.
Storage pool given to the file system group.
System devices associated with the group.
Disks on which the metavolume resides.
EXAMPLE #2
id
20
name
ufsg1
To list all file system groups, type:
$ fs_group -list
acl in_use type member_of fs_set
0
n
100
18
Where:
Value
member_of
98
Definition
Groups which the file system group belong to.
Command Line Interface Reference for File 8.1
The fs Commands
EXAMPLE #3
To display information for the file system group, ufsg1, type:
$ fs_group -info ufsg1
id
= 22
name
= ufsg1
acl
= 0
in_use
= False
type
= group
fs_set
= ufs1
pool
=
stor_devs =
000187940268-0006,000187940268-0007,000187940268-0008,000187940268-0009
disks
= d3,d4,d5,d6
EXAMPLE #1 provides a description of command output.
EXAMPLE #4
To add file system, ufs2, to the file system group, ufsg1, type:
$ fs_group -xtend ufsg1 ufs2
id
= 22
name
= ufsg1
acl
= 0
in_use
= False
type
= group
fs_set
= ufs1,ufs2
pool
=
stor_devs =
000187940268-0006,000187940268-0007,000187940268-0008,000187940268-0009,00018794
0268-000A,000187940268-000B,000187940268-000C,000187940268-000D
disks
= d3,d4,d5,d6,d7,d8,d9,d10
EXAMPLE #1 provides a description of command output.
EXAMPLE #5
To remove file system, ufs2, from the file system group, ufsg1, type:
$ fs_group -shrink ufsg1 ufs2
id
= 22
name
= ufsg1
acl
= 0
in_use
= False
type
= group
fs_set
= ufs1
pool
=
stor_devs =
000187940268-0006,000187940268-0007,000187940268-0008,000187940268-0009
disks
= d3,d4,d5,d6
EXAMPLE #1 provides a description of command output.
fs_group
99
The fs Commands
EXAMPLE #6
To delete file system group, ufsg1, type:
$ fs_group -delete ufsg1
id
name
acl
in_use
type
fs_set
stor_devs
disks
=
=
=
=
=
=
=
=
22
ufsg1
0
False
group
EXAMPLE #1 provides a description of command output.
100
Command Line Interface Reference for File 8.1
The fs Commands
fs_rdf
Manages the remote data facility (RDF) functionality for a file system
residing on RDF drives.
SYNOPSIS
fs_rdf {<fs_name>|id=<fs_id>}
-Mirror {on|off|refresh}
| -Restore [-Force]
| -info
Note: RDF is supported only on a VNX attached to a Symmetrix.
DESCRIPTION
OPTIONS
The fs_rdf command turns mirroring on and off for an RDF file
system and displays information about RDF relationships.
-Mirror {on|off|refresh}
The on option resumes the link between the RDF drives of a file
system thereby enabling mirroring for the RDF file system. The off
option halts mirroring between the file systems, and the refresh
option does an immediate mirror on then off which refreshes the file
system image.
-Restore [-Force]
Restores a file system from the R2 side (remote) when remote
TimeFinder/FS FarCopy is used. The -Restore can only be executed
on the R1 side. The -Force option must be used when restoring a file
system with enabled.
-info
Displays information about RDF relationships.
SEE ALSO
EXAMPLE #1
Using SRDF/S with VNX for Disaster Recovery, Using TimeFinder/FS,
NearCopy, and FarCopy on VNX for File, and Using VNX File-Level
Retention.
To turn on mirroring for ufs1_snap1 from the R1 Control Station,
type:
$ fs_rdf ufs1_snap1 -Mirror on
id
name
acl
in_use
type
volume
=
=
=
=
=
=
20
ufs1_snap1
0
False
uxfs
v168
fs_rdf
101
The fs Commands
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Fri Apr 23 16:29:23 EDT 2004
stor_devs =
002804000190-0052,002804000190-0053,002804000190-0054,002804000190-0055
disks
= rootd33,rootd34,rootd35,rootd36
RDF Information:
remote_symid
remote_sym_devname
ra_group_number
dev_rdf_type
dev_ra_status
dev_link_status
rdf_mode
rdf_pair_state
rdf_domino
adaptive_copy
adaptive_copy_skew
num_r1_invalid_tracks
num_r2_invalid_tracks
dev_rdf_state
remote_dev_rdf_state
rdf_status
link_domino
prevent_auto_link_recovery
link_config
suspend_state
consistency_state
adaptive_copy_wp_state
prevent_ra_online_upon_pwron
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
002804000218
2
R1
READY
READY
SYNCHRONOUS
SYNCINPROG
DISABLED
DISABLED
65535
0
736440
READY
WRITE_DISABLED
0
DISABLED
DISABLED
NA
DISABLED
NA
ENABLED
Where:
Value
id
name
acl
in_use
type
volume
pool
rw_servers
ro_servers
rw_vdms
ro_vdms
102
Definition
ID of a file system that is assigned automatically.
Name assigned to a file system.
Access control value for a file system.
Whether a file system is registered into the mount table.
Type of file system. nas_fs provides a description of the types.
Volume on which a file system resides.
Storage pool for the file system.
Servers with read-write access to a file system.
Servers with read-only access to a file system.
VDM servers with read-write access to a file system.
VDM servers with read-only access to a file system.
Command Line Interface Reference for File 8.1
The fs Commands
The remote RDF file system.
The system devices associated with a file system.
The disks on which the metavolume resides.
The serial number of the system containing the target volume.
The system device name of the remote device in an RDF pair.
The RA group number (1-n).
The type of RDF device. Possible values are: R1 and R2.
RA status. Possible values are: READY, NOT_READY, WRITE_DISABLED, STATUS_NA,
STATUS_MIXED.
dev_link_status
Link status. Possible values are: READY, NOT_READY, WRITE_DISABLED, NA, MIXED.
rdf_mode
The RDF mode. Possible values are: SYNCHRONOUS, SEMI_SYNCHRONOUS,
ADAPTIVE_COPY, MIXED.
rdf_pair_state
The composite state of the RDF pair. Possible values are:
INVALID, SYNCINPROG, SYNCHRONIZED, SPLIT, SUSPENDED, FAILED_OVER,
PARTITIONED, R1_UPDATED, R1_UPDINPROG, MIXED.
rdf_domino
The RDF device domino. Possible values are: ENABLED, DISABLED, MIXED.
adaptive_copy
Possible values are: DISABLED, WP_MODE, DISK_MODE, MIXED.
adaptive_copy_skew
Number of invalid tracks when in Adaptive copy mode.
num_r1_invalid_tracks Number of invalid tracks on the source (R1) device.
num_r2_invalid_tracks Number of invalid tracks on the target (R2) device.
dev_rdf_state
Specifies the composite RDF state of the RDF device. Possible values are: READY,
NOT_READY, WRITE_DISABLED, NA, MIXED.
remote_dev_rdf_state
Specifies the composite RDF state of the remote RDF device. Possible values are: READY,
NOT_READY, WRITE_DISABLED, NA, MIXED.
rdf_status
Specifies the RDF status of the device. Possible values are:
READY, NOT_READY, WRITE_DISABLED, NA, MIXED.
link_domino
RDF link domino. Possible values are: ENABLED, DISABLED.
prevent_auto_link_reco When enabled, prevents the automatic resumption of data copy across the RDF links as
very
soon as the links have recovered. Possible values are: ENABLED, DISABLED.
link_config
Possible values are: CONFIG_ESCON, CONFIG_T3.
suspend_state
Specifies the status of R1 devices in a consistency group. Possible states are: NA,
OFFLINE, OFFLINE_PEND, ONLINE_MIXED.
consistency_state
Specifies state of an R1 device related to consistency groups. Possible states are:
ENABLED, DISABLED.
adaptive_copy_wp_state Specifies state of the adaptive copy mode. Possible states are: NA, OFFLINE,
OFFLINE_PEND, ONLINE_MIXED.
prevent_ra_online_upon Specifies the state of the RA director coming online after power on. Possible states are:
_pwron
ENABLED, DISABLED.
backup_of
stor_devs
disks
remote_symid
remote_sym_devname
ra_group_number
dev_rdf_type
dev_ra_status
fs_rdf
103
The fs Commands
EXAMPLE #2
To display RDF-related information for ufs1_snap1 from the R2
Control Station, type:
$ fs_rdf ufs1_snap1 -info
id
= 20
name
= ufs1_snap1
acl
= 0
in_use
= False
type
= uxfs
volume
= v168
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Fri Apr 23 16:29:23 EDT 2004
stor_devs =
002804000190-0052,002804000190-0053,002804000190-0054,002804000190-0055
disks
= rootd33,rootd34,rootd35,rootd36
RDF Information:
remote_symid
remote_sym_devname
ra_group_number
dev_rdf_type
dev_ra_status
dev_link_status
rdf_mode
rdf_pair_state
rdf_domino
adaptive_copy
adaptive_copy_skew
num_r1_invalid_tracks
num_r2_invalid_tracks
dev_rdf_state
remote_dev_rdf_state
rdf_status
link_domino
prevent_auto_link_recovery
link_config
suspend_state
consistency_state
adaptive_copy_wp_state
prevent_ra_online_upon_pwron
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
002804000218
2
R1
READY
READY
SYNCHRONOUS
SYNCINPROG
DISABLED
DISABLED
65535
0
696030
READY
WRITE_DISABLED
0
DISABLED
DISABLED
NA
DISABLED
NA
ENABLED
EXAMPLE #1 provides a description of command output.
104
Command Line Interface Reference for File 8.1
The fs Commands
EXAMPLE #3
To turn the mirroring off for ufs1_snap1 on the R1 Control Station,
type:
$ fs_rdf ufs1_snap1 -Mirror off
remainder(MB) = 20548..17200..13110..8992..4870..746 0
id
= 20
name
= ufs1_snap1
acl
= 0
in_use
= False
type
= uxfs
volume
= v168
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Fri Apr 23 16:29:23 EDT 2004
stor_devs =
002804000190-0052,002804000190-0053,002804000190-0054,002804000190-0055
disks
= rootd33,rootd34,rootd35,rootd36
RDF Information:
remote_symid
remote_sym_devname
ra_group_number
dev_rdf_type
dev_ra_status
dev_link_status
rdf_mode
rdf_pair_state
rdf_domino
adaptive_copy
adaptive_copy_skew
num_r1_invalid_tracks
num_r2_invalid_tracks
dev_rdf_state
remote_dev_rdf_state
rdf_status
link_domino
prevent_auto_link_recovery
link_config
suspend_state
consistency_state
adaptive_copy_wp_state
prevent_ra_online_upon_pwron
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
002804000218
2
R1
READY
NOT_READY
SYNCHRONOUS
SUSPENDED
DISABLED
DISABLED
65535
0
0
READY
WRITE_DISABLED
0
DISABLED
DISABLED
OFFLINE
DISABLED
NA
ENABLED
EXAMPLE #1 provides a description of command output.
fs_rdf
105
The fs Commands
EXAMPLE #4
To perform a mirror refresh for ufs1_snap1 on the R1 Control Station,
type:
$ fs_rdf ufs1_snap1 -Mirror refresh
remainder(MB) = 1 0
id
= 20
name
= ufs1_snap1
acl
= 0
in_use
= False
type
= uxfs
volume
= v168
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Fri Apr 23 16:29:23 EDT 2004
stor_devs =
002804000190-0052,002804000190-0053,002804000190-0054,002804000190-0055
disks
= rootd33,rootd34,rootd35,rootd36
RDF Information:
remote_symid
remote_sym_devname
ra_group_number
dev_rdf_type
dev_ra_status
dev_link_status
rdf_mode
rdf_pair_state
rdf_domino
adaptive_copy
adaptive_copy_skew
num_r1_invalid_tracks
num_r2_invalid_tracks
dev_rdf_state
remote_dev_rdf_state
rdf_status
link_domino
prevent_auto_link_recovery
link_config
suspend_state
consistency_state
adaptive_copy_wp_state
prevent_ra_online_upon_pwron
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
002804000218
2
R1
READY
NOT_READY
SYNCHRONOUS
SUSPENDED
DISABLED
DISABLED
65535
0
0
READY
WRITE_DISABLED
0
DISABLED
DISABLED
OFFLINE
DISABLED
NA
ENABLED
EXAMPLE #1 provides a description of command output.
106
Command Line Interface Reference for File 8.1
The fs Commands
EXAMPLE #5
To restore the file system ufs1_snap1 from the R1 Control Station,
type:
$ /nas/sbin/rootfs_rdf ufs1_snap1 -Restore
remainder(MB) = 1 0
id
= 20
name
= ufs1_snap1
acl
= 0
in_use
= False
type
= uxfs
volume
= v168
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Fri Apr 23 16:29:23 EDT 2004
stor_devs =
002804000190-0052,002804000190-0053,002804000190-0054,002804000190-0055
disks
= rootd33,rootd34,rootd35,rootd36
RDF Information:
remote_symid
remote_sym_devname
ra_group_number
dev_rdf_type
dev_ra_status
dev_link_status
rdf_mode
rdf_pair_state
rdf_domino
adaptive_copy
adaptive_copy_skew
num_r1_invalid_tracks
num_r2_invalid_tracks
dev_rdf_state
remote_dev_rdf_state
rdf_status
link_domino
prevent_auto_link_recovery
link_config
suspend_state
consistency_state
adaptive_copy_wp_state
prevent_ra_online_upon_pwron
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
002804000218
2
R1
READY
READY
SYNCHRONOUS
SYNCHRONIZED
DISABLED
DISABLED
65535
0
0
READY
WRITE_DISABLED
0
DISABLED
DISABLED
NA
DISABLED
NA
ENABLED
EXAMPLE #1 provides a description of command output.
fs_rdf
107
The fs Commands
fs_timefinder
Manages the TimeFinder/FS functionality for the specified file
system or file system group.
SYNOPSIS
fs_timefinder {<fs_name>|id=<fs_id>}
-Mirror {on|off|refresh [-Force]}[-star]
| [-name <name>] -Snapshot [-volume <volume_name>][-option <options>][-star]
| -Restore [-Force][-option <options>][-star]
Note: TimeFinder/FS is supported only on a VNX attached to a Symmetrix.
DESCRIPTION
OPTIONS
The fs_timefinder command creates a copy of a file system or file
system group that can be placed into a mirrored mode with its
original file system. The Symmetrix must already have business
continuance volumes (BCVs) configured to the same size as the
volumes on the VNX. After the copy of the file system has been
made, it can be mounted on any Data Mover.
-Mirror {on|off|refresh}
on places the unmounted file system copy, created by using the
-Snapshot option, into a mirrored mode with its original file system.
The file system copy is frozen and remains unavailable to users until
mirrored mode is turned off.
The refresh option initiates an immediate -Mirror on then off for the
unmounted file system copy, thereby refreshing the file system copy.
[-Force]
The file system copy should not be mounted read-write when
placed into mirrored mode or when refreshed. If the file system
copy is mounted read-write, the -Force option can be used to
force a refresh if the metavolume is an STD type. The -Force
option requires root command and must be executed by using
/nas/sbin/rootfs_timefinder.
[-star]
The -star option allows the fs_timefinder command to run on
STAR SRDF configuration.
108
Command Line Interface Reference for File 8.1
The fs Commands
!
CAUTION
Performing a mirror refresh may be time consuming, relative to the
amount of data that has changed in the file system.
[-name <name>] -Snapshot [-star]
Creates a copy of a file system and assigns an optional name to the
file system copy. If a name is not specified, one is assigned by default.
If no options are provided, a name and metavolume are
automatically assigned. Use nas_fs to delete the copy of the file
system.
!
CAUTION
Creating a copy by using -Snapshot may be time consuming,
relative to the size of a file system.
[-volume <volume_name>]
Assigns a metavolume to a file system copy. The metavolume
must be created by using the nas_volume -Clone command prior
to executing this option. The metavolume must be a BCV type
and have the same characteristics as the metavolume of the
original file system.
[-option <options>]
Specifies the following comma-separated options:
mirror=on
Leaves the file system copy in mirrored mode.
disktype=<disktype>
For systems with both local and R1BCVs, specifies the type of
volume to use when creating a snapshot. In a TimeFinder/FS
FarCopy configuration, use disktype=R1BCV for creating a
snapshot of the PFS on the local VNX for file. For creating a
snapshot of an imported FarCopy snapshot on the remote VNX
for file, use disktype=STD. This option is supported only for
RAID group based disk volumes and cannot be combined with
the "mapped_pool=" option.
By default, the system uses the first available R1BCV or BCV, or
R1STD or STD device.
Use the disktype= option to designate which to use if there are R1
devices in your configuration.
fs_timefinder
109
The fs Commands
pool=<mapped_pool>
Specifies the mapped pool to use when creating a snapshot from
that pool. This option is supported only for mapped pool disk
volumes and cannot be combined with the disktype= option.
A mapped pool is a VNX for file storage pool that is dynamically
generated when diskmark is run. It is a one-to-one mapping with
either a VNX for block storage pool or a Symmetrix Storage
Group.
Note: If the pool= option is used when creating a snapshot, the disk volume
will be selected only from this pool. If the pool does not have enough disk
volumes to create a snapshot for the source file system, the fs_timefinder
command reports an error.
[-star]
Allows fs_timefinder command to run on STAR SRDF
configsuration.
-Restore
Restores a file system to its original location by using the unmounted
file system copy created with the -Snapshot option. The original file
system must not have any associated SnapSure checkpoints. The -star
option allows the fs_timefinder command to run on STAR SRDF
configuration.
!
CAUTION
Restoring a file system may be time consuming, relative to the
amount of data that has changed in the file system.
[-Force]
Forces a restore of a file system copy that is mounted on the
metavolume as read-only, or if the volume is an STD type.
[-option <options>]
Specifies the following comma-separated options:
mirror=on
Places the file system copy in mirrored mode.
[-star]
Allows fs_timefinder command to run on STAR SRDF
configsuration.
110
Command Line Interface Reference for File 8.1
The fs Commands
SEE ALSO
EXAMPLE #1
Using TimeFinder/FS, NearCopy, and FarCopy on VNX for File, fs_ckpt,
fs_group, and nas_fs.
To create a TimeFinder/FS copy of the PFS, type:
$ fs_timefinder ufs1 -Snapshot
operation in progress (not interruptible)...
remainder(MB) =
43688..37205..31142..24933..18649..12608..7115..4991..4129..3281..2457..1653..81
5..0
operation in progress (not interruptible)...id
= 18
name
= ufs1
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= mtv1
pool
=
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
backups
= ufs1_snap1
stor_devs =
000187940268-0006,000187940268-0007,000187940268-0008,000187940268-0009
disks
= d3,d4,d5,d6
disk=d3
stor_dev=000187940268-0006
addr=c0t1l0-48-0
server=server_2
disk=d3
stor_dev=000187940268-0006
addr=c16t1l0-33-0
server=server_2
disk=d4
stor_dev=000187940268-0007
addr=c0t1l1-48-0
server=server_2
disk=d4
stor_dev=000187940268-0007
addr=c16t1l1-33-0
server=server_2
disk=d5
stor_dev=000187940268-0008
addr=c0t1l2-48-0
server=server_2
disk=d5
stor_dev=000187940268-0008
addr=c16t1l2-33-0
server=server_2
disk=d6
stor_dev=000187940268-0009
addr=c0t1l3-48-0
server=server_2
disk=d6
stor_dev=000187940268-0009
addr=c16t1l3-33-0
server=server_2
id
= 19
name
= ufs1_snap1
acl
= 0
in_use
= False
type
= uxfs
worm
= off
volume
= v456
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Thu Oct 28 14:13:30 EDT 2004
stor_devs =
000187940268-0180,000187940268-0181,000187940268-0182,000187940268-0183
disks
= rootd378,rootd379,rootd380,rootd381
fs_timefinder
111
The fs Commands
Where:
Value
name
acl
in_use
type
worm
volume
pool
rw_servers
ro_servers
rw_vdms
ro_vdms
backups
backup_of
stor_devs
disks
Definition
Name assigned to the file system.
Access control value for a file system. nas_acl provides information.
If a file system is registered into the mount table of a Data Mover.
Type of file system. -list provides a description of the types.
Whether is enabled.
Volume on which the file system resides.
Storage pool for the file system.
Servers with read-write access to a file system.
Servers with read-only access to a file system.
VDM servers with read-write access to a file system.
VDM servers with read-only access to a file system.
Name of associated backups.
File system that the file system copy is made from.
System devices associated with a file system. The storage device output is the result of the Symmetrix
hardware system.
Disks on which the metavolume resides.
EXAMPLE #2
To create a TimeFinder/FS copy of the PFS, ufs1, and leave a file
system copy in mirrored mode, type:
$ fs_timefinder ufs1 -Snapshot -option mirror=on
operation in progress (not interruptible)...id
= 18
name
= ufs1
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= mtv1
pool
=
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
backups
= ufs1_snap1
stor_devs =
000187940268-0006,000187940268-0007,000187940268-0008,000187940268-0009
disks
= d3,d4,d5,d6
disk=d3
stor_dev=000187940268-0006
addr=c0t1l0-48-0
server=server_2
disk=d3
stor_dev=000187940268-0006
addr=c16t1l0-33-0
server=server_2
disk=d4
stor_dev=000187940268-0007
addr=c0t1l1-48-0
server=server_2
disk=d4
stor_dev=000187940268-0007
addr=c16t1l1-33-0
server=server_2
disk=d5
stor_dev=000187940268-0008
addr=c0t1l2-48-0
server=server_2
disk=d5
stor_dev=000187940268-0008
addr=c16t1l2-33-0
server=server_2
disk=d6
stor_dev=000187940268-0009
addr=c0t1l3-48-0
server=server_2
disk=d6
stor_dev=000187940268-0009
addr=c16t1l3-33-0
server=server_2
112
Command Line Interface Reference for File 8.1
The fs Commands
id
= 19
name
= ufs1_snap1
acl
= 0
in_use
= False
type
= mirrorfs
worm
= off
volume
= v456
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Thu Oct 28 14:19:03 EDT 2004
remainder = 0 MB (0%)
stor_devs =
000187940268-0180,000187940268-0181,000187940268-0182,000187940268-0183
disks
= rootd378,rootd379,rootd380,rootd381
EXAMPLE #1 provides a description of command output.
EXAMPLE #3
To turn mirroring off for a file system copy, ufs1_snap1, type:
$ fs_timefinder ufs1_snap1 -Mirror off
operation in progress (not interruptible)...
remainder(MB) = 0
operation in progress (not interruptible)...id
= 18
name
= ufs1
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= mtv1
pool
=
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
backups
= ufs1_snap1
stor_devs =
000187940268-0006,000187940268-0007,000187940268-0008,000187940268-0009
disks
= d3,d4,d5,d6
disk=d3
stor_dev=000187940268-0006
addr=c0t1l0-48-0
server=server_2
disk=d3
stor_dev=000187940268-0006
addr=c16t1l0-33-0
server=server_2
disk=d4
stor_dev=000187940268-0007
addr=c0t1l1-48-0
server=server_2
disk=d4
stor_dev=000187940268-0007
addr=c16t1l1-33-0
server=server_2
disk=d5
stor_dev=000187940268-0008
addr=c0t1l2-48-0
server=server_2
disk=d5
stor_dev=000187940268-0008
addr=c16t1l2-33-0
server=server_2
disk=d6
stor_dev=000187940268-0009
addr=c0t1l3-48-0
server=server_2
disk=d6
stor_dev=000187940268-0009
addr=c16t1l3-33-0
server=server_2
id
= 19
fs_timefinder
113
The fs Commands
name
= ufs1_snap1
acl
= 0
in_use
= False
type
= uxfs
worm
= off
volume
= v456
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Thu Oct 28 14:21:50 EDT 2004
stor_devs =
000187940268-0180,000187940268-0181,000187940268-0182,000187940268-0183
disks
= rootd378,rootd379,rootd380,rootd381
EXAMPLE #1 provides a description of command output.
EXAMPLE #4
To turn mirroring on for a file system copy, ufs1_snap1, type:
$ fs_timefinder ufs1_snap1 -Mirror on
operation in progress (not interruptible)...id
= 18
name
= ufs1
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= mtv1
pool
=
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
backups
= ufs1_snap1
stor_devs =
000187940268-0006,000187940268-0007,000187940268-0008,000187940268-0009
disks
= d3,d4,d5,d6
disk=d3
stor_dev=000187940268-0006
addr=c0t1l0-48-0
server=server_2
disk=d3
stor_dev=000187940268-0006
addr=c16t1l0-33-0
server=server_2
disk=d4
stor_dev=000187940268-0007
addr=c0t1l1-48-0
server=server_2
disk=d4
stor_dev=000187940268-0007
addr=c16t1l1-33-0
server=server_2
disk=d5
stor_dev=000187940268-0008
addr=c0t1l2-48-0
server=server_2
disk=d5
stor_dev=000187940268-0008
addr=c16t1l2-33-0
server=server_2
disk=d6
stor_dev=000187940268-0009
addr=c0t1l3-48-0
server=server_2
disk=d6
stor_dev=000187940268-0009
addr=c16t1l3-33-0
server=server_2
id
name
acl
in_use
type
worm
114
=
=
=
=
=
=
19
ufs1_snap1
0
False
mirrorfs
off
Command Line Interface Reference for File 8.1
The fs Commands
volume
= v456
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Thu Oct 28 14:21:50 EDT 2004
remainder = 0 MB (0%)
stor_devs =
000187940268-0180,000187940268-0181,000187940268-0182,000187940268-0183
disks
= rootd378,rootd379,rootd380,rootd381
EXAMPLE #1 provides a description of command output.
EXAMPLE #5
To perform a mirror refresh on ufs1_snap1, type:
$ fs_timefinder ufs1_snap1 -Mirror refresh
operation in progress (not interruptible)...
remainder(MB) = 4991..4129..3281..2457..1653..815..0
operation in progress (not interruptible)...id
= 18
name
= ufs1
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= mtv1
pool
=
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
backups
= ufs1_snap1
stor_devs =
000187940268-0006,000187940268-0007,000187940268-0008,000187940268-0009
disks
= d3,d4,d5,d6
disk=d3
stor_dev=000187940268-0006
addr=c0t1l0-48-0
server=server_2
disk=d3
stor_dev=000187940268-0006
addr=c16t1l0-33-0
server=server_2
disk=d4
stor_dev=000187940268-0007
addr=c0t1l1-48-0
server=server_2
disk=d4
stor_dev=000187940268-0007
addr=c16t1l1-33-0
server=server_2
disk=d5
stor_dev=000187940268-0008
addr=c0t1l2-48-0
server=server_2
disk=d5
stor_dev=000187940268-0008
addr=c16t1l2-33-0
server=server_2
disk=d6
stor_dev=000187940268-0009
addr=c0t1l3-48-0
server=server_2
disk=d6
stor_dev=000187940268-0009
addr=c16t1l3-33-0
server=server_2
id
name
acl
in_use
type
worm
volume
pool
=
=
=
=
=
=
=
=
19
ufs1_snap1
0
False
uxfs
off
v456
fs_timefinder
115
The fs Commands
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Thu Oct 28 14:25:21 EDT 2004
stor_devs =
000187940268-0180,000187940268-0181,000187940268-0182,000187940268-0183
disks
= rootd378,rootd379,rootd380,rootd381
EXAMPLE #1 provides a description of command output.
EXAMPLE #6
To restore the file system copy, ufs1_snap1, to its original location,
type:
$ /nas/sbin/rootfs_timefinder ufs1_snap1 -Restore -Force
operation in progress (not interruptible)...
remainder(MB) = 0
operation in progress (not interruptible)...id
= 19
name
= ufs1_snap1
acl
= 0
in_use
= False
type
= uxfs
worm
= off
volume
= v456
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Thu Oct 28 14:25:21 EDT 2004
stor_devs =
000187940268-0180,000187940268-0181,000187940268-0182,000187940268-0183
disks
= rootd378,rootd379,rootd380,rootd381
id
= 18
name
= ufs1
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= mtv1
pool
=
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
backups
= ufs1_snap1
stor_devs =
000187940268-0006,000187940268-0007,000187940268-0008,000187940268-0009
disks
= d3,d4,d5,d6
disk=d3
stor_dev=000187940268-0006
addr=c0t1l0-48-0
server=server_2
disk=d3
stor_dev=000187940268-0006
addr=c16t1l0-33-0
server=server_2
116
Command Line Interface Reference for File 8.1
The fs Commands
disk=d4
disk=d4
disk=d5
disk=d5
disk=d6
disk=d6
stor_dev=000187940268-0007
stor_dev=000187940268-0007
stor_dev=000187940268-0008
stor_dev=000187940268-0008
stor_dev=000187940268-0009
stor_dev=000187940268-0009
EXAMPLE #7
addr=c0t1l1-48-0
addr=c16t1l1-33-0
addr=c0t1l2-48-0
addr=c16t1l2-33-0
addr=c0t1l3-48-0
addr=c16t1l3-33-0
server=server_2
server=server_2
server=server_2
server=server_2
server=server_2
server=server_2
To create a snapshot for a mapped pool, type:
$ fs_timefinder ufs1 -name ufs1_snap1 -Snapshot -option
pool=bcv_sg
operation in progress (not interruptible)...
remainder(MB) = ..14184..0
operation in progress (not interruptible)...id
= 87
name
= ufs1
acl
= 0
in_use
= False
type
= uxfs
worm
= off
volume
= mtv1
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backups
= ufs1_snap1
auto_ext = no,thin=no
deduplication
= unavailable
stor_devs = 000194900546-0037
disks
= d11
id
= 88
name
= ufs1_snap1
acl
= 0
in_use
= False
type
= uxfs
worm
= off
volume
= v456
pool
= bcv_sg
member_of = root_avm_fs_group_49
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
backup_of = ufs1 Fri Oct 1 12:03:10 EDT 2010
auto_ext = no,thin=no
deduplication
= unavailable
thin_storage
= False
tiering_policy = thickfp2
mirrored = False
stor_devs = 000194900546-003C
fs_timefinder
117
The fs Commands
disks
= rootd16
Where:
Value
auto_ext
deduplication
Definition
Indicates whether auto-extension and thin provisioning are enabled.
Deduplication state of the file system. The file data is transferred to the storage which performs the
deduplication and compression on the data. The states are:
• On — Deduplication on the file system is enabled.
• Suspended — Deduplication on the file system is suspended. Deduplication does not perform any
new space reduction but the existing files that were reduced in space remain the same.
• Off — Deduplication on the file system is disabled. Deduplication does not perform any new space
reduction and the data is now reduplicated.
thin_storage
Indicates whether the block system uses thin provisioning. Values are: True, False, Mixed.
tiering_policy Indicates the tiering policy in effect. If the initial tier and the tiering policy are the same, the values are:
Auto-Tier, Highest Available Tier, Lowest Available Tier. If the initial tier and the tiering policy are not the
same, the values are: Auto-Tier/No Data Movement, Highest Available Tier/No Data Movement, Lowest
Available Tier/No Data Movement.
mirrored
Indicates whether the disk is mirrored.
118
Command Line Interface Reference for File 8.1
4
The get and set
Commands
This chapter lists the VNX Command Set provided for managing,
configuring, and monitoring of File Movers. The commands are
network protocol applications, prefixed with get or set, and appear
alphabetically. The command line syntax (Synopsis), a description of
the options, and an example of usage are provided for each
command. Commands included are:
◆
◆
get_attributes .................................................................................... 120
set_attributes..................................................................................... 126
The get and set Commands
119
The get and set Commands
get_attributes
Reads the attributes of the specified file on the primary storage and
verifies the VNX FileMover API configuration.
SYNOPSIS
get_attributes
[-u <username>] [-p <password>][-d][-h <handle>] [-w] [-a] [-S <CA_cert_file>]
[-V <http_version>] <primary_server> <primary_file>
DESCRIPTION
get_attributes is executed on the Control Station after starting the
VNX FileMover API service to return CIFS, NFS, and all vendor
extended attributes in a text format. get_attributes uses the program
“digest” to actually transport the command over the network.
Note: get_attributes is not part of Control Station CLI framework.
PREREQUISITES
OPTIONS
Starts the HTTP server for VNX FileMover by using server_http, and
creates user account for the specified VNX FileMover using
server_user unless user authentication is set to none.
-u <username>
Specifies the account name for digest or basic access authentication. If
no authentication is used, just provides dhsm_user.
-p <password>
Specifies the password for the digest or the basic access
authentication account. If no authentication is used, just provides
dhsm_user.
-d
Displays offline attributes of a deduped file.
-h <handle>
Gets the status of files by handle instead of <primary_file>.
-w
Gets the status of Write Once Read Many (WORM) in file attribute
data.
-a
Includes the Access Control List (ACL) information along with the
standard file attributes.
120
Command Line Interface Reference for File 8.1
The get and set Commands
-S <CA_cert_file>
Uses SSL to perform server certificate verification with the Certificate
Authority (CA) certificated in <CA_cert_file>.
The file name that has the certificate of the CA used to create the Data
Movers certificate for SSL authentication. This is only needed if SSL is
turned on the FileMover API service.
-V <http_version>
Uses the specified version, HTTP/1.0 or HTTP/1.1. Default is
HTTP/1.0.
<primary_server>
Specifies the IP address of the Data Mover, which is hosting the
primary file. Must use IP address, not the NW host name.
<primary_file>
Specifies the full path to the file on the Data Mover.
SEE ALSO
Using VNX FileMover and server_http, server_user, and
server_certificate.
EXAMPLE #1
To verify offline status using VNX FileMover API, type:
$ /nas/tools/dhsm/get_attributes 10.5.8.111 /fs1/pax.tar
<?dhsm?>
<DHSM_GET_ATTRS PATH="nfs:/pfs/file1"/>
HTTP/1.0 200 OK
Date: Sat, 20 Aug 2005 16:36:31 GMT
Content-type: text/xml; charset=UTF-8
<?xml version="1.0" encoding="UTF-8"?>
<?dhsm?>
<RESPONSE>
<OFFLINE_ATTRS
OFFLINE_PATH="http://dhsm-w2k/dir1/File4"
OFFLINE_MTIME="1124555122"
OFFLINE_ETAG="&quot;08c0aea3a5c51:d6a&quot;"
/>
<STANDARD_ATTRS
HANDLE="4294969278-19-1124494767"
ONLINE_CTIME="1124555598000000"
… same as before.
/>
<RESULT SUCCESS="True" VALUE=""/>
</RESPONSE>
get_attributes
121
The get and set Commands
Where:
Value
10.5.8.111
/fs1/pax.tar
Definition
Indicates the IP address of the Data Mover which hosts the primary file.
Indicates the path to the primary file.
EXAMPLE #2
To verify offline status of a deduped file with -d option, type:
$ /nas/tools/dhsm/get_attributes -d 128.221.252.2
/afs/3-1.log
New Command length is 65
spawn telnet 128.221.252.2 5080
Trying 128.221.252.2...
Connected to server_2 (128.221.252.2).
Escape character is '^]'.
POST /dhsm HTTP/1.0
Content-type: text/xml
Content-length: 65
<?dhsm?>
<DHSM_GET_ATTRS PATH="/afs/3-1.log"
DEDUPE="True"/>
HTTP/1.1 200 OK
Connection: Close
Content-Type: text/xml; charset=UTF-8
Server: EMC File Mover service
Date: Mon, 28 Jan 2013 02:00:54 GMT
<?xml version="1.0" encoding="UTF-8"?>
<?dhsm?>
<RESPONSE>
<OFFLINE_ATTRS
OFFLINE_PATH="dart://rde/AQAAAA/cd/e1/Adl3gRkNrbbpS7+hdRXX84qYoWvUAAAAAAAAAAA
AAAAAAwAAAAAAACtrGw;100143ab"
OPTIMIZED_HANDLE="29-1359022552"
ORIGINAL_BYTES_USED="2859008"
DEDUPE_LINKCOUNT="1"
DEDUPE_FSIZE="390821"
OFFLINE_BLOCK_COUNT="348"
BLOCKS_NEEDED_FOR_REINGEST="59"
PE_ID="DHSM API"
/>
<STANDARD_ATTRS
HANDLE="4294968227-27-1359022531"
ONLINE_CTIME="1359022551000451"
UID="0"
GID="0"
ATIME="1359022531"
MTIME="1359022531"
CTIME="1359022551"
CREATE_TIME="1359022531"
122
Command Line Interface Reference for File 8.1
The get and set Commands
DOS_ATTRS="2080"
PARENT_INODE="2"
FSIZE="2845467"
BLOCK_SIZE="8192"
BLOCKS="784"
BYTES_USED="401408"
INODE="27"
DEVICE="931"
NLINK="1"
MODE="0644"
FILE_TYPE="File"
/>
<RESULT SUCCESS="True" VALUE=""/>
</RESPONSE>
EXAMPLE #3
To verify offline status of a deduped file without -d option, type:
$ /nas/tools/dhsm/get_attributes 128.221.252.2
/afs/3-1.log
New Command length is 52
spawn telnet 128.221.252.2 5080
Trying 128.221.252.2...
Connected to server_2 (128.221.252.2).
Escape character is '^]'.
POST /dhsm HTTP/1.0
Content-type: text/xml
Content-length: 52
<?dhsm?>
<DHSM_GET_ATTRS PATH="/afs/3-1.log"
/>
HTTP/1.1 200 OK
Connection: Close
Content-Type: text/xml; charset=UTF-8
Server: EMC File Mover service
Date: Mon, 28 Jan 2013 02:01:35 GMT
<?xml version="1.0" encoding="UTF-8"?>
<?dhsm?>
<RESPONSE>
<STANDARD_ATTRS
HANDLE="4294968227-27-1359022531"
ONLINE_CTIME="1359022551000451"
UID="0"
GID="0"
ATIME="1359022531"
MTIME="1359022531"
CTIME="1359022551"
CREATE_TIME="1359022531"
DOS_ATTRS="2080"
PARENT_INODE="2"
FSIZE="2845467"
get_attributes
123
The get and set Commands
BLOCK_SIZE="8192"
BLOCKS="784"
BYTES_USED="401408"
INODE="27"
DEVICE="931"
NLINK="1"
MODE="0644"
FILE_TYPE="File"
/>
<RESULT SUCCESS="True" VALUE=""/>
</RESPONSE>
EXAMPLE #4
To read the status of a given primary storage, type:
$ /nas/tools/dhsm/get_attributes -u dhsm_user -p
bad_password 10.5.8.111 /
Sending 105 bytes ***
POST /dhsm HTTP/1.0
Content-type: text/xml
Content-length: 38
<?dhsm?>
<DHSM_GET_ATTRS PATH="/" />
HTTP/1.1 401 Unauthorised
Connection: Close
Content-Length: 0
WWW-Authenticate: Basic realm="DHSM_Authorization"
Server: EMC File Mover service
Date: Mon, 01 Oct 2007 17:34:09 GMT
basic challenge
open_connection: server IP 10.5.8.111
open_connection: streaming socket open
open_connection: bind successful
open_connection: connect successful
open_connection: local port = 55315, local addr = 10.5.8.111
Sending 160 bytes ***
POST /dhsm HTTP/1.0
Authorization: Basic ZGhzbV91c2VyOmJhZF9wYXNzd29yZA==
Content-type: text/xml
Content-length: 38
<?dhsm?>
<DHSM_GET_ATTRS PATH="/" />
***
SENT 160 HEADER
HTTP/1.1 401 Unauthorised
Connection: Close
Content-Length: 0
WWW-Authenticate: Basic realm="DHSM_Authorization"
Server: EMC File Mover service
Date: Mon, 01 Oct 2007 17:34:09 GMT
124
Command Line Interface Reference for File 8.1
The get and set Commands
Did not get response - either the username and password were not presented for
responding to a server challenge OR the response to the challenge was challenged
again
EXAMPLE #1 provides a description of command output.
get_attributes
125
The get and set Commands
set_attributes
Changes a file on primary storage into a Stub File or a WORM file.
SYNOPSIS
set_attributes
[-m <migration_method[default|full|partial|read_pass_through]>]
[-u <username>][-p <password>] [-v <offline_mtime>] [-i <info>] [-d <pe_id>]
[-s <file_size>] [-c <online_ctime>][-e <entity_tag>]
[-w <worm_expiration_time>] [-S <CA_cert_file>] [-V <http_version>]
<primary_server> <primary_file>
DESCRIPTION
set_attributes uses the program “digest” to actually transport the
command over the network to set EMC specific attributes, which are
not available in CIFS or NFS.
Note: set_attributes is not part of Control Station CLI framework.
PREREQUISITES
OPTIONS
Before running the command, firstly enables VNX FileMover
operations on a file system by using fs_dhsm, starts the HTTP server
for VNX FileMover by using pserver_http, and creates user account
for the specified VNX FileMover using rserver_userunless user
authentication is set to none.
-m <migration_method>
Chooses one of the methods Data Mover will use when a client
accesses data on a stub file.
The methods include the following: Full – on the first access to the
stub file, the Data Mover will migrate all of the file’s data back from
the secondary store. Partial – On every access to the stub file, the Data
Mover will migrate 128k of data from the secondary store. Read_pass
through – This is the default value. The data will be read from the
secondary storage device and given to the client and no data blocks
will be migrated back to the Data Mover.
-u <username>
Specifies the account name for digest or basic access authentication. If
no authentication is used, just provides dhsm_user.
-p <password>
Specifies the password for the digest or the basic access
authentication account. If no authentication is used, just provides
dhsm_user.
126
Command Line Interface Reference for File 8.1
The get and set Commands
-v <offline_mtime>
Sets <offline_mtime> verifier with a integer value. By default the
script will search the mount command to find the stub file’s verifier
value.
-i <info>
Sets offline attribute INFO.
-d <pe_id>
Sets offline attribute PE_ID.
-s <file_size>
Specifies the logical size of the file in bytes.
-c <online_ctime>
Sets <online_ctime> verifier with integer values. By default the script
will call get_attributes to find the stub file’s verifier value.
-e <entity_tag>
Sets offline attribute ETAG. This should be used instead of the -v
option for HTTP connections to a secondary server.
-w <worm_expiration_time>
Sets the stub file Write Once Read Memory (WORM) expiration time.
<worm_expiration_time> is an integer in seconds since Jan 1, 1970
GMT. A <worm_expiration_time> value of 0 indicates infinite
expiration time.
-S <CA_cert_file>
Uses SSL to perform server certificate verification with the Certificate
Authority (CA) certificated in <CA_cert_file>.
The file name that has the certificate of the CA used to create the Data
Movers’ certificate for SSL authentication. This is only needed if SSL
is turned on the FileMover API service.
-V <http_version>
Uses the specified version, HTTP/1.0 or HTTP/1.1. Default is
HTTP/1.0.
<primary_server>
Specifies the IP address of the Data Mover, which is hosting the
primary file. Must use IP address, not the NW host name.
<primary_file>
Specifies the full path to the file on the Data Mover.
SEE ALSO
Using VNX FileMover and server_http, server_user, and
server_certificate.
set_attributes
127
The get and set Commands
EXAMPLE #1
To create a stub file on the primary storage, type:
$ /nas/tools/dhsm/set_attributes -v 1191008770 10.5.8.111
/fs1/pax.tar nfs://io2/fs1ata/pax.tar
open_connection: server IP 10.5.8.111
open_connection: streaming socket open
open_connection: bind successful
open_connection: connect successful
open_connection: local port = 55315, local addr = 10.5.8.111
Sending 260 bytes ***
POST /dhsm HTTP/1.0
Content-type: text/xml
Content-length: 192
<?dhsm?>
<DHSM_SET_OFFLINE_ATTRS
ONLINE_CTIME="1191008731000171"
HANDLE="4294967320-16-1191008731"
OFFLINE_PATH="nfs://io2/fs1ata/pax.tar"
OFFLINE_MTIME="1191008770"
/>
***
SENT 260 HEADER
HTTP/1.1 200 OK
Connection: Close
Content-Type: text/xml; charset=UTF-8
Server: EMC File Mover service
Date: Fri, 28 Sep 2007 20:12:03 GMT
<?xml version="1.0" encoding="UTF-8"?>
<?dhsm?>
<RESPONSE>
<RESULT SUCCESS="True" VALUE=""/>
</RESPONSE>
Note: Make sure HTTP service for VNX FileMover is started by using
server_http.
EXAMPLE #2
To create a stub file on a secondary server for HTTP connections,
type:
$ /nas/tools/dhsm/set_attributes -u dhsm_user -p
dhsm_user -e f5040c-14a000-c986cd80 -V HTTP/1.1
10.5.8.111 /fs1/pax.tar http://linc57/pax.tar
FILE == /fs1/pax.tar
open_connection: server IP 10.5.8.111
open_connection: streaming socket open
open_connection: bind successful
open_connection: connect successful
open_connection: local port = 55315, local addr = 10.5.8.111
Sending 275 bytes ***
128
Command Line Interface Reference for File 8.1
The get and set Commands
POST /dhsm HTTP/1.1
Host:10.5.8.111
Content-type: text/xml
Content-length: 190
<?dhsm?>
<DHSM_SET_OFFLINE_ATTRS
ONLINE_CTIME="1191010323000002"
HANDLE="4294967320-16-1191008731"
OFFLINE_PATH="http://linc57/pax.tar"
ETAG="f5040c-14a000-c986cd80"
/>
***
SENT 275 HEADER
HTTP/1.1 200 OK
Connection: Keep-Alive
Content-Type: text/xml; charset=UTF-8
Keep-Alive: max=299, timeout=60
Server: EMC File Mover service
Date: Tue, 02 Oct 2007 13:37:24 GMT
Transfer-Encoding: Chunked
<?xml version="1.0" encoding="UTF-8"?>
<?dhsm?>
<RESPONSE>
<RESULT SUCCESS="True" VALUE=""/>
</RESPONSE>
set_attributes
129
The get and set Commands
130
Command Line Interface Reference for File 8.1
5
The nas Commands
This chapter lists the VNX Command Set provided for managing,
configuring, and monitoring of NAS database. The commands are
prefixed with nas and appear alphabetically. The command line
syntax (Synopsis), a description of the options, and an example of
usage are provided for each command. Commands included are:
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
nas_acl................................................................................................
nas_automountmap.........................................................................
nas_ca_certificate .............................................................................
nas_cel................................................................................................
nas_checkup......................................................................................
nas_ckpt_schedule ...........................................................................
nas_connecthome.............................................................................
nas_config .........................................................................................
nas_copy............................................................................................
nas_cs .................................................................................................
nas_dbtable .......................................................................................
nas_devicegroup ..............................................................................
nas_disk .............................................................................................
nas_diskmark....................................................................................
nas_emailuser ...................................................................................
nas_environment..............................................................................
nas_event...........................................................................................
nas_fs .................................................................................................
nas_fsck .............................................................................................
nas_halt..............................................................................................
nas_inventory ...................................................................................
nas_license.........................................................................................
nas_logviewer...................................................................................
The nas Commands
133
136
138
141
154
159
166
174
178
183
190
194
198
206
210
213
221
228
266
270
275
279
281
131
The nas Commands
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
132
nas_message ..................................................................................... 284
nas_mview ........................................................................................ 285
nas_pool ............................................................................................ 294
nas_quotas ......................................................................................... 311
nas_rdf ............................................................................................... 328
nas_replicate ..................................................................................... 349
nas_rp ................................................................................................ 368
nas_server ......................................................................................... 376
nas_slice............................................................................................. 391
nas_stats ............................................................................................ 395
nas_storage ....................................................................................... 402
nas_task ............................................................................................. 418
nas_version ....................................................................................... 422
nas_volume....................................................................................... 424
Command Line Interface Reference for File 8.1
The nas Commands
nas_acl
Manages the access control level table.
SYNOPSIS
nas_acl
-list
| -info {-user|-group|-owner} <numerical_id>
| -delete {-user|-group} <numerical_id>
|[-name <name>] -create {-user|-group} <numerical_id> level=<acl_level>
| -modify {-user|-group} <numerical_id>
{[num_id=<numerical_id>][,level=<acl_level>]}
DESCRIPTION
nas_acl creates, lists, and displays information for access control level
entries within the table, and deletes the specified group or entries.
The access control level table is created and recognized in the NAS
database and contains assigned levels for users and groups. A user
must be defined in the /etc/passwd file prior to being assigned an
entry in the table. Creating an access control level entry defines the
access level allowed for the user or group once a value has been
established for an object.
Note: root privileges are required to create, modify, or delete the access
control level table. The root user is permitted access to all objects.
OPTIONS
-list
Lists the access control level table.
-info {-user|-group|-owner} <numerical_id>
Displays information for the user, group, or index entry of the owner
as specified by the <numerical_id>.
-delete {-user|-group} <numerical_id>
Deletes the entry for the specified user or group from the access
control level table.
-create {-user|-group} <numerical_id> level= <acl_level>
Creates an access control level entry for the specified user or group.
The <numerical_id> can be a user ID (UID) or group ID (GID).
Note: Before executing this command, the user or group must exist in the
Control Station in the /etc/passwd file or the /etc/group file.
nas_acl
133
The nas Commands
The <acl_level> is a single-digit (between numbers 2 and 9) input
representing available access control levels. Levels 2, 3, and 4 which
are established by default are:
2 — admin — Is the most privileged level and includes privileges
allowed from the operator and observer levels.
3 — operator — Includes privileges from the observer level.
4 — observer — The least privileged.
Levels 5—9 are available for configuration.
[-name <name>]
The name is case-sensitive and indicates a name by which the
entry is referred.
Once a value has been set, the level assigned the user or group is
checked in the ACL table and the level of access to the object is
determined.
-modify {-user|-group} <numerical_id>
{[num_id=<numerical_id>] [,level=<acl_level>] }
Modifies the <numerical_id> and level for an access control level
entry.
SEE ALSO
EXAMPLE #1
Controlling Access to System Objects on VNX, nas_fs, nas_volume,
nas_rp, and nas_storage.
Before creating access control level entries, su to root. To create
entries in the access control level table, type:
# nas_acl -name user1 -create -user 211 level=3
done
# nas_acl -name user2 -create -user 212 level=2
done
# nas_acl -name user3 -create -user 213 level=4
done
# nas_acl -name user4 -create -user 214 level=2
done
# nas_acl -name user5 -create -user 215 level=3
done
# nas_acl -name user6 -create -user 216 level=4
done
134
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #2
index
1
2
3
4
5
6
7
type
user
user
user
user
user
user
user
To display the access control level table, type:
$ nas_acl -list
level
num_id
name
admin
201
nasadmin
operator 211
user1
admin
212
user2
observer 213
user3
admin
214
user4
operator 215
user5
observer 216
user6
Where:
Value
index
type
level
num_id
name
Definition
Access control level table index entry number.
User or group for the entry.
Level of access permitted.
Numerical ID for identifying the entry.
Name given to the entry.
EXAMPLE #3
To display information for an access control level entry, type:
$ nas_acl -info -user 211
id
name
level
user_id
=
=
=
=
2
user1
operator
211
Where:
Value
id
name
level
user_id
Definition
Index entry.
Name given for the entry.
Level of access permitted.
Also known as the num_id.
EXAMPLE #4
To modify an access control level entry, type:
# nas_acl -modify -user 211 level=7
done
EXAMPLE #5
To delete an access control level entry, type:
# nas_acl -delete -user 211
done
nas_acl
135
The nas Commands
nas_automountmap
Manages the automount map file.
SYNOPSIS
nas_automountmap
-list_conflict <infile> [-out <outfile>]
| -create [-in <infile>] [-out <outfile>]
DESCRIPTION
OPTIONS
nas_automountmap creates and displays an automount map that
contains all permanently exported file systems that are used by the
automount daemon.
-list_conflict <infile>
Prints a list of the mount points that are used more than once.
[-out <outfile>]
Prints a conflicting list and saves it to an <outfile>.
-create
Creates an automount map and prints it to the screen only.
[-in <infile>] [-out <outfile>]
Merges an automount map with an existing map <infile> and
outputs it to an <outfile>.
[-out <outfile>]
Creates an automount map and outputs it to an <outfile>.
EXAMPLE #1
To create an automount map, type:
$ nas_automountmap -create
ufs1 -rw,intr,nosuid 127.0.0.1,10.172.128.47,128.221.253.2,128.221.252.2:/ufs1
ufs2 -rw,intr,nosuid 127.0.0.1,10.172.128.47,128.221.253.2,128.221.252.2:/ufs2
EXAMPLE #2
To create an automount map and save it to a file, type:
$ nas_automountmap -create -out automountmap
$ more automountmap
ufs1 -rw,intr,nosuid 127.0.0.1,10.172.128.47,128.221.253.2,128.221.252.2:/ufs1
ufs2 -rw,intr,nosuid 127.0.0.1,10.172.128.47,128.221.253.2,128.221.252.2:/ufs2
EXAMPLE #3
To print a conflicting list, type:
$ nas_automountmap -list_conflict automountmap
Conflicting lists:
ufs1 -rw,intr,suid 172.16.21.202:/ufs1
ufs1_172.16.21.203 -rw,intr,suid 172.16.21.203:/ufs1
136
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #4
To merge an automount map file with an existing map file, type:
$ nas_automountmap -create -in automountmap -out
automountmap1
nas_automountmap
137
The nas Commands
nas_ca_certificate
Manages the Control Station as a Certificate Authority (CA) for
VNX’s Public Key Infrastructure (PKI).
SYNOPSIS
nas_ca_certificate
-display
| -generate
DESCRIPTION
nas_ca_certificate generates a public/private key set and a CA
certificate for the Control Station. When the Control Station is serving
as a CA, it must have a private key with which to sign the certificates
it generates for the Data Mover. The Control Station CA certificate
contains the corresponding public key, which is used by clients to
verify the signature on a certificate received from the Data Mover.
nas_ca_certificate also displays the text of the CA certificate so you
can copy it and distribute it to network clients. In order for a network
client to validate a certificate sent by a Data Mover that has been
signed by the Control Station, the client needs the Control Station CA
certificate (specifically the public key from the CA certificate) to
verify the signature of the Data Mover’s certificate.
The initial Control Station public/private key set and CA certificate
are generated automatically during a VNX software 5.6 install or
upgrade. A new Control Station public/private key set and CA
certificate is not required unless the CA key set is compromised or the
CA certificate expires. The Control Station CA certificate is valid for 5
years.
You must be root to execute the -generate option from the /nas/sbin
directory.
Once a Control Station CA certificate is generated, you must perform
several additional tasks to ensure that the new certificate is integrated
into VNX’s PKI framework. The Security Configuration Guide for File
and the Unisphere online help for the PKI interface explain these
tasks.
OPTIONS
-display
Displays the Control Station CA certificate. The certificate text is
displayed on the terminal screen. Alternatively, you can redirect it to
a file.
138
Command Line Interface Reference for File 8.1
The nas Commands
-generate
Generates a new CA public/private key set and certificate for the
Control Station. This certificate is valid for 5 years from the date it is
generated.
SEE ALSO
EXAMPLE #1
server_certificate.
To generate a new Control Station CA certificate, type:
# /nas/sbin/nas_ca_certificate -generate
New keys and certificate were successfully generated.
EXAMPLE #2
To display the Control Station’s CA certificate, type:
# /nas/sbin/nas_ca_certificate -display
Note: Clients need only the certificate text enclosed by BEGIN
CERTIFICATE and END CERTIFICATE although most clients can handle the
entire output.
Certificate:
Data:
Version: 3 (0x2)
Serial Number: 3 (0x3)
Signature Algorithm: sha1WithRSAEncryption
Issuer: O=Celerra Certificate Authority, CN=eng173100
Validity
Not Before: Mar 23 21:07:40 2007 GMT
Not After : Mar 21 21:07:40 2012 GMT
Subject: O=Celerra Certificate Authority, CN=eng173100
Subject Public Key Info:
Public Key Algorithm: rsaEncryption
RSA Public Key: (2048 bit)
Modulus (2048 bit):
00:da:b2:37:86:05:a3:73:d5:9a:04:ba:db:05:97:
d2:12:fe:1a:79:06:19:eb:c7:2c:c2:51:93:7f:7a:
93:59:37:63:1e:53:b3:8d:d2:7f:f0:e3:49:42:22:
f4:26:9b:b4:e4:a6:40:6d:8d:e7:ea:07:8e:ca:b7:
7e:88:71:9d:11:27:5a:e3:57:16:03:a7:ee:19:25:
07:d9:42:17:b4:eb:e6:97:61:13:54:62:03:ec:93:
b7:e6:f1:7f:21:f0:71:2d:c4:8a:8f:20:d1:ab:5a:
6a:6c:f1:f6:2f:26:8c:39:32:93:93:67:bb:03:a7:
22:29:00:11:e0:a1:12:4b:02:79:fb:0f:fc:54:90:
30:65:cd:ea:e6:84:cc:91:fe:21:9c:c1:91:f3:17:
1e:44:7b:6f:23:e9:17:63:88:92:ea:80:a5:ca:38:
9a:b3:f8:08:cb:32:16:56:8b:c4:f7:54:ef:75:db:
36:7e:cf:ef:75:44:11:69:bf:7c:06:97:d1:87:ff:
5f:22:b5:ad:c3:94:a5:f8:a7:69:21:60:5a:04:5e:
00:15:04:77:47:03:ec:c5:7a:a2:bf:32:0e:4d:d8:
dc:44:fa:26:39:16:84:a7:1f:11:ef:a3:37:39:a6:
35:b1:e9:a8:aa:a8:4a:72:8a:b8:c4:bf:04:70:12:
nas_ca_certificate
139
The nas Commands
b3:31
Exponent: 65537 (0x10001)
X509v3 extensions:
X509v3 Subject Key Identifier:
35:06:F2:FE:CC:21:4B:92:DA:74:C9:47:CE:BB:37:21:5E:04:E2:E6
X509v3 Authority Key Identifier:
keyid:35:06:F2:FE:CC:21:4B:92:DA:74:C9:47:CE:BB:37:21:5E:04:E2:E6
DirName:/O=Celerra Certificate Authority/CN=eng173100
serial:00
X509v3 Basic Constraints:
CA:TRUE
X509v3 Subject Alternative Name:
DNS:eng173100
Signature Algorithm: sha1WithRSAEncryption
09:c3:13:26:16:be:44:56:82:5d:0e:63:07:19:28:f3:6a:c4:
f3:bf:93:25:85:c3:55:48:4e:07:84:1d:ea:18:cf:8b:b8:2d:
54:13:25:2f:c9:75:c1:28:39:88:91:04:df:47:2c:c0:8f:a4:
ba:a6:cd:aa:59:8a:33:7d:55:29:aa:23:59:ab:be:1d:57:f6:
20:e7:2b:68:98:f2:5d:ed:58:31:d5:62:85:5d:6a:3f:6d:2b:
2d:f3:41:be:97:3f:cf:05:8b:7e:f5:d7:e8:7c:66:b2:ea:ed:
58:d4:f0:1c:91:d8:80:af:3c:ff:14:b6:e7:51:73:bb:64:84:
26:95:67:c6:60:32:67:c1:f7:66:f4:79:b5:5d:32:33:3c:00:
8c:75:7d:02:06:d3:1a:4e:18:0b:86:78:24:37:18:20:31:61:
59:dd:78:1f:88:f8:38:a0:f4:25:2e:c8:85:4f:ce:8a:88:f4:
4f:12:7e:ee:84:52:b4:91:fe:ff:07:6c:32:ca:41:d0:a6:c0:
9d:8f:cc:e8:74:ee:ab:f3:a5:b9:ad:bb:d7:79:67:89:34:52:
b4:6b:39:db:83:27:43:84:c3:c3:ca:cd:b2:0c:1d:f5:20:de:
7a:dc:f0:1f:fc:70:5b:71:bf:e3:14:31:4c:7e:eb:b5:11:9c:
96:bf:fe:6f
-----BEGIN CERTIFICATE----MIIDoDCCAoigAwIBAgIBAzANBgkqhkiG9w0BAQUFADA8MSYwJAYDVQQKEx1DZWxl
cnJhIENlcnRpZmljYXRlIEF1dGhvcml0eTESMBAGA1UEAxMJZW5nMTczMTAwMB4X
DTA3MDMyMzIxMDc0MFoXDTEyMDMyMTIxMDc0MFowPDEmMCQGA1UEChMdQ2VsZXJy
YSBDZXJ0aWZpY2F0ZSBBdXRob3JpdHkxEjAQBgNVBAMTCWVuZzE3MzEwMDCCASIw
DQYJKoZIhvcNAQEBBQADggEPADCCAQoCggEBANqyN4YFo3PVmgS62wWX0hL+GnkG
GevHLMJRk396k1k3Yx5Ts43Sf/DjSUIi9CabtOSmQG2N5+oHjsq3fohxnREnWuNX
FgOn7hklB9lCF7Tr5pdhE1RiA+yTt+bxfyHwcS3Eio8g0ataamzx9i8mjDkyk5Nn
uwOnIikAEeChEksCefsP/FSQMGXN6uaEzJH+IZzBkfMXHkR7byPpF2OIkuqApco4
mrP4CMsyFlaLxPdU73XbNn7P73VEEWm/fAaX0Yf/XyK1rcOUpfinaSFgWgReABUE
d0cD7MV6or8yDk3Y3ET6JjkWhKcfEe+jNzmmNbHpqKqoSnKKuMS/BHASszECAwEA
AaOBrDCBqTAdBgNVHQ4EFgQUNQby/swhS5LadMlHzrs3IV4E4uYwZAYDVR0jBF0w
W4AUNQby/swhS5LadMlHzrs3IV4E4uahQKQ+MDwxJjAkBgNVBAoTHUNlbGVycmEg
Q2VydGlmaWNhdGUgQXV0aG9yaXR5MRIwEAYDVQQDEwllbmcxNzMxMDCCAQAwDAYD
VR0TBAUwAwEB/zAUBgNVHREEDTALggllbmcxNzMxMDAwDQYJKoZIhvcNAQEFBQAD
ggEBAAnDEyYWvkRWgl0OYwcZKPNqxPO/kyWFw1VITgeEHeoYz4u4LVQTJS/JdcEo
OYiRBN9HLMCPpLqmzapZijN9VSmqI1mrvh1X9iDnK2iY8l3tWDHVYoVdaj9tKy3z
Qb6XP88Fi3711+h8ZrLq7VjU8ByR2ICvPP8UtudRc7tkhCaVZ8ZgMmfB92b0ebVd
MjM8AIx1fQIG0xpOGAuGeCQ3GCAxYVndeB+I+Dig9CUuyIVPzoqI9E8Sfu6EUrSR
/v8HbDLKQdCmwJ2PzOh07qvzpbmtu9d5Z4k0UrRrOduDJ0OEw8PKzbIMHfUg3nrc
8B/8cFtxv+MUMUx+67URnJa//m8=
-----END CERTIFICATE-----
140
Command Line Interface Reference for File 8.1
The nas Commands
nas_cel
Performs management of remotely linked VNX or a linked pair of
Data Movers.
SYNOPSIS
nas_cel
-list
| -delete {<cel_name>|id=<cel_id>} [-Force]
| -info {<cel_name>|id=<cel_id>}
| -update {<cel_name>|id=<cel_id>}
| -modify {<cel_name>|id=<cel_id>}
{[-passphrase <passphrase>][-name <new_name>][-ip <ipaddr>]}
| -create <cel_name> -ip <ipaddr> -passphrase <passphrase>
| -interconnect <interconnect_options>
DESCRIPTION
nas_cel manages the linking of the remote VNX to the local VNX.
nas_cel also creates the trusted relationship between source and
destination VNX Control Stations in configurations such as EMC
VNX Replicator™.
For VNX Replicator only, nas_cel -interconnect also builds the
connection (interconnect) between a pair of Data Movers.
Linked VNX systems are acknowledged:
OPTIONS
◆
Automatically during the installation
◆
When executing the nas_cel -create
◆
When performing a nas_rdf -init to set up the Symmetrix Remote
Data Facility (SRDF) relationship between two VNX.
-list
Lists all VNX linked to the current VNX. The hostname of the Control
Station active during installation appears as the <cel_name>.
Note: The ID of the object is an integer and is assigned automatically. The
name of the VNX might be truncated if it is too long for the display. To view
the full name, use the -info option with the VNX ID.
-delete <cel_name>|id=<cel_id>} [-Force]
Deletes the relationship of the remote VNX, and removes its entry
from the NAS database on the local VNX.
The -Force option applies to SRDF and EMC MirrorView™/S
configurations only. If the VNX to be deleted is part of an SRDF or
nas_cel
141
The nas Commands
MirrorView/S configuration, -delete must be specified with the
-Force option; otherwise, an error is generated. You cannot use -Force
if the specified VNX is also being used by VNX Replicator, file system
copy (for example, with nas_copy), or TimeFinder/FS NearCopy or
FarCopy. If the deletion is necessary, clean up these configurations
before performing the forced deletion.
-info {<cel_name>|id=<cel_id>}
Displays information for the remote VNX. To view the <cel_id> of
configured VNX, use -list.
-update {<cel_name>|id=<cel_id>}
Updates the local VNX entry with the local Control Station's
hostname and IP address configuration. It also updates the local Data
Mover-to-Data Mover authentication setup.
For the remote VNX, updates all Data Movers that were down or
experiencing errors during the -create or -modify and restores them
to service by using the configuration required for Data Mover
authentication.
Note: Data Mover authentication is used in iSCSI replication as the
mechanism enabling two Data Movers (local or remote) to authenticate
themselves and perform the requested operations. The -update option
communicates with each Data Mover and either updates the configuration,
or creates the configuration if it is being done for the first time.
-modify {<cel_name>|id=<cel_id>}
{[-passphrase <passphrase>][-name <new_name>][-ip
<ipaddr>]}
Changes the current passphrase, name, or IP address of the remote
VNX to the new passphrase, name, or IP address in the local VNX
database and modifies the remote Data Mover authentication setup
by communicating with each Data Mover in the cabinet. The
passphrase must have 6 to 15 characters.
-create <cel_name> -ip <ipaddr> -passphrase <passphrase>
Builds the trusted relationship between one VNX and another VNX
in a configuration such as VNX Replicator, SRDF, and MirrorView/S.
The -create must be executed twice to ensure communication from
both sides, first on the source VNX (to identify the destination VNX)
and then on the destination VNX (to identify the source VNX). You
must assign a name when you create the relationship (for example, a
name that identifies the remote VNX in a local entry). The IP address
specified represents the appropriate remote VNX’s primary Control
142
Command Line Interface Reference for File 8.1
The nas Commands
Station (in slot 0); the passphrase specified is used to manage the
remote VNX. The passphrase must have 6 to 15 characters and be the
same between the source and destination VNXs to enable
communication.
INTERCONNECT
OPTIONS
Type nas_cel -interconnect to display interconnect options:
-interconnect
{ -create <name>
-source_server <movername>
-destination_system {<cel_name>|id=<cel_id>}
-destination_server <movername>
-source_interfaces {<name_service_interface_name>|ip=<ipaddr>}
[,{<name_service_interface_name>|ip=<ipaddr>},...]
-destination_interfaces {<name_service_interface_name>|
ip=<ipaddr>}[,{<name_service_interface_name>|ip=<ipaddr>},...]
[-bandwidth <bandwidthSched>]
| -modify {<name>|id=<interConnectId>}
{[-source_interfaces {<name_service_interface_name>|ip=<ipaddr>},...]
[-destination_interfaces
{<name_service_interface_name>|ip=<ipaddr>},...]
[-bandwidth <bandwidthSched>]
[-name <newName>]}
| -pause {<name>|id=<interConnectId>}
| -resume {<name>|id=<interConnectId>}
| -delete {<name>|id=<interConnectId>}
| -info {<name>|id=<interConnectId>|-all}
| -list [-destination_system {<cel_name>|id=<cel_id>}]
| -validate {<name>|id=<interconnectId>}}
An interconnect supports VNX Replicator sessions by defining the
communications path between a given Data Mover pair located on
the same cabinet or different cabinets. The interconnect configures a
list of local (source) and peer (destination) interfaces (using IP
addresses and interface names), and a bandwidth schedule for all
replication sessions using the interconnect. Only one interconnect can
be established for a given Data Mover pair in any direction.
Note: You must delete all user-defined interconnects configured for a Data
Mover before you can rename it. After you rename the Data Mover, you must
re-create the source and peer interconnects with the new Data Mover name
and then restart any associated replication sessions.
To fully establish an interconnect, nas_cel -interconnect must be
issued twice, once from each side (the local side and its peer side).
Both sides of the interconnect must exist before VNX Replicator
sessions for local or remote replication can use the interconnect. Only
nas_cel
143
The nas Commands
the local side of an interconnect on which the source replication object
resides is specified when creating the replication session. Loopback
interconnects are created and named automatically and can be
viewed using nas_cel -interconnect -list. You cannot create, modify,
or delete loopback interconnects.
-create <name>
Assigns a name, up to 255 characters, to the appropriate side of the
interconnect. The name must be unique for each Data Mover. Make
the name meaningful, identifying servers and, for remote replication,
VNX names or sites.
Remote replication naming example:
s2CelA_s3CelB or NYs3_LAs4 (local side)
s3CelB_s2CelA or LAs4_NYs3 (peer side)
Local replication naming example:
s2_s3 (source side on local system)
s3_s2 (peer side on the same system)
-source_server <moverName>
Specifies the name of an available local Data Mover to use for the
local side of the interconnect.
-destination_system {<cel_name>|id=<cel_id>}
Specifies the name or ID of the VNX where the peer Data Mover
resides.
-destination_server <movername>
Specifies the name of an available Data Mover, on the same or
different system, to use for the peer side of the interconnect.
-source_interfaces
{<name_service_interface_name>|ip=<ipaddr>}
[,{<name_service_interface_name>|ip=<ipaddr>},...]
Configures a list of interfaces available for the local side of the
interconnect. You can define the list by using IP addresses (IPv4
or IPv6) or name service interface names or a combination of
both, but how you specify an interface determines how it must be
specified by the replication session later (by name service
interface name or IP address).
If you define an interface by using an IP address, ensure that the
source interface list uses the same IPv4/IPv6 protocol. An IPv4
interface cannot connect to an IPv6 interface and vice versa. Both
sides of the connection must use the same protocol.
144
Command Line Interface Reference for File 8.1
The nas Commands
For each network protocol type (IPv4/IPv6) specified in the
source interface list, at least one interface from the same type
must be specified in the destination interfaces list and vice versa.
For example, if the source interface list includes one or more IPv6
addresses, the destination interface list must also include at least
one IPv6 address.
The name service interface name is a fully qualified name given
to a network interface that must resolve to a single IP address (for
example, using a DNS server).
Note: To prevent potential errors during interface selection (especially
after a failover/switchover), it is highly recommended that you specify
the same local and peer interface lists when configuring each side of the
interconnect.
-destination_interfaces
{<name_service_interface_name>| ip=<ipaddr>}
[,{<name_service_interface_name>|ip= <ipaddr>},...]
Configures a list of interfaces available on the peer side of the
interconnect. You can define the list by using IP addresses (IPv4
or IPv6) or name service interface names or a combination of
both, but how you specify each interface determines how it is
specified by the replication session.
If you define an interface using an IP address, ensure that the
source interface list uses the same IPv4/IPv6 protocol. An IPv4
interface cannot connect to an IPv6 interface and vice versa. Both
sides of the connection must use the same protocol.
For each network protocol type (IPv4/IPv6) specified in the
destination interface list, at least one interface from the same type
must be specified in the source interfaces list and vice versa. For
example, if the source interface list includes one or more IPv6
addresses, the destination interface list must also include at least
one IPv6 address. The name service interface name is a fully
qualified name given to a network interface that must resolve to a
single IP address (for example, using a DNS server).
[-bandwidth <bandwidthSched>]
Specifies a schedule to control the interconnect bandwidth used
on specific days, or times instead of using all available bandwidth
at all times for the interconnect (the default).
nas_cel
145
The nas Commands
Note: The bandwidth schedule executes based on Data Mover time, not
Control Station time.
The schedule applies to all VNX Replicator sessions using the
interconnect. Specify a schedule with one or more
comma-separated entries, most specific to least specific, as
follows:
[{Su|Mo|Tu|We|Th|Fr|Sa}][HH:00-HH:00][/Kbps]
,[ <next_entry>],[...]
Example:
MoTuWeThFr07:00-18:00/2000,/8000 means use a limit of 2000
Kb/s from 7 a.m. to 6 p.m. Monday through Friday; otherwise,
use a bandwidth limit of 8000 Kb/s.
-interconnect -modify{<name>|id=<interConnectId>}
Modifies one or more of the following characteristics of an
existing interconnect, as specified by the name or ID for the
appropriate side of the interconnect.
Note: You cannot modify the peer side of an interconnect configured on a
remote system; you must modify it from that system. Also, you cannot
modify an interface in use by a replication session.
[-source_interfaces{<name_service_interface_name>|
ip= <ipAddr>},… ]
Modifies the list of interfaces (name service interface name or IP
addresses or both) available for use on the local side of an
interconnect. The new list of interfaces completely replaces the
previous list.
Note: To avoid problems with interface selection, any changes made to
the interface lists should be reflected on both sides of an interconnect.
146
Command Line Interface Reference for File 8.1
The nas Commands
[-destination_interfaces{<name_service_interface_name
>|ip= <ipAddr>},…]
Modifies the list of interfaces (name service interface name or IP
addresses or both) available for use on the peer side of an
interconnect. The new list of interfaces completely replaces the
previous list.
[-bandwidth <bandwidth>]
Modifies the existing bandwidth schedule for the specified
interconnect, or creates a schedule if none existed previously. The
schedule allocates the interconnect bandwidth for specific days or
times or both instead of using all available bandwidth at all times
for the interconnect (the default). The schedule applies to all
replication sessions using the interconnect. Specify a schedule
with one or more comma-separated entries, most specific to least
specific, as follows:
[{Su|Mo|Tu|We|Th|Fr|Sa}][HH:00-HH:00][/Kbps]
,[ <next_entry>],[...]
Example:
MoTuWeThFr07:00-18:00/2000,/8000 means use a limit of 2000
Kb/s from 7 a.m. to 6 p.m. Monday through Friday; otherwise,
use a bandwidth limit of 8000 Kb/s.
[-name <newName>]
Changes the name of the specified interconnect to a new name.
-interconnect -pause {<name>|id=<interConnectId>}
Halts data transmission over the existing Data Mover interconnect
until you resume transmission over the interconnect or delete the
interconnect. This affects all replication sessions using the specified
interconnect.
-interconnect -resume {<name>|id= <interConnectId>}
Resumes data transmission over the Data Mover interconnect,
making the interconnect available for use by replication sessions.
-interconnect -delete {<name>|id= <interConnectId>}
Deletes the Data Mover interconnect, thereby making the
interconnect unavailable for use by any replication sessions. You
cannot delete an interconnect if it is in use by a replication session.
You can delete a paused interconnect.
nas_cel
147
The nas Commands
-interconnect -info {<name>|id=<interConnectId>| -all}
Displays information about the specified interconnect or about all
interconnects known to the local system.
-interconnect -list [-destination_system <cel_name>
|id=<cel_id>]
By default, lists the interconnects available on the local VNX.
Specifying the name or ID of a remote VNX also lists the
interconnects available on that VNX.
-interconnect -validate {<name>|id= <interconnectId>}
Verifies the interconnect, verifying that authentication is configured
properly by opening the connection between the Data Mover pair.
Validation is done for loopback, local, and remote configuration.
SEE ALSO
EXAMPLE #1
Using VNX Replicator, nas_copy, nas_replicate, and nas_task.
To create an entry for the remote VNX, type:
$ nas_cel -create cs110 -ip 172.24.102.240 -passphrase
nasdocs
progress (not interruptible)...
3
cs110
0
operation in
id
=
name
=
owner
=
device
=
channel
=
net_path
= 172.24.102.240
VNX_id = APM000438070430000
passphrase = nasdocs
Where:
Value
id
name
owner
device
channel
net_path
VNX_id
passphrase
148
Definition
ID of the remote VNX on the local VNX.
Name assigned in the local view to the remote VNX.
ACL ID assigned automatically.
R2 device mounted by the local Control Station to read the database of the remote Control Station in the
SRDF environment. This value is unique to the Symmetrix system.
Pair of devices used in the rdf channel. One is used for writing messages to the remote (wdev), the other to
read messages from them. This value is unique to the Symmetrix system.
IP address of the remote VNX.
Unique VNX ID number.
Used for authentication with a remote VNX.
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #2
id
0
3
name
cs100
cs110
For the VNX for block, to list all remote VNXs, type:
$ nas_cel -list
owner mount_dev channel
0
0
net_path
172.24.102.236
172.24.102.240
CMU
APM000420008180000
APM000438070430000
For the VNX with a Symmetrix system, to list all remote VNXs, type:
id
0
1
name
cs30
cs40
$ nas_cel -list
owner mount_dev channel
0
500
/dev/sdj1 /dev/sdg
net_path
CMU
172.24.172.152 0028040001900006
172.24.172.151 0028040002180000
Where:
Value
id
name
owner
mount_dev
channel
net_path
CMU
Definition
ID of the remote VNX on the local VNX.
Name assigned in the local view to the remote VNX.
ACL ID assigned automatically.
Mounted database from the remote VNX in the SRDF environment. This value is unique to the Symmetrix
system.
RDF channel from where information is read and written. This value is unique to the Symmetrix system.
IP address of the remote VNX.
VNX Management Unit (unique VNX ID number).
EXAMPLE #3
To display information for the remote VNX, cs110, type:
$ nas_cel -info cs110
id
= 3
name
= cs110
owner
= 0
device
=
channel
=
net_path
= 172.24.102.240
VNX_id = APM000438070430000
passphrase = nasdocs
EXAMPLE #1 provides information for a description of command
outputs.
EXAMPLE #4
To update the Control Station entry for cs110, type:
$ nas_cel -update cs110
operation in progress (not interruptible)...
id
= 3
name
= cs110
owner
= 0
nas_cel
149
The nas Commands
device
=
channel
=
net_path
= 172.24.102.240
VNX_id = APM000438070430000
passphrase = nasdocs
EXAMPLE #1 provides information for a description of command
outputs.
EXAMPLE #5
To modify the passphrase and name for the remote Control Station
cs110, type:
$ nas_cel -modify cs110 -passphrase nasdocs_replication
-name cs110_target
progress (not interruptible)...
3
cs110_target
0
operation in
id
=
name
=
owner
=
device
=
channel
=
net_path
= 172.24.102.240
VNX_id = APM000438070430000
passphrase = nasdocs_replication
EXAMPLE #1 provides information for a description of command
outputs.
EXAMPLE #6
To delete the Control Station entry of the remote VNX, cs110_target,
type:
$ nas_cel -delete cs110_target
operation in progress (not interruptible)...
id
= 3
name
= cs110_target
owner
= 0
device
=
channel
=
net_path
= 172.24.102.240
VNX_id = APM000438070430000
passphrase = nasdocs_replication
EXAMPLE #1 provides information for a description of command
outputs.
150
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #7
To create an interconnect NYs3_LAs2 between Data Mover server_3
and remote Data Mover server_2, and use a bandwidth limit of 2000
Kb/s from 7 a.m. to 6 p.m. Monday through Friday; otherwise, use a
bandwidth limit of 8000 Kb/s, type:
$ nas_cel -interconnect -create NYs3_LAs2 -source_server
server_3
-destination_system cs110 -destination_server server_2
-source_interfaces
ip=10.6.3.190 -destination_interfaces ip=10.6.3.173
-bandwidth
MoTuWeThFr07:00-18:00/2000,/8000
operation in progress (not interruptible)...
id
= 30003
name
= NYs3_LAs2
source_server
= server_3
source_interfaces
= 10.6.3.190
destination_system
= cs110
destination_server
= server_2
destination_interfaces
= 10.6.3.173
bandwidth schedule
= MoTuWeThFr07:00-18:00/2000,/8000
crc enabled
= yes
number of configured replications = 0
number of replications in transfer = 0
status
= The interconnect is OK.
Where:
Value
id
name
source_server
source_interfaces
Definition
ID of the interconnect.
Name of the interconnect.
Name of an available local Data Mover to use for the local side of the interconnect.
IP addresses available for the local side of the interconnect (at least one, or a name service
interface name).
destination_system
Control Station names of the VNX systems available for use in a remote replication
session. Local System is the default.
destination_server
Name of an available peer Data Mover to use for the peer side of the interconnect.
destination_interface IP addresses available for the peer side of the interconnect (at least one, or a name service
s
interface name). For loopback interconnects, the interface is fixed at 127.0.0.1.
bandwidth schedule
Bandwidth schedule with one or more comma-separated entries, most specific to least
specific.
crc enabled
Indicates that the Cyclic Redundancy Check (CRC) method is in use for verifying the
integrity of data sent over the interconnect. CRC is automatically enabled and cannot be
disabled.
nas_cel
151
The nas Commands
Value
number of configured
replications
number of
replications in
transfer
status
EXAMPLE #8
Definition
Number of replication sessions currently configured.
Number of replications are currently in transfer.
Status of the interconnect.
To modify the bandwidth schedule of the interconnect NYs3_LAs2,
type:
$ nas_cel -interconnect -modify NYs3_LAs2 -bandwidth
MoWeFr07:00-18:00/2000,TuTh07:00-18:00/4000,/8000
operation in progress (not interruptible)...
id
= 30003
name
= NYs3_LAs2
source_server
= server_3
source_interfaces
= 10.6.3.190
destination_system
= cs110
destination_server
= server_2
destination_interfaces
= 10.6.3.173
bandwidth schedule
=
MoWeFr07:00-18:00/2000,TuTh07:00-18:00/4000,/8000
crc enabled
= yes
number of configured replications = 0
number of replications in transfer = 0
status
= The interconnect is OK.
EXAMPLE #7 provides a description of the command outputs.
EXAMPLE #9
id
20001
30001
30003
name
loopback
loopback
NYs3_LAs2
EXAMPLE #10
To list available interconnects, type:
$ nas_cel -interconnect -list
source_server
destination_system
server_2
cs100
server_3
cs100
server_3
cs110
destination_server
server_2
server_3
server_2
To pause the interconnect with id=30003, type:
$ nas_cel -interconnect -pause id=30003
done
EXAMPLE #11
To resume the interconnect NYs3_LAs2, type:
$ nas_cel -interconnect -resume NYs3_LAs2
done
EXAMPLE #12
To validate the interconnect NYs3_LAs2, type:
$ nas_cel -interconnect -validate NYs3_LAs2
152
Command Line Interface Reference for File 8.1
The nas Commands
NYs3_LAs2: validating 9 interface pairs: please wait...ok
EXAMPLE #13
To display the detailed information about the interconnect
NYs3_LAs2, type:
$ nas_cel -interconnect -info NYs3_LAs2
id
= 30003
name
= NYs3_LAs2
source_server
= server_3
source_interfaces
= 10.6.3.190
destination_system
= cs110
destination_server
= server_2
destination_interfaces
= 10.6.3.173
bandwidth schedule
=
MoWeFr07:00-18:00/2000,TuTh07:00-18:00/4000,/8000
crc enabled
= yes
number of configured replications = 0
number of replications in transfer = 0
status
= The interconnect is OK.
EXAMPLE #7 provides a description of the command outputs.
EXAMPLE #14
To delete interconnect NYs3_LAs2, type:
$ nas_cel -interconnect -delete NYs3_LAs2
operation in progress (not interruptible)...
id
= 30003
name
= NYs3_LAs2
source_server
= server_3
source_interfaces
= 10.6.3.190
destination_system
= cs110
destination_server
= server_2
destination_interfaces
= 10.6.3.173
bandwidth schedule
=
MoWeFr07:00-18:00/2000,TuTh07:00-18:00/4000,/8000
crc enabled
= no
number of configured replications = 0
number of replications in transfer = 0
status
= The interconnect is OK.
EXAMPLE #7 provides a description of the command outputs.
nas_cel
153
The nas Commands
nas_checkup
Provides a system health checkup for the VNX.
SYNOPSIS
nas_checkup
[-version|-help|-rerun]
DESCRIPTION
nas_checkup runs scheduled and unscheduled health checks on the
VNX, reports problems that are found and the actions needed to fix
the problem, and acts as a system health monitor.
The scheduled run time for the nas_checkup command is every 2
weeks by default. If a warning or error is discovered during this time,
an alert is posted on the Unisphere.
Set up email notification for warnings or errors in the Unisphere
Notifications page, or modify and load the sample nas_checkup
event configuration file.
If a problem is discovered that requires EMC Service Personnel
assistance, nas_checkup will notify EMC.
OPTIONS
No arguments
Runs a series of system health checks on the VNX and reports the
problems that are found and the actions needed to fix the problem.
Note: No email, callhome, or Unisphere alert is posted when the health check
is run unscheduled.
-version
Displays the version of health check that is run on the VNX.
-help
Provides help.
-rerun
Reruns the checks that produce error messages in the previous health
checkup. It does not rerun the checks that produce warning or
information messages. If there are no checks that produce error
messages, then the -rerun switch generates a message that there is
nothing to rerun.
154
Command Line Interface Reference for File 8.1
The nas Commands
CHECKS
Nas_checkup runs a subset of the available checks based on the
configuration of your system. The complete list of available checks
are:
Control Station Checks:
Check if minimum free space exists
Check if minimum free space exists ns
Check if enough free space exists
Check if enough free space exists ns
Check if NAS Storage API is installed correctly
Check if NAS Storage APIs match
Check if NBS clients are started
Check if NBS configuration exists
Check if NBS devices are accessible
Check if NBS service is started
Check if standby is up
Check if Symapi data is present
Check if Symapi is synced with System
Check integrity of NASDB
Check if primary is active
Check all callhome files delivered
Check if NAS partitions are mounted
Data Mover Checks:
Check boot files
Check if hardware is supported
Check if primary is active
Check if root filesystem has enough free space
Check if using standard DART image
Check MAC address
Check network connectivity
Check status
System Checks:
Check disk emulation type
Check disk high availability access
Check disks read cache enabled
Check disks and storage processors write cache enabled
Check if access logix is enabled
Check if FLARE is committed
Check if FLARE is supported
Check if microcode is supported
Check no disks or storage processors are failed over
Check that no disks or storage processors are faulted
Check that no hot spares are in use
Check that no hot spares are rebuilding
Check control lun size
Check if storage processors are read cache enabled
nas_checkup
155
The nas Commands
FILES
The files associated with system health checkups are:
/nas/log/nas_
checkup-run.<timestamp>.log
Contains information about the checks that were run, problems found, and
actions needed to fix the problem.
/nas/log/nas_checkup.<timestam Produced when a scheduled nas_checkup is run and contains the same
p>.log
information as the nas_checkup-run.<timestamp>.log.
Contains the overall results of nas_checkup.
/nas/log/syslog
/nas/site/checkup_eventlog.cfg Provides a sample nas_checkup event configuration file. This is the file to be
modified to add email addresses and load the file.
SEE ALSO
EXAMPLE #1
Check Version:
Check Command:
Check Log
:
Configuring Events and Notifications on VNX for File.
To run a health check on the VNX, type:
$ nas_checkup
5.6.23.1
/nas/bin/nas_checkup
/nas/log/checkup-run.070611-064115.log
-------------------------------------Checks------------------------------------Control Station: Checking if file system usage is under limit.............. Pass
Control Station: Checking if file systems have enough space to upgrade..... Pass
Control Station: Checking if NAS Storage API is installed correctly........ Pass
Control Station: Checking if NBS clients are started....................... Pass
Control Station: Checking if NBS configuration exists...................... Pass
Control Station: Checking if NBS devices are accessible.................... Pass
Control Station: Checking if NBS service is started........................ Pass
Control Station: Checking if standby is up................................. N/A
Control Station: Checking if Symapi data is present........................ Pass
Control Station: Checking if Symapi is synced with Storage System.......... Pass
Control Station: Checking integrity of NASDB............................... Pass
Control Station: Checking all callhome files delivered..................... Pass
Control Station: Checking resolv conf...................................... Pass
Control Station: Checking if NAS partitions are mounted.................... Pass
Control Station: Checking ipmi connection.................................. Pass
Control Station: Checking nas site eventlog configuration.................. Pass
Control Station: Checking nas sys mcd configuration........................ Pass
Control Station: Checking nas sys eventlog configuration................... Pass
Control Station: Checking logical volume status............................ Pass
Control Station: Checking ups is available................................. Fail
Data Movers
: Checking boot files....................................... Pass
Data Movers
: Checking if primary is active............................. Pass
Data Movers
: Checking if root filesystem has enough free space......... Pass
Data Movers
: Checking if using standard DART image..................... Pass
Data Movers
: Checking network connectivity............................. Pass
Data Movers
: Checking status........................................... Pass
Data Movers
: Checking dart release compatibility....................... Pass
156
Command Line Interface Reference for File 8.1
The nas Commands
Data Movers
: Checking dart version compatibility....................... Pass
Data Movers
: Checking server name...................................... Pass
Data Movers
: Checking unique id........................................ Pass
Data Movers
: Checking CIFS file server configuration................... N/A
Data Movers
: Checking domain controller connectivity and configuration. N/A
Data Movers
: Checking DNS connectivity and configuration............... N/A
Data Movers
: Checking connectivity to WINS servers..................... N/A
Data Movers
: Checking connectivity to NTP servers...................... N/A
Data Movers
: Checking connectivity to NIS servers...................... Pass
Data Movers
: Checking virus checker server configuration............... N/A
Data Movers
: Checking if workpart is OK................................ Pass
Data Movers
: Checking if free full dump is available...................
?
Data Movers
: Checking if each primary data mover has standby........... Fail
Storage System : Checking disk emulation type.............................. Pass
Storage System : Checking disk high availability access.................... Pass
Storage System : Checking disks read cache enabled......................... Pass
Storage System : Checking disks and storage processors write cache enabled. Pass
Storage System : Checking if access logix is enabled....................... Pass
Storage System : Checking if FLARE is committed............................ Pass
Storage System : Checking if FLARE is supported............................ Pass
Storage System : Checking if microcode is supported........................ Pass
Storage System : Checking no disks or storage processors are failed over... Pass
Storage System : Checking that no disks or storage processors are faulted.. Pass
Storage System : Checking that no hot spares are in use.................... Pass
Storage System : Checking that no hot spares are rebuilding................ Pass
Storage System : Checking minimum control lun size......................... Pass
Storage System : Checking maximum control lun size......................... Fail
Storage System : Checking system lun configuration......................... Pass
Storage System : Checking if storage processors are read cache enabled..... Pass
Storage System : Checking if auto assign are disabled for all luns......... Pass
Storage System : Checking if auto trespass are disabled for all luns....... Pass
Storage System : Checking backend connectivity............................. Pass
-------------------------------------------------------------------------------One or more warnings are shown below. It is recommended that you follow the
instructions below to correct the problem then try again.
-----------------------------------Information---------------------------------Control Station: Check ups is available
Symptom: The following UPS emcnasUPS_i0 emcnasUPS_i1 is(are)
not available
Data Movers: Check if each primary data mover has standby
Symptom: The following primary Data Movers server_2, server_3 does
not have a standby Data Mover configured. It is recommended that each
primary Data Mover have a standby configured for it with automatic
failover policy for high availability.
Storage System: Check maximum control lun size
Symptom:
nas_checkup
157
The nas Commands
* The size of control LUN 5 is 32 GB. It is larger than the
recommended size of 14 GB. The additional space will be reserved by
the system.
-------------------------------------------------------------------------------------------------------------------Warnings-----------------------------------Data Movers: Check if free full dump is available
Symptom: Cannot get workpart structure. Command failed.
* Command: /nas/sbin/workpart -r
* Command output: open: Permission denied
* Command exit code: 2
Action : Contact EMC Customer Service and refer to EMC Knowledgebase
emc146016. Include this log with your support request.
--------------------------------------------------------------------------------
EXAMPLE #2
Check Version:
Check Command:
To display help for nas_checkup, type:
$ nas_checkup -help
5.6.23.1
/nas/bin/nas_checkup
usage: nas_checkup
[ -help | -version ]
EXAMPLE #3
Check Version:
Check Command:
To display the version of nas_checkup utility, type:
$ nas_checkup -version
5.6.23.1
/nas/bin/nas_checkup
DIAGNOSTICS
nas_checkup returns one of the following exit statuses:
◆
0 — No problems found
◆
1 — nas_checkup posted information
◆
2 — nas_checkup discovered a warning
◆
3 — nas_checkup discovered an error
◆
255 — Any other error
Examples of errors that could cause a 255 exit status include, but
are not limited to:
• If nas_checkup is run when another instance of nas_checkup
is running
• If nas_checkup is run by someone other than root or the
administrator group (generally nasadmin)
• If nas_checkup is run on the standby Control Station
158
Command Line Interface Reference for File 8.1
The nas Commands
nas_ckpt_schedule
Manages SnapSure checkpoint scheduling for the VNX.
SYNOPSIS
nas_ckpt_schedule
-list
| -info {-all|<name>|id=<id>}
| -create <name>
-filesystem {<name>|id=<id>} [-description <description>]
-recurrence {
once [-start_on <YYYY-MM-DD>] -runtimes <HH:MM>
[-ckpt_name <ckpt_name>]
| daily [-every <number_of_days>]
[-start_on <YYYY-MM-DD>][-end_on <YYYY-MM-DD>]
-runtimes <HH:MM>[,...]
{-keep <number_of_ckpts>|-ckpt_names <ckpt_name>[,...]}
| weekly [-every <number_of_weeks>]
-days_of_week {Mon|Tue|Wed|Thu|Fri|Sat|Sun}[,...]
[-start_on <YYYY-MM-DD>][-end_on <YYYY-MM-DD>]
-runtimes <HH:MM>[,...]
{-keep <number_of_ckpts>|-ckpt_names <ckpt_name>[,...]}
| monthly [-every <number_of_months>] -days_of_month <1-31>[,...]
[-start_on <YYYY-MM-DD>][-end_on <YYYY-MM-DD>]
-runtimes <HH:MM>[,...]
{-keep <number_of_ckpts>|-ckpt_names <ckpt_name>[,...]}}
[{-cvfsname_prefix <prefix>|-time_based_cvfsname }]
| -modify {<name>|id=<id>}
[-name <new_name>]
[{-cvfsname_prefix <prefix>| -time_based_cvfsname}]
[-description <description>]
[-recurrence {daily|weekly|monthly}]
[-every {number_of_days|number_of_weeks|<number_of_months}]
[-days_of_week {Mon|Tue|Wed|Thu|Fri|Sat|Sun}[,...]]
[-days_of_month <1-31>[,...]][ -start_on <YYYY-MM-DD>]
[-end_on <YYYY-MM-DD>][ -runtimes <HH:MM>[,...]]
| -delete {<name>|id=<id>}
| -pause {<name>|id=<id>}
| -resume {<name>|id=<id>}
DESCRIPTION
OPTIONS
nas_ckpt_schedule creates and lists the schedules for the SnapSure
checkpoints. Schedules can be run once, daily, weekly, or monthly
and can be modified, paused, resumed, and deleted.
-list
Lists all checkpoint schedules on the system, the name of the
schedule, the next run date, the state, and the description.
nas_ckpt_schedule
159
The nas Commands
-info {-all|<name>|id=<id>}
Lists detailed information for all schedules or for the specified
schedule.
-create <name> -filesystem {<name>|id=<id>}
[-description <description>] -recurrence {
Creates a checkpoint schedule for the file system that is specified by
<name> or <id>. The schedule name in -create <name> must be
unique. The -description option provides a label for the schedule.
The -recurrence option specifies if the checkpoint operation occurs
once, daily, weekly, or monthly.
Note: It is recommended that a time interval of at least 15 minutes in between
the creation of two checkpoints on the same production file system. Using
VNX SnapSure provides information on checkpoint scheduling.
once [-start_on <YYYY-MM-DD>] -runtime <HH:MM>
[-ckpt_name <ckpt_name>]
If once is specified, the hours and minutes for the snapshot to be
run must be specified. A start date and name may be optionally
assigned to the checkpoint.
For a one-time checkpoint schedule, only one runtime can be
provided. For one-time schedules, the option -ckpt_name can
specify a name for the single checkpoint; if omitted, the default
naming is used (<schedule_name>_<fs_name>_<num>) where
<num> is a four digit integer beginning with 0001.
|daily [-every <number_of_days>] [-start_on
<YYYY-MM-DD>][-end_on <YYYY-MM-DD>]
-runtimes <HH:MM>[,...]
{-keep <number_of_ckpts>|-ckpt_names
<ckpt_name>[,...]}
If daily is specified, the checkpoint is taken every day unless
-every is specified indicating the number of days between runs.
The -start_on option indicates the day when the checkpoints will
start and -end_on indicates the day when they end.
The -runtimes option specifies one or more times to take a
checkpoint on each scheduled day. The -keep option specifies the
maximum number of checkpoints to be kept at any one time
(using default checkpoint naming). <number_of_ckpts> should
be equal to the number of checkpoint names specified for a
schedule. The -ckpt_name option assigns one or more specific
names to each checkpoint as it is taken.
160
Command Line Interface Reference for File 8.1
The nas Commands
|weekly [-every <number_of_weeks>] -days_of_week
{Mon|Tue|Wed|Thu|Fri|Sat|Sun} [,...][-start_on
<YYYY-MM-DD>][-end_on <YYYY-MM-DD>]-runtimes
<HH:MM>[,...]
{-keep <number_of_ckpts>|-ckpt_names
<ckpt_name>[,...]}
If weekly is specified, the checkpoint is taken every week unless
the -every option is specified indicating the number of weeks
between runs. The -days_of_week option specifies one or more
days during the week on which to run the schedule. The -start_on
option indicates the day when the checkpoints will start and
-end_on indicates the day when they end.
The -runtimes option specifies one or more times to take a
checkpoint on each scheduled day. The -keep option specifies the
maximum number of checkpoints to be kept at any one time
(using default checkpoint naming). The -ckpt_name option
assigns one or more specific names to each checkpoint as it is
taken.
|monthly [-every <number_of_months>] -days_of_month
<1-31>[,...][-start_on <YYYY-MM-DD>][-end_on
<YYYY-MM-DD>]-runtimes <HH:MM>[,...]
{-keep <number_of_ckpts> |-ckpt_names <ckpt_name>[,...]}}
If monthly is specified, the checkpoint is taken every month
unless the -every is specified indicating the number of months
between runs. The -days_of_month option specifies one or more
days during the month on which to run the schedule. <days> is
specified as an integer 1 through 31. The -start_on option
indicates the day when the checkpoints will start and -end_on
indicates the day when they end.
The -runtimes option specifies one or more times to take a
checkpoint on each scheduled day. The -keep option specifies
either the maximum number of checkpoints to be kept at any one
time (using default checkpoint naming) or using the -ckpt_name
option, one or more specific names to assign each checkpoint as it
is taken.
Note: The schedule that is set takes effect immediately unless -start_on is
specified. Daily, weekly, and monthly schedules run indefinitely unless
-end_on is included.
The -cvfsname_prefix option specifies the customized prefix of a
CVFS name. This prefix along with the cvfsname_delimiter and
nas_ckpt_schedule
161
The nas Commands
the cvfs_starting_index make up the CVFS name. The
-time_based_cvfsname option specifies the CVFS name based on
the creation time of the CVFS. It is the default method for
generating CVFS names and will be used if the prefix is not
specified.
Note: The prefix must be a PFS-wide unique string and can contain up to
20 ASCII characters. The prefix must not include intervening spaces,
colons (:), or slashes (/).
-modify {<name>|id=<id>} [-name <new_name>]
[{-cvfsname_prefix <prefix>| -time_based_cvfsname}]
[-description <description>] [-recurrence
{daily|weekly|monthly}] [-every <number_of_days>|
<number_of_weeks>|<number_of_months>]
[-days_of_week {Mon|Tue|Wed|Thu|Fri|Sat|Sun}[,...]]
[-days_of_month <1-31>[,...]][-start_on <YYYY-MM-DD>]
[-end_on <YYYY-MM-DD> ][ -runtimes <HH:MM>[,...]]
Modifies the scheduled checkpoint entry as specified.
-delete {<name>|id=<id>}
Deletes the specified checkpoint schedule. This operation does not
delete any checkpoints.
-pause {<name>|id=<id>}
Pauses the specified checkpoint schedule, including checkpoint
creations.
-resume {<name>|id=<id>}
Resumes a paused checkpoint schedule.
162
Command Line Interface Reference for File 8.1
The nas Commands
SEE ALSO
EXAMPLE #1
Using VNX SnapSure.
To create a checkpoint schedule that creates a checkpoint of the file
system ufs1 daily at 8 a.m. and 8 p.m. starting on 11/13/06 with the
last run on 12/13/07, and keep 7 checkpoints, type:
$ nas_ckpt_schedule -create ufs1_ckpt_sched1 -filesystem
ufs1 -description "Daily
Checkpoint schedule for ufs1" -recurrence daily -every 1
-start_on 2006-11-13
-end_on 2007-12-13 -runtimes 8:00,20:00 -keep 7
-cvfsname_prefix daily
This command returns no output.
EXAMPLE #2
To create a checkpoint schedule that creates a checkpoint of the file
system ufs1 weekly on Mondays at 6 p.m., starting on 11/13/06 with
the last run on 12/13/07, and name new checkpoints
ufs1_ckpt_mon1, ufs1_ckpt_mon2, ufs1_ckpt_mon3,
ufs1_ckpt_mon4, type:
$ nas_ckpt_schedule -create ufs1_ckpt_sched2 -filesystem
ufs1 -description "Weekly Checkpoint schedule for ufs1"
-recurrence weekly -every 1 -days_of_week Mon -start_on
2006-11-13 -end_on 2007-12-13 -runtimes 18:00 -ckpt_names
ufs1_ckpt_mon1,ufs1_ckpt_mon2,ufs1_ckpt_mon3,ufs1_ckpt_m
on4 -cvfsname_prefix weekly
This command returns no output.
EXAMPLE #3
To create a checkpoint schedule that creates a checkpoint of the file
system ufs1 every other 15th of the month at 7 p.m., and keep 12
checkpoints, type:
$ nas_ckpt_schedule -create ufs1_ckpt_sched3 -filesystem
ufs1 -description
"Monthly Checkpoint schedule for ufs1" -recurrence
monthly -every 2 -days_of_month
15 -runtimes 19:00 -keep 12 -cvfsname_prefix monthly
This command returns no output.
EXAMPLE #4
To create a checkpoint schedule that creates a checkpoint of the file
system ufs1 once at 3:09 p.m., type:
$ nas_ckpt_schedule -create ufs1_ckpt_sched4 -filesystem
ufs1 -description
nas_ckpt_schedule
163
The nas Commands
"One-time Checkpoint Schedule for ufs1" -recurrence once
-runtimes 15:09
This command returns no output.
EXAMPLE #5
To list all checkpoint schedules, type:
id
name
description
state
next run
=
=
=
=
=
$ nas_ckpt_schedule -list
6
ufs1_ckpt_sched2
Weekly Checkpoint schedule for ufs1
Pending
Mon Nov 13 18:00:00 EST 2006
id
name
description
state
next run
=
=
=
=
=
80
ufs1_ckpt_sched4
One-time Checkpoint Schedule for ufs1
Pending
Tue Nov 14 15:09:00 EST 2006
id
name
description
state
next run
=
=
=
=
=
5
ufs1_ckpt_sched1
Daily Checkpoint schedule for ufs1
Pending
Mon Nov 13 20:00:00 EST 2006
id
name
description
state
next run
=
=
=
=
=
7
ufs1_ckpt_sched3
Monthly Checkpoint schedule for ufs1
Pending
Wed Nov 15 19:00:00 EST 2006
EXAMPLE #6
To modify the recurrence of the checkpoint schedule
ufs1_ckpt_sched3 to run every 10th of the month, type:
$ nas_ckpt_schedule -modify ufs1_ckpt_sched3 -recurrence
monthly -every 1 -days_of_month 10
This command returns no output.
EXAMPLE #7
To get detailed information about checkpoint schedule, type:
$ nas_ckpt_schedule -info ufs1_ckpt_sched3
id = 7
name = ufs1_ckpt_sched3
description = Monthly Checkpoint schedule for ufs1
CVFS name prefix = monthly
tasks = Checkpoint ckpt_ufs1_ckpt_sched3_001 on filesystem id=25, Checkpoint
ckpt_ufs1_ckpt_sched3_002 on filesystem id=25, Checkpoint
ckpt_ufs1_ckpt_sched3_003 on filesystem id=25, Checkpoint
164
Command Line Interface Reference for File 8.1
The nas Commands
ckpt_ufs1_ckpt_sched3_004 on filesystem
ckpt_ufs1_ckpt_sched3_005 on filesystem
ckpt_ufs1_ckpt_sched3_006 on filesystem
ckpt_ufs1_ckpt_sched3_007 on filesystem
ckpt_ufs1_ckpt_sched3_008 on filesystem
ckpt_ufs1_ckpt_sched3_009 on filesystem
ckpt_ufs1_ckpt_sched3_010 on filesystem
ckpt_ufs1_ckpt_sched3_011 on filesystem
ckpt_ufs1_ckpt_sched3_012 on filesystem
next run = Sun Dec 10 19:00:00 EST 2006
state = Pending
recurrence = every 1 months
start on = Mon Nov 13 16:47:51 EST 2006
end on =
at which times = 19:00
on which days of week =
on which days of month = 10
EXAMPLE #8
id=25,
id=25,
id=25,
id=25,
id=25,
id=25,
id=25,
id=25,
id=25
Checkpoint
Checkpoint
Checkpoint
Checkpoint
Checkpoint
Checkpoint
Checkpoint
Checkpoint
To pause a checkpoint schedule, type:
$ nas_ckpt_schedule -pause ufs1_ckpt_sched1
This command returns no output.
EXAMPLE #9
To resume a checkpoint schedule, type:
$ nas_ckpt_schedule -resume ufs1_ckpt_sched1
This command returns no output.
EXAMPLE #10
To delete a checkpoint schedule, type:
$ nas_ckpt_schedule -delete ufs1_ckpt_sched2
This command returns no output.
nas_ckpt_schedule
165
The nas Commands
nas_connecthome
Configures email, FTP, modem, HTTPS and ESRS transport
mechanisms for transporting Callhome event files to user-configured
destinations.
SYNOPSIS
nas_connecthome
-info
| -test {-email_1|-email_2|-ftp_1|-ftp_2|-modem_1|-modem_2|-https|-esrs}
| -modify [-modem_priority {Disabled|1|2|3}]
[-modem_number <phone_number>]
[-modem_number_2 <phone_number>]
[-ftp_priority {Disabled|1|2|3}]
[-ftp_server {<hostname>|<ip_addr>}]
[-ftp_port <port>]
[-ftp_user <username>]
[-ftp_passwd [<passwd>]]
[-ftp_folder <path>]
[-ftp_ipprotocol {IPv4|IPv6}]
[-ftp_mode {active|passive}]
[-ftp_server_2 {<hostname>|<ip_addr>}]
[-ftp_port_2 <port>]
[-ftp_user_2 <username>]
[-ftp_passwd_2 [<passwd>]]
[-ftp_folder_2 <path>]
[-ftp_ipprotocol_2 {IPv4|IPv6}]
[-ftp_mode_2 {active|passive}]
[-email_priority {Disabled|1|2|3}]
[-email_from <email_addr>]
[-email_to {<email_addr>[,<email_addr>]}
[-email_subject <email_subject>]
[-email_server {<hostname>|<ip_addr>}]
[-email_ipprotocol {IPv4|IPv6}]
[-email_server_2 {<hostname>|<ip_addr>}]
[-email_ipprotocol_2 {IPV4|IPV6}]
[-esrs_priority {Disabled|1|2|3}]
[ -https_priority {Disabled|1|2|3}]
[ -https_url {url}
[ -https_ipprotocol {IPv4|IPv6}
[-dial_in_number <phone_number>]
[-serial_number <serial_number>]
[-site_id <site_id>]
[-encryption_enabled {yes|no}]
[-dial_in_enabled {yes|no}]
[-service_info]
| -help
166
Command Line Interface Reference for File 8.1
The nas Commands
DESCRIPTION
nas_connecthome pauses and resumes the ConnectHome service,
displays and configures parameters for email, FTP, modem, HTTPS,
and ESRS, which are mechanisms used for transmitting event files.
nas_connecthome enables a user to configure primary and optional
secondary destinations for each transport mechanism.
nas_connecthome also tests connectivity to the destination
configured for a transport mechanism.
This command must be executed from /nas/sbin/.
OPTIONS
-info
Displays the enabled and disabled configuration parameters for all
transport mechanisms:
-test
{-email_1|-email_2|-ftp_1|-ftp_2|-modem_1|-modem_2|-https|esrs}
Tests connectivity to the destination configured and enabled for the
specified transport mechanism.
-modify
Modifies the following configuration parameters for any or all
transport mechanisms:
[-modem_priority {Disabled|1|2|3}]
Enables modem as a Primary, Secondary, or Tertiary transport
mechanism. Specifying Disabled removes modem as a transport
mechanism.
[-modem_number <phone_number>]
Sets or modifies the primary phone number of the modem.
Note: Specifying "" (empty double quotes) disables the use of the existing
phone number.
[-modem_number_2 <phone_number>]
Sets or modifies the secondary phone number of the modem.
Note: Specifying "" (empty double quotes) disables the use of the existing
phone number for this transport mechanism.
[-ftp_priority {Disabled|1|2|3}]
Enables FTP as a Primary, Secondary, or Tertiary transport
mechanism. Specifying Disabled removes FTP as a transport
mechanism.
nas_connecthome
167
The nas Commands
[-ftp_server {<hostname>|<ip_addr>}]
Sets or modifies the hostname or IP address of the primary FTP
server and corresponding port. The allowable input is IPv4
address, IPv6 address, or domain name.
[-ftp_port <port>]
Sets or modifies the port of the primary FTP server and
corresponding port. The valid input is an integer between 1 and
65535. If an empty string " " is provided for this option, the port
number is reset to the default value 21.
[-ftp_user <username>]
Sets or modifies the username of the login account on the primary
FTP server.
Note: Specifying "" (empty double quotes) reverts to the default value of
onalert.
[-ftp_passwd [<passwd>]]
Sets or modifies the password of the login account on the primary
FTP server.
Note: Specifying "" (empty double quotes) reverts to the default value of
EMCCONNECT.
[-ftp_folder <path>]
Sets or modifies the path to the folder on the primary FTP server
where the event files have to be deposited.
Note: Specifying "" (empty double quotes) reverts to the default value of
incoming.
[-ftp_ipprotocol {IPV4|IPV6}]
Sets or modifies the transfer mode of the primary FTP transport
mechanism. If an IPv4 address is provided to FTP server, the
corresponding IP protocol is changed to IPv4 automatically. If an
IPv6 address is used, the IP protocol is changed to IPv6. When
hostname is specified, no IP protocol change is made.
[-ftp_mode {active|passive}]
Sets or modifies the transfer mode of the primary FTP transport
mechanism.
168
Command Line Interface Reference for File 8.1
The nas Commands
Note: Specifying "" (empty double quotes) reverts to the default value of
active.
[-ftp_server_2 <hostname>[<ip_addr>]]
Sets or modifies the hostname or IP address of the secondary FTP
server and corresponding port. The allowable input is IPv4
address, IPv6 address, or domain name.
[-ftp_port_2 <port>]
Sets or modifies the port of the secondary FTP server and
corresponding port. The valid input is an integer between 1 and
65535. If an empty string “” is provided for this option, the port
number is reset to the default value of 21.
[-ftp_user_2 <username>]
Sets or modifies the username of the login account on the
secondary FTP server.
Note: Specifying "" (empty double quotes) reverts to the default value of
onalert.
[-ftp_passwd_2 [<passwd>]]
Sets or modifies the password of the login account on the
secondary FTP server.
Note: Specifying "" (empty double quotes) reverts to the default value of
EMCCONNECT.
[-ftp_folder_2 <path>]
Sets or modifies the path of the folder on the secondary FTP
server where the event files have to be deposited.
Note: Specifying "" (empty double quotes) reverts to the default value of
incoming.
[-ftp_ipprotocol_2 {IPv4|IPv6}]
Sets or modifies the transfer mode of the secondary FTP transport
mechanism.
[-ftp_mode_2 { active|passive}]
Sets or modifies the transfer mode of the secondary FTP transport
mechanism.
nas_connecthome
169
The nas Commands
Note: Specifying "" (empty double quotes) reverts to the default value of
active.
[-email_priority {Disabled|1|2|3 }]
Enables email as a Primary, Secondary, or Tertiary transport
mechanism. Specifying Disabled removes email as a transport
mechanism.
[-email_from <email_addr>]
Sets or modifies the sender's email address. The maximum
number of characters that can be specified is 63.
Note: Specifying "" (empty double qoutes) reverts to the default value of
connectemc@emc.com.
[-email_to <email_addr>[,<email_addr> ]]
Sets or modifies the destination email addresses that receive the
event files. Multiple email addresses can be specified with a
comma separating each address. The maximum number of
characters that can be specified is 255.
Note: Specifying "" (empty double quotes) reverts to the default value of
emailalert@emc.com.
[-email_subject <email_subject>]
Sets or modifies the subject of the email message.
Note: Specifying "" (empty double quotes) reverts to the default value of
CallHome Alert.
[-email_server {<hostname>|<ip_addr>}]
Sets or modifies the primary email server that accepts and routes
email messages.
Note: Specifying "" (empty double quotes) disables the use of the existing
email server for this transport mechanism.
[-email_ipprotocol {IPv4|IPv6}]
Sets or modifies the secondary email server that accepts and
routes email messages.
170
Command Line Interface Reference for File 8.1
The nas Commands
[-email_server_2 {<hostname>|<ip_addr>}]
Sets or modifies the secondary email server that accepts and
routes email messages.
Note: Specifying "" (empty double quotes) disables the use of the existing
email server for this transport mechanism.
[-email_ipprotocol_2 {IPv4|IPv6}]
Sets or modifies the secondary email server that accepts and
routes email messages.
[-esrs_priority {Disabled|1|2|3}]
Enables ESRS as a Primary, Secondary, or Tertiary transport
mechanism. Specifying Disabled removes ESRS as a transport
mechanism.
[-https_priority {Disabled|1|2|3}]
Enables HTTPS as a Primary, Secondary, or Tertiary transport
mechanism. Specifying Disabled removes HTTPS as a transport
mechanism.
[-https_url]
Indicates URL of the monitoring station.
[-https_ipprotocol {IPv4|IPv6}]
Sets or modifies the transfer mode of the secondary HTTPS
transport mechanism.
[-dial_in_number <phone_number>]
Sets or modifies the dial-in phone number of the modem.
Note: Specifying "" (empty double quotes) does not disable the number
or restore its default value. The empty string is stored as is.
[-serial_number <serial_number>]
Sets or modifies the VNX serial number, if it was not
automatically detected.
Note: Specifying "" (empty double quotes) does not disable the number
or restore its default value. The empty string is stored as is.
[-site_id <site_id>]
Sets or modifies the site ID.
nas_connecthome
171
The nas Commands
Note: Specifying "" (empty double quotes) does not disable the number
or restore its default value. The empty string is stored as is.
[-encryption_enabled {yes|no}]
Enables or disables the encryption of the CallHome payload
during transmission.
Note: Specifying "" (empty double quotes) reverts to the default value of
yes.
[-dial_in_enabled {yes|no}]
Enables or disables dial-in login sessions.
Note: Specifying "" (empty double quotes) reverts to the default value of
yes.
SEE ALSO
EXAMPLE #1
Configuring Events and Notifications on VNX for File.
To display configuration information, type:
# /nas/sbin/nas_connecthome -info
ConnectHome Configuration:
Encryption Enabled
= yes
Dial In :
Enabled
= yes
Modem phone number
= 9123123123
Site ID
= MY SITE
Serial number
= APM00054703223
ESRS :
Priority
Email :
Priority
Sender Address
Recipient Address(es)
Subject
Primary :
Email Server
Secondary :
Email Server
FTP :
Priority
Primary :
FTP Server
FTP Port
172
= 1
=
=
=
=
1
admin@yourcompany.com
emailalert@emc.com
CallHome Alert
= backup.mailhub.company.com
=
= 2
= 1.2.3.4
= 22
Command Line Interface Reference for File 8.1
The nas Commands
FTP User Name
FTP Password
FTP Remote Folder
FTP Transfer Mode
Secondary :
FTP Server
FTP Port
FTP User Name
FTP Password
FTP Remote Folder
FTP Transfer Mode
Modem :
Priority
Primary :
Phone Number
BT Tymnet
Secondary :
Phone Number
BT Tymnet
EXAMPLE #2
=
=
=
=
onalert
**********
incoming
active
=
=
=
=
=
=
1.2.4.4
22
onalert
**********
incoming
active
= Disabled
=
= no
=
= no
To test the primary email server, type:
# /nas/sbin/nas_connecthome -test -email_1
--------------------------------------------------------ConnectEMC 2.0.27-bl18 Wed Aug 22 10:24:32 EDT 2007
RSC API Version: 2.0.27-bl18
Copyright (C) EMC Corporation 2003-2007, all rights reserved.
--------------------------------------------------------Reading configuration file: ConnectEMC.ini.
Run Service begin...
Test succeeded for Primary Email.
EXAMPLE #3
To modify the configuration information, type:
# /nas/sbin/nas_connecthome -modify -esrs_priority 1
--------------------------------------------------------ConnectEMC 2.0.27-bl18 Wed Aug 22 10:24:32 EDT 2007
RSC API Version: 2.0.27-bl18
Copyright (C) EMC Corporation 2003-2007, all rights reserved.
--------------------------------------------------------Reading configuration file: ConnectEMC.ini.
Run Service begin...
Modify succeeded for Primary ESRS.
nas_connecthome
173
The nas Commands
nas_config
Manages a variety of configuration settings on the Control Station,
some of which are security based.
SYNOPSIS
nas_config
-IPalias {-list
| -create [-name <device_name>] <numeric_id>
| -delete [-name <device_name>] <numeric_id>}
| -ssl
| -sessiontimeout [<number_in_minutes>|off]
| -password [-min <6..15>] [-retries <max_allowed>] [-newchars <min_num>]
[-digits <min_num>] [-spechars <min_num>] [-lcase <min_num>] [-ucase
<min_num>]
| -password -default
DESCRIPTION
nas_config -IPalias configures different IP addresses to point to the
same network device allowing use of IP aliasing to manage the
Control Station. This enables communication with the primary
Control Station using a single IP address regardless of whether the
primary Control Station is running in slot 0 or slot 1.
nas_config -ssl generates an X.509 digital certificate on the Control
Station. Unisphere uses the Secure Sockets Layer (SSL) protocol to
create a secure connection between a user’s Web browser and the
Control Station’s Apache Web server. When a VNX system is initially
installed, a generic certificate is generated. After configuring the
Control Station’s network configuration (hostname, DNS domain
name, or IP address) and before using the Unisphere, a new
certificate should be generated.
nas_config -sessiontimeout sets a system-wide value that
automatically times out a Control Station shell session after a
specified period of inactivity.
nas_config -password supports a password quality policy by
requiring that passwords chosen by VNX users adhere to certain
rules.
You must be root to execute this command from the /nas/sbin
directory.
174
Command Line Interface Reference for File 8.1
The nas Commands
OPTIONS
-IPalias -list
Lists IP aliases configured on the VNX.
-IPalias -create [-name <device_name>] <numeric_id>
Creates an IP alias for the Control Station.
<device_name> is the name for a specified device:
◆
If you specify a device name that device must have an IP address.
◆
If you do not specify a device name, the system uses the external
network interface.
<numeric_id> is a user-defined number, and can be an integer
between 0 and 255. The system allows up to 256 aliases for any
device.
-delete [-name <device_name>] <numeric_id>
Deletes an IP alias for the Control Station.
-ssl
Installs a SSL certificate on the Control Station and restarts the HTTP
server.
-sessiontimeout [<number_in_minutes>|off]
Displays the current session timeout value in minutes.
<number_in_minutes> sets the number of minutes a Control Station
shell sesssion can be inactive before it is timed out. Possible values
are 5 to 240 minutes. The default value is 60 minutes. Session timeout
is enabled by default. To disable session timeout, type off or 0 to
indicate zero minutes.
The -sessiontimeout option enables the native timeout properties of
the underlying shells on the Control Station. The relevant shell man
page provides a description of how the mechanism works.
-password
Prompts for specific password policy definitions. The current value
for each policy definition is shown in brackets.
[-min <6..15>] defines the minimum length of the new
password. The default length is eight characters. The length has
to be a value between 6 and 15 characters.
[-retries <max_allowed>] defines the number of attempts a
user can make to define an acceptable new password before the
command fails. The default value is three attempts.
nas_config
175
The nas Commands
[-newchars <min_num>] defines the minimum number of
characters that must be in the new password that were not
included in the old password. The default value is three
characters.
[-digits <min_num>] defines the minimum number of digits
that must be included in the new password. The default value is
one digit.
[-spechars <min_num>] defines the minimum number of
special characters (such as ! @ # $ % & ^ and *) that must be
included in the new password. The default value is 0.
[-lcase <min_num>] defines the minimum number of
lowercase characters that must be included in the new password.
The default value is 0.
[-ucase <min_num>] defines the minimum number of uppercase
characters that must be included in the new password. The
default value is 0.
-password -default
Resets the password policy definitions to their default values.
SEE ALSO
EXAMPLE #1
Do you
Please
Do you
Please
want slot_0
enter an IP
want slot_0
enter a new
Security Configuration Guide for File.
To create an IP alias for the Control Station, type:
# /nas/sbin/nas_config -IPalias -create 0
IP address <1.2.3.4> as your alias [yes or no]: no
address to use as an alias: 1.2.3.6
IP address <1.2.3.4> as your alias [yes or no]: yes
IP address for slot_0: 1.2.3.6
EXAMPLE #2
To view the IP alias that you created, type:
# /nas/sbin/nas_config -IPalias -list
alias IPaddress state
eth2:0 1.2.3.6 UP
EXAMPLE #3
To delete an IP alias, type:
# /nas/sbin/nas_config -IPalias -delete 0
All current sessions using alias eth2:0 will terminate
Do you want to continue [yes or no]: yes
done
176
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #4
To generate and install a certificate for the Apache Web server on the
Control Station, type:
# /nas/sbin/nas_config -ssl
Installing a new SSL certificate requires restarting the Apache web server.
Do you want to proceed? [y/n]: y
New SSL certificate has been generated and installed successfully.
EXAMPLE #5
To change the session timeout value from the default value of 60
minutes to 100 minutes, type:
# /nas/sbin/nas_config -sessiontimeout 100
done
EXAMPLE #6
To disable session timeout, type:
# /nas/sbin/nas_config -sessiontimeout 0
done
or
# /nas/sbin/nas_config -sessiontimeout off
done
EXAMPLE #7
To set specific password policy definitions, type:
# /nas/sbin/nas_config -password
Minimum length for a new password (Between 6 and 15): [8]
Number of attempts to allow before failing: [3]
Number of new characters (not in the old password): [3]
Number of digits that must be in the new password: [1]
Number of special characters that must be in a new password: [0]
Number of lower case characters that must be in password: [0]
Number of upper case characters that must be in password: [0]
EXAMPLE #8
To set the minimum length of a new password to 10 characters, type:
# /nas/sbin/nas_config -password -min 10
EXAMPLE #9
To reset the current password policy definitions to their default
values, type:
# /nas/sbin/nas_config -password -default
nas_config
177
The nas Commands
nas_copy
Creates a replication session for a one-time copy of a file system. This
command is available with VNX Replicator only.
SYNOPSIS
nas_copy
-name <sessionName>
-source
{-fs {<name>|id=<fsId>}|-ckpt {<ckptName>|id=<ckptId>}
-destination
{-fs {id=<dstFsId>|<existing_dstFsName>}
|-pool {id=<dstStoragePoolId>}|<dstStoragePool>}
[-storageSystem <dstStorageSerialNumber>]}
[-from_base {<ckpt_name>|id=<ckptId>}]
-interconnect {<name>|id=<interConnectId>}
[-source_interface {<nameServiceInterfaceName>|ip=<ipaddr>}]
[-destination_interface {<nameServiceInterfaceName>|ip=<ipaddr>}]
[-overwrite_destination]
[-refresh]
[-full_copy]
[-background]
DESCRIPTION
nas_copy from the Control Station on the source side performs a
one-time copy of a source read-only file system or a checkpoint file
system.
Note: Depending on the size of the data in the source, this command may
take some time to complete. Once a copy session begins, you can monitor it
or interrupt it if necessary using the nas_task command. You can list all
replication sessions, including copy sessions, using the nas_replicate -list
command.
OPTIONS
-name <sessionName> -source -fs {<name>|id=<fsId>|-ckpt {
<ckptName>|id=<ckptId>} -destination {-fs
{<existing_dstFsName>| id=<dstFsId>[-pool
<dstStoragePool>| id=<dstStoragePoolId>}] [-from_base
{<ckpt_Name>| id=<ckptId>}] -interconnect {<name>|id=
<interConnectId>}
Creates a VNX Replicator session that performs a one-time copy of a
source read-only file system or a checkpoint file system.
The session name assigned must be unique for the Data Mover pair
as defined by the interconnect. The naming convention
<source_fs_or_ckpt_name>_replica<#> is used if a read-only file
system or checkpoint at the destination already has the same name as
178
Command Line Interface Reference for File 8.1
The nas Commands
the source. An integer between 1 and 4 is assigned according to how
many replicas of that file system or checkpoint already exist.
The -source specifies the name or ID of an existing read-only file
system or checkpoint file system as the source for this copy session.
This is to be used as a common base for the initial transfer. The
checkpoint is identified by checkpoint name or checkpoint file system
ID. This option is intended to accommodate upgrade situations to
VNX Replicator.
The -destination specifies either an existing destination file system or
the storage needed to create the destination file system automatically,
as part of the copy operation. An existing destination file system
must be read-only and the same size as the source. Specifying a
storage pool or ID creates the destination file system automatically, as
read-only, using the same name and size as the source file system.
[-storageSystem <dstStorageSerialNumber>]
When the destination file system is to be created from a pool, it
specifies the system for the destination file system to reside. Use
the nas_storage -list command to obtain the serial number of the
system.
[-from_base {ckpt_name>|id=<ckptId>}]
Specifies an existing source file system checkpoint to be used as a
common base for the initial data transfer. The checkpoint is identified
by the checkpoint name or ID.
The -interconnect specifies the local (source) side of an established
Data Mover interconnect to use for this copy session. Use the nas_cel
-interconnect -list command on the source VNX to list the
interconnects available to VNX Replicator sessions.
[-source_interface {<nameServiceInterfaceName> |ip=
<ipAddr>}]
Instructs the copy session to use a specific local interface defined
for the interconnect on the source VNX instead of selecting the
local interface supporting the lowest number of sessions (the
default). If this local interface was defined for the interconnect
using a name service interface name, specify the name service
interface name; if it was defined using an IP address, specify the
IP address. The source_interfaces field of the output from the
nas_cel -interconnect -info command shows how the source
interface was defined. This option does not apply to a loopback
interconnect, which always uses 127.0.0.1.
nas_copy
179
The nas Commands
[-destination_interface
{<nameServiceInterfaceName>|ip=<ipaddr>}]
Instructs the copy session to use a specific interface defined for
the interconnect on the destination VNX instead of selecting the
peer interface supporting the lowest number of sessions (the
default). If this peer interface was defined for the interconnect
using a name service interface name, specify the name service
interface name; if it was defined using an IP address, specify the
IP address. The destination_interfaces field of the output from
the nas_cel -interconnect -info command shows how the peer
interface was defined. This option does not apply to a loopback
interconnect, which always uses 127.0.0.1.
[-overwrite_destination]
For an existing destination, discards any changes made to the
destination object and restores it from the established common
base (differential copy). If this option is not specified and an
existing destination object contains different content than the
established common base, an error is returned.
[-refresh {<name>|id=<session_id>}
Updates a destination checkpoint that has the same name as the
copied checkpoint. This option does not refresh the source object;
it refreshes only the destination for a existing checkpoint. If you
specify this option and no checkpoint exists with the same name,
the command returns an error.
[-full_copy]
For an existing destination object, if a common base checkpoint
exists, it performs a full copy of the source checkpoint to the
destination, instead of a differential copy. If this option is not
specified and an existing destination object has different content
than the established common base, an error is returned.
[-background]
Executes the command in an asynchronous mode. Use the
nas_task command to check the status of the command.
SEE ALSO
EXAMPLE #1
180
nas_cel, nas_replicate, nas_task.
To create a one-time copy of a checkpoint file system with session
name ufs1_replica1 with the source checkpoint ufs_ckpt1 and
destination pool clar_r5_performance on the interconnect
NYs3_LAs2, source interface 10.6.3.190, and destination interface
10.6.3.173, type:
Command Line Interface Reference for File 8.1
The nas Commands
$ nas_copy -name ufs1_replica1 -source -ckpt ufs1_ckpt1
-destination -pool clar_r5_performance -interconnect
NYs3_LAs2 -source_interface 10.6.3.190
-destination_interface 10.6.3.173
OK
EXAMPLE #2
To create a one-time copy of a read-only file system for the session
ufs1_replica1 with source file system ufs1 and overwrite an existing
destination file system ufs1 on the interconnect NYs3_LAs2, source
interface 10.6.3.190, and destination interface 10.6.3.173, type:
$ nas_copy -name ufs1_replica1 -source -fs ufs1
-destination -fs ufs1 -interconnect NYs3_LAs2
-source_interface 10.6.3.190 -destination_interface
10.6.3.173 -overwrite_destination
OK
EXAMPLE #3
To initiate a differential copy of ufs_ckpt2 to the ufs1_destination file
system using ufs1_ckpt1 as the common base, using the -from_base
option, type:
$ nas_copy -name ufs1_replica1 -source -ckpt -ufs1_ckpt2
-destination -fs ufs1_destination -from_base ufs1_ckpt1
-interconnect NYs3_LAs2
OK
!
CAUTION
Using the -from_base option overrides any common base that may
exist. Ensure that the specified checkpoint represents the correct
state of the destination file system.
EXAMPLE #4
To refresh the destination of the replication session ufs1_replica1 for
the source checkpoint ufs1_ckpt1 and destination file system ufs1 on
the interconnect NYs3_LAs2, type:
$ nas_copy -name ufs1_replica1 -source -ckpt ufs1_ckpt1
-destination -fs ufs1 -interconnect NYs3_LAs2 -refresh
OK
EXAMPLE #5
To perform a full copy of the source checkpoint to the destination for
the replication session ufs1_replica1 with the source file system ufs1
and destination file system ufs1 on the interconnect NYs3_LAs2,
type:
$ nas_copy -name ufs1_replica1 -source -fs ufs1
-destination -fs ufs1 -interconnect NYs3_LAs2
-overwrite_destination -full_copy -background
nas_copy
181
The nas Commands
Info 26843676673: In Progress: Operation is still running.
the Task Status screen for results.
182
Command Line Interface Reference for File 8.1
Check task id 4177 on
The nas Commands
nas_cs
Manages the configuration properties of the Control Station.
SYNOPSIS
nas_cs
-info [-timezones]
| -set [-hostname <hostname>]
[-ip4address <ipv4_address>]
[-ip4netmask <ipv4_netmask>]
[-ip4gateway <ipv4_gateway>]]
[-ip6address <ipv6_address[/prefix_length]>]
[-ip6gateway <ipv6_gateway>]
[-dns_domain <dns_domain_name>]
[-search_domain <domain_name>[,...]]
[-dns_servers <dns_server>[,...]]
[-session_monitor_timeout <days>]
[-session_idle_timeout <minutes>]
[-time <yyyymmddhhmm [ss]>]
[-timezone <time_zone_str>]
[-ntp_servers <ntp_server>[,...]]
| -clear [-ip4gateway]
[-ip6address]
[-ip6gateway]
[-dns]
[-search_domain]
[-session_monitor_timeout]
[-session_idle_timeout]
[-ntp_servers]
[-dns_domain]
[-dns_servers]
| -reboot
DESCRIPTION
OPTIONS
nas_cs sets, clears, and lists the Control Station configuration. nas_cs
can be used to reboot the Control Station.
-info [-timezones]
Displays the Control Station configuration. The -timezones option
displays all supported time zones on the Control Station.
-set [-hostname <hostname>]
Sets the user-configurable parameters of the Control Station
configuration. Sets the hostname of the primary Control Station. To
specify a hostname, the maximum number of characters is 64,
excluding white spaces and dot characters.
[-ip4address {<ipv4_address>]
Sets the IPv4 network address of the primary Control Station. The
IPv4 address must be a valid address.
nas_cs
183
The nas Commands
[-ip4netmask <ipv4_netmask>]
Sets the subnet mask for a valid IPv4 network address on the
primary Control Station.
[-ip4gateway <ipv4_gateway>]
Sets the IPv4 network address of the gateway machine for the
primary Control Station on the network. The IPv4 address must
be a valid address.
[-ip6address <ipv6_addr[/prefix_length]>]
Sets the IPv6 network address of the primary Control Station. The
IPv6 address must be a valid address. The /prefix_length option
sets the integer value, between 8 and 128, for the prefix length of
the IPv6 address of the primary Control Station.
[-ip6gateway <ipv6_gateway>]
Sets the IPv6 network address of the gateway machine for the
primary Control Station on the network. The IPv6 address must
be a valid address.
[-dns_domain <dns_domain_name>]
Sets the Domain Name System of which the primary Control
Station is a member. It can accept valid domain names.
[-search_domain <domain_name>[,...]]
Sets the order in which DNS domains are searched to resolve a
comma-separated list of valid domain names.
[-dns_servers <dns_server>[,...]]
Sets the IP addresses of the DNS servers of the domain. It is a
comma-separated list of valid IPv4 or IPv6 addresses and can
have multiple DNS addresses.
[-session_monitor_timeout <days>]
Sets the limit for the number of days until when a valid login is
allowed to run queries on the primary Control Station. Any active
management session requires a login on the primary Control
Station.
[-session_idle_timeout <minutes>]
Sets the timeout period in minutes for an inactive administrator
session to become invalid.
[-time <yyyymmddhhmm [ss]>]
Sets the current system date and time in the format
<yyyymmddhhmm [ss]>].
184
Command Line Interface Reference for File 8.1
The nas Commands
[-timezone <time_zone_str>]
Sets a valid time zone value on the primary Control Station.
[-ntp_servers <ntp_server>[,...]]
Sets the IP addresses of the NTP server used by the primary
Control Station. It is a comma-separated list of valid IPv4 or IPv6
addresses, FQDN or unqualified name of NTP server. Multiple
NTP servers can be specified.
-keyid <key_identifier> [,...]
Sets the NTP server authentication. It is a comma-separated list of
positive integers with range of 1-65534. It is an ordered list in the
same order as -ntp_servers. If authentication is not needed for a
particular NTP server, the keyid field for it should be 0 to indicate
that no key will be used. In that case, the corresponding keyvalue
should also be left empty. If keyid is entered with leading zeroes,
they are ignored.
-keyvalue <key_value> [,...]
Sets the NTP server authentication. It is a comma-separated list of
keys with length of each key between 1-16 characters. This key
value is obtained from the NTP server and is a message digest
key used to authenticate NTP server response with symmetric
key cryptography (refer RFC5905 and RFC 1321). For each keyid
specified for a NTP server, there should be corresponding
keyvalue. If NTP server authentication is not needed for a server,
keyid should be made 0 and keyvalue should be empty. Valid
keyvalues are ASCII characters with ASCII codes 0x20 to 0x7E
excluding the space character(ASCII code 0x20) and the #
character(ASCII code 0x23).
-clear [-ip4gateway]
Clears the user-configurable parameters of the Control Station
configuration. Clears the IPv4 network address of the gateway
machine for the primary Control Station on the network.
[-ip6address]
Clears the IPv6 network address of the primary Control Station.
[-ip6gateway]
Clears the IPv6 network address of the gateway machine for the
primary Control Station on the network.
[-dns]
Clears the Domain Name System of which the primary Control
Station is a member.
nas_cs
185
The nas Commands
[-search_domain]
Clears the order in which DNS domains are searched to resolve a
list of valid domain names.
[-session_monitor_timeout]
Clears the limit for the number of days until when a valid login is
allowed to run queries on the primary Control Station. Any active
management session requires a login on the primary Control
Station.
[-session_idle_timeout]
Clears the timeout period in minutes for an inactive administrator
session to become invalid.
[-ntp_servers]
Clears the IP addresses of the NTP server used by the primary
Control Station.
[-dns_domain]
Clears the domain name of which the Control Station is a
member. This option will clear only the DNS domain name
unlike option -clear -dns which clears both DNS domain and
DNS servers.
[-dns_servers]
Clears the IP adresses of the DNS servers of the domain. This
option will clear only the DNS servers unlike option -clear -dns
which clears both DNS domain and DNS servers.
-reboot
Reboots the primary Control Station.
186
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #1
To display the configuration properties of the primary Control
Station, type:
$ nas_cs -i
Host Name
= navinas21381
Version
= 7.1.9-3
Location
= system:VNX5500:FNM001103001862007|controlStation::0
Status
= OK
Standby Location
=
Standby Status
=
IPv4 Address
= 10.244.213.81
IPv4 Netmask
= 255.255.255.0
IPv4 Gateway
= 10.244.213.1
IPv6 Address
=
IPv6 Gateway
=
DNS Domain
= clrcase.lab.emc.com
DNS Domain search order =
DNS Servers
= 10.244.210.18,10.244.163.18
Session Monitor Timeout = 0 Days
Session Idle Timeout
= 0 Minutes
NTP Servers (Key,Value) = 2001:db8:cafe:1111::246 (5,AbCd)
1.2.3.4 (3,xYz)
ntp.lab.emc.com (1,s3Cr3t)
System Time
= Mon Sep 05 05:46:48 EDT 2011
EXAMPLE #2
To set the hostname, IPv4 network address, subnet mask, and IPv4
gateway for the primary Control Station, type:
$ nas_cs -set -hostname eng24416 -ip4address
172.24.244.16 -ip4netmask 255.255.255.0 -ip4gateway
128.221.252.0
OK
EXAMPLE #3
To set the IPv6 address and the IPv6 gateway for the primary Control
Station, type:
$ nas_cs -set -ip6address
2002:ac18:af02:f4:20e:cff:fe6e:d524/64 -ip6gateway
2002:ac18:af02:f4:20e:cff:fe6e:d527
OK
EXAMPLE #4
To set the DNS domain, search domains, and DNS servers for the
primary Control Station, type:
$ nas_cs -set -dns_domain eng.lss.emc.com -search_domain
lss.emc.com,rtp.lab.emc.com -dns_servers
172.24.175.172,172.24.175.173
OK
nas_cs
187
The nas Commands
EXAMPLE # 5
To set the DNS server with IPv6 address, type:
$ nas_cs -set -dns_servers
2002:ac18:af02:f4:20e:cff:fe6e:d524,2002:ac18:af02:f4:
20e:cff:fe6e:d345,2002:ac18:af02:f4:20e:cff:fe6e:c213
OK
EXAMPLE #6
To set NTP server without authentication, type:
$ nas_cs -set -ntp_servers
2002:ac18:af02:f4:20e:cff:fe6e:d524,2002:ac18:af02:f4:
20e:cff:fe6e:d345,2002:ac18:af02:f4:20e:cff:fe6e:c213
OK
EXAMPLE #7
To set the session monitor timeout and session idle timeout for the
primary Control Station, type:
$ nas_cs -set -session_monitor_timeout 2
-session_idle_timeout 30
OK
EXAMPLE #8
To set the date, time, timezone, and NTP servers for the primary
Control Station, type:
$ nas_cs -set -time 200811070205 -timezone
America/New_York -ntp_server 128.221.252.0
OK
EXAMPLE # 9
To delete the domain name of the control station, type:
$ nas_cs -clear -dns_domain
OK
EXAMPLE # 10
To delete DNS servers of the control station, type:
$ nas_clear -clear -dns_servers
OK
EXAMPLE #11
To clear the IPv4 gateway for the primary Control Station, type:
$ nas_cs -clear -ip4gateway
OK
EXAMPLE #12
To clear the IPv6 network address and IPv6 gateway for the primary
Control Station, type:
$ nas_cs -clear -ip6address -ip6gateway
OK
EXAMPLE #13
To clear the DNS domain and DNS server configuration for the
primary Control Station, type:
$ nas_cs -clear -dns
188
Command Line Interface Reference for File 8.1
The nas Commands
OK
EXAMPLE #14
To clear the domain search configuration for the primary Control
Station, type:
$ nas_cs -clear -search_domain
OK
EXAMPLE #15
To clear the NTP server configuration for the primary Control Station,
type:
$ nas_cs -clear -ntp_servers
OK
EXAMPLE #16
To prevent sending settings to domain, type:
$ nas_cs -set -ntp_servers
10.109.187.23,10.109.187.24,10.109.187.25
-no_domain_push
OK
EXAMPLE #17
To reboot the primary Control Station, type:
$ nas_cs -reboot
OK
nas_cs
189
The nas Commands
nas_dbtable
Displays the table records of the Control Station.
SYNOPSIS
nas_dbtable
To execute the command against a database that is on the Data Mover area:
-info -mover <movername> -db <dbname>
-query <tablename> -mover <movername> -db <dbname>
-filter {(<fieldname> <operator> <value> [{-and|-or}
<fieldname>{<|<=|>|>=|=|.CONTAIN.};<value>]...] ]
-list -mover <movername>
DESCRIPTION
Displays the table records of the specified Data Mover. It also filters
the records of a particular field, and lists those records by using
primary or secondary key values.
To execute the command against a database that is on the Control Station area:
-info -cs_path <cs_pathname> -db <dbname>
-query <tablename> -cs_path <cs_pathname> -db <dbname>
-filter {(<fieldname> <operator> <value> [{-and|-or}
<fieldname>{<|<=|>|>=|=|.CONTAIN.};<value>]...] ]
-list -cs_path <cs_pathname>
DESCRIPTION
Displays the table records of the Control Station. It also filters the
records of a particular field, and lists those records by using primary
or secondary key values.
The database located in the Data Mover can be read directly. The
backup of the database is read on the Control Station. If the database
is inconsistent, the nas_dbtable command allows you to manually
verify the backup of the database before restoring it.
The Data Mover table uses the standard XML interface of the
administration commands. The application can structure each table
data and keys as a set of fields. Each field has a unique name, type,
and size.
The table structure is stored in the db.<base name> file. It is backed
up and restored with the database. The DBMS reader uses this
description of the table structure to read and display the records from
the backup database.
190
Command Line Interface Reference for File 8.1
The nas Commands
DATA MOVER OPTIONS
-info -mover <movername> -db <dbname>
Displays the schema of a table or the list of fields and keys. It also
displays the number of records of the table so that the user can know
if it is reasonable to dump the entire table.
-query <tablename> -mover <movername> -db <dbname>
Displays the records of a table. Selects the records to display on the
value of some fields or secondary keys.
-filter { (<fieldname><operator><value> [{-and|-or}
<fieldname>{<|<=|>|>=|=|.CONTAIN.};<value>]...] ]
Filters the records of a particular field, and lists the records using
primary or secondary key values. The default with multiple filters is
the -and option. Only the = operator is supported in the first
implementation.
Note: The keys are used when the -and option is used. Multiple fields with
the -or option parses the table, and applies a filter on each record.
The <fieldname> argument is the name of a secondary key or
field. If the secondary key is declared as a sequence of fields, it is
used by specifying either the value of its fields or value. If the
secondary key is not declared in the schema, then rename the key
and its value as filter.
The <value> argument is the value of the field encoded in character.
CONTROL STATION
OPTIONS
-info -cs_path <cs_pathname> -db <dbname>
Displays the schema of a table or the list of fields and keys. It also
displays the number of records of the table so that the user can know
if it is reasonable to dump the entire table.
-query <tablename> -cs_path <cs_pathname> -db <dbname>
Displays the records of the table. Selects the records to display on the
value of some fields or secondary keys.
-filter { (<fieldname><operator><value> [{-and|-or}
<fieldname>{<|<=|>|>=|=|.CONTAIN.};<value>]...] ]
Filters the records of a particular field, and lists the records using
primary or secondary key values. The default with multiple filters is
the -and option. Only the = operator is supported in the first
implementation.
nas_dbtable
191
The nas Commands
Note: Keys are used when the -filter option contains all components of the
key, and the -and option is used. With the -or option, it is necessary to parse
all the records.
The <fieldname> argument is the name of a secondary key or
field. If the secondary key is declared as a sequence of fields, it is
used by specifying either the value of its fields or the secondary
key value. If the secondary key is not declared in the schema,
rename the key and its value as filter.
The <value> argument is the value of the field encoded in
character.
-list -cs_path <cs_pathname>
Displays the list of databases and tables within a particular directory
of the Control Station area.
SEE ALSO
server_dbms
EXAMPLE #1
To display the Secmap schema of the Data Mover, type:
$ nas_dbtable -info -mover <movername> -db Secmap
Database identification
=======================
Base Name
Table Name
= Secmap
= Mapping
Primary Key Schema
==================
sid
= SID
Secondary Key Components
========================
xid
= xidType, fxid
Data Schema
===========
origin
AD
192
= Enumeration
Unknown
:
Secmap
:
Localgroup :
Etc
:
Nis
:
0
16
32
48
64
: 80
Command Line Interface Reference for File 8.1
The nas Commands
xidType
=
fxid
cdate
gid
name
=
=
=
=
EXAMPLE #2
Usrmap
: 96
Ldap
: 112
Ntx
: 128
Enumeration
unknown_name : -2
unknown_sid : -1
unknown_type : 0
user
: 1
group
: 2
Unsigned Integer size : 4
Date
Unsigned Integer size : 4
String, length container size : 2
To filter the records of the Secmap schema, type:
sid
$ nas_dbtable -query Mapping -mover <movername> -db
Secmap -filter fxid=10011
= S-1-5-15-2b3be507-6bc5c62-3f32a78a-8cc
origin
xidType
fxid
cdate
gid
name
=
=
=
=
=
=
Record count
Last key
= 1
= 1050000000000051500000007e53b2b625cbc068aa7323fcc080000
Nis
user
10011
Fri Sep 11 17:39:09 2009
107
DVT2KA\MaxUsers00000011
nas_dbtable
193
The nas Commands
nas_devicegroup
Manages an established MirrorView/Synchronous (MirrorView/S)
consistency group, also known as a device group.
SYNOPSIS
nas_devicegroup
-list
| -info {<name>|id=<id>|-all} [-sync [yes|no]]
| -acl <acl_value> {<name>|id=<id>}
| -suspend {<name>|id=<id>}
| -resume {<name>|id=<id>}
DESCRIPTION
nas_devicegroup lists the device group information for a
MirrorView/S configuration, gets detailed information about a
consistency group, specifies an access control level value for the
group, suspends MirrorView/S operations, or resumes operations of
the device group.
A MirrorView/S with a VNX configuration involves source and
destination VNXs attached to old versions of systems. MirrorView/S
performs synchronous mirroring of source storage logical units
(LUNs) representing production images, where the mirrored LUNs
are part of a MirrorView/S consistency group.
On the source VNX, you must be root to issue the -acl, -suspend, and
-resume options.
nas_devicegroup must be run from a Control Station in slot 0; it will
report an error if run from a Control Station in slot 1.
OPTIONS
-list
Displays a list of available configured MirrorView/S device groups.
-info {<name>|id=<id>|-all} [-sync [yes|no]]
Displays detailed information about the MirrorView/S configuration
for a specific device group or for all groups.
[-sync [yes|no]]
The -sync option first synchronizes the Control Station’s view
with the VNX for block before displaying configuration
information. The default is yes.
194
Command Line Interface Reference for File 8.1
The nas Commands
-acl <acl_value> {<name>|id=<id>}
Sets an access control level value that defines the owner of the
system, and the level of access allowed for users and groups defined
in the access control level table. The nas_acl command provides more
information.
!
CAUTION
The access control level value for the group should not be changed
from the default setting. A change in access control level value can
prevent MirrorView/S from functioning properly.
-suspend {<name>|id=<id>}
Temporarily halts mirroring from the source to the destination,
thereby suspending the link. Changes can still be made to the source
LUNs, but are not applied to the destination LUNs until operations
are resumed.
-resume {<name>|id=<id>}
Resumes device group operations and restarts mirroring,
synchronizing the destination LUNs with the source LUNs.
ID
2
SEE ALSO
Using MirrorView/Synchronous with VNX for Disaster Recovery, nas_acl,
and nas_logviewer.
SYSTEM OUTPUT
The number associated with the storage device is dependent on the
attached system of the system; for MirrorView/S, VNX for block
displays a prefix of APM before a set of integers, for example,
APM00033900124-0019. The VNX for block supports the following
system-defined AVM storage pools for MirrorView/S only: cm_r1,
cm_r5_performance, cm_r5_economy, cmata_archive, and cmata_r3.
EXAMPLE #1
To list the configured MirrorView/S device groups that are available,
type:
name
mviewgroup
EXAMPLE #2
$ nas_devicegroup -list
owner
storage ID
acl type
500
APM00053001549 0
MVIEW
To display detailed information for a MirrorView/S device group,
type:
$ nas_devicegroup -info mviewgroup
Sync with CLARiiON backend ...... done
name
= mviewgroup
description
=
uid
= 50:6:1:60:B0:60:27:20:0:0:0:0:0:0:0:0
nas_devicegroup
195
The nas Commands
state
role
condition
recovery policy
number of mirrors
mode
owner
mirrored disks
local clarid
remote clarid
mirror direction
=
=
=
=
=
=
=
=
=
=
=
Synchronized
Primary
Active
Automatic
16
SYNC
500
APM00053001549
APM00053001552
local -> remote
Where:
Value
Sync with CLARiiON
system
name
description
uid
state
Definition
Indicates that a sync with the VNX for block was performed to retrieve the most recent
information. This does not appear if you specify -info -sync no.
Name of the device group.
Brief description of device group.
UID assigned, based on the system.
State of the device group (for example, Consistent, Synchronized, Out-of-Sync,
Synchronizing, Scrambled, Empty, Incomplete, or Local Only).
Whether the current system is the Primary (source) or Secondary (destination).
Whether the group is functioning (Active), Inactive, Admin Fractured (suspended),
Waiting on Sync, System Fractured (which indicates link down), or Unknown.
Type of recovery policy (Automatic is the default and recommended value for group
during system configuration; if Manual is set, use -resume after a link down failure).
Number of mirrors in group.
MirrorView mode (always SYNC in this release).
User whom the object is assigned to, and is indicated by the index number in the access
control level table. nas_acl provides information.
Comma-separated list of disks that are mirrored.
APM number of local VNX for block storage array.
APM number of remote VNX for block storage array.
On primary system, local to remote (on primary system); on destination system, local
from remote.
role
condition
recovery policy
number of mirrors
mode
owner
mirrored disks
local clarid
remote clarid
mirror direction
EXAMPLE #3
name
description
uid
state
role
196
To display detailed information about a MirrorView/S device group
without synchronizing the Control Station’s view with the VNX for
block, type:
=
=
=
=
=
$ nas_devicegroup -info id=2 -sync no
mviewgroup
50:6:1:60:B0:60:27:20:0:0:0:0:0:0:0:0
Consistent
Primary
Command Line Interface Reference for File 8.1
The nas Commands
condition
recovery policy
number of mirrors
mode
owner
mirrored disks
local clarid
remote clarid
mirror direction
EXAMPLE #4
=
=
=
=
=
=
=
=
=
Active
Automatic
16
SYNC
500
APM00053001549
APM00053001552
local -> remote
To halt operation of the specified device group, as a root user, type:
# nas_devicegroup -suspend mviewgroup
Sync with CLARiiON backend ...... done
STARTING an MV 'SUSPEND' operation.
Device group: mviewgroup ............ done
The MV 'SUSPEND' operation SUCCEEDED.
done
EXAMPLE #5
To resume operations of the specified device group, as a root user,
type:
# nas_devicegroup -resume mviewgroup
Sync with CLARiiON backend ...... done
STARTING an MV 'RESUME' operation.
Device group: mviewgroup ............ done
The MV 'RESUME' operation SUCCEEDED.
done
nas_devicegroup
197
The nas Commands
nas_disk
Manages the disk table.
SYNOPSIS
nas_disk
-list
| -delete <disk_name> [[-perm]|[-unbind]]
| -info {<disk_name>|id=<disk_id>}
| -rename <old_name> <new_name>
DESCRIPTION
OPTIONS
nas_disk displays a list of known disks and renames, deletes, or
displays information for the specified disk.
-list
Lists the disk table.
Note: The ID of the object is an integer and is assigned automatically. The
name of the disk might be truncated if it is too long for the display. To display
the full name, use the -info option with the disk ID.
-delete <disk_name> [[ -perm]|[-unbind]
Deletes an entry from the disk table. In a VNX, restores the VNX for
block LUN name to its default value.
Unless -perm is specified, the disk is still identified as a VNX disk
and can be discovered and marked again using server_devconfig.
The -perm option removes the entry from the disk table and deletes
the diskmark. The disk is then available to be deployed for use by
another platform. The -unbind option removes the LUN from the
VNX Storage group (if EMC Access Logix™ is enabled). The -unbind
option permanently destroys the LUN and its contents. If this is the
last LUN using a RAID group, then the RAID group will be deleted.
-info {<disk_name>|id=<disk_id>}
Displays information for a specific <disk_name> or <disk_id> such
as size, type, and Access control level (ACL).
-rename <old_name> <new_name>
Renames a disk to <new_name>.
Note: If a VNX for block LUN uses the default name, renames it in the format
VNX_<VNX-hoxstname>_<lun-id>_<VNX-dvol-name>.
198
Command Line Interface Reference for File 8.1
The nas Commands
SEE ALSO
SYSTEM OUTPUT
EXAMPLE #1
id
1
2
3
4
5
6
7
8
9
10
11
12
13
inuse
y
y
y
y
y
y
n
n
n
n
n
n
n
VNX System Operations and server_devconfig.
The number associated with the storage device is dependent on the
attached system. VNX for block displays a prefix of alphabetic
characters before a set of integers, for example,
FCNTR074200038-0019. Symmetrix systems display as a set of
integers, for example, 002804000190-003C.
To list the disk table for a VNX for block, type:
$ nas_disk -list
storageID-devID
FCNTR074200038-0000
FCNTR074200038-0001
FCNTR074200038-0002
FCNTR074200038-0003
FCNTR074200038-0004
FCNTR074200038-0005
FCNTR074200038-0010
FCNTR074200038-0012
FCNTR074200038-0011
FCNTR074200038-0013
FCNTR074200038-0014
FCNTR074200038-0015
FCNTR074200038-0015
sizeMB
11263
11263
2047
2047
2047
32767
1023
1023
1023
1023
10239
10239
10239
type
CLSTD
CLSTD
CLSTD
CLSTD
CLSTD
CLSTD
CLSTD
CLSTD
CLSTD
CLSTD
MIXED
MIXED
MIXED
name
root_disk
root_ldisk
d3
d4
d5
d6
d7
d8
d9
d10
d11
d12
d13
servers
1,2
1,2
1,2
1,2
1,2
1,2
1,2
1,2
1,2
1,2
1,2
1,2
1,2
Where:
Value
id
inuse
sizeMB
storageID-devID
type
Definition
ID of the disk (assigned automatically).
Used by any type of volume or file system.
Total size of disk.
ID of the system and device associated with the disk.
Type of disk contingent on the system attached; CLSTD, CLATA, CMSTD, CLEFD, CMEFD,
CMATA, MIXED (indicates tiers used in the pool contain multiple disk types), Performance,
Capacity, Extreme_performance, Mirrored_mixed, Mirrored_performance, Mirrored_capacity,
and Mirrored_extreme_performance are VNX disk types and STD, BCV, R1BCV, R2BCV,
R1STD, R2STD, ATA, R1ATA, R2ATA,BCVA, R1BCA, R2BCA, EFD, FTS, R1FTS,R2FTS,
BCVF, R1BCF, R2BCF,BCVMIXED, R1MIXED, R2MIXED, R1BCVMIXED, and R2BCVMIXED
are Symmetrix disk types.
Name of the disk; ’dd’ in a disk name indicates a remote disk.
Servers that have access to this disk.
name
servers
EXAMPLE #2
id
1
2
inuse
y
y
sizeMB
11507
11507
To list the disk table for the system with a Symmetrix system, type:
$ nas_disk -list
storageID-devID
type
000190100530-00FB
STD
000190100530-00FC
STD
name
root_disk
root_ldisk
servers
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
nas_disk
199
The nas Commands
3
y
4
y
5
y
6
y
7
n
8
n
9
n
10
n
11
n
12
n
13
n
14
n
15
n
16
n
17
n
18
n
19
n
20
n
21
n
22
n
23
n
24
n
[....]
155
n
156
n
157
n
158
n
159
n
160
n
161
n
162
n
163
n
164
n
165
n
166
n
167
n
168
n
169
n
170
n
171
n
172
n
173
n
174
n
175
n
176
n
177
n
178
n
179
n
180
n
181
n
182
n
183
n
200
2076
2076
2076
65536
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
000190100530-00FD
000190100530-00FE
000190100530-00FF
000190100530-04D4
000190100530-0102
000190100530-0103
000190100530-0104
000190100530-0105
000190100530-0106
000190100530-0107
000190100530-0108
000190100530-0109
000190100530-010A
000190100530-010B
000190100530-010C
000190100530-010D
000190100530-010E
000190100530-010F
000190100530-0110
000190100530-0111
000190100530-0112
000190100530-0113
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
STD
d3
d4
d5
d6
d7
d8
d9
d10
d11
d12
d13
d14
d15
d16
d17
d18
d19
d20
d21
d22
d23
d24
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
28560
000190100530-0196
000190100530-0197
000190100530-0198
000190100530-0199
000190100530-019A
000190100530-019B
000190100530-019C
000190100530-019D
000190100530-019E
000190100530-019F
000190100530-01A0
000190100530-01A1
000190100530-01A2
000190100530-01A3
000190100530-01A4
000190100530-01A5
000190100530-01A6
000190100530-01A7
000190100530-01A8
000190100530-01A9
000190100530-01AA
000190100530-01AB
000190100530-01AC
000190100530-01AD
000190100530-01AE
000190100530-01AF
000190100530-01B0
000190100530-01B1
000190100530-01B2
STD
STD
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
BCV
d155
d156
rootd157
rootd158
rootd159
rootd160
rootd161
rootd162
rootd163
rootd164
rootd165
rootd166
rootd167
rootd168
rootd169
rootd170
rootd171
rootd172
rootd173
rootd174
rootd175
rootd176
rootd177
rootd178
rootd179
rootd180
rootd181
rootd182
rootd183
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
Command Line Interface Reference for File 8.1
The nas Commands
184
185
186
187
188
189
190
191
192
193
194
195
196
197
198
199
200
201
202
203
204
n
n
n
n
n
n
n
n
n
n
n
n
n
n
n
n
n
n
n
n
y
28560
28560
28560
11507
11507
11507
11507
11507
11507
11507
11507
11507
11507
11507
11507
11507
11507
11507
11507
11507
11507
000190100530-01B3
000190100530-01B4
000190100530-01B5
000190100530-051D
000190100530-051E
000190100530-051F
000190100530-0520
000190100530-0521
000190100530-0522
000190100530-0523
000190100530-0524
000190100530-0525
000190100530-0526
000190100530-0527
000190100530-0528
000190100530-0529
000190100530-052A
000190100530-052B
000190100530-052C
000190100530-052D
000190100530-052E
BCV
BCV
BCV
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
EFD
rootd184
rootd185
rootd186
d187
d188
d189
d190
d191
d192
d193
d194
d195
d196
d197
d198
d199
d200
d201
d202
d203
d204
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
1,2,3,4,5,6,7,8
Note: This is a partial listing due to the length of the output.
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #3
To view information for disk d7 for a system with a VNX for block,
type:
$ nas_disk -info d7
id
= 7
name
= d7
acl
= 0
in_use
= True
pool
= TP1
size (MB) = 273709
type
= Mixed
protection= RAID5(4+1)
stor_id
= FCNTR074200038
stor_dev = 0012
volume_name = d7
storage_profiles = TP1
thin
= True
tiering_policy = Auto-tier
compressed= False
mirrored = False
servers
= server_2,server_3,server_4,server_5
server = server_2
addr=c0t1l2
server = server_2
addr=c32t1l2
server = server_2
addr=c16t1l2
server = server_2
addr=c48t1l2
nas_disk
201
The nas Commands
server
server
server
server
server
server
server
server
server
server
server
server
=
=
=
=
=
=
=
=
=
=
=
=
server_3
server_3
server_3
server_3
server_4
server_4
server_4
server_4
server_5
server_5
server_5
server_5
addr=c0t1l2
addr=c32t1l2
addr=c16t1l2
addr=c48t1l2
addr=c0t1l2
addr=c32t1l2
addr=c16t1l2
addr=c48t1l2
addr=c0t1l2
addr=c32t1l2
addr=c16t1l2
addr=c48t1l2
Where:
Value
id
name
acl
in_use
pool
size (MB)
type
protection
stor_id
stor_dev
volume_name
storage_profiles
thin
tiering_policy
compressed
mirrored
servers
addr
EXAMPLE #4
Definition
ID of the disk (assigned automatically).
Name of the disk.
Access control level value of the disk.
Used by any type of volume or file system.
Name of the storage pool in use.
Total size of the disk.
Type of disk contingent on the system attached; VNX for block disk types are CLSTD,
CLATA, CMSTD, CLEFD, CLSAS, CMEFD, CMATA, MIXED (indicates tiers used in the pool
contain multiple disk types), Performance, Capacity, Extreme_performance,
Mirrored_mixed, Mirrored_performance, Mirrored_capacity, and
Mirrored_extreme_performance.
The type of disk protection that has been assigned.
ID of the system associated with the disk.
ID of the device associated with the disk.
Name of the volume residing on the disk.
The storage profiles to which the disk belongs.
Indicates whether the block system uses thin provisioning. Values are: True, False.
Indicates the tiering policy in effect. If the initial tier and the tiering policy are the same, the
values are: Auto-Tier, Highest Available Tier, Lowest Available Tier. If the initial tier and the
tiering policy are not the same, the values are: Auto-Tier/No Data Movement, Highest
Available Tier/No Data Movement, Lowest Available Tier/No Data Movement.
For VNX for block, indicates whether data is compressed. Values are: True, False, Mixed
(indicates some of the LUNs, but not all, are compressed).
Indicates whether the disk is mirrored.
Lists the servers that have access to this disk.
Path to system (SCSI address).
To view information for disk d205 for the system with a Symmetrix
system, type:
$ nas_disk -info d205
202
Command Line Interface Reference for File 8.1
The nas Commands
id
= 205
name
= d205
acl
= 0
in_use
= True
pool
= SG0
size (MB) = 28560
type
= Mixed
protection= RAID1
symm_id
= 000190100530
symm_dev = 0539
volume_name = d205
storage_profiles = SG0_000192601245
thin
= True
tiering_enabled = True
compression = True
mirrored = False
servers
=
server_2,server_3,server_4,server_5,server_6,server_7,server_8,server_9
server = server_2
addr=c0t14l0
FA=03A
FAport=0
server = server_2
addr=c16t14l0
FA=04A
FAport=0
server = server_3
addr=c0t14l0
FA=03A
FAport=0
server = server_3
addr=c16t14l0
FA=04A
FAport=0
server = server_4
addr=c0t14l0
FA=03A
FAport=0
server = server_4
addr=c16t14l0
FA=04A
FAport=0
server = server_5
addr=c0t14l0
FA=03A
FAport=0
server = server_5
addr=c16t14l0
FA=04A
FAport=0
server = server_6
addr=c0t14l0
FA=03A
FAport=0
server = server_6
addr=c16t14l0
FA=04A
FAport=0
server = server_7
addr=c0t14l0
FA=03A
FAport=0
server = server_7
addr=c16t14l0
FA=04A
FAport=0
server = server_8
addr=c0t14l0
FA=03A
FAport=0
server = server_8
addr=c16t14l0
FA=04A
FAport=0
server = server_9
addr=c0t14l0
FA=03A
FAport=0
server = server_9
addr=c16t14l0
FA=04A
FAport=0
Where:
Value
id
name
acl
in_use
pool
size (MB)
type
protection
symm_id
Definition
ID of the disk (assigned automatically).
Name of the disk.
Access control level value of the disk.
Used by any type of volume or file system.
Name of the storage pool in use.
Total size of disk.
Type of disk contingent on the system attached; Symmetrix disk types are STD, BCV,
R1BCV, R2BCV, R1STD, R2STD, ATA, R1ATA, R2ATA, BCVA, R1BCA, R2BCA, EFD, FTS,
R1FTS,R2FTS, BCVF, R1BCF, R2BCF, BCVMIXED, R1MIXED, R2MIXED, R1BCVMIXED,
and R2BCVMIXED. If multiple disk volumes are used, the type is Mixed.
The type of disk protection that has been assigned.
ID of the Symmetrix system associated with the disk.
nas_disk
203
The nas Commands
symm_dev
volume_name
storage_profiles
thin
tiering_enabled
compressed
mirrored
servers
addr
EXAMPLE #5
ID of the Symmetrix device associated with the disk.
Name of the volume residing on the disk.
The storage profiles to which the disk belongs.
Indicates whether the system uses thin provisioning. Values are: True, False, Mixed.
Indicates whether a tiering policy is being used.
For VNX with Symmetrix backend, indicates whether data is compressed. Values are:
True, False, Mixed (indicates some of the LUNs, but not all, are compressed).
Indicates whether the disk is mirrored.
Lists the servers that have access to this disk.
Path to system (SCSI address).
To view information for disk d3454 (FTS device created using eDisk
configured in external provisioning mode) for the system with a
Symmetrix system, type:
$ nas_disk -info id=3454
id
= 3454
name
= d3454
acl
= 0
in_use
= False
size (MB) = 28772
type
=FTS
protection= Unprotected
symm_id
= 000196900010
symm_dev = 0031
volume_name = d3454
storage_profiles = symm_fts
thin
= False
compression = False
mirrored = False
servers
= server_2,server_3
server = server_2
addr=c0t1l3
FA=02E
server = server_2
addr=c16t1l3
FA=01E
server = server_2
addr=c32t1l3
FA=01E
server = server_2
addr=c48t1l3
FA=02E
server = server_3
addr=c0t1l3
FA=02E
server = server_3
addr=c16t1l3
FA=01E
server = server_3
addr=c32t1l3
FA=01E
server = server_3
addr=c48t1l3
FA=02E
FAport=0
FAport=0
FAport=0
FAport=0
FAport=0
FAport=0
FAport=0
FAport=0
EXAMPLE #4 provides a description of command outputs.
EXAMPLE #6
To rename a disk in the system with a VNX for block, type:
$ nas_disk -rename d7 disk7
id
name
acl
in_use
204
=
=
=
=
7
disk7
0
True
Command Line Interface Reference for File 8.1
The nas Commands
size (MB) = 273709
type
= CLSTD
protection= RAID5(4+1)
stor_id
= FCNTR074200038
stor_dev = 0012
volume_name = disk7
storage_profiles = clar_r5_performance
virtually_provisioned = False
mirrored = False
servers
= server_2,server_3,server_4,server_5
server = server_2
addr=c0t1l2
server = server_2
addr=c32t1l2
server = server_2
addr=c16t1l2
server = server_2
addr=c48t1l2
server = server_3
addr=c0t1l2
server = server_3
addr=c32t1l2
server = server_3
addr=c16t1l2
server = server_3
addr=c48t1l2
server = server_4
addr=c0t1l2
server = server_4
addr=c32t1l2
server = server_4
addr=c16t1l2
server = server_4
addr=c48t1l2
server = server_5
addr=c0t1l2
server = server_5
addr=c32t1l2
server = server_5
addr=c16t1l2
server = server_5
addr=c48t1l2
EXAMPLE #4 provides a description of command outputs.
EXAMPLE #7
To delete a disk entry from the disk table for the system with a VNX
for block, type:
$ nas_disk -delete d24
id
= 24
name
= d24
acl
= 0
in_use
= False
size (MB) = 456202
type
= CLATA
protection= RAID5(6+1)
stor_id
= FCNTR074200038
stor_dev = 0023
storage_profiles = clarata_archive
virtually_provisioned = False
mirrored = False
servers
= server_2,server_3,server_4,server_5
EXAMPLE #4 provides a description of command outputs.
nas_disk
205
The nas Commands
nas_diskmark
Queries the system, manages and lists the SCSI devices configuration.
SYNOPSIS
nas_diskmark
-mark {-all|<movername>} [-discovery {y|n}] [-monitor {y|n}]
[-Force {y|n}]
| -list {-all|<movername>}
DESCRIPTION
!
nas_diskmark queries the available system device and tape device
configuration, saves the device configuration into the Data Movers
database, and lists SCSI devices. This command also manages NAS
database configuration related to advanced data services from
back-end storage system.
CAUTION
It is recommended that all Data Movers have the same device
configuration. When adding devices to the device table for a single
Data Mover only, certain actions such as standby failover are not
successful unless the standby Data Mover has the same disk device
configuration as the primary Data Mover.
The -all option executes the command for all Data Movers.
LUN migration for VNX
Symmetrix systems
When a newly created LUN having an ID same as that of some
removed device is detected, the command may report a conflict error.
After a LUN is removed at the backend and then a new LUN is
created with the same ID, Control Station cannot be aware of its
deletion at first. The error occurs because the new LUN has duplicate
storage ID and device ID with stale disk volume. This case only
applies in Symmetrix backend.
For example, During LUN migration, where a Symmetrix device is
moved from source storage group (SG) to destination SG, the LUN ID
of this device in the source SG should be maintained even in the
destination SG. Else, this will reflect in a conflict error on the Control
Station during running nas_diskmark.
OPTIONS
-mark {-all|<movername>}
Queries SCSI devices and saves them into the device table database
on the Data Mover.
206
Command Line Interface Reference for File 8.1
The nas Commands
Modifies VNX for block LUN names to the
VNX_<VNX-hostname>_<lun-id>_<VNX-dvol-name> format, if the
LUNs use the default Unisphere name.
!
CAUTION
The time taken to complete this command may be lengthy,
dependent on the number and type of attached devices.
[-discovery {y|n}]
Enables or disables the storage discovery operation.
!
CAUTION
Disabling the -discovery option should be done only under the
direction of an EMC Customer Service Engineer.
[-monitor {y|n}]
Displays the progress of the query and discovery operations.
[-Force {y|n}]
Overrides the health check failures and changes the storage
configuration.
!
CAUTION
Use the -Force option only when directed to do so, as high
availability can be lost when changing storage configuration.
-list {-all|<movername>}
Lists the SCSI devices for the specified Data Mover or all Data
Movers.
EXAMPLE #1
To query SCSI devices on server_2 and display the progress of the
query operation, type:
$ nas_diskmark -mark server_2 -monitor y
Discovering storage (may take several minutes)
server_2:
chain 0 ..........
chain 16 ........
chain 32 ........
chain 48 ..........
chain 96 ..........
chain 112 ..........
Verifying disk reachability
nas_diskmark
207
The nas Commands
Verifying
Verifying
Verifying
Verifying
Verifying
done
file system reachability
local domain
disk health
gate keepers
device group
EXAMPLE #2
server_2 :
chain= 0,
stor_id=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
$ nas_diskmark -list server_2
chain 0 :
scsi-0
HK190807090011 VNX_id= HK1908070900110032
0/0 type= disk sz= 11263 val= 1 info= DGC RAID 5 03243200000032NI
0/1 type= disk sz= 11263 val= 2 info= DGC RAID 5 03243300010033NI
0/2 type= disk sz= 2047 val= 3 info= DGC RAID 5 03243400020034NI
0/3 type= disk sz= 2047 val= 4 info= DGC RAID 5 03243500030035NI
0/4 type= disk sz= 2047 val= 5 info= DGC RAID 5 03243600040036NI
0/5 type= disk sz= 32767 val= 6 info= DGC RAID 5 03243700050037NI
1/0 type= disk sz= 274811 val= 7 info= DGC RAID 5 03244400100044NI
1/1 type= disk sz= 274811 val= -5 info= DGC RAID 5 03244500110045NI
1/2 type= disk sz= 274811 val= 8 info= DGC RAID 5 03244600120046NI
1/3 type= disk sz= 274811 val= -5 info= DGC RAID 5 03244700130047NI
1/4 type= disk sz= 274811 val= 9 info= DGC RAID 5 03245600140056NI
1/5 type= disk sz= 274811 val= -5 info= DGC RAID 5 03245700150057NI
1/6 type= disk sz= 274811 val= 10 info= DGC RAID 5 03245800160058NI
1/7 type= disk sz= 274811 val= -5 info= DGC RAID 5 03245900170059NI
1/8 type= disk sz= 274811 val= 99 info= DGC RAID 5 03245A0018005ANI
1/9 type= disk sz= 274811 val= -5 info= DGC RAID 5 03245B0019005BNI
1/10 type= disk sz= 274811 val= 97 info= DGC RAID 5 03245C001A005CNI
1/11 type= disk sz= 274811 val= -5 info= DGC RAID 5 03245D001B005DNI
1/12 type= disk sz= 274811 val= 13 info= DGC RAID 5 03245E001C005ENI
1/13 type= disk sz= 274811 val= -5 info= DGC RAID 5 03245F001D005FNI
1/14 type= disk sz= 274811 val= 14 info= DGC RAID 5 032460001E0060NI
1/15 type= disk sz= 274811 val= -5 info= DGC RAID 5 032461001F0061NI
server_2 : chain 1
no drives on chain
server_2 : chain 2
no drives on chain
server_2 : chain 3
no drives on chain
server_2 : chain 4
no drives on chain
server_2 : chain 5
no drives on chain
server_2 : chain 6
no drives on chain
server_2 : chain 7
no drives on chain
server_2 : chain 8
no drives on chain
208
To list the SCSI devices for server_2, type:
:
:
:
:
:
:
:
:
Command Line Interface Reference for File 8.1
The nas Commands
server_2 : chain 9 :
no drives on chain
server_2 : chain 10 :
no drives on chain
server_2 : chain 11 :
no drives on chain
server_2 : chain 12 :
no drives on chain
server_2 : chain 13 :
no drives on chain
server_2 : chain 14 :
no drives on chain
server_2 : chain 15 :
no drives on chain
.
Note: This is a partial listing due to the length of the output.
nas_diskmark
209
The nas Commands
nas_emailuser
Manages email notifications for serious system events.
SYNOPSIS
nas_emailuser
-info
| -test
| -modify
[-enabled {yes|no}]
[-to <email_addr> [,...]]
[-cc <email_addr> [,...]]
[-email_server <email_server>]
[-subject_prefix <email_subject>]
[-from <email_addr>]
| -init
DESCRIPTION
OPTIONS
nas_emailuser enables, configures, and tests email notifications for
serious system events.
-info
Displays the configuration for email notifications.
-test
Generates a test event that sends a test email notification to the email
addresses configured in -to and -cc. The recipient email address must
be configured prior to testing email notification.
Note: After the -test option is run, all the configured recipients must be asked
to confirm whether they received the test email with the correct system
identification information.
-modify
Modifies one or more of the following configuration parameters:
[-enabled {yes|no}]
Enables email notification if yes is specified. The recipient email
address must be configured prior to enabling email notification.
Disables email notification if no is specified.
[-to <email_addr> [,...]]
Configures one or more recipient email addresses. The email
addresses are comma-separated, enclosed in single-quotes, and
follow the mailbox@fully_qualified_domain_name format. For
example, ’storage_admin@yourcompany.com’,
’backup_admin@yourcompany.com’.
210
Command Line Interface Reference for File 8.1
The nas Commands
Refer the following email address format guidelines for
configuring email addresses. An email address can contain:
• A maximum of 63 characters; the field can contain a maximum
of 255 characters:
– ASCII characters: a through z, A through Z, 0 through 9, ! #
$ % & * + - / = ? ^ _ ` {|,} ~ are allowed; a period, if it is not
the first or last character in the mailbox
– Alphanumeric strings
• Single quotes, if they are escaped in the format:
– - your\'email@yourcompany.com
– -'first'\''email@yourcompany.com,second'\''email@yourcom
pany.com'
[-cc <email_addr> [,...]]
Configures a list of carbon-copy recipients. The email addresses
are comma-separated, enclosed in single-quotes, and follow the
mailbox@fully_qualified_domain_name format. For example,
’storage_admin@yourcompany.com’. For the email address
character set and format guidelines, refer the -to option.
[-email_server <email_server>]
Configures the email server that accepts and routes the email
notifications. <email_server> specifies an IP address or the fully
qualified domain name, which can have 1 to 63 characters. The IP
addresses 0.0.0.0 and 255.255.255.255 are not allowed.
[-subject_prefix <email_subject>]
Specifies the email subject prefix. The subject prefix for the email
notification can be from 1 to 63 characters long, is enclosed in
quotes, and should contain printable ASCII characters. You can
customize the subject prefix for specific needs like email filtering.
The default subject is "System Notification."
[-from <email_addr>]
Configures the sender’s email address. If the sender’s email
address is not specfied, a default email address of the format
root@<hostname> is configured. The email address follows the
mailbox@fully_qualified_domain_name format. For example,
’storage_admin@yourcompany.com’. For the email address
character set and format guidelines, refer the -to option.
nas_emailuser
211
The nas Commands
-init
Initializes the default state; displays a status message if the
feature has already been initialized. The -init option must be used
only when directed.
SEE ALSO
EXAMPLE #1
Configuring Events and Notifications on VNX for File.
To configure email notifications using email server 10.6.50.122 from
administrator to support, while copying engineering and
documentation, type:
$ nas_emailuser -modify -to
szg30@fire2.hosts.pvt.dns,support1@nasdocs.emc.com,docum
entation@nasdocs.emc.com
OK
EXAMPLE #2
To display information on email notifications, type:
$ nas_emailuser -info
Service Enabled
= Yes
Recipient Address(es)
=
szg30@fire2.hosts.pvt.dns,support1@nasdocs.emc.com,documentation@nasdocs.emc.com
Carbon copy Address(es) =
Email Server
= 10.241.168.23
Subject Prefix
= System Notification
Sender Address
=
EXAMPLE #3
To test email notifications, type:
$ nas_emailuser -test
OK
EXAMPLE #4
To disable email notification, type:
$ nas_emailuser -modify -enabled no
OK
212
Command Line Interface Reference for File 8.1
The nas Commands
nas_environment
Reports the inlet air temperatures and input power to the user.
SYNOPSIS
nas_environment -info
{
| -system [-present|-average]
| -dme [enclosure_id] [-intemp [f|c]|-power] [-present]|[-average]
| -array [-present|-average]
| -shelf {<shelf_id>|<-all>} [-intemp [f|c]|-power]
[-present|-average]
| -battery [a|b] [-present|-average]
| -spe [-intemp [f|c]|-power] [-present|-average]
| -all
}
DESCRIPTION
OPTIONS
nas_environment -info displays the inlet air temperatures of the
Data Mover enclosures, disk-array enclosures, the input power of the
Data Mover enclosures, disk-array enclosures, and standby power
supply through the CLI and Unisphere GUI.
-system
Displays the present or average input power information of the
system, which includes file and block on VNX systems, and file only
on gateway systems.
-present
Displays the current value, which is a sum of the present input
power for all supported systems. The current value is computed
as the 30 second average of the power consumption sampled
every three seconds.
-average
Displays the average value. It requires an hour to calculate the
correct value. N/A is displayed if there is less than one hour
worth of data. The average value is computed as the 60 minute
rolling average of the present power consumption values.
-dme
Displays the present or average inlet air temperature and input
power information on a specified Data Mover enclosure. If a specific
enclosure_id is not specified, all Data Mover enclosure information is
displayed.
nas_environment
213
The nas Commands
enclosure_id
Specifies a Data Mover enclosure_id on which to display
information.
-intemp [f|c]
Displays the inlet air temperature information. The f flag
indicates Fahrenheit. The default value or c flag indicates Celsius.
-power
Displays the input power information.
-present
Displays the current value.
-average
Displays the average value. It requires an hour to calculate the
correct value. N/A is displayed if there is less than one hour
worth of data.
-array
Displays the present or average input power information on the
array.
-present
Displays the current value.
-average
Displays the average value. It requires an hour to calculate the
correct value. N/A is displayed if there is less than one hour
worth of data.
-shelf
Allows to input a value for a selected enclosure. It displays the
present and average inlet air temperature and input power
information on a specified disk-array enclosure. If a specific
enclosure_id is not specified, all disk-array enclosure information is
displayed.
<shelf_id>
Specifies a disk array enclosure_id on which to display
information.
-power
Displays the input power information.
-present
Displays the current value.
214
Command Line Interface Reference for File 8.1
The nas Commands
-average
Displays the average value. It requires an hour to calculate the
correct value. N/A is displayed if there is less than one hour
worth of data.
-battery
Displays the present and average input power information on a
specified standby power supply. If no -a or -b is specified, then the
information is displayed on both standby power supplies.
-a
Specifies a standby power supply A on which to display
information.
-b
Specifies a standby power supply B on which to display
information.
-present
Displays the current value.
-average
Displays the average value. It requires an hour to calculate the
correct value. N/A is displayed if there is not one hour worth of
data.
-spe
Displays the present and average input power information on a
specified standby power supply.
-intemp [f|c]
Displays the inlet air temperature information. The f flag
indicates Fahrenheit. The default value or c flag indicates Celsius.
-power
Displays the input power information.
-present
Displays the current value.
-average
Displays the average value. It requires an hour to calculate the
correct value. N/A is displayed if there is less than one hour
worth of data.
nas_environment
215
The nas Commands
-all
Displays the following:
EXAMPLE #1
◆
System input power
◆
Data Mover enclosure inlet air temperatures and input power
◆
Array input power
◆
Disk-array enclosure inlet air temperatures and input power
◆
Storage processor enclosure inlet air temperatures and input
power
◆
Standby power supply input power
To view the present and average input power information for file and
block on systems or file only on gateway system, type:
$ nas_environment -info -system
System = Celerra ns 600 APM 000237001650000
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
EXAMPLE #2
To view the average inlet air temperature on Data Mover enclosure 1
in degrees Farenheit, type:
$ nas_environment -info -dme 1 -intemp f -average
Component = DME 0 Data Mover 1
Temperature Status = OK
Rolling average (degrees) = 53F
Temperature Status = OK
Rolling average (degrees) = 53F
EXAMPLE #3
To view the average inlet air temperature on Data Mover enclosure 1
in degrees Celsius, type:
$ nas_environment -info -dme1 -intemp c -average
Data Mover Enclosure 1
Status: Valid
Inlet Air Temperature
Rolling average (degrees Celsius): 11.3
EXAMPLE #4
To view the present system information, type:
$ nas_environment -info -system -present
System = Celerra ns 600 APM 000237001650000
Power Status = OK
Present (watts) = 150
216
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #5
To view the information array (input power and inlet temperature),
type:
$ nas_environment -info -array
Component = CLARiiON CX600 APM0023700165
Power Status = OK
Present (watts) = 230
Rolling average (watts) = 245
EXAMPLE #6
To view the present and average inlet air temperature on all shelves,
type:
$ nas_environment -info -shelf -all
Component = Shelf 0/0 Shelf 0/0
Power Status = OK
Present (watts) = 150
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
Temperature Status = OK
Present (degrees) = 12C
Rolling average (degrees) = 11C
Component = Shelf 0/1 Shelf 0/1
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
Temperature Status = OK
Present (degrees) = 12C
Rolling average (degrees) = 11C
Component = Shelf 1/0 Shelf 1/0
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
Temperature Status = OK
Present (degrees) = 12C
Rolling average (degrees) = 11C
Component = Shelf 1/1 Shelf 1/1
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
Temperature Status = OK
Present (degrees) = 12C
Rolling average (degrees) = 11C
EXAMPLE #7
To view the present and average inlet air temperature on shelf 1,
enclosure 1, type:
nas_environment
217
The nas Commands
$ nas_environment -info -shelf 1/1 -average
Component = Shelf 1/1 Shelf 1/1
Power Status = OK
Rolling average (watts) = 150
Temperature Status = OK
Rolling average (degrees) = 11C
EXAMPLE #8
To view the present and average inlet air temperature on all spes,
type:
$ nas_environment -info -spe
Component = SPE 0 SPE 0
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
Temperature Status = OK
Present (degrees) = 12C
Rolling average (degrees) = 11C
EXAMPLE #9
To view the present information for all batteries, type:
$ nas_environment -info -battery
Component = Shelf 0/0 SP A
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
Component = Shelf 0/0 SP B
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
EXAMPLE #10
To view the present information for all batteries, type:
$ nas_environment -info -battery a -average
Component = Shelf 0/0 SP A
Power Status = OK
Rolling average (watts) = 150
EXAMPLE #11
To view all the components, type:
$ nas_environment -info -all
Component = Shelf 0/0 SP A
Power Status = OK
Rolling average (watts) = 150
EXAMPLE #12
To view the present and average inlet air temperature of the dme,
shelf, spe components, type:
$ nas_environment -info -all
Component = Celerra ns600 APM000237001650000
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
218
Command Line Interface Reference for File 8.1
The nas Commands
Component = DME 0 Data Mover 0
Power Status = OK
Present (watts) = 200
Rolling average (watts) = 333
Temperature Status = OK
Present (degrees) = 12C
Rolling Average (degrees) = 11C
Component = DME 0 Data Mover 1
Power Status = Error 13690667102: Not Present
Present (watts) = N/A
Rolling average (watts) = N/A
Temperature Status = Error 13690667102: Unsupported
Present (watts) = N/A
Rolling average (watts) = N/A
Component = DME 0 Data Mover 2
Power Status = Error 13690667102: Uninitialized
Present (watts) = 150
Average (watts) = N/A
Temperature Status = Error 13690667102: Uninitialized
Present (degrees) = 12C
Average (degrees) = N/A
Component = DME 0 Data Mover 3
Power Status = Error 13690667102: Failed
Present (watts) = 150
Average (watts) = N/A
Temperature Status = Error 13690667102: Failed
Present (degrees) = 12C
Average (degrees) = N/A
Component = Shelf 0/0
Power Status = OK
Present (watts): 150
Rolling average (watts): 150
Present (watts): 150
Rolling average (watts): 150
Temperature Status = OK
Present (degrees) = 12C
Rolling average (degrees) = 11C
Component = Shelf 0/1
Power Status = Error 13690667102: Invalid
Present (watts) = N/A
Rolling average (watts) = N/A
Temperature Status = Error 13690667102: Invalid
Present (degrees) = N/A
Rolling average (degrees) = N/A
Component = Shelf 1/0
Power Status = OK
Present (watts) = 150
nas_environment
219
The nas Commands
Rolling average (watts) = 150
Temperature Status = OK
Present (degrees) = 12C
Rolling average (degrees) = 11C
Component = Shelf 1/1
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
Temperature Status = OK
Present (degrees) = 12C
Rolling average (degrees) = 11C
Component = CLARiiON CX600 APM00023700165
Power Status = OK
Present (watts) = 230
Rolling average (watts) = 245
Component = SPE 0 SPE 0
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
Temperature Status = OK
Present (degrees) = 12C
Rolling average (degrees) = 11C
Component = Shelf 0/0 SP A
Power Status = OK
Present (watts) = 150
Power Status = OK
Present (watts) = 150
Rolling average (watts) = 150
Component = Shelf 0/0 SP B
Power Status = Error 13690667102: Not Present
Present (watts) = N/A
Rolling average (watts) = N/A
220
Command Line Interface Reference for File 8.1
The nas Commands
nas_event
Provides a user interface to system-wide events.
SYNOPSIS
nas_event
-Load {-info|<file_name>}
| -Unload <file_name>
| -list
-action {-info|{trap|logfile|mail|callhome|exec|udprpc|tcprpc|terminate}
[-component {<component>|<id>}][-facility {<facility>
| <id>}][-severity {<severity>|<id>}]]|[-id]}
|-component {-info|{<component>|<id>} [-facility {<facility>|<id>}]|[-id]}
|-severity {-info|<severity>|<id>} [-component {<component>|<id>}
[-facility {<facility>|<id>}]]|[-id]
|-keyword <keyword> [-component {<component>|<id>}
[-facility{<facility>|<id>}][-severity {<severity>|<id>}]]|[-id]
DESCRIPTION
nas_event provides a user interface to system-wide events within the
VNX. The VNX includes a default event configuration file that
contains a mapping of facilities that generate events, and the
associated actions triggered by those events.
To list the default configuration files, type:
$ nas_event -Load -info
Using a text editor, a new event configuration file can be created and
loaded into the database to add an event.
VNX facilities generate events that trigger specific actions. An event
consists of:
!
◆
An assigned ID for the event and the facility issuing the event
◆
The high water mark for the severity of the event
◆
A description of the event
◆
The system-defined action to take when the event occurs
CAUTION
The callhome events in the system are carefully reviewed and
configured to provide the right level of support. Do not add, delete,
or modify lines that specify the callhome action in the event
configuration files in the /nas/sys directory. User-defined event
configuration files should not use the callhome action.
nas_event
221
The nas Commands
OPTIONS
-Load {-info|<file_name>}
Loads the event configuration file <file_name> into the system. The
-info option displays the currently loaded configuration files.
-Unload <file_name>
Unloads the configuration file <file_name> from the system.
!
CAUTION
The /nas/sys/nas_eventlog.cfg configuration file must not be
deleted, as this can lead to data loss. Unloading or modifying
configuration files that are loaded by default can affect logging,
alerts, notifications, and system operations.
-list
The -list option displays components and facilities that generate
events, and the actions that are taken as a result. You can search for an
event, facility, or action by using a keyword. Component, facility, and
severity can be specified by either the text name or ID. The output is
displayed with parameter names in the form $(paraname,
typeIndicator, fmtStr).
-action
{-info|{trap|logfile|mail|callhome|exec|udprpc|tcprpc
|
terminate}
With the -info option, lists all the possible actions associated with
events. If one of the actions trap, logfile, mail, callhome, exec,
udprpc, tcprpc, or terminate is specified, lists the possible events
that trigger the specified action. These events are categorized by
component and facility:
[-component {<component>|<id>}][-facility
{<facility> |<id>}]
Lists the possible events in the specified component that
trigger the given action. If facility is specified, lists the events
in the specified component and facility that trigger the given
action.
[-severity {<severity>|<id>}]
Lists the possible events with the specified severity that
trigger the given action.
[-id]
Lists the output with the MessageID number in addition to
BaseID, Severity, and Brief_Description.
222
Command Line Interface Reference for File 8.1
The nas Commands
-component
{-info|{<component>|<id>}[-facility{<facility>
|<id>}]
With the -info option, lists the ids and names of all the
components. If the component is specified, lists the ids and names
of all the facilities under that component. Specifying facility lists
the events that can be generated by the specified facility and
component.
[-id]
Lists the output with the message ID number in addition to
BaseID and Brief_Description.
-severity {-info|<severity>|<id>}]
With the -info option, lists the severity levels. If severity is
specified, lists the events with the specified severity level.
[-component {<component>|<id>} [-facility
<facility>|<id>]
Lists the events filtered by the given severity and component.
If facility is specified lists the events further filtered by the
given facility.
[-id]
Lists the output with the MessageID number in addition to
BaseID, Severity, and Brief_Description.
Note: To receive email notifications sent to multiple recipients,
specify the email addresses within the quotes and separate them with
a comma.
-keyword <keyword>
Lists all events that match the specified keyword.
[-component {<component>|<id>}]
[-facility{<facility> |<id>}]
Lists events filtered by the specified keyword and component.
If facility is specified, lists the events further filtered by the
given facility.
[-severity {<severity>|<id>}]
Lists events filtered by the specified severity.
[-id]
Lists the output with the MessageID number in addition to
BaseID, Severity, and Brief_Description.
nas_event
223
The nas Commands
SEE ALSO
EXAMPLE #1
Configuring Events and Notifications on VNX for File.
After using a text editor to create an event configuration file, to load
the new configuration file into the NAS database, type:
$ nas_event -Load /nas/site/new_eventlog.cfg
EventLog : will load /nas/site/new_eventlog.cfg...done
EXAMPLE #2
To verify that the configuration file was loaded, type:
$ nas_event -Load -info
Loaded config. files:
1:
/nas/sys/nas_eventlog.cfg
2:
/nas/http/webui/etc/web_client_eventlog.cfg
3:
/nas/site/new_eventlog.cfg
EXAMPLE #3
To list actions, type:
$ nas_event -list -action
action
terminate
trap
exec
mail
callhome
logfile
EXAMPLE #4
To list the events that trigger the mail action, type:
$ nas_event -list -action mail
CS_PLATFORM(6)
|--> EventLog(130)
BaseID
Severity
50
EMERGENCY(0)
51
ALERT(1)
52
CRITICAL(2)
EXAMPLE #5
Brief_Description
${text,8,%s}
${text,8,%s}
${text,8,%s}
To list the components, type:
$ nas_event -list -component -info
Id
1
2
5
6
Component
DART
CS_CORE
XML_API
CS_PLATFORM
EXAMPLE #6
To list the facilities under the component DART, type:
$ nas_event -list -component DART -info
DART(1)
|->Id
24
26
224
Facility
ADMIN
CAM
Command Line Interface Reference for File 8.1
The nas Commands
27
36
40
43
45
46
51
52
54
56
58
62
64
68
70
72
73
75
77
78
81
83
84
86
93
96
101
102
107
108
111
115
117
118
122
144
146
148
CFS
DRIVERS
FSTOOLS
IP
KERNEL
LIB
NDMP
NFS
SECURITY
SMB
STORAGE
TIMESYNC
UFS
LOCK
SVFS
XLT
NETLIB
MGFS
VRPL
LDAP
VC
RCPD
VMCAST
CHAMII
USRMAP
DHSM
ACLUPD
FCP
VCS
REP
DPSVC
SECMAP
WINS
DNS
DBMS
PERFSTATS
CEPP
DEDUPE
EXAMPLE #7
To list the events generated by DART in the facility with the ID 146,
type:
$ nas_event -list -component DART -facility 146
DART(1)
|--> CEPP(146)
BaseID
Severity
1
NOTICE(5)
2
ERROR(3)
Brief_Description
CEPP server ${ipaddr,8,%s} of pool ${pool,8,%s} is
${status,8,%s}. Vendor ${vendor,8,%s}, ntStatus
0x${ntstatus,2,%x}.
Error on CEPP server ${ipaddr,8,%s} of pool
${pool,8,%s}: ${status,8,%s}. Vendor ${vendor,8,%s},
ntStatus 0x${ntstatus,2,%x}.
nas_event
225
The nas Commands
3
4
NOTICE(5)
NOTICE(5)
EXAMPLE #8
The CEPP facility is started.
The CEPP facility is stopped.
To list events with severity 4 generated by component CS_CORE and
facility DBMS, and to display the MessageID in the output, type:
$ nas_event -list -severity 4 -component CS_CORE
-facility DBMS -id
CS_CORE(2)
|--> DBMS(122)
MessageID
86444212226
BaseID
2
86444212227
3
EXAMPLE #9
Brief_Description
Db: Compact${compact_option,8,%s}: ${db_name,8,%s}:
Failed: ${db_status,8,%s}.
Db Env: ${db_env,8,%s}: Log Remove: Failed:
${db_status,8,%s}.
To list events filtered by the keyword freeblocks, type:
$ nas_event -list -keyword freeblocks
DART(1)
|--> DBMS(122)
BaseID
Severity
2
CRITICAL(2)
3
ALERT(1)
EXAMPLE #10
Brief_Description
Only ${freeblocks,3,%llu} free blocks in the root
file system (fsid ${fsid,2,%u}) of the VDM
${vdm,8,%s}.
The root file system (fsid ${fsid,2,%u}) of the
VDM ${vdm,8,%s} is full. There are only
${freeblocks,3,%llu} free blocks.
To list events with the keyword data generated in DART with the
severity level 6, type:
$ nas_event -list -keyword data -component DART -severity
6
DART(1)
|--> USRMAP(93)
BaseID
Severity
1
INFO(6)
4
INFO(6)
8
INFO(6)
9
INFO(6)
DART(1)
|--> SECMAP(115)
BaseID
Severity
1
INFO(6)
5.6 format has started.
2
INFO(6)
226
Brief_Description
The Usermapper database has been created.
The Usermapper database has been destroyed.
The migration of the Usermapper database to the
VNX version 5.6 format has
started.
The Usermapper database has been successfully
migrated.
Brief_Description
The migration of the secmap database to the VNX version
The secmap database has been successfully migrated.
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #11
To unload the event configuration file, type:
$ nas_event -Unload /nas/site/new_eventlog.cfg
EventLog : will unload /nas/site/new_eventlog.cfg... done
EXAMPLE #12
To receive email notifications that are sent to multiple recipients, add
the following line to your /nas/sys/eventlog.cfg file:
disposition severity=0-3, mail "nasadmin@nasdocs.emc.com,
helpdesk@nasdocs.emc.com"
EXAMPLE #13
To list the events that trigger a particular trap action, type:
$ nas_event -l -a trap | more
CS_PLATFORM(6)
|--> BoxMonitor(131)
BaseID Severity Brief_Description
1 CRITICAL(2) EPP failed to initialize.
3 CRITICAL(2) Failed to create ${threadname,8,%s} thread.
4 CRITICAL(2) SIB Read failure: ${string,8,%s}
..
CS_PLATFORM(6)
|--> SYR(143)
BaseID Severity Brief_Description
5 INFO(6) The SYR file ${src_file_path,8,%s} with
${dest_extension,8,%s} extension is attached.
nas_event
227
The nas Commands
nas_fs
Manages local file systems for the VNX.
SYNOPSIS
nas_fs
-list [-all]
| -delete <fs_name> [-option <options>][-Force]
| -info [-size] {-all|<fs_name>|id=<fs_id>} [-option <options>]
| -rename <old_name> <new_name> [-Force]
| -size <fs_name>
| -user_files <fs_name>
| -translate <fs_name> -access_policy start
-to {MIXED} -from {NT|NATIVE|UNIX|SECURE}
| -translate <fs_name> -access_policy status
| -xtend <fs_name> {<volume_name>|size=<integer>[T|G|M|%][pool=<pool>]
[storage=<system_name>]} [-option <options>]
| -modify <fs_name> -auto_extend {no|yes [-thin {no|yes}]}
[-hwm <50-99>%][-max_size <integer>[T|G|M]]
| -modify <fs_name> -worm [-default_retention {<integer>{Y|M|D}|infinite}]
[-min_retention {<integer>{Y|M|D}|infinite}]
[-max_retention {<integer>{Y|M|D}|infinite}]
| -modify <fs_name> -worm [-auto_lock {enable[-policy_interval
<integer>{M|D|H}]|disable}]
[-auto_delete {enable|disable}]
[-policy_interval <integer>{M|H|D}]
| -modify <fs_name> -worm -reset_epoch <year>
| -Type <type> <fs_name> -Force
| [-name <name>][-type <type>] –create <volume_name>
[samesize=<fs_name>[:cel=<cel_name>]]
[worm={enterprise|compliance|off} [-default_retention {<integer>{Y|M|D}
|infinite}] [-min_retention {<integer>{Y|M|D}|infinite}]
[-max_retention {<integer>{Y|M|D}|infinite}]]
[log_type={common|split}] [fast_clone_level={1|2}][-option <options>]
| [-name <name>][-type <type>] –create {size=<integer>[T|G|M]
| samesize=<fs_name>[:cel=<cel_name>]}
pool=<pool> [storage=<system_name>][worm={enterprise|compliance|off}
[-default_retention {<integer>{Y|M|D}|infinite}]
[-min_retention {<integer>{Y|M|D}|infinite}]
[-max_retention {<integer>{Y|M|D}|infinite}]]
[log_type={common|split}] [fast_clone_level={1|2}]
[-auto_extend {no|yes} [-thin {no|yes}]
[-hwm <50-99>%][-max_size <integer>[T|G|M]]}]
[-option <options>]
| [-name <name>] -type nmfs -create
DESCRIPTION
228
nas_fs creates, deletes, extends, and lists file systems. nas_fs displays
the attributes of a file system, translates the access policy, enables,
automatic file system extension and thin provisioning capabilities,
Command Line Interface Reference for File 8.1
The nas Commands
manages retention periods, enables automatic file locking and
automatic file deletion, and manages access control level values.
OPTIONS
-list [-all]
Displays a list of file systems and their attributes such as the name,
ID, usage, type, access control level setting, the residing volume, and
the server. The -all option displays all file systems including
system-generated internal file systems. For example, Replicator
internal checkpoints.
Note: The ID is an integer and is assigned automatically, but not always
sequentially, depending on ID availability. The name of a file system might be
truncated if it is more than 19 characters. To display the full file system name,
use the -info option with a file system ID.
File system types are:
• 1=uxfs (default)
• 5=rawfs (unformatted file system)
• 6=mirrorfs (mirrored file system)
• 7=ckpt (checkpoint)
• 8=mgfs (migration file system)
• 100=group file system
• 102=nmfs (nested mount file system)
Note: File system types uxfs, mgfs, nmfs, and rawfs are created by using
nas_fs. Other file system types are created either automatically or with their
specific commands.
-delete <fs_name>
Deletes the file system specified by file system name or ID. A file
system cannot be deleted when it is mounted or part of a group.
[-option <options>]
Specifies the following comma-separated options:
volume
Deletes the file system’s underlying volume structure.
Note: If a checkpoint is created with a volume that has been specified by
size, the underlying volume is deleted when the checkpoint is deleted. If
a file system using a storage pool is deleted, the underlying volume
structure is also deleted.
nas_fs
229
The nas Commands
[-Force]
Forces the deletion of a file system with SnapSure checkpoints
known as the PFS, when a task scheduler such as an automated
scheduler for SnapSure is running or is enabled.
-info [-size] {-all|<fs_name>|id=<fs_id>}
Displays the attributes of a single file system, or all file systems,
including the configuration of associated disks and replication
sessions that are stopped or configured on the file system. If a file
system is mounted, data is reported from the NAS database and the
Data Mover. If a file system is unmounted, data is reported from the
NAS database only.
The -size option also displays the total size of the file system and the
block count in megabytes.
[-option <options>]
Specifies the following comma-separated options:
mpd
Displays the current directory type and translation status for the
specified Multi-Protocol Directory (MPD) file system.
-rename <old_name> <new_name>
Changes the file system name from <old_name> to <new_name>.
[-Force]
Forces the rename of the file system with SnapSure checkpoints
known as the PFS.
-size <fs_name>
Displays the total size of the file system and the block count in
megabytes. The total size of a file system relates to the mounted or
unmounted status of a file system.
-acl <acl_value> <fs_name>
Sets an access control level value that defines the owner of a file
system, and the level of access allowed for users and groups defined
in the access control level table. The nas_acl command provides more
information.
-translate <fs_name> -access_policy start -to {MIXED}
-from {NT |NATIVE|UNIX|SECURE}
Synchronizes the UNIX and Windows permissions on the specified
file system. Prior to executing the -translate option by using
server_mount, mount the specified file system with the MIXED
access-checking policy. The <fs_name> must be a uxfs file system
type mounted as read/write.
230
Command Line Interface Reference for File 8.1
The nas Commands
The policy specified in the -from option instructs the VNX about
which operating system (UNIX or Windows) to derive permissions
from, when migrating to the MIXED or MIXED_COMPAT
access-checking policy (set with server_mount). For example, if you
type UNIX in the -from option, all ACLs are regenerated from the
UNIX mode bits. The policy typed in the -from option does not relate
to the policy previously used by the file system object.
-translate <fs_name> -access_policy status
Prints the status of the access policy translation for the specified file
system.
-xtend <fs_name> <volume_name>
Adds the specified volume to the mounted file system.
-xtend <fs_name> size=<integer>[T|G|M|%]
Adds the volume as specified by its desired size to the file system or
checkpoint. Type an integer within the range of 1 to 1024, then specify
T for terabytes, G for gigabytes (default), M for megabytes, or type an
integer representing the percentage of a file system’s size followed by
the percent sign. The extended volume added to the file system by
the system will have a size equal to or greater than the total size
specified.
!
CAUTION
When executing this command, extends should be performed
incrementally by using like volumes to reduce time consumption.
[pool=<pool>]
Applies the specified storage pool rule set to the volume that
has been added to the mounted file system.
Note: The storage pool is a rule set that contains automatically
created volumes and defines the type of disk volumes used and how
they are aggregated.
[storage=<system_name>]
Specifies the system on which the checkpoint resides. If a
system is not specified, the default system is the one on which
the file system resides. If the file system spans multiple
systems, the default is to use all the systems on which the file
system resides. Use nas_storage -list to obtain attached
system names.
nas_fs
231
The nas Commands
[-option <options>]
Specifies the following comma-separated options:
slice={y|n}
Specifies whether the disk volumes used by the file system may
be shared with other file systems that use a slice. The slice=y
option allows the file system to share disk volumes with other file
systems. The slice=n option gives the new file system exclusive
access to the disk volumes it uses, and is relevant when using
TimeFinder/FS.
When symm_std, symm_std_rdf_src, symm_ata,
symm_ata_rdf_src, symm_ata_rdf_tgt, and symm_std_rdf_tgt
symm_fts, symm_fts_rdf_tgt, and symm_fts_rdf_src pools are
specified, the default is not to slice the volumes, which is
overridden with slice=y. For symm_efd, the default is slice=y,
because TimeFinder/FS is not supported with Flash (EFD) disk
types.
When clar_r1, clar_r5_performance, clar_r5_economy, clar_r6,
clarata_r3, clarata_r6, clarata_r10, clarata_archive, cm_r1,
cm_r5_performance, cm_r5_economy, cm_r6, cmata_r3,
cmata_archive, cmata_r6, cmata_r10, clarsas_archive, clarsas_r6,
clarsas_r10, clarefd_r5, clarefd_r10, cmsas_archive, cmsas_r6,
cmsas_r10, and cmefd_r5 pools are specified, the default for
standard AVM pools is to slice the volumes (slice=y), which is
overridden by using slice=n. The default for mapped pools is not
to slice the volumes (slice=n). Use nas_pool to change the default
slice option.
-modify <fs_name> -auto_extend {no|yes [-thin {yes|no}]}
[-hwm <50-99>%][-max_size <integer> [T|G|M]]]
For an AVM file system, turns automatic file system extension and
thin provisioning on or off, and sets a high water mark and
maximum size for the file system. When file system extension is
turned on, the file system is automatically extended up to the
maximum size specified when the high water mark is reached. The
default for -auto_extend is no.
Thin provisioning reports the maximum file system size to the CIFS,
NFS, and FTP users, even if the actual size of the file system is
smaller. If thin provisioning is disabled, the true file system size and
maximum file system sizes are reported to the system administrator.
Thin provisioning requires that a maximum file system size also be
set.
232
Command Line Interface Reference for File 8.1
The nas Commands
If a file system is created in a storage pool that contains:
◆
Only thick LUNs, use the nas_fs command's -thin option to
enable thin provisioning on the file system.
◆
Only thin LUNs, using the nas_fs command's -thin option is not
recommended. It is redundant, but allowed, for a thin file system
to be built on thin LUNs.
◆
Both thick and thin LUNs, the file system may be built on either
thick LUNs, thin LUNs, or both thick and thin LUNs. Using the
nas_fs command's -thin option may be redundant if the file
system uses thin LUNs.
Automatic File System Extension cannot be used for any file system
that is part of an RDF configuration (for example, file systems on
Data Movers configured with an RDF standby). Do not use the nas_fs
command with the -auto_extend option for file systems associated
with RDF configurations.
[-hwm <50-99>%]
Specifies the size threshold that must be reached before the file
system is automatically extended. Type an integer between the
range of 50 and 99 to represent the percentage of file system
usage. The default is 90.
[-max_size <integer> [T|G|M]]
Sets the maximum file system size to which a file system can be
extended. Type an integer and specify T for terabytes, G for
gigabytes (default), or M for megabytes. If the -max_size option is
not specified, then it defaults to the maximum limit of the file
system size which is 16 terabytes.
-modify <fs_name> -worm [-default_retention
{<integer>{Y|M|D}|infinite}]
[-min_retention {<integer>{Y|M|D}|infinite}]
[-max_retention {<integer>{Y|M|D}|infinite}]
For an FLR-enabled file system, manages retention periods.
[-default_retention {<integer>{Y|M|D}|infinite}]
Sets a default retention period that is used in an FLR-enabled file
system when a file is locked and a retention period is not
specified. This value must be greater than or equal to the
-min_retention option, and less than or equal to the
-max_retention option. Type an integer and specify Y for years,
M for months, or D for days. The default value is infinite. Setting
infinite means that the files can never be deleted.
nas_fs
233
The nas Commands
[-min_retention {<integer>{Y|M|D}|infinite}]
Sets the minimum retention period that files on an FLR-enabled
file system can be locked and protected from deletion. This value
must be less than or equal to the -max_retention option. Type an
integer and specify Y for years, M for months, or D for days. The
default value is one day. Setting infinite means that the files can
never be deleted.
[-max_retention {<integer>{Y|M|D}|infinite}]
Sets the maximum retention period that files on an FLR-enabled
file system can be locked and protected from deletion. Type an
integer and specify Y for years, M for months, or D for days. The
default value is infinite. Setting infinite means that the files can
never be deleted.
-modify <fs_name> -worm [-auto_lock {enable
[-policy_interval <integer>{M|D|H}]|disable}]
[-auto_delete {enable|disable}]
[-policy_interval <integer>{M|D|H}]
For an FLR-enabled file system, manages automatic file locking and
automatic file deletion.
[-auto_lock {enable|disable}]
Specifies whether automatic file locking for all files in an
FLR-enabled file system is on or off. When enabled, auto-locked
files are set with the default retention period value.
[-policy_interval <integer>{M|D|H}]
Specifies an interval for how long to wait after files are modified
before the files are automatically locked in an FLR-enabled file
system. Type an integer and specify M for minutes, D for days, or
H for hours. The policy interval has a minimum value of 1 minute
and a maximum value of 366 days. The default value is 1 hour.
[-auto_delete {enable|disable}]
Specifies whether automatically deleting locked files from an
FLR-enabled file system once the retention period has expired is
on or off.
-modify <fs_name> -worm -reset_epoch <year>
For an FLR-enabled file system, specifies the base year used for
calculating the retention date of a file beyond 2038. Type an
integer between the range of 2000 and 2037. The default value is
2003. The maximum value for the retention period is December
234
Command Line Interface Reference for File 8.1
The nas Commands
31, 2104 11:59:59 p.m. Trying to set a date beyond this value
generates an error. Refer to Using VNX File-Level Retention for
additional information.
-Type <type> <fs_name> -Force
Changes the file system type from the one of <fs_name> to the new
specified <type>.
!
CAUTION
Converting uxfs to rawfs is prevented.
!
CAUTION
The conversion from rawfs to uxfs will fail with "Error 3105:
invalid file system specified” because a uxfs is not available on the
rawfs. However, if the user initially creates a rawfs, and restores an
NDMP volume backup on the rawfs, then the conversion from the
rawfs to a uxfs will be successful.
CREATING A FILE
SYSTEM
File systems can be created by using:
◆
A volume specified by name
◆
A volume specified by its size and desired storage pool
◆
An existing local or remote file system with the samesize option
◆
An existing local or remote file system with the samesize option
and by using space from the available storage pool
[-name <name>][-type <type>] -create <volume_name>
Creates a file system on the specified volume and assigns an optional
name to the file system. If a name is not specified, one is assigned
automatically.
A file system name cannot:
◆
Begin with a dash (-)
◆
Be comprised entirely of integers
◆
Be a single integer
◆
Contain the word root or contain a colon (:)
The -type option assigns the file system type to be uxfs (default),
mgfs, or rawfs.
nas_fs
235
The nas Commands
[samesize=<fs_name>[:cel=<cel_name>]]
Specifies that the new file system must be created with the same
size as the specified local or remote file system. When using the
samesize option by using the options described below, the slice=
must be set to y.
[worm={enterprise|compliance|off}]
Enables storage capability on a new file system. The option can
only be specified when creating a new file system; existing file
systems cannot be converted. After a file system is enabled, it is
persistently marked as such until the time it is deleted.
Note: The compliance file system cannot be deleted if it has protected files.
!
CAUTION
The Enterprise version of this feature is intended for
self-regulated archiving. The administrator is considered as a
trusted user and the capability does not protect the archived
data from the administrator’s actions. If the administrator
attempts to delete the file system, the file system issues a
warning message and prompts the administrator to confirm the
operation. This version is not intended for high-end
compliance applications such as pharmaceuticals, aerospace, or
finance.
As part of enabling file-level retention (worm) on a new file
system, you can also set these retention period options:
[-default_retention {<integer>{Y|M|D}|infinite}]
Sets a default retention period that is used in an FLR-enabled file
system when a file is locked and a retention period is not
specified. This value must be greater than or equal to the
-min_retention option, and less than or equal to the
-max_retention option. Type an integer and specify Y for years, M
for months, D for days, or infinite. The default value is infinite
which means that the files can never be deleted.
[-min_retention {<integer>{Y|M|D}|infinite}]
Setsthe minimum retention period that files on an FLR-enabled
file system can be locked and protected from deletion. This value
must be less than or equal to the -max_retention option. Type an
integer and specify Y for years, M for months, D for days, or
infinite. The default value is 1 day. Setting infinite means that
the files can never be deleted.
236
Command Line Interface Reference for File 8.1
The nas Commands
[-max_retention {<integer>{Y|M|D}|infinite}]
Sets the maximum retention period that files on an FLR-enabled
file system can be locked and protected from deletion. Type an
integer and specify Y for years, M for months, D for days, or
infinite. The default value is infinite which means that the files
can never be deleted.
log_type={common|split}
Specifies the type of log file associated with the file system. Log
files can be either shared (common) or uniquely assigned to
individual file systems(split). For SRDF Async or STAR feature,
split option is strongly recommended to avoid fsck before
mounting a BCV file system on SiteB or SiteC.
[fast_clone_level={1|2}]
fast_clone_level=2 enables ability to create fast clone of a fast
clone (also called as the second level fast clone) on the file system.
fast_clone_level=1 enables ability to create a fast clone. File level
retention and fast clone creation cannot be enabled together on a
file system. Enabling split log implies fast_clone_level=2, if file
level retention is not enabled on the file system. Replication
sessions cannot be created between two different fast_clone_level
capable file systems.
Note: fast_clone_level=1 indicates that a fast clone can be created on
the file system and it is the default option if nothing is specified.
[-option <options>]
Specifies the following comma-separated options:
nbpi=<number>
The number of bytes per inode block. The default is 8192 bytes.
mover=<movername>
Assigns an optional Data Mover to build a file system. If no Data
Mover is assigned, the system will automatically pick the first
available Data Mover to build the file system.
slice={y|n}
Specifies whether the disk volumes used by the new file system
may be shared with other file systems by using a slice. The
slice=y option allows the file system to share disk volumes with
other file systems. The slice=n option ensures that the new file
system has exclusive access to the disk volumes it uses, and is
relevant when using TimeFinder/FS.
nas_fs
237
The nas Commands
When symm_std, symm_std_rdf_src, symm_ata,
symm_ata_rdf_src, symm_ata_rdf_tgt, and symm_std_rdf_tgt
symm_fts, symm_fts_rdf_tgt, and symm_fts_rdf_src pools are
specified, the default is not to slice the volumes. When slice=y is
specified, it overrides the default. For symm_efd, the default is
slice=y, because TimeFinder/FS is not supported with Flash disk
types.
When clar_r1, clar_r5_performance, clar_r5_economy, clar_r6,
clarata_r3, clarata_r6, clarata_r10, clarata_archive, cm_r1,
cm_r5_performance, cm_r5_economy, cm_r6, cmata_r3,
cmata_archive, cmata_r6, cmata_r10, clarsas_archive, clarsas_r6,
clarsas_r10, clarefd_r5, clarefd_r10, cmsas_archive, cmsas_r6,
cmsas_r10, and cmefd_r5 pools are specified, the default for
standard AVM pools is to slice the volumes (slice=y), which is
overridden by using slice=n. The default for mapped pools is not
to slice the volumes (slice=n). Use nas_pool to change the default
slice option.
id=<desired_id>
Specifies the ID to be assigned to the new file system. If a file
system already exists with the specified ID, a warning is
displayed indicating that the ID is not available, and the new file
system is assigned the next available ID.
[-name <name>][-type <type>] -create {size=
<integer>[T|G|M]|samesize=<fs_name>[:cel=<cel_name>]}
pool=<pool>
Creates a file system on the volume specified by its desired size and
storage pool or by using the same size as a specified local or remote
file system. Also assigns an optional name and file system type to a
file system. If a name is not specified, one is assigned automatically. A
file system name can be up to 240 characters, but cannot begin with a
dash (-), be comprised entirely of integers or be a single integer,
contain the word root or contain a colon (:). Available file system
types are uxfs (default), mgfs, or rawfs.
When using the samesize option by using the options described
below, the slice= should be set to y. The new file system is created
with the same size as the specified file system.
The pool option specifies a rule set for the new file system that
contains automatically created volumes and defines the type of disk
volumes used and how they are aggregated. Storage pools are system
defined (storage pool description provides more information) or user
defined.
238
Command Line Interface Reference for File 8.1
The nas Commands
[worm={enterprise|compliance|off}]
Enables the storage capability on the new file system. The
capability can only be specified when creating a new file system;
existing file systems cannot be converted. After a file system is
enabled, it is persistently marked as such until the time when it is
deleted.
!
CAUTION
The Enterprise version of this feature is intended for
self-regulated archiving. The administrator is considered as a
trusted user and this feature does not protect the archived data
from the administrator’s actions. If the administrator attempts
to delete a file system, the file system issues a warning message
and prompts the administrator to confirm the operation. This
version of this feature is not intended for high-end compliance
applications such as pharmaceuticals, aerospace, or finance.
As part of enabling file-level retention (worm) on a new file
system, you can also set these retention period options:
[-default_retention {<integer>{Y|M|D}|infinite}]
Sets a default retention period that is used in an FLR-enabled file
system when a file is locked and a retention period is not
specified. This value must be greater than or equal to the
-min_retention option, and less than or equal to the
-max_retention option. Type an integer and specify Y for years, M
for months, D for days, or infinite. The default value is infinite
which means that the files can never be deleted.
[-min_retention {<integer>{Y|M|D}|infinite}]
Setsthe minimum retention period that files on an FLR-enabled
file system can be locked and protected from deletion. This value
must be less than or equal to the -max_retention option. Type an
integer and specify Y for years, M for months, D for days, or
infinite. The default value is 1 day. Setting infinite means that
the files can never be deleted.
[-max_retention {<integer>{Y|M|D}|infinite}]
Sets the maximum retention period that files on an FLR-enabled
file system can be locked and protected from deletion. Type an
integer and specify Y for years, M for months, D for days, or
infinite. The default value is infinite which means that the files
can never be deleted.
nas_fs
239
The nas Commands
[storage=<system_name>]
Specifies the system on which the file system resides. Use
nas_storage -list to obtain a list of the available system names.
[-auto_extend {no|yes} [-thin {no|yes}]
For an AVM file system, turns automatic file system extension
and thin provisioning on or off, and sets a high water mark and
maximum size for the file system. When automatic file system
extension is turned on, the file system is automatically extended
up to the maximum size specified when the high water mark is
reached. The default for -auto_extend is no.
Thin provisioning reports the maximum file system size to the
CIFS, NFS, and FTP users, even if the actual size of the file system
is smaller. If disabled, the true file system size and maximum file
system sizes are reported to the system administrator. Thin
provisioning requires that a maximum file system size also be set.
If a file system is created in a storage pool that contains:
• Only thick LUNs, use the nas_fs command's -thin option to
enable thin provisioning on the file system.
• Only thin LUNs, using the nas_fs command's -thin option is
not recommended. It is redundant, but allowed, for a thin file
system to be built on thin LUNs.
• Both thick and thin LUNs, the file system may be built on
either thick LUNs, thin LUNs, or both thick and thin LUNs.
Using the nas_fs command's -thin option may be redundant if
the file system uses thin LUNs.
Note: SRDF pools are not supported.
[-hwm <50-99>%]
Specifies the size threshold that must be reached before the file
system is automatically extended. Type an integer between the
range of 50 and 99 to represent the percentage of file system
usage. The default is 90.
[-max_size <integer> [T|G|M]]
Sets the maximum file system size to which a file system can be
extended. Type an integer and specify T for terabytes, G for
gigabytes (default), or M for megabytes. If the -max_size option is
not specified, then it defaults to the maximum limit of the file
240
Command Line Interface Reference for File 8.1
The nas Commands
system size which is 16 terabytes. Maximum size must be set to
enable thin provisioning. The maximum size is what is presented
to users as the file system size through thin provisioning.
[-option <options>]
Specifies the following comma-separated options:
nbpi=<number>
The number of bytes per inode block. The default is 8192 bytes.
mover=<movername>
Assigns an optional Data Mover on which to build a file system.
If no Data Mover is assigned, the system will automatically pick
the first available Data Mover to build the file system.
slice={y|n}
Specifies whether the disk volumes used by the new file system
may be shared with other file systems by using a slice. The
slice=y option allows the file system to share disk volumes with
other file systems. The slice=n option ensures that the new file
system has exclusive access to the disk volumes it uses, and is
relevant when using TimeFinder/FS.
When symm_std, symm_std_rdf_src, symm_ata,
symm_ata_rdf_src, symm_ata_rdf_tgt, and symm_std_rdf_tgt
symm_fts, symm_fts_rdf_tgt, and symm_fts_rdf_src pools are
specified, the default is not to slice the volumes, which is
overridden with slice=y. For symm_efd, the default is slice=y,
because TimeFinder/FS is not supported with Flash disk types.
When clar_r1, clar_r5_performance, clar_r5_economy, clar_r6,
clarata_r3, clarata_r6, clarata_r10, clarata_archive, cm_r1,
cm_r5_performance, cm_r5_economy, cm_r6, cmata_r3,
cmata_archive, cmata_r6, cmata_r10, clarsas_archive, clarsas_r6,
clarsas_r10, clarefd_r5, clarefd_r10, cmsas_archive, cmsas_r6,
cmsas_r10, and cmefd_r5 pools are specified, the default for
standard AVM pools is to slice the volumes (slice=y), which is
overridden by using slice=n. The default for mapped pools is not
to slice the volumes (slice=n). Use nas_pool to change the default
slice option.
[-name <name>] -type nmfs -create
Creates a nested mount file system (NMFS) that can be used to
combine multiple uxfs file systems into a single virtual file
system. The NMFS can then be mounted and exported as a single
share or mount point.
nas_fs
241
The nas Commands
SEE ALSO
Managing Volumes and File Systems with VNX Automatic Volume
Management, Managing Volumes and File Systems for VNX Manually,
Using VNX File-Level Retention, Controlling Access to System Objects on
VNX, Using VNX Replicator, fs_ckpt, fs_timefinder, nas_acl, nas_rdf,
nas_volume, server_export, server_mount, fs_dedupe, and
server_mountpoint.
SYSTEM OUTPUT
The number associated with the storage device is dependent on the
attached system. VNX for block displays a prefix of APM before a set
of integers, for example, APM00033900124-0019. Symmetrix systems
appear as 002804000190-003C. The outputs displayed in the examples
use a VNX for block.
VNX for block supports the following system-defined storage pools:
clar_r1, clar_r5_performance, clar_r5_economy, clar_r6, clarata_r3,
clarata_r6, clarata_r10, clarata_archive, cm_r1, cm_r5_performance,
cm_r5_economy, cm_r6, cmata_r3, cmata_archive, cmata_r6,
cmata_r10, clarsas_archive, clarsas_r6, clarsas_r10, clarefd_r5,
clarefd_r10, cmsas_archive, cmsas_r6, cmsas_r10, and cmefd_r5.
VNXs with a Symmetrix system support the following
system-defined storage pools: symm_std_rdf_src, symm_std,
symm_ata, symm_ata_rdf_src, symm_ata_rdf_tgt,
symm_std_rdf_tgt, symm_efd, symm_fts, symm_fts_rdf_tgt, and
symm_fts_rdf_src.
For user-defined storage pools, the difference in output is in the disk
type. Disk types when using a Symmetrix are STD, R1STD, R2STD,
BCV, R1BCV, R2BCV, ATA, R1ATA, R2ATA, BCVA, R1BCA,
R2BCA, EFD, FTS, R1FTS, R2FTS, R1BCF, R2BCF, and BCVF.
Disk types when using VNX for block are: CLSTD, CLEFD, CLATA,
MIXED (indicates that tiers used in the pool contain multiple disk
types), Performance, Capacity, and Extreme_performance and for
VNX for block involving mirrored disks are: CMEFD, CMSTD,
CMATA, Mirrored_mixed, Mirrored_performance,
Mirrored_capacity, and Mirrored_extreme_performance.
EXAMPLE #1
id
name
acl
in_use
type
worm
242
=
=
=
=
=
=
To create a file system named ufs1 on metavolume mtv1, type:
$ nas_fs -name ufs1 -create mtv1
37
ufs1
0
False
uxfs
enterprise with no protected files
Command Line Interface Reference for File 8.1
The nas Commands
worm_clock= Clock not initialized
worm Max Retention Date= NA
worm Default Retention Period= infinite
worm Minimum Retention Period= 1 Day
worm Maximum Retention Period= infinite
FLR Auto_lock= off
FLR Policy Interval= 3600 seconds
FLR Auto_delete= off
FLR Epoch Year= 2003
volume
= mtv1
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication = off
stor_devs = APM00042000818-0012,APM00042000818-0014
disks
= d7,d9
Where:
Value
id
name
acl
in_use
type
worm
volume
pool
rw_servers
ro_servers
rw_vdms
ro_vdms
worm_clock
worm Max Retention
Date
worm Default
Retention Period
Definition
Automatically assigned ID of a file system.
Name assigned to a file system.
Access control value assigned to the file system.
If a file system is registered into the mount table of a Data Mover.
Type of file system. -list provides information for a description of the types.
Write Once Read Many (WORM) state of file system. It states whether the state is disabled or
set to either compliance or enterprise.
Volume on which a file system resides.
Storage pool for the file system.
Servers with read/write access to a file system.
Servers with read-only access to a file system.
VDM servers with read/write access to a file system.
VDM servers with read-only access to a file system.
Software clock maintained by the file system. The clock functions only when the file system is
mounted read/write.
Time when the protected files expire. The file system can be deleted only after this date. The
special values returned are:
• 3 — The file system is FLR-E with protected files.
• 2 — The file system is scanning for max_retention period.
• 1 — The default value (no protected files created).
• 0 — Infinite retention period (if the server is up and running).
Specifies a default retention period that files on an FLR-enabled file system will be locked and
protected from deletion. If you do not set either a minimum retention period or a maximum
retention period, this default value is used when file-level retention is enabled.
nas_fs
243
The nas Commands
worm Minimum
Retention Period
worm Maximum
Retention Period
FLR Auto_lock
FLR Policy Interval
FLR Auto_delete
FLR Epoch Year
volume
pool
rw_servers
ro_servers
rw_vdms
ro_vdms
auto_ext
deduplication
stor_devs
disks
Specifies the minimum retention period that files on an FLR-enabled file system will be locked
and protected from deletion.
Specifies the maximum retention period that files on an FLR-enabled file system will be locked
and protected from deletion.
Specifies whether automatic file locking for all files in an FLR-enabled file system is on or off.
Specifies an interval for how long to wait after files are modified before the files are automatically
locked and protected from deletion.
Specifies whether locked files are automatically deleted once the retention period has expired.
Specifies the base year used for calculating the retention date of a file beyond 2038. When a file
is locked with its atime set to a value greater than the FLR Epoch Year value, the file’s retention
date is set to the file’s atime value. When a file is locked with its atime set to a value less than the
FLR Epoch Year value, the file’s retention date is set to 2038 + (YEAR(atime) - 1970).
Volume on which a file system resides.
Storage pool for the file system.
Servers with read/write access to a file system.
Servers with read-only access to a file system.
VDM servers with read/write access to a file system.
VDM servers with read-only access to a file system.
Indicates whether auto-extension and thin provisioning are enabled.
Deduplication state of the file system. The file data is transferred to the storage which performs
the deduplication and compression on the data. The states are:
• On — Deduplication on the file system is enabled.
• Suspended — Deduplication on the file system is suspended. Deduplication does not
perform any new space reduction but the existing files that were reduced in space remain the
same.
• Off — Deduplication on the file system is disabled. Deduplication does not perform any new
space reduction and the data is now reduplicated.
System devices associated with a file system.
Disks on which the metavolume resides.
Note: The deduplication state is unavailable when the file system is
unmounted.
EXAMPLE #2
To display information about a file system by using the file system
ID, 14, using the clar_mapped_pool VNX mapped pool, type:
$ nas_fs -info id=14
id
= 14
name
= ufs2_flre
acl
= 0
in_use
= True
type
= uxfs
worm
= enterprise with no protected files
worm_clock= Fri Jul 29 07:56:42 EDT 2011
worm Max Retention Date= No protected files created
worm Default Retention Period= 10 Years
worm Minimum Retention Period= 30 Days
244
Command Line Interface Reference for File 8.1
The nas Commands
worm Maximum Retention Period= 10 Years
FLR Auto_lock= off
FLR Policy Interval= 3600 seconds
FLR Auto_delete= off
FLR Epoch Year= 2003
volume
= v117
pool
= clar_mapped_pool
member_of = root_avm_fs_group_50
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= Off
thin_storage = True
tiering_policy = Auto-tier
compressed= False
mirrored = False
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
disk=d16
stor_dev=BB005056830430-0019 addr=c0t1l9
server=server_2
disk=d16
stor_dev=BB005056830430-0019 addr=c16t1l9
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c0t1l6
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c16t1l6
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c0t1l5
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c16t1l5
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c0t1l0
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c16t1l0
server=server_2
Where:
Value
Definition
thin_storage
Indicates whether the VNX for block system uses thin provisioning. Values are: True, False, Mixed.
tiering_policy Indicates the tiering policy in effect. If the initial tier and the tiering policy are the same, the values are:
Auto-Tier, Highest Available Tier, Lowest Available Tier. If the initial tier and the tiering policy are not the
same, the values are: Auto-Tier/No Data Movement, Highest Available Tier/No Data Movement, Lowest
Available Tier/No Data Movement.
compressed
Indicates whether data is compressed. Values are True, False, Mixed (indicates some of the LUNs, but
not all, are compressed).
mirrored
Indicates whether the disk is mirrored.
EXAMPLE #3
To display a list of file systems, type:
$ nas_fs -list
id
1
2
3
inuse type acl
n
1
0
y
1
0
n
5
0
volume
20
50
83
name
server
root_fs_1
root_fs_common
1
root_fs_ufslog
nas_fs
245
The nas Commands
5
6
7
8
9
10
11
13
14
n
n
n
n
y
n
y
y
y
5
5
5
5
1
5
1
1
1
0
0
0
0
0
0
0
0
0
EXAMPLE #4
103
104
105
106
22
108
112
115
117
root_fs_d3
root_fs_d4
root_fs_d5
root_fs_d6
root_fs_2
root_panic_reserve
ufs1
ufs1_flr
ufs2_flre
1
1
1
1
To list all the file systems including internal checkpoints, type:
$ nas_fs -list -all
id
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
212
213
214
230
231
342
343
986
987
988
989
1343
1344
1345
1346
1347
1348
1349
246
inuse type acl
n
1
0
y
1
0
y
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
n
1
0
y
1
0
n
5
0
n
5
0
y
1
0
y
101 0
n
1
0
y
1
0
y
1
0
y
1
0
y
1
0
n
11 0
y
7
0
y
1
0
y
5
0
n
11 0
y
7
0
y
7
0
y
1
0
n
11 0
y
7
0
y
7
0
volume
24
26
28
30
32
34
36
38
40
42
44
46
48
50
52
54
87
90
315
0
318
346
349
560
563
0
1722
1725
1728
0
2351
2351
2354
0
2358
2358
name
server
root_fs_1
root_fs_2
1
root_fs_3
2
root_fs_4
root_fs_5
root_fs_6
root_fs_7
root_fs_8
root_fs_9
root_fs_10
root_fs_11
root_fs_12
root_fs_13
root_fs_14
root_fs_15
root_fs_common
2,1
root_fs_ufslog
root_panic_reserve
v2src1
1
root_avm_fs_group_3
v2dst1
v2srclun1
1
v2dstlun1
2
root_fs_vdm_srcvdm1 1
root_fs_vdm_srcvdm2 1
vpfs986
gstest
1
src1
1
dst1
1
vpfs1343
root_rep_ckpt_342_2 1
root_rep_ckpt_342_2 1
root_fs_vdm_srcvdm1 1
vpfs1347
root_rep_ckpt_1346_ 1
root_rep_ckpt_1346_ 1
Command Line Interface Reference for File 8.1
The nas Commands
1350
1354
1358
1359
1360
1361
1362
1363
1364
1365
1366
1367
1368
1369
1370
1371
1372
1376
1380
1381
1382
1383
1384
1385
1386
1387
1388
1389
1390
1391
1392
1393
1394
1395
1396
1397
1398
1399
1400
1401
1402
1403
1404
1405
1406
y
n
n
y
y
n
n
n
n
y
y
y
n
n
n
y
y
y
n
y
y
y
n
y
y
y
n
y
y
n
y
y
n
y
y
y
y
y
n
y
y
n
y
y
y
1
1
11
7
7
1
11
7
7
1
7
7
11
7
7
1
1
1
11
7
7
1
11
7
7
1
1
1
1
11
7
7
11
7
7
7
1
1
11
7
7
11
7
7
7
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
2367
2374
0
2383
2383
2385
0
2388
2388
2392
2383
2383
0
2395
2395
2399
2401
2406
0
2414
2414
2416
0
2419
2419
2423
2425
2427
2429
0
2432
2432
0
2435
2435
2432
2439
2441
0
2444
2444
0
2447
2447
2444
fs1
fs1_replica1
vpfs1358
root_rep_ckpt_1350_
root_rep_ckpt_1350_
fs1_replica2
vpfs1362
root_rep_ckpt_1361_
root_rep_ckpt_1361_
fs1365
root_rep_ckpt_1350_
root_rep_ckpt_1350_
vpfs1368
root_rep_ckpt_1354_
root_rep_ckpt_1354_
root_fs_vdm_v1
f1
root_fs_vdm_v1_repl
vpfs1380
root_rep_ckpt_1372_
root_rep_ckpt_1372_
f1_replica1
vpfs1384
root_rep_ckpt_1383_
root_rep_ckpt_1383_
cworm
cworm1
fs2
fs3
vpfs1391
root_rep_ckpt_1389_
root_rep_ckpt_1389_
vpfs1394
root_rep_ckpt_1390_
root_rep_ckpt_1390_
fs2_ckpt1
fs4
fs5
vpfs1400
root_rep_ckpt_1398_
root_rep_ckpt_1398_
vpfs1403
root_rep_ckpt_1399_
root_rep_ckpt_1399_
fs4_ckpt1
v9
v9
v9
1
v9
v9
1
v40
2
v40
v40
v41
v41
v41
1
2
2
2
2
2
2
2
2
2
2
2
2
2
2
Note: NDMP and Replicator internal checkpoints can be identified by
specific prefixes in the filename. Using VNX SnapSure provides more
information for internal checkpoints naming formats.
nas_fs
247
The nas Commands
EXAMPLE #5
To create a uxfs file system named ufs20 on system BB005056830430,
with a size of 1 GB, by using the clar_r5_performance pool and
allowing the file system to share disk volumes with other file
systems, type:
$ nas_fs -name ufs20 -type uxfs -create size=1G
pool=clar_r5_performance storage=BB005056830430 -option
slice=y
id
= 15
name
= ufs20
acl
= 0
in_use
= False
type
= uxfs
worm
= off
volume
= v119
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= unavailable
stor_devs =
BB005056830430-0018,BB005056830430-0017,BB005056830430-0014,BB005056830430-0011
disks
= d15,d14,d11,d8
Where:
Value
member_of
Definition
File system group to which the file system belongs.
EXAMPLE #1 provides a description of command output.
EXAMPLE #6
To create a rawfs file system named ufs3 with the same size as the file
system ufs1 by using the clar_r5_performance pool and allowing the
file system to share disk volumes with other file systems, type:
$ nas_fs -name ufs3 -type rawfs -create samesize=ufs1
pool=clar_r5_performance storage=APM00042000818 -option
slice=y
id
=
name
=
acl
=
in_use
=
type
=
worm
=
volume
=
pool
=
member_of =
rw_servers=
248
39
ufs3
0
False
rawfs
off
v173
clar_r5_performance
root_avm_fs_group_3
Command Line Interface Reference for File 8.1
The nas Commands
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication = unavailable
stor_devs = APM00042000818-001F,APM00042000818-001D,APM00042000818-0019,APM00042
000818-0016
disks
= d20,d18,d14,d11
EXAMPLE #1 and EXAMPLE #3 provide a description of command
outputs.
EXAMPLE #7
To create a uxfs file system named ufs4, with a size of 100 GB, by
using the clar_r5_performance pool, with file-level retention set to
enterprise, 4096 bytes per inode, and server_3 for file system
building, type:
$ nas_fs -name ufs4 -create size=100G
pool=clar_r5_performance worm=enterprise -option
nbpi=4096,mover=server_3
id
= 16
name
= ufs4
acl
= 0
in_use
= False
type
= uxfs
worm
= enterprise with no protected files
worm_clock= Clock not initialized
worm Max Retention Date= NA
worm Default Retention Period= infinite
worm Minimum Retention Period= 1 Day
worm Maximum Retention Period= infinite
FLR Auto_lock= off
FLR Policy Interval= 3600 seconds
FLR Auto_delete= off
FLR Epoch Year= 2003
volume
= v121
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= unavailable
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
To ensure retention of protected files, it can also be set to compliance by typing:
nas_fs
249
The nas Commands
$ nas_fs -name ufs4 -create size=100G
pool=clar_r5_performance worm=compliance -option
nbpi=4096,mover=server_3
id
= 17
name
= ufs4
acl
= 0
in_use
= False
type
= uxfs
worm
= compliance with no protected files
worm_clock= Clock not initialized
worm Max Retention Date= NA
worm Default Retention Period= infinite
worm Minimum Retention Period= 1 Day
worm Maximum Retention Period= infinite
FLR Auto_lock= off
FLR Policy Interval= 3600 seconds
FLR Auto_delete= off
FLR Epoch Year= 2003
volume
= v123
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= unavailable
stor_devs =
BB005056830430-0018,BB005056830430-0017,BB005056830430-0014,BB005056830430-0011
disks
= d15,d14,d11,d8
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #8
To create a file system named ufs30, with a size of 1 GB, by using the
clar_r5_performance pool, with file-level retention set to enterprise,
a minimum retention period of 30 days, and a maximum retention
period of 10 years, type:
$ nas_fs -name ufs30 -create size=1G
pool=clar_r5_performance worm=enterprise -min_retention
30D -max_retention 10Y
id
= 18
name
= ufs30
acl
= 0
in_use
= False
type
= uxfs
worm
= enterprise with no protected files
worm_clock= Clock not initialized
worm Max Retention Date= NA
worm Default Retention Period= 10 Years
250
Command Line Interface Reference for File 8.1
The nas Commands
worm Minimum Retention Period= 30 Days
worm Maximum Retention Period= 10 Years
FLR Auto_lock= off
FLR Policy Interval= 3600 seconds
FLR Auto_delete= off
FLR Epoch Year= 2003
volume
= v125
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= unavailable
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #9
To display information about file system ufs4, type:
$ nas_fs -info ufs4
id
= 16
name
= ufs4
acl
= 0
in_use
= False
type
= uxfs
worm
= enterprise with no protected files
worm_clock= Clock not initialized
worm Max Retention Date= NA
worm Default Retention Period= infinite
worm Minimum Retention Period= 1 Day
worm Maximum Retention Period= infinite
FLR Auto_lock= off
FLR Policy Interval= 3600 seconds
FLR Auto_delete= off
FLR Epoch Year= 2003
volume
= v121
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= unavailable
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
nas_fs
251
The nas Commands
disks
= d16,d13,d12,d7
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #10
To create a uxfs file system named ufs40, with a size of 10 GB, by
using the clar_r5_performance pool, and an ID of 8000 assigned to
ufs1, type:
$ nas_fs -name ufs40 -type uxfs -create size=10G
pool=clar_r5_performance -option slice=y,id=8000
id
= 8000
name
= ufs40
acl
= 0
in_use
= False
type
= uxfs
worm
= off
volume
= v127
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= unavailable
stor_devs =
BB005056830430-0018,BB005056830430-0017,BB005056830430-0014,BB005056830430-0011
disks
= d15,d14,d11,d8
EXAMPLE #11
To create a uxfs file system named ufs41, with a size of 10 GB, by
using the clar_r5_performance pool, and an ID of 8000 assigned to
ufs1, type:
$ nas_fs -name ufs41 -type uxfs -create size=10G
pool=clar_r5_performance -option slice=y,id=8000
id
=
name
=
acl
=
in_use
=
type
=
worm
=
volume
=
pool
=
member_of =
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext =
252
8001
ufs41
0
False
uxfs
off
v129
clar_r5_performance
root_avm_fs_group_3
no,thin=no
Command Line Interface Reference for File 8.1
The nas Commands
deduplication
= unavailable
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
Warning 17716815881: unavailable id : 8000.
Note: The warning output is displayed if the desired ID is not available.
Because id=8000 was used in Example 10, the system set the id to 8001
instead.
EXAMPLE #12
To view the size of ufs1, type:
$ nas_fs -size ufs1
total = 945 avail = 945 used = 1 ( 0% ) (sizes in MB) ( blockcount = 2097152 )
volume: total = 1024 (sizes in MB) ( blockcount = 2097152 ) avail = 944 used =
80 ( 8% )
When a file system is mounted, the size info for the volume and a file
system, as well as the number of blocks that are used are displayed.
Where:
Value
total
blockcount
EXAMPLE #13
Definition
Total size of the file system.
Total number of blocks used.
To rename a file system from ufs1 to ufs5, type:
$ nas_fs -rename ufs1 ufs5
id
= 11
name
= ufs5
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= v112
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= Off
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
disk=d16
stor_dev=BB005056830430-0019 addr=c0t1l9
server=server_2
disk=d16
stor_dev=BB005056830430-0019 addr=c16t1l9
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c0t1l6
server=server_2
nas_fs
253
The nas Commands
disk=d13
disk=d12
disk=d12
disk=d7
disk=d7
stor_dev=BB005056830430-0016
stor_dev=BB005056830430-0015
stor_dev=BB005056830430-0015
stor_dev=BB005056830430-0010
stor_dev=BB005056830430-0010
addr=c16t1l6
addr=c0t1l5
addr=c16t1l5
addr=c0t1l0
addr=c16t1l0
server=server_2
server=server_2
server=server_2
server=server_2
server=server_2
EXAMPLE #1 and EXAMPLE #3 provide a description of command
outputs.
EXAMPLE #14
To extend the file system, ufs1, with the volume, emtv2b, type:
$ nas_fs -xtend ufs1 emtv2b
id
= 38
name
= ufs1
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= v171
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication = off
stor_devs = APM00042000818-001F,APM00042000818-001D,APM00042000818-0019,APM00042
000818-0016,APM00042000818-001C
disks
= d20,d18,d14,d11,d17
disk=d20
stor_dev=APM00042000818-001F addr=c0t1l15
server=server_2
disk=d20
stor_dev=APM00042000818-001F addr=c32t1l15
server=server_2
disk=d18
stor_dev=APM00042000818-001D addr=c0t1l13
server=server_2
disk=d18
stor_dev=APM00042000818-001D addr=c32t1l13
server=server_2
disk=d14
stor_dev=APM00042000818-0019 addr=c0t1l9
server=server_2
disk=d14
stor_dev=APM00042000818-0019 addr=c32t1l9
server=server_2
disk=d11
stor_dev=APM00042000818-0016 addr=c0t1l6
server=server_2
disk=d11
stor_dev=APM00042000818-0016 addr=c32t1l6
server=server_2
disk=d17
stor_dev=APM00042000818-001C addr=c0t1l12
server=server_2
disk=d17
stor_dev=APM00042000818-001C addr=c32t1l12
server=server_2
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #15
To extend the file system named ufs5, with the specified size of 1 GB,
by using clar_r5_performance pool, type:
$ nas_fs -xtend ufs5 size=1G pool=clar_r5_performance
id
name
acl
254
= 11
= ufs5
= 0
Command Line Interface Reference for File 8.1
The nas Commands
in_use
= True
type
= uxfs
worm
= off
volume
= v112
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= Off
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
disk=d16
stor_dev=BB005056830430-0019 addr=c0t1l9
server=server_2
disk=d16
stor_dev=BB005056830430-0019 addr=c16t1l9
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c0t1l6
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c16t1l6
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c0t1l5
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c16t1l5
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c0t1l0
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c16t1l0
server=server_2
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #16
To set the access control level to 1432 for the file system ufs5, type:
$ nas_fs -acl 1432 ufs5
id
= 11
name
= ufs5
acl
= 1432, owner=nasadmin, ID=201
in_use
= True
type
= uxfs
worm
= off
volume
= v112
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= Off
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
disk=d16
stor_dev=BB005056830430-0019 addr=c0t1l9
server=server_2
disk=d16
stor_dev=BB005056830430-0019 addr=c16t1l9
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c0t1l6
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c16t1l6
server=server_2
nas_fs
255
The nas Commands
disk=d12
disk=d12
disk=d7
disk=d7
stor_dev=BB005056830430-0015
stor_dev=BB005056830430-0015
stor_dev=BB005056830430-0010
stor_dev=BB005056830430-0010
addr=c0t1l5
addr=c16t1l5
addr=c0t1l0
addr=c16t1l0
server=server_2
server=server_2
server=server_2
server=server_2
Note: The value 1432 specifies nasadmin as the owner and gives users with
an access level of at least observer read access only, users with an access level
of at least operator read/write access, and users with an access level of at
least admin read/write/delete access.
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #17
To set the maximum retention period for file system ufs2_flre to 11
years, type:
$ nas_fs -modify ufs2_flre -worm -max_retention 11Y
id
= 14
name
= ufs2_flre
acl
= 0
in_use
= True
type
= uxfs
worm
= enterprise with no protected files
worm_clock= Fri Jul 29 11:14:27 EDT 2011
worm Max Retention Date= No protected files created
worm Default Retention Period= 10 Years
worm Minimum Retention Period= 30 Days
worm Maximum Retention Period= 11 Years
FLR Auto_lock= off
FLR Policy Interval= 3600 seconds
FLR Auto_delete= off
FLR Epoch Year= 2003
volume
= v117
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= Off
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
disk=d16
stor_dev=BB005056830430-0019 addr=c0t1l9
server=server_2
disk=d16
stor_dev=BB005056830430-0019 addr=c16t1l9
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c0t1l6
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c16t1l6
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c0t1l5
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c16t1l5
server=server_2
256
Command Line Interface Reference for File 8.1
The nas Commands
disk=d7
disk=d7
stor_dev=BB005056830430-0010 addr=c0t1l0
stor_dev=BB005056830430-0010 addr=c16t1l0
server=server_2
server=server_2
EXAMPLE #1 provides a description of command outputs.
To set the maximum retention period for file system ufs2_flre to 11
years, type:
$ nas_fs -modify ufs2_flre -worm -max_retention 11Y
id
= 14
name
= ufs2_flre
acl
= 0
in_use
= True
type
= uxfs
worm
= enterprise with no protected files
worm_clock= Fri Jul 29 11:14:27 EDT 2011
worm Max Retention Date= No protected files created
worm Default Retention Period= 10 Years
worm Minimum Retention Period= 30 Days
worm Maximum Retention Period= 11 Years
FLR Auto_lock= off
FLR Policy Interval= 3600 seconds
FLR Auto_delete= off
FLR Epoch Year= 2003
volume
= v117
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= Off
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
disk=d16
stor_dev=BB005056830430-0019 addr=c0t1l9
server=server_2
disk=d16
stor_dev=BB005056830430-0019 addr=c16t1l9
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c0t1l6
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c16t1l6
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c0t1l5
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c16t1l5
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c0t1l0
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c16t1l0
server=server_2
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #18
To reset the FLR epoch year for file system ufs2_flre to 2000, type:
$ nas_fs -modify ufs2_flre -worm -reset_epoch 2000
id
= 14
nas_fs
257
The nas Commands
name
= ufs2_flre
acl
= 0
in_use
= True
type
= uxfs
worm
= enterprise with no protected files
worm_clock= Fri Jul 29 11:18:36 EDT 2011
worm Max Retention Date= No protected files created
worm Default Retention Period= 10 Years
worm Minimum Retention Period= 30 Days
worm Maximum Retention Period= 11 Years
FLR Auto_lock= off
FLR Policy Interval= 3600 seconds
FLR Auto_delete= off
FLR Epoch Year= 2000
volume
= v117
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= Off
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
disk=d16
stor_dev=BB005056830430-0019 addr=c0t1l9
server=server_2
disk=d16
stor_dev=BB005056830430-0019 addr=c16t1l9
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c0t1l6
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c16t1l6
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c0t1l5
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c16t1l5
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c0t1l0
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c16t1l0
server=server_2
EXAMPLE #19
To enable FLR automatic file locking with a policy interval of 30
minutes for file system ufs2_flre, type:
$ nas_fs -modify ufs2_flre -worm -auto_lock enable
-policy_interval 30M
id
= 14
name
= ufs2_flre
acl
= 0
in_use
= True
type
= uxfs
worm
= enterprise with no protected files
worm_clock= Fri Jul 29 12:14:44 EDT 2011
worm Max Retention Date= No protected files created
worm Default Retention Period= 10 Years
worm Minimum Retention Period= 30 Days
worm Maximum Retention Period= 11 Years
FLR Auto_lock= on
258
Command Line Interface Reference for File 8.1
The nas Commands
FLR Policy Interval= 1800 seconds
FLR Auto_delete= off
FLR Epoch Year= 2000
volume
= v117
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= Off
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
disk=d16
stor_dev=BB005056830430-0019 addr=c0t1l9
server=server_2
disk=d16
stor_dev=BB005056830430-0019 addr=c16t1l9
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c0t1l6
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c16t1l6
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c0t1l5
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c16t1l5
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c0t1l0
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c16t1l0
server=server_2
EXAMPLE #20
To enable FLR automatic file deletion for file system ufs2_flre, type
:$ nas_fs -modify ufs2_flre -worm -auto_delete enable
id
= 40
name
= ufs4
acl
= 0
in_use
= True
type
= uxfs
worm
= enterprise with no protected files
worm_clock= Wed Jul 6 11:11:13 UTC 2011
worm Max Retention Date= No protected files created
worm Default Retention Period= 1 Year
worm Minimum Retention Period= 1 Day
worm Maximum Retention Period= 1 Year
FLR Auto_lock= on
FLR Policy Interval= 1800 seconds
FLR Auto_delete= on
FLR Epoch Year= 2000
volume
= v175
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication = Off
stor_devs = APM00042000818-001F,APM00042000818-001D,APM00042000818-0019,APM00042
nas_fs
259
The nas Commands
000818-0016
disks
= d20,d18,d14,d11
EXAMPLE #21
To start the conversion of the file system, ufs2, and to conform to the
MIXED access policy mode, type:
$ nas_fs -translate ufs2 -access_policy start -to MIXED
-from NT
id
= 38
name
= ufs2
acl
= 1432, owner=nasadmin, ID=201
in_use
= True
type
= uxfs
worm
= off
volume
= v171
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication = off
stor_devs = APM00042000818-001F,APM00042000818-001D,APM00042000818-0019,APM00042
000818-0016,APM00042000818-001C
disks
= d20,d18,d14,d11,d17
disk=d20
stor_dev=APM00042000818-001F addr=c0t1l15
server=server_2
disk=d20
stor_dev=APM00042000818-001F addr=c32t1l15
server=server_2
disk=d18
stor_dev=APM00042000818-001D addr=c0t1l13
server=server_2
disk=d18
stor_dev=APM00042000818-001D addr=c32t1l13
server=server_2
disk=d14
stor_dev=APM00042000818-0019 addr=c0t1l9
server=server_2
disk=d14
stor_dev=APM00042000818-0019 addr=c32t1l9
server=server_2
disk=d11
stor_dev=APM00042000818-0016 addr=c0t1l6
server=server_2
disk=d11
stor_dev=APM00042000818-0016 addr=c32t1l6
server=server_2
disk=d17
stor_dev=APM00042000818-001C addr=c0t1l12
server=server_2
disk=d17
stor_dev=APM00042000818-001C addr=c32t1l12
server=server_2
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #22
To display the status of access policy conversion for ufs2, type:
$ nas_fs -translate ufs2 -access_policy status
status=In progress
percent_inode_scanned=90
EXAMPLE #23
To create a nested mount file system, nmfs1, type:
$ nas_fs -name nmfs1 -type nmfs -create
id
name
acl
260
= 8002
= nmfs1
= 0
Command Line Interface Reference for File 8.1
The nas Commands
in_use
= False
type
= nmfs
worm
= off
volume
= 0
pool
=
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= unavailable
stor_devs =
disks
=
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #24
To delete ufs1, type:
$ nas_fs -delete ufs41
name
= ufs41
acl
= 0
in_use
= False
type
= uxfs
worm
= off
volume
= v129
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
deduplication
= unavailable
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #25
To create a file system named ufs3, with a size of 1 GB, by using the
clar_r5_performance pool, a maximum size of 10 GB and with
auto-extend and thin provisioning enabled, type:
$ nas_fs -name ufs3 -create size=1G
pool=clar_r5_performance -auto_extend yes -max_size 10G
-thin yes
id
name
acl
in_use
type
worm
volume
pool
member_of
=
=
=
=
=
=
=
=
=
8003
ufs3
0
False
uxfs
off
v133
clar_r5_performance
root_avm_fs_group_3
nas_fs
261
The nas Commands
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = hwm=90%,max_size=10240M,thin=yes
deduplication
= unavailable
stor_devs =
BB005056830430-0018,BB005056830430-0017,BB005056830430-0014,BB005056830430-0011
disks
= d15,d14,d11,d8
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #26
To disable thin provisioning on ufs3, type:
$ nas_fs -modify ufs3 -thin no
id
= 8003
name
= ufs3
acl
= 0
in_use
= False
type
= uxfs
worm
= off
volume
= v133
pool
= clar_r5_performance
member_of = root_avm_fs_group_3
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = hwm=90%,max_size=10240M,thin=no
deduplication
= unavailable
stor_devs =
BB005056830430-0018,BB005056830430-0017,BB005056830430-0014,BB005056830430-0011
disks
= d15,d14,d11,d8
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #27
id
=
name
=
acl
=
in_use
=
type
=
worm
=
volume
=
pool
=
member_of =
rw_servers=
ro_servers=
rw_vdms
=
262
To query the current directory type and translation status for MPD,
type:
$ nas_fs -info ufs5 -option mpd
11
ufs5
1432, owner=nasadmin, ID=201
True
uxfs
off
v112
clar_r5_performance
root_avm_fs_group_3
server_2
Command Line Interface Reference for File 8.1
The nas Commands
ro_vdms
=
auto_ext = no,thin=no
deduplication
= Off
stor_devs =
BB005056830430-0019,BB005056830430-0016,BB005056830430-0015,BB005056830430-0010
disks
= d16,d13,d12,d7
disk=d16
stor_dev=BB005056830430-0019 addr=c0t1l9
server=server_2
disk=d16
stor_dev=BB005056830430-0019 addr=c16t1l9
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c0t1l6
server=server_2
disk=d13
stor_dev=BB005056830430-0016 addr=c16t1l6
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c0t1l5
server=server_2
disk=d12
stor_dev=BB005056830430-0015 addr=c16t1l5
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c0t1l0
server=server_2
disk=d7
stor_dev=BB005056830430-0010 addr=c16t1l0
server=server_2
Multi-Protocol Directory Information
Default_directory_type
= DIR3
Needs_translation
= False
Translation_state
= Never
Has_translation_error
= False
Where:
Value
Definition
Default_directory_type The default directory type for the file system. Available types are: DIR3 and COMPAT.
Needs_translation
If true, then the file system may contain more than one directory type. If false, then all
directories are of the file system default directory type.
Translation_state
The current state of the translation thread. Available states are: never, not requested,
pending, queued, running, paused, completed, and failed.
Has_translation_error Indicated if the most recent translation encountered any errors.
Default_directory_type
DIR3
DIR3
Needs_translation state
False
True
COMPAT
False
COMPAT
True
File system
Is MPD. No action required.
Requires translation or file system maintenance.
Contact EMC Customer Service.
Is COMPAT and requires translation.
Contact EMC Customer Service.
Requires translation.
Contact EMC Customer Service.
The state where both Default_directory_type=DIR3 and
Needs_transalation=False assure that this file system's directories are
all in MPD format, and there are no directories of the obsolete
single-protocol format.
Any other combination of states, for example,
Needs_transalation=True, indicates that there could be non-MPD
nas_fs
263
The nas Commands
directories in the file system which may not be compatible with a
future release.
EXAMPLE #28
To display the information about the file system ufs3 and a valid
fast_clone_level of 1 or 2, type::
$ nas_fs -info ufs3
id
= 478
name
= ufs2_flre
acl
= 0
in_use
= False
type
= uxfs
worm
= off
volume
= v1168
pool
= clarsas_archive
member_of = root_avm_fs_group_32
rw_servers=
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
fast_clone_level= unavailable
deduplication
= unavailable
stor_devs =
APM00112101832-0019,APM00112101832-0028,APM00112101832-0027,APM00112101832-0022
disks
= d25,d19,d32,d16
EXAMPLE #29
To display the information about a file system using the file system
ufs4 using Symmetrix backend mapped pool, type:
$ nas_fs -info ufs4
id
= 32
name
= ufs4
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= v644
pool
= symm_mapped_pool
member_of = root_avm_fs_group_21
rw_servers= server_2
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = hwm=50%,max_size=1024M,thin=yes
fast_clone_level = 1
deduplication
= Off
compressed= Mixed
frontend_io_quota = maxiopersec 500,maxmbpersec 500
stor_devs = 000196900016-0553
disks
= d524
264
Command Line Interface Reference for File 8.1
The nas Commands
disk=d524
disk=d524
disk=d524
disk=d524
disk=d524
disk=d524
disk=d524
disk=d524
stor_dev=000196900016-0553
stor_dev=000196900016-0553
stor_dev=000196900016-0553
stor_dev=000196900016-0553
stor_dev=000196900016-0553
stor_dev=000196900016-0553
stor_dev=000196900016-0553
stor_dev=000196900016-0553
addr=c4t3l5-0-0
addr=c20t3l5-0-0
addr=c36t3l5-0-0
addr=c52t3l5-0-0
addr=c68t3l5-0-0
addr=c84t3l5-0-0
addr=c100t3l5-0-0
addr=c116t3l5-0-0
server=server_2
server=server_2
server=server_2
server=server_2
server=server_2
server=server_2
server=server_2
server=server_2
Where:
Value
compressed
frontend_io_quota
Definition
For VNX with Symmetrix backend, indicates whether data is compressed. Values are: True,
False, Mixed (indicates some of the LUNs, but not all, are compressed).
For VNX with Symmetrix backend, indicates if Frotend IO Quota is configured on this mapped
pool, could also have value as False (indicates Frontend IO Quota is not configured on
mapped SG in Symmetrix backend).
nas_fs
265
The nas Commands
nas_fsck
Manages fsck and aclchk utilities on specified file systems.
SYNOPSIS
nas_fsck
-list
| -info {-all|<fs_name>|id=<fs_id>}
| -start {<fs_name>|id=<fs_id>} [-aclchkonly][-monitor][-mover <mover_name>]
[-Force]
DESCRIPTION
nas_fsck uses the fsck and aclchk utilities to perform a check for
consistency and errors on the specified file system. nas_fsck also lists
and displays the status of the fsck and aclchk utilities. File systems
must be mounted read-write to use these utilities.
Depending on the size of the file system, the fsck utility may use a
significant portion of the system's memory and may affect overall
system performance. Hence, it should not be run on a server under
heavy load as it is possible that the server may run out of resources.
In most cases, the user will be notified if sufficient memory is not
available to run a file system check. In these cases, one of the
following can be done:
◆
◆
◆
OPTIONS
Start the file system during off-peak hours.
Reboot the server and let nas_fsck run on reboot.
Run nas_fsck on a different server if the file system is
unmounted.
-list
Displays a list of all the file systems undergoing fsck or aclchk.
-info {-all|<fs_name>|id=<fs_id>}
Queries the Data Mover and displays information about the status of
the fsck or aclchk utilities for the specified file system.
-start {<fs_name>|id=<fs_id>}
Starts the fsck and the aclchk utilities on the specified file system.
266
Command Line Interface Reference for File 8.1
The nas Commands
!
CAUTION
If file system check is started on a mounted file system, the file
system will be unavailable for the duration of the check. NFS
clients will display the message NFS server not responding and
CIFS clients will lose connectivity with the server and will have to
remap shares.
[-aclchkonly]
Initiates the aclchk utility only, which checks and corrects any
errors in the ACL database and removes duplicate ACL
information stored on the specified file system. The aclchkonly
option can only be used on a file system that is not exported. The
default is for both fsck and aclchk.
Note: The NDMP backup process must be stopped on the Data Mover
before using the nas_fsck -aclchkonly command.
[-monitor]
Displays the status of fsck and aclchk until the command
completes.
Note: For a mounted file system, a <movername> is not required because
the fsck and aclchk utilities are run on the Data Mover where the file
system is mounted.
[-Force]
Forces a fsck or aclchk to be run on an enabled file system.
SEE ALSO
EXAMPLE #1
Managing Volumes and File Systems for VNX Manually and nas_fs.
To start file system check on ufs1 and monitor the progress, type:
$ nas_fsck -start ufs1 -monitor
id
= 27
name
= ufs1
volume
= mtv1
fsck_server
= server_2
inode_check_percent
= 10..20..30..40..60..70..80..100
directory_check_percent
= 0..0..100
used_ACL_check_percent
= 100
free_ACL_check_status
= Done
cylinder_group_check_status = In Progress..Done
nas_fsck
267
The nas Commands
Where:
Value
id
name
volume
fsck_server
inode_check_percent
directory_check_percent
used_ACL_check_percent
free_ACL_check_status
cylinder_group_check_status
EXAMPLE #2
Definition
Automatically assigned ID of a file system.
Name assigned to the file system.
Volume on which the file system resided.
Name of the Data Mover where the utility is being run.
Percentage of inodes in the file system checked and fixed.
Percentage of directories in the file system checked and fixed.
Percentage of used ACLs that have been checked and fixed.
Status of the ACL check.
Status of the cylinder group check.
To start ACL check on ufs1, type:
$ nas_fsck -start ufs1 -aclchkonly
ACLCHK: in progress for file system ufs1
EXAMPLE #3
To start a file system check on ufs2 using Data Mover server_5, type:
$ nas_fsck -start ufs2 -mover server_5
name
= ufs2
id
= 23
volume
= v134
fsck_server
= server_5
inode_check_percent
= 40
directory_check_percent
= 0
used_ACL_check_percent
= 0
free_ACL_check_status
= Not Started
cylinder_group_check_status = Not Started
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #4
id
23
27
type
1
1
state
FSCK
ACLCHK
To list all current file system checks, type:
$ nas_fsck
volume
134
144
-list
name
ufs2
ufs1
Where:
Value
id
type
state
volume
name
server
268
Definition
Automatically assigned ID of a file system.
Type of file system.
Utility being run.
Volume on which the file system resided.
Name assigned to the file system.
Server on which fsck is being run.
Command Line Interface Reference for File 8.1
server
4
1
The nas Commands
EXAMPLE #5
To display information about file system check for ufs2 that is
currently running, type:
$ nas_fsck -info ufs2
name
= ufs2
id
= 23
volume
= v134
fsck_server
= server_5
inode_check_percent
= 100
directory_check_percent
= 100
used_ACL_check_percent
= 100
free_ACL_check_status
= Done
cylinder_group_check_status = In Progress
EXAMPLE #1 provides a description of command outputs.
EXAMPLE #6
To display information about all file system checks that are currently
running, type:
$ nas_fsck -info -all
name
= ufs2
id
= 23
volume
= v134
fsck_server
= server_5
inode_check_percent
= 30
directory_check_percent
= 0
used_ACL_check_percent
= 0
free_ACL_check_status
= Not Started
cylinder_group_check_status = Not Started
name
id
volume
fsck_server
inode_check_percent
directory_check_percent
used_ACL_check_percent
free_ACL_check_status
cylinder_group_check_status
=
=
=
=
=
=
=
=
=
ufs1
27
mtv1
server_2
100
0
0
Not Started
Not Started
EXAMPLE #1 provides a description of command outputs.
nas_fsck
269
The nas Commands
nas_halt
Performs a controlled halt of all Control Stations, Data Movers, and
Storage Processors on a Unified/File VNX system.
SYNOPSIS
nas_halt [-f] [-sp] now
DESCRIPTION
nas_halt performs an orderly shutdown of the Control Stations, Data
Movers, and optionally, the Storage Processors. nas_halt can only be
executed by root users and must be run from the /nas/sbin directory.
Note: For Gateway servers, only root users can use nas_halt to power off the
Gateway.
OPTIONS
-f
Performs an immediate halt for a Unified/File VNX system.
-sp
Powers off Control Stations, Data Movers and Storage Processors on
a Unified/File VNX system.
now
Performs an interactive halt for a Unified/File VNX system.
SEE ALSO
EXAMPLE #1
VNX System Operations and server_cpu.
To perform a halt of the VNX Control Stations and Data Movers, type:
# /nas/sbin/nas_halt now
********************************
WARNING! *******************************
You are about to HALT this VNX including all of its Control Stations
and Data Movers. DATA will be UNAVAILABLE when the system is halted.
Note that this command does *not* halt the storage array.
ARE YOU SURE YOU WANT TO CONTINUE? [yes or no] : yes
Sending the halt signal to the Master Control Daemon...: Done
May 3 11:12:54 cs100 EMCServer: nas_mcd: Check and halt other CS...: Done
May 3 11:13:26 cs100 JSERVER: *** Java Server is exiting ***
May 3 11:13:31 cs100 ucd-snmp[11218]: Received TERM or STOP signal... shutting
down...
May 3 11:13:31 cs100 snmpd: snmpd shutdown succeeded
May 3 11:13:32 cs100 setup_enclosure: Executing -dhcpd stop option
May 3 11:13:32 cs100 snmptrapd[11179]: Stopping snmptrapd
270
Command Line Interface Reference for File 8.1
The nas Commands
May 3 11:13:32 cs100 EV_AGENT[13721]: Signal TERM received
May 3 11:13:32 cs100 EV_AGENT[13721]: Agent is going down
May 3 11:13:40 cs100 DHCPDMON: Starting DHCPD on CS 0
May 3 11:13:41 cs100 setup_enclosure: Executing -dhcpd start option
May 3 11:13:41 cs100 dhcpd: Internet Software Consortium DHCP Server V3.0pl1
May 3 11:13:41 cs100 dhcpd: Copyright 1995-2001 Internet Software Consortium.
May 3 11:13:41 cs100 dhcpd: All rights reserved.
May 3 11:13:41 cs100 dhcpd: For info, please visit
http://www.isc.org/products/DHCP
May 3 11:13:41 cs100 dhcpd: Wrote 0 deleted host decls to leases file.
May 3 11:13:41 cs100 dhcpd: Wrote 0 new dynamic host decls to leases file.
May 3 11:13:41 cs100 dhcpd: Wrote 0 leases to leases file.
May 3 11:13:41 cs100 dhcpd: Listening on
LPF/eth2/00:00:f0:9d:04:13/128.221.253.0/24
May 3 11:13:41 cs100 dhcpd: Sending on
LPF/eth2/00:00:f0:9d:04:13/128.221.253.0/24
May 3 11:13:41 cs100 dhcpd: Listening on
LPF/eth0/00:00:f0:9d:01:e5/128.221.252.0/24
May 3 11:13:41 cs100 dhcpd: Sending on
LPF/eth0/00:00:f0:9d:01:e5/128.221.252.0/24
May 3 11:13:41 cs100 dhcpd: Sending on
Socket/fallback/fallback-net
May 3 11:13:59 cs100 mcd_helper: : Failed to umount /nas (0)
May 3 11:13:59 cs100 EMCServer: nas_mcd: Failed to gracefully shutdown MCD and
halt servers. Forcing halt and reboot...
May 3 11:13:59 cs100 EMCServer: nas_mcd: Halting all servers...
May 3 11:15:00 cs100 get_datamover_status: Data Mover server_5: COMMAND doesnt
match.
EXAMPLE #2
To perform an immediate halt of the VNX Control Stations and Data
Movers, type:
# /nas/sbin/nas_halt -f now
Sending the halt signal to the Master Control Daemon...: Done
May 3 11:12:54 cs100 EMCServer: nas_mcd: Check and halt other CS...: Done
May 3 11:13:26 cs100 JSERVER: *** Java Server is exiting ***
May 3 11:13:31 cs100 ucd-snmp[11218]: Received TERM or STOP signal... shutting
down...
May 3 11:13:31 cs100 snmpd: snmpd shutdown succeeded
May 3 11:13:32 cs100 setup_enclosure: Executing -dhcpd stop option
May 3 11:13:32 cs100 snmptrapd[11179]: Stopping snmptrapd
May 3 11:13:32 cs100 EV_AGENT[13721]: Signal TERM received
May 3 11:13:32 cs100 EV_AGENT[13721]: Agent is going down
May 3 11:13:40 cs100 DHCPDMON: Starting DHCPD on CS 0
May 3 11:13:41 cs100 setup_enclosure: Executing -dhcpd start option
May 3 11:13:41 cs100 dhcpd: Internet Software Consortium DHCP Server V3.0pl1
May 3 11:13:41 cs100 dhcpd: Copyright 1995-2001 Internet Software Consortium.
May 3 11:13:41 cs100 dhcpd: All rights reserved.
May 3 11:13:41 cs100 dhcpd: For info, please visit
http://www.isc.org/products/DHCP
May 3 11:13:41 cs100 dhcpd: Wrote 0 deleted host decls to leases file.
nas_halt
271
The nas Commands
May 3 11:13:41 cs100 dhcpd: Wrote 0 new dynamic host decls to leases file.
May 3 11:13:41 cs100 dhcpd: Wrote 0 leases to leases file.
May 3 11:13:41 cs100 dhcpd: Listening on
LPF/eth2/00:00:f0:9d:04:13/128.221.253.0/24
May 3 11:13:41 cs100 dhcpd: Sending on
LPF/eth2/00:00:f0:9d:04:13/128.221.253.0/24
May 3 11:13:41 cs100 dhcpd: Listening on
LPF/eth0/00:00:f0:9d:01:e5/128.221.252.0/24
May 3 11:13:41 cs100 dhcpd: Sending on
LPF/eth0/00:00:f0:9d:01:e5/128.221.252.0/24
May 3 11:13:41 cs100 dhcpd: Sending on
Socket/fallback/fallback-net
May 3 11:13:59 cs100 mcd_helper: : Failed to umount /nas (0)
May 3 11:13:59 cs100 EMCServer: nas_mcd: Failed to gracefully shutdown MCD and
halt servers. Forcing halt and reboot...
May 3 11:13:59 cs100 EMCServer: nas_mcd: Halting all servers...
May 3 11:15:00 cs100 get_datamover_status: Data Mover server_5: COMMAND doesnt
match.
EXAMPLE #3
To perform a halt of the VNX Control Stations, Data Movers, and
Storage Processors on a unified system, type:
# /nas/sbin/nas_halt -sp now
********************************
WARNING! *******************************
You are about to HALT this Celerra including all of its Control Stations
and Data Movers. DATA will be UNAVAILABLE when the system is halted.
-sp switch will HALT the Storage Processors on a unified system.
ARE YOU SURE YOU WANT TO CONTINUE? [ yes or no ] : yes
Unified system detected.
Sending the halt signal to the Master Control Daemon...: Done
Feb 27 20:13:35 emcnas_i0 EMCServer: nas_mcd: Check and halt other CS...: Done
Feb 27 20:13:36 emcnas_i0 CIM: Server
**** Shutting down CIM Server *****
Feb 27 20:13:37 emcnas_i0 CIM: Server
**** Shutting down slp daemon *****
Feb 27 20:14:10 emcnas_i0 JSERVER: *** Java Server is exiting ***
Feb 27 20:14:19 emcnas_i0 mcd_helper: Cleaned up old APL share memory 71499780
Feb 27 20:14:19 emcnas_i0 mcd_helper: Cleaned up old APL semaphores 1376257
1409026 1441795
Feb 27 20:14:21 emcnas_i0 setup_enclosure: Executing -dhcpd stop option
Feb 27 20:14:34 emcnas_i0 DHCPDMON: Starting DHCPD on CS 0
Feb 27 20:14:36 emcnas_i0 setup_enclosure: Executing -dhcpd start option
Feb 27 20:14:36 emcnas_i0 dhcpd: Internet Systems Consortium DHCP Server
V3.0.5-RedHat
Feb 27 20:14:36 emcnas_i0 dhcpd: Copyright 2004-2006 Internet Systems Consortium.
Feb 27 20:14:36 emcnas_i0 dhcpd: All rights reserved.
Feb 27 20:14:36 emcnas_i0 dhcpd: For info, please visit
http://www.isc.org/sw/dhcp/
Feb 27 20:14:36 emcnas_i0 dhcpd: Wrote 0 deleted host decls to leases file.
Feb 27 20:14:36 emcnas_i0 dhcpd: Wrote 0 new dynamic host decls to leases file.
Feb 27 20:14:36 emcnas_i0 dhcpd: Wrote 8 leases to leases file.
272
Command Line Interface Reference for File 8.1
The nas Commands
Feb 27 20:14:36 emcnas_i0 dhcpd: Listening on
LPF/eth2/a0:36:9f:17:bb:03/128.221.253/24
Feb 27 20:14:36 emcnas_i0 dhcpd: Sending on
LPF/eth2/a0:36:9f:17:bb:03/128.221.253/24
Feb 27 20:14:36 emcnas_i0 dhcpd: Listening on
LPF/eth0/00:1e:67:6a:50:a9/128.221.252/24
Feb 27 20:14:36 emcnas_i0 dhcpd: Sending on
LPF/eth0/00:1e:67:6a:50:a9/128.221.252/24
Feb 27 20:14:36 emcnas_i0 dhcpd: Sending on
Socket/fallback/fallback-net
Feb 27 20:14:50 emcnas_i0 mcd_helper: Failed to umount /nas (0)
Feb 27 20:14:50 emcnas_i0 EMCServer: nas_mcd: Failed to gracefully shutdown MCD
and halt servers. Forcing halt...
Feb 27 20:14:50 emcnas_i0 EMCServer: nas_mcd: Halting all servers...
Feb 27 20:15:03 emcnas_i0 get_backend_status: Navicli getagent aborted, possibly
due to a bad network setup!
Feb 27 20:15:06 emcnas_i0 get_datamover_status: Data Mover server_3: COMMAND
doesnt match.
Feb 27 20:15:06 emcnas_i0 get_datamover_status: Data Mover server_2: COMMAND
doesnt match.
Feb 27 20:15:15 emcnas_i0 EMCServer: nas_mcd: Shutting down SPs...
Feb 27 20:15:17 emcnas_i0 mcd_helper: Shutting down SPB ...
Feb 27 20:15:20 emcnas_i0 mcd_helper: Navicli shutdownsp returned SUCCESS
Feb 27 20:16:04 emcnas_i0 kernel: ND id 5 rcv tmo, crnt_tm 101357, last rcv_tm
101297, que 27
Feb 27 20:16:04 emcnas_i0 kernel: ND dev nde enter nd_conn, id 5, err -32
Feb 27 20:16:07 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 3fddd80, port
43283 err(-115)
Feb 27 20:16:14 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 2fddd80, port
43283 err(-115)
Feb 27 20:16:19 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 3fcdd80, port
43283 err(-113)
Feb 27 20:16:20 emcnas_i0 mcd_helper: SPB is inaccessible and probably down
Feb 27 20:16:25 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 2fcdd80, port
43283 err(-113)
Feb 27 20:16:32 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 3fddd80, port
43283 err(-113)
Feb 27 20:16:40 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 2fddd80, port
43283 err(-113)
Feb 27 20:16:49 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 3fcdd80, port
43283 err(-113)
Feb 27 20:16:59 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 2fcdd80, port
43283 err(-113)
Feb 27 20:17:10 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 3fddd80, port
43283 err(-113)
Feb 27 20:17:20 emcnas_i0 mcd_helper: Shutting down SPA ...
Feb 27 20:17:21 emcnas_i0 mcd_helper: Navicli shutdownsp returned SUCCESS
Feb 27 20:17:22 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 2fddd80, port
43283 err(-113)
Feb 27 20:17:35 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 3fcdd80, port
43283 err(-113)
Feb 27 20:17:49 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 2fcdd80, port
43283 err(-113)
nas_halt
273
The nas Commands
Feb 27 20:18:04 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 3fddd80, port
43283 err(-113)
Feb 27 20:18:20 emcnas_i0 kernel: ND: id 5 can't conn socket, addr 2fddd80, port
43283 err(-113)
274
Command Line Interface Reference for File 8.1
The nas Commands
nas_inventory
Provides detailed information about hardware components in the
system.
SYNOPSIS
nas_inventory
{
-list [-location]
| {-info <location>|-all}
| -tree
}
DESCRIPTION
OPTIONS
nas_inventory displays detailed information about the hardware
components that are configured on a system.
-list
Displays a list of all hardware components and their associated name,
type, status, and system ID.
[-location]
Displays the location string for each component in the output.
The location string is a unique identifier for the component.
Specifies the location string with enclosed double quotes (" ") and
displays a list of detailed information for the specific component
for which the string is the unique ID.
-info <location_string>|-all
Displays a list of all the properties for a component, including the
component name, type, status, variant, associated system, serial
number, part number, and history.
The -all option lists detailed information for all components in the
system.
-tree
Displays a hierarchical tree of components, including the status of
each component.
nas_inventory
275
The nas Commands
EXAMPLE #1
Component
Battery A
VNX NS40G
FCNTR083000055001A
CLARiiON CX4-240
FCNTR083000055
DME 0 Data Mover 2
DME 0 Data Mover 2
Ethernet Module
DME 0 Data Mover 2
SFP BE0
DME 0 Data Mover 2
SFP BE1
DME 0 Data Mover 2
SFP FE0
To display a list of components on the system, type:
$ nas_inventory -list
Type
Status
Battery
OK
VNX
Warning
System ID
CLARiiON CX4-240FCNTR083000055
VNX NS40GFCNTR083000055001A
CLARiiON
OK
Data Mover OK
CLARiiON CX4-240 FCNTR083000055
VNX NS40G FCNTR083000055001A
Module
OK
VNX NS40G FCNTR083000055001A
SFP
OK
VNX NS40G FCNTR083000055001A
SFP
OK
VNX NS40G FCNTR083000055001A
SFP
OK
VNX NS40G FCNTR083000055001A
Where:
Value
Component
Type
Status
System ID
EXAMPLE #2
Definition
Description of the component.
The type of component. Possible types are: battery, blower, VNX, Control Station, Data
Mover, and disk.
The current status of the component. Status is component type specific. There are several
possible status values, each of which is associated with a particular component type.
The identifier for the VNX or the storage ID of the system containing the component.
To display a list of components and component locations, type:
$ nas_inventory -list -location
Component
Type
Status System ID
Location
Battery A
Battery OK
CLARiiON CX4-240 FCNTR083000055
system:NS40G:FCNTR083000055001A|clariionSystem:CX4-240:FCNTR083000055|sps::A
Celerra NS40G FCNTR083000055001A Celerra
Warning Celerra NS40G
FCNTR083000055001A
system:NS40G:FCNTR083000055001A
CLARiiON CX4-240 FCNTR083000055 CLARiiON OK
CLARiiON CX4-240 FCNTR083000055
system:NS40G:FCNTR083000055001A|clariionSystem:CX4-240:FCNTR083000055
DME 0 Data Mover 2
Data Mover OK
Celerra NS40G FCNTR083000055001A
system:NS40G:FCNTR083000055001A|enclosure:xpe:0|mover:NS40:2
DME 0 Data Mover 2 Ethernet Module Module OK
Celerra NS40G FCNTR083000055001A
system:NS40G:FCNTR083000055001A|enclosure:xpe:0|mover:NS40:2|module:ethernet:
DME 0 Data Mover 2 SFP BE0
SFP
OK
Celerra NS40G FCNTR083000055001A
system:NS40G:FCNTR083000055001A|enclosure:xpe:0|mover:NS40:2|sfp::BE0
DME 0 Data Mover 2 SFP BE1
SFP
OK
Celerra NS40G FCNTR083000055001A
system:NS40G:FCNTR083000055001A|enclosure:xpe:0|mover:NS40:2|sfp::BE1
DME 0 Data Mover 2 SFP FE0
SFP
OK
Celerra NS40G FCNTR083000055001A
system:NS40G:FCNTR083000055001A|enclosure:xpe:0|mover:NS40:2|sfp::FE0
276
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #3
To list information for a specific component, type:
$ nas_inventory -info "system:NS40G:FCNTR083000055001A|
clariionSystem:CX4-240:FCNTR083000055|iomodule::B0"
Location = system:NS40G:FCNTR083000055001A|clariionSystem:CX4-240:
FCNTR083000055|iomodule::B0
Component Name = IO Module B0
Type
= IO Module
Status
= OK
Variant
= 4 PORT FIBRE IO MODULE
Storage System = CLARiiON CX4-240 FCNTR083000055
Serial Number = CF2YW082800426
Part Number
= 103-054-100C
History
= EMC_PART_NUMBER:103-054-100C
EMC_ARTWORK_REVISION:C01
EMC_ASSEMBLY_REVISION:C03
EMC_SERIAL_NUMBER:CF2YW082800426
VENDER_PART_NUMBER:N/A
VENDER_ARTWORK_NUMBER:N/A
VENDER_ASSEMBLY_NUMBER:N/A
VENDER_SERIAL_NUMBER:N/A
VENDOR_NAME:N/A
LOCATION_OF_MANUFACTURE:N/A
YEAR_OF_MANUFACTURE:N/A
MONTH_OF_MANUFACTURE:N/A
DAY_OF_MONTH_OF_MANUFACTURE:N/A
ASSEMBLY_NAME:4 PORT FIBRE IO MODULE
Note: The location string must be enclosed in double quotes.
Where:
Value
Location
Component
Type
Status
Variant
Storage System
Serial Number
Part Number
History
Definition
The unique identifier of the component and where the component is located in the component
hierarchy.
The description of the component.
The type of component. Possible types are: battery, blower, VNX for file, VNX for block, Control
Station, Data Mover, and disk.
The current condition of the component. Status is component type specific. There are several
possible status values, each of which is associated with a particular component type.
The specific type of hardware.
The model and serial number of the system.
The serial number of the hardware component.
The part number of the hardware component.
If available, the history information of the component. Possible values are: part number, serial
number, vendor, date of manufacture, and CPU information.
nas_inventory
277
The nas Commands
EXAMPLE #4
To display components in a tree structure, type:
$ nas_inventory -tree
Component
Type
Celerra NS40G FCNTR083000055001A
Celerra
CLARiiON CX4-240 FCNTR083000055
CLARiiON
Battery A
Battery
IO Module A0
IO Module
IO Module A1
IO Module
IO Module A2
IO Module
IO Module A3
IO Module
IO Module A4
IO Module
IO Module B0
IO Module
IO Module B1
IO Module
IO Module B2
IO Module
IO Module B3
IO Module
IO Module B4
IO Module
Power Supply A0
Power Supply
Power Supply A1
Power Supply
Power Supply B0
Power Supply
Power Supply B1
Power Supply
278
Command Line Interface Reference for File 8.1
Status
Warning
OK
OK
OK
OK
Empty
Empty
Empty
OK
OK
Empty
Empty
Empty
OK
OK
OK
OK
The nas Commands
nas_license
Enables software packages.
SYNOPSIS
nas_license
-list
| -create <package_name>[|<key_code>]
| -delete <package_name>
| -init
DESCRIPTION
nas_license enables software packages that are available for use with
the system. The <key_code> is supplied by EMC.
All entries are case-sensitive.
OPTIONS
No arguments
Displays a usage message that contains all available and valid
software packages that can be installed.
-list
Displays the site_key as a string and any software packages for which
a license has been installed. The site_key is a permanent license and
cannot be deleted. The failover license is legacy and should only be
used only when instructed to do so by EMC personnel.
Note: Licenses installed on the Control Station are read by the system. The
site_key is a unique identifier which gets generated the first time nas_license
is run. The site_key is also used to decode the key_code supplied by EMC
personnel for special packages.
-create <package_name>[=<key_code>]
Installs the license for the indicated <package_names>. Valid
<package_names> are:
site key
nfs
cifs
failover
snapsure
advancedmanager
replicator
filelevelretention
nas_license
279
The nas Commands
Note: These packages do not require key_code as they can be enabled from
the GUI. Special packages are supplied along with the required key_code by
the EMC Customer Support Representative. The package failover requires
key_code.
-delete <package_name>
Deletes the license for the specified <package_name>.
-init
Initializes the database and re-creates the license file by using the
site_key that is already installed. The license file is located at
/nas/site as nas_license. It contains license keys in an encrypted
format. The -init option should be run only if the license file
containing all the license information has been lost and the following
error message is received:
license table is not initialized
Once the license file has been re-created, the rest of the entries, if
present, should be re-added by using the -create option.
EXAMPLE #1
To install a license for the snapsure software package, type:
$ nas_license -create snapsure
done
EXAMPLE #2
key
site_key
advancedmanager
nfs
cifs
snapsure
replicator
filelevelretention
EXAMPLE #3
To display all software packages with currently installed licenses,
type:
$ nas_license -list
status
value
online
42 de 6f d1
online
online
online
online
online
online
To delete a license for specified software package, type:
$ nas_license -delete snapsure
done
EXAMPLE #4
To initialize the database and re-create the license file, type:
$ nas_license -init
done
280
Command Line Interface Reference for File 8.1
The nas Commands
nas_logviewer
Displays the content of nas_eventlog generated log files.
SYNOPSIS
nas_logviewer <file_name>
[-f][-v|-t]
DESCRIPTION
OPTIONS
nas_logviewer displays the event log and other logs created by
nas_eventlog. The log files may be system generated, or created by
the user. Information in the log file is read from oldest to newest.
No arguments
Displays the contents of the specified logfile.
-f
Monitors the growth of the log by entering into an endless loop,
pausing and reading the log as it is being generated. To exit, press
Ctrl-C together.
[-v|-t]
Displays the log files in verbose or terse format.
SEE ALSO
EXAMPLE #1
Configuring Events and Notifications on VNX for File and server_log.
To view the contents of the sys_log file, type:
$ nas_logviewer /nas/log/sys_log|more
May 12 18:01:57 2007:CS_PLATFORM:NASDB:INFO:300:::::nasdb_backup:
checkpoint in progress
May 12 18:02:59 2007:CS_PLATFORM:NASDB:INFO:305:::::nasdb_backup:
Checkpoint done
May 12 18:03:00 2007:CS_PLATFORM:NASDB:ERROR:202:::::NAS database
detectedMay 12 18:03:12 2007
:CS_PLATFORM:NASDB:INFO:306:::::nasdb_backup: NAS DB Backup done
May 12 19:01:52 2007:CS_PLATFORM:NASDB:INFO:300:::::nasdb_backup:
checkpoint in progress
May 12 19:02:50 2007:CS_PLATFORM:NASDB:INFO:305:::::nasdb_backup:
Checkpoint done
May 12 19:02:51 2007:CS_PLATFORM:NASDB:ERROR:202:::::NAS database
detectedMay 12 19:03:02 2007
:CS_PLATFORM:NASDB:INFO:306:::::nasdb_backup: NAS DB Backup done
May 12 20:01:57 2007:CS_PLATFORM:NASDB:INFO:300:::::nasdb_backup:
checkpoint in progress
May 12 20:02:58 2007:CS_PLATFORM:NASDB:INFO:305:::::nasdb_backup:
Checkpoint done
May 12 20:02:59 2007:CS_PLATFORM:NASDB:ERROR:202:::::NAS database
detectedMay 12 20:03:10 2007
NAS_DB
NAS_DB
error
NAS_DB
NAS_DB
error
NAS_DB
NAS_DB
error
nas_logviewer
281
The nas Commands
:CS_PLATFORM:NASDB:INFO:306:::::nasdb_backup: NAS DB Backup done
May 12 21:01:52 2007:CS_PLATFORM:NASDB:INFO:300:::::nasdb_backup: NAS_DB
checkpoint in progress
May 12 21:02:51 2007:CS_PLATFORM:NASDB:INFO:305:::::nasdb_backup: NAS_DB
Checkpoint done
May 12 21:02:52 2007:CS_PLATFORM:NASDB:ERROR:202:::::NAS database error
detectedMay 12 21:03:03 2007
:CS_PLATFORM:NASDB:INFO:306:::::nasdb_backup: NAS DB Backup done
Note: This is a partial listing due to the length of the outputs.
EXAMPLE #2
May
May
May
May
May
May
May
May
May
May
May
May
May
May
12
12
12
12
12
12
12
12
12
12
12
12
12
12
18:01:57
18:02:59
18:03:00
18:03:12
19:01:52
19:02:50
19:02:51
19:03:02
20:01:57
20:02:58
20:02:59
20:03:10
21:01:52
21:02:51
To display the contents of the log files in terse format, type:
$ nas_logviewer -t /nas/log/sys_log
2007:96108871980:nasdb_backup: NAS_DB checkpoint in progress
2007:96108871985:nasdb_backup: NAS_DB Checkpoint done
2007:83223969994:NAS database error detected
2007:96108871986:nasdb_backup: NAS DB Backup done
2007:96108871980:nasdb_backup: NAS_DB checkpoint in progress
2007:96108871985:nasdb_backup: NAS_DB Checkpoint done
2007:83223969994:NAS database error detected
2007:96108871986:nasdb_backup: NAS DB Backup done
2007:96108871980:nasdb_backup: NAS_DB checkpoint in progress
2007:96108871985:nasdb_backup: NAS_DB Checkpoint done
2007:83223969994:NAS database error detected
2007:96108871986:nasdb_backup: NAS DB Backup done
2007:96108871980:nasdb_backup: NAS_DB checkpoint in progress
2007:96108871985:nasdb_backup: NAS_DB Checkpoint done
EXAMPLE #3
To display the contents of the log files in verbose format, type:
$ nas_logviewer -v /nas/log/sys_log|more
logged time = May 12 18:01:57 2007
creation time = May 12 18:01:57 2007
slot id =
id = 96108871980
severity = INFO
component = CS_PLATFORM
facility = NASDB
baseid = 300
type = EVENT
brief discription = nasdb_backup: NAS_DB checkpoint in progress
full discription = The Celerra configuration database is being checkpointed.
recommended action = No action required.
logged time = May 12 18:02:59 2007
creation time = May 12 18:02:59 2007
slot id =
id = 96108871985
severity = INFO
component = CS_PLATFORM
facility = NASDB
282
Command Line Interface Reference for File 8.1
The nas Commands
baseid = 305
type = EVENT
brief description = nasdb_backup: NAS_DB Checkpoint done
full description = The NAS DB backup has completed a checkpoint of the current
reparation for performing a backup of NAS system data.
recommended action = No action required.
EXAMPLE #4
To monitor the growth of the current log, type:
$ nas_logviewer -f /nas/log/sys_log|more
May 12 18:01:57 2007:CS_PLATFORM:NASDB:INFO:300:::::nasdb_backup:
checkpoint in progress
May 12 18:02:59 2007:CS_PLATFORM:NASDB:INFO:305:::::nasdb_backup:
Checkpoint done
May 12 18:03:00 2007:CS_PLATFORM:NASDB:ERROR:202:::::NAS database
detectedMay 12 18:03:12 2007
:CS_PLATFORM:NASDB:INFO:306:::::nasdb_backup: NAS DB Backup done
May 12 19:01:52 2007:CS_PLATFORM:NASDB:INFO:300:::::nasdb_backup:
checkpoint in progress
May 12 19:02:50 2007:CS_PLATFORM:NASDB:INFO:305:::::nasdb_backup:
Checkpoint done
May 12 19:02:51 2007:CS_PLATFORM:NASDB:ERROR:202:::::NAS database
detectedMay 12 19:03:02 2007
:CS_PLATFORM:NASDB:INFO:306:::::nasdb_backup: NAS DB Backup done
May 12 20:01:57 2007:CS_PLATFORM:NASDB:INFO:300:::::nasdb_backup:
checkpoint in progress
May 12 20:02:58 2007:CS_PLATFORM:NASDB:INFO:305:::::nasdb_backup:
Checkpoint done
May 12 20:02:59 2007:CS_PLATFORM:NASDB:ERROR:202:::::NAS database
detectedMay 12 20:03:10 2007
:CS_PLATFORM:NASDB:INFO:306:::::nasdb_backup: NAS DB Backup done
May 12 21:01:52 2007:CS_PLATFORM:NASDB:INFO:300:::::nasdb_backup:
checkpoint in progress
May 12 21:02:51 2007:CS_PLATFORM:NASDB:INFO:305:::::nasdb_backup:
Checkpoint done
May 12 21:02:52 2007:CS_PLATFORM:NASDB:ERROR:202:::::NAS database
detectedMay 12 21:03:03 2007
:CS_PLATFORM:NASDB:INFO:306:::::nasdb_backup: NAS DB Backup done
NAS_DB
NAS_DB
error
NAS_DB
NAS_DB
error
NAS_DB
NAS_DB
error
NAS_DB
NAS_DB
error
nas_logviewer
283
The nas Commands
nas_message
Displays message description.
SYNOPSIS
nas_message
-info <MessageId>
DESCRIPTION
OPTIONS
nas_message provides detailed descriptions to a dedicated message.
A brief description, full description, and recommended user action of
the message are displayed.
-info <MessageId>
Displays detailed descriptions of the error message, including
severity, component, facility, BaseID, and recommended user action.
The message parameters are displayed in the form ${stateDesc,8,%s}
and not as parameter values. The <MessageId> must be a positive
integer.
SEE ALSO
EXAMPLE #1
Celerra Network Server Error Messages Guide.
To display detailed descriptions for error message 13421838337, type:
$ nas_message -info 13421838337
MessageID = 13421838337
BaseID
= 1
Severity = ERROR
Component = CS_CORE
Facility = default
Type
= STATUS
Brief_Description = Operation not permitted${arg0,8,%s}
Full_Description
= The operation is not permitted due to an ACL or ownership
issue on the specified object.
Recommended_Action = Check ownership or ACL of the object in question.If
appropriate change the setting to resolve the conflict. Refer to the nas_acl and
chmod man page.
284
Command Line Interface Reference for File 8.1
The nas Commands
nas_mview
Performs MirrorView/Synchronous (MirrorView/S) operations on a
system attached to an older version of VNX for block.
SYNOPSIS
nas_mview
-info
| -init <cel_name>
| -activate
| -restore
DESCRIPTION
nas_mview retrieves MirrorView/S cabinet-level information,
initializes the source and destination systems for MirrorView/S,
activates a failover to a destination VNX for file, or restores the source
site after a failover.
MirrorView/S is supported on a system attached to an older version
of VNX for block array serving as the boot storage, not the secondary
storage. nas_mview must be run from a Control Station in slot 0; it
will report an error if run from a Control Station in slot 1.
nas_mview must be issued as root from the /nas/sbin directory. For
the -init and -info options, log in with your administrative username
and use the su root command to log in as root. For the -activate and
-restore options, you must log in to the destination system using the
remote administration account (for example, dradmin) and log in as
root.
OPTIONS
-info
Displays disaster recovery information such as the MirrorView/S
device group eligible, displays the MirrorView/S Data Mover
configuration for the current system.
-init <cel_name>
Initializes the MirrorView/S relationship between the source and
destination systems based on if the configuration is active/passive
(unidirectional) or active/active’ (bidirectional).
Note: The apostrophe in active/active’ indicates that both sites have source
LUNs mirrored at the other site.
nas_mview
285
The nas Commands
The passphrase-protected relationship between the source and
destination systems in the MirrorView/S configuration must be built
prior to initialization using the nas_cel -create command:
◆
On the destination Control Station in a MirrorView/S
active/passive configuration, use the -init option to specify the
name of the source system.
◆
On the Control Station of each system in a MirrorView/S
active/active’ configuration, use the -init option to specify the
name of the remote system. The active/active configuration is a
bidirectional configuration in which a VNX for file can serve both
as source and destination for another system.
-activate
Executed from the destination system using the remote
administration account, initiates a failover from the source to the
destination system. The activation works as follows:
◆
If the source is available, the -activate option swaps the
primary-secondary role for all mirrors in the MirrorView/S
device group and makes the destination LUNs read/write. The
standby Data Movers acquire the IP and MAC addresses, file
systems, and export tables of their source counterparts.
◆
If the original source site is unavailable, the destination LUNs are
promoted to the primary role, making them visible to the
destination VNX for file. The original source LUNs cannot be
converted to backup images; they stay visible to the source VNX
for file, and the original destination site is activated with new
source (primary) LUNs only. If the source cannot be shut down in
a disaster scenario, any writes occurring after the forced
activation will be lost during a restore.
-restore
Issued from the destination system using the remote administration
account, restores a source system after a MirrorView/S failover, and
fails back the device group to the source system.
The restore process begins by checking the state of the device group.
If the device group state is Local Only (where each mirror has only
the source LUN), the device group will be fully synchronized and
rebuilt before the failback can occur. If the device group condition is
fractured, an incremental synchronization is performed before the
failback occurs. Source devices are then synchronized with the data
on the original destination devices, I/O access is shut down, the
original destination Data Movers are rebooted as remote standbys,
286
Command Line Interface Reference for File 8.1
The nas Commands
and the mirrored devices are failed back. When the source side is
restored, the source Data Movers and their services are restarted.
If the restore fails, the source Control Station is not reachable on the
data network. To complete the restore, access the source, log in as
root, and type /nasmcd/sbin/nas_mview -restore.
SEE ALSO
SYSTEM OUTPUT
Using MirrorView/Synchronous with VNX for Disaster Recovery, nas_cel,
and nas_checkup.
The number associated with the storage device reflects the attached
system; for MirrorView/S, VNX for block displays a prefix of APM
before a set of integers, for example, APM00033900124-0019.
The VNX for block supports the following system-defined AVM
storage pools for MirrorView/S only: cm_r1, cm_r5_performance,
cm_r5_economy, cmata_archive, cmata_r3, cm_r6, and cmata_r6.
EXAMPLE #1
To initialize a destination VNX for file in an active/passive
configuration to communicate with source site source_cs, from the
destination Control Station, type:
# /nas/sbin/nas_mview -init source_cs
Celerra with MirrorView/Synchronous Disaster Recovery
Initializing source_cs
-->
target_cs
Contacting source_cs for remote storage info
Local storage system: APM00053001549
Remote storage system: APM00053001552
Enter the
Username:
Password:
Password:
Global CLARiiON account information
emc
***
Retype your response to validate
***
Discovering storage on source_cs (may take several minutes)
Setting security information for APM00053001549
Discovering storage APM00053001552 (may take several minutes)
Discovering storage (may take several minutes)
Contacting source_cs for remote storage info
Gathering server information...
Contacting source_cs for server capabilities...
Analyzing server information...
Source servers available to be configured for remote DR
-------------------------------------------------------
nas_mview
287
The nas Commands
1.
server_2:source_cs
2.
server_3:source_cs [ local standby ]
v.
Verify standby server configuration
q.
Quit initialization process
c.
Continue initialization
Select a source_cs server: 1
Destination servers available to act as remote standby
-----------------------------------------------------1.
server_2:target_cs [ unconfigured standby ]
2.
server_3:target_cs [ unconfigured standby ]
b.
Back
Select a target_cs server: 1
Source servers available to be configured for remote DR
------------------------------------------------------1.
server_2:source_cs [ remote standby is server_2:target_cs ]
2.
server_3:source_cs [ local standby ]
v.
Verify standby server configuration
q.
Quit initialization process
c.
Continue initialization
Select a source_cs server: 2
Destination servers available to act as remote standby
-----------------------------------------------------server_2:target_cs [ is remote standby for server_2:source_cs ]
2.
server_3:target_cs [ unconfigured standby ]
b.
Back
Select a target_cs server: 2
Source servers available to be configured for remote DR
------------------------------------------------------1.
server_2:source_cs [ remote standby is server_2:target_cs ]
2.
server_3:source_cs [ remote standby is server_3:target_cs ]
v.
Verify standby server configuration
q.
Quit initialization process
c.
Continue initialization
Select a source_cs server: c
Standby configuration validated OK
Enter user information for managing remote site source_cs
Username: dradmin
Password: *******
Retype your response to validate
Password: *******
Active/Active configuration
Initializing (source_cs-->target_cs)
Do you wish to continue? [yes or no] yes
Updating MirrorView configuration cache
288
Command Line Interface Reference for File 8.1
The nas Commands
Setting up server_3 on source_cs
Setting up server_2 on source_cs
Creating user account dradmin
Setting acl for server_3 on target_cs
Setting acl for server_2 on target_cs
Updating the Celerra domain information
Creating device group mviewgroup on source_cs
done
EXAMPLE #2
To get information about a source MirrorView configuration (for
example, on new_york configured as active/passive), type:
# /nas/sbin/nas_mview -info
***** Device Group Configuration *****
name
= mviewgroup
description
=
uid
= 50:6:1:60:B0:60:26:BC:0:0:0:0:0:0:0:0
state
= Consistent
role
= Primary
condition
= Active
recovery policy
= Automatic
number of mirrors
= 16
mode
= SYNC
owner
= 0
mirrored disks
=
root_disk,root_ldisk,d5,d8,d10,d11,d24,d25,d26,d27,d29,d30,d31,d32,d33,d39,
local clarid
= APM00053001552
remote clarid
= APM00053001549
mirror direction
= local -> remote
***** Servers configured with RDFstandby *****
id
=
name
=
acl
=
type
=
slot
=
member_of =
standby
=
RDFstandby=
status
:
defined =
actual =
1
server_2
1000, owner=nasadmin, ID=201
nas
2
id
name
acl
type
slot
member_of
2
server_3
1000, owner=nasadmin, ID=201
standby
3
=
=
=
=
=
=
server_3, policy=auto
slot=2
enabled
online, active
nas_mview
289
The nas Commands
standbyfor=
RDFstandby=
status
:
defined =
actual =
server_2
slot=3
enabled
online, ready
***** Servers configured as standby *****
No servers configured as standby
Where:
Value
Device group configuration:
name
description
uid
state
role
condition
recovery policy
number of mirrors
mode
owner
mirrored disks
local clarid
remote clarid
mirror direction
290
Definition
Name of the consistency (device) group.
Brief description of device group.
UID assigned, based on the system.
State of the device group (for example, Consistent, Synchronized, Out-of-Sync, Synchronizing,
Scrambled, Empty, Incomplete, or Local Only).
Whether the current system is the Primary (source) or Secondary (destination) for this group.
Whether the group is functioning (Active), Inactive, Admin Fractured (suspended), Waiting on
Sync, System Fractured (which indicates link down), or Unknown.
Type of recovery policy (Automatic is the default and recommended value for group during
system configuration; if Manual is set, you must use -resume after a link down failure).
Number of mirrors in group.
MirrorView mode (always SYNC in this release).
ACL ID assigned (0 indicates no control). nas_acl provides information.
Comma-separated list of disks that are mirrored.
APM number of local VNX for block storage array.
APM number of remote VNX for block storage array.
On primary system, local to remote (on primary system); on destination system, local from
remote.
Command Line Interface Reference for File 8.1
The nas Commands
Value
Servers configured with
RDFstandby/ Servers
configured as standby:
id
name
acl
type
slot
member_of
standby
RDFstandby
standbyfor
status
Definition
Server ID
Server name
ACL value and owner
Server type (for example, nas or standby)
Slot number for this Data Mover
If applicable, shows membership information.
If this Data Mover is configured with local standbys, the server that is the local standby and any
policy information.
If this Data Mover is configured with a remote RDF standby, the slot number of the destination
Data Mover that serves as the RDF standby.
If this Data Mover is also configured as a local standby, the server numbers for which it is a local
standby.
Indicates whether the Data Mover is defined and online/ready.
EXAMPLE #3
To activate a failover, log in to destination Control Station using
dradmin account, su to root, and type:
# /nas/sbin/nas_mview -activate
Sync with CLARiiON backend ...... done
Validating mirror group configuration ...... done
Is source site source_cs ready for complete shut down (power OFF)? [yes or no] yes
Contacting source site source_cs, please wait... done
Shutting down remote site source_cs ...................................... done
Sync with CLARiiON backend ...... done
STARTING an MV 'FAILOVER' operation.
Device group: mviewgroup ............ done
The MV 'FAILOVER' operation SUCCEEDED.
Failing over Devices ... done
Adding NBS access for server_2 ........ done
Adding NBS access for server_3 ........ done
Activating the target environment ... done
server_2 : going offline
rdf : going active
replace in progress ...done
failover activity complete
server_3 : going offline
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
nas_mview
291
The nas Commands
commit
commit
commit
in progress (not interruptible)...done
in progress (not interruptible)...done
in progress (not interruptible)...done
done
EXAMPLE #4
To restore, log in to the destination Control Station using dradmin
account, as a root user, and type:
# /nas/sbin/nas_mview -restore
Sync with CLARiiON backend ...... done
Validating mirror group configuration ...... done
Contacting source site source_cs, please wait... done
Running restore requires shutting down source site source_cs.
Do you wish to continue? [yes or no] yes
Shutting down remote site source_cs ....... done
Is source site source_cs ready for storage restoration ? [yes or no] yes
Sync with CLARiiON backend ...... done
STARTING an MV 'RESUME' operation.
Device group: mviewgroup ............ done
The MV 'RESUME' operation SUCCEEDED.
Percent synchronized: 100
Updating device group ... done
Is source site ready for network restoration ? [yes or no] yes
Restoring servers ...... done
Waiting for servers to reboot ...... done
Removing NBS access for server_2 .. done
Removing NBS access for server_3 .. done
Waiting for device group ready to failback .... done
Sync with CLARiiON backend ...... done
STARTING an MV 'FAILBACK' operation.
Device group: mviewgroup ............ done
The MV 'FAILBACK' operation SUCCEEDED.
Restoring remote site source_cs ...... failed
Error 5008: -1:Cannot restore source_cs. Please run restore on site source_cs.
Then on the Source Control Station, as a root user, type:
# /nasmcd/sbin/nas_mview -restore
Stopping NAS services. Please wait...
Powering on servers ( please wait ) ...... done
Sync with CLARiiON backend ...... done
STARTING an MV 'SUSPEND' operation.
Device group: mviewgroup ............ done
The MV 'SUSPEND' operation SUCCEEDED.
server_2 : going standby
rdf : going active
replace in progress ...done
failover activity complete
server_3 : going standby
rdf : going active
292
Command Line Interface Reference for File 8.1
The nas Commands
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
commit in progress (not interruptible)...done
Sync with CLARiiON backend ...... done
STARTING an MV 'RESUME' operation.
Device group: mviewgroup ............ done
The MV 'RESUME' operation SUCCEEDED.
Restarting NAS services ...... done
commit in progress (not interruptible)...done
commit in progress (not interruptible)...done
done
nas_mview
293
The nas Commands
nas_pool
Manages the user-defined and system-defined storage pools for the
system.
SYNOPSIS
nas_pool
-list
| -info {<name>|id=<id>|-all}[-storage <system_name>]
| -size {<name>|id=<id>|-all} [-mover <mover>][-slice {y|n}]
[-storage <system_name>]
| -create [-name <name>][-acl <acl>][-description <desc>]
[-volumes <volume_name>[,<volume_name>,...]]
[-default_slice_flag {y|n}] [-is_greedy {y|n}]
| -create [-name <name>][-acl <acl>][-description <desc>]
[-default_slice_flag {y|n}] [-is_greedy {y|n}]
-size <integer>[M|G|T][-storage <system_name>]
-template <system_pool_name> [-num_stripe_members <num>]
[-stripe_size <num>]
| -modify {<name>|id=<id>} [-storage <system_name>] [-name <name>]
[-acl <acl>] [-description <desc>][-default_slice_flag {y|n}]
[-is_dynamic {y|n}][-is_greedy {y|n}]
| -delete {<name>|id=<id>} [-deep] [-storage <system_name>]
| -xtend {<name>|id=<id>} [-storage <system_name>]
-volumes <volume_name>[,<volume_name>,...]
| -xtend {<name>|id=<id>} -size <integer> [M|G|T][-storage <system_name>]
| -shrink {<name>|id=<id>} [-storage <system_name>] -volumes
<volume_name>[,<volume_name>,...][-deep]
DESCRIPTION
nas_pool creates, deletes, extends, shrinks, lists, displays, manages
the access control level, and modifies a user-defined storage pool.
nas_pool extends, shrinks, lists, displays, and modifies
system-defined storage pools.
OPTIONS
-list
Lists all storage pools on the system.
-info {<name>|id=<id>|-all} [-storage <system_name>]
Displays detailed information for the specified storage pool, or all
storage pools. The -storage option can be used to differentiate pools
when the same pool name is used in multiple systems.
-size {<name>|id=<id>|-all}
Displays the size for the specified storage pool, or all storage pools.
294
Command Line Interface Reference for File 8.1
The nas Commands
[-mover <mover>]
Displays size information that is visible to the physical Data
Mover or the virtual Data Mover (VDM).
[-slice {y|n}]
If y is typed, displays size information when volumes in the
storage pool are sliced. If n is typed, displays size information
when volumes in the storage pool are not sliced. The -slice option
defaults to the value of default_slice_flag for the storage pool.
[-storage <system_name>]
Displays size information for members that reside on a specified
system.
-create
Creates a user-defined storage pool.
[-name <name>]
Assigns a name to the new storage pool. If no name is specified,
assigns one by default.
[-acl <acl>]
Sets an access control level value that defines the owner of the
storage pool, and the level of access allowed for users and groups
defined in the access control level table. The nas_acl command
provides more information.
[-description <desc>]
Assigns a comment to the storage pool.
[-volumes <volume_name>[,<volume_name>,...]
Designates the members to be added to the storage pool. The
members can be any meta, slice, stripe, or disk volumes.
[-default_slice_flag {y|n}]
If set to y (default), then members of the storage pool might be
sliced when space is allocated from the storage pool. If set to n,
members of the storage pool will not be sliced when space is
dispensed from the storage pool and the volumes specified
cannot be built on a slice.
[-is_greedy {y|n}]
If set to n (default), the system uses space from the user-defined
storage pool's existing member volumes in the order that the
volumes were added to the pool to create a new file system or
extend an existing file system.
nas_pool
295
The nas Commands
If set to y, the user-defined storage pool uses space from the
least-used member volume to create a new file system. When
there is more than one least-used member volume available, AVM
selects the member volume that contains the most disk volumes.
For example, if one member volume contains four disk volumes
and another member volume contains eight disk volumes, AVM
selects the one with eight disk volumes. If there are two or more
member volumes that have the same number of disk volumes,
AVM selects the one with the lowest ID.
[-size <integer> {M|G|T}]
Creates a storage pool with the size specified. M specifies
megabytes, G specifies gigabytes (default), and T specifies
terabytes. The maximum size that you can specify for a storage
pool is the maximum supported storage capacity for the system.
[-storage <system_name>]
Specifies the system on which one or more volumes will be
created, to be added to the storage pool.
[-template <system_pool_name>]
Specifies a system pool name, required when the -size option is
specified. The user pool will be created using the profile attributes
of the specified system pool template.
[-num_stripe_members <num>]
Specifies the number of stripe members for user pool creation by
size. The -num_stripe_members option works only when both
-size and -template options are specified. It overrides the number
of stripe members attribute of the specified system pool template.
[-stripe_size <num>]
Specifies the stripe size for user pool creation by size. The
-stripe_size option works only when both -size and -template
options are specified. It overrides the stripe size attribute of the
specified system pool template.
-modify {<name>|id=<id>} [-storage <system_name>]
Modifies the attributes of the specified user-defined or
system-defined storage pool. The -storage option can be used to
differentiate pools when the same pool name is used in multiple
systems.
Managing Volumes and File Systems with VNX Automatic Volume
Management lists the available system-defined storage pools.
296
Command Line Interface Reference for File 8.1
The nas Commands
[-name <name>]
Changes the name of the storage pool to the new name.
[-acl <acl>]
Sets an access control level value that defines the owner of the
storage pool, and the level of access allowed for users and groups
defined in the access control level table. The nas_acl command
provides more information.
[-description <desc>]
Changes the comment for the storage pool.
[-default_slice_flag {y|n}]
If set to y (default), then members of the storage pool might be
sliced when space is dispensed from the storage pool. If set to n,
members of the storage pool will not be sliced when space is
dispensed from the storage pool and the volumes specified
cannot be built on a slice.
[-is_dynamic {y|n}]
Allows a system-defined storage pool to automatically extend or
shrink member volumes.
Note: The -is_dynamic option is for system-defined storage pools only.
[-is_greedy {y|n}]
For system-defined storage pools, if set to y, then the storage pool
attempts to create new member volumes before using space from
existing member volumes. A system-defined storage pool that is
not greedy (set to n), consumes all the space existing in the
storage pool before trying to add additional member volumes. A
y or n value must be specified when modifying a system-defined
storage pool.
For user-defined storage pools, if set to n (default), the system
uses space from the user-defined storage pool's existing member
volumes in the order that the volumes were added to the pool to
create a new file system.
For user-defined storage pools, if set to y, the system uses space
from the least-used member volume in the user-defined storage
pool to create a new file system. When there is more than one
least-used member volume available, AVM selects the member
volume that contains the most disk volumes. For example, if one
member volume contains four disk volumes and another member
volume contains eight disk volumes, AVM selects the one with
nas_pool
297
The nas Commands
eight disk volumes. If there are two or more member volumes
that have the same number of disk volumes, AVM selects the one
with the lowest ID.
For both system-defined and user-defined pools when extending
a file system, the is_greedy attribute is ignored unless there is not
enough free space on the existing volumes that the file system is
using to meet the requested extension size.
-delete {<name>|id=<id>} [-storage <system_name>]
Deletes a storage pool. Storage pools cannot be deleted if any
members are in use. After deletion, the storage pool no longer exists
on the system; however, members of the storage pool are not deleted.
The -storage option can be used to differentiate pools when the same
pool name is used in multiple systems.
[-deep]
Deletes the storage pool and also recursively deletes each
member of the storage pool. Each storage pool member is deleted
unless it is in use or is a disk volume.
-xtend {<name>|id=<id>} [-storage <system_name>]
-volumes <volume_name>[, <volume_name>,...]
Adds one or more unused volumes to a storage pool. The -storage
option can be used to differentiate pools when the same pool name is
used in multiple systems. If the default_slice_value is set to n,
member volumes cannot contain slice volumes (for compatibility
with TimeFinder/FS).
Note: Extending a storage pool by volume is for user-defined storage pools
only.
-xtend {<name>|id=<id>} -size <integer> [M|G|T]
Extends the specified storage pool with one or more volumes of the
size equal to or greater than the size specified. When specifying the
volume by size, type an integer between 1 and 1024, then specify T
for terabytes, G for gigabytes (default), or M for megabytes.
[-storage <system_name>]
Specifies the system on which one or more volumes will be
created, to be added to the storage pool.
Note: To successfully extend a system-defined storage pool by size, the
is_dynamic attribute must be set to n, and there must be enough
available disk volumes to satisfy the request.
298
Command Line Interface Reference for File 8.1
The nas Commands
-shrink {<name>|id=<id>} [-storage <system_name>]
-volumes <volume_name>[,<volume_name>,...][-deep]
Shrinks the storage pool by the specified unused volumes. The
-storage option can be used to differentiate pools when the same pool
name is used in multiple systems. When the -deep option is used to
shrink a user-defined storage pool, it removes the specified member
volumes from the pool, and recursively deletes any unused volumes
unless it is a disk volume. If the -deep option is not used to shrink a
user-defined storage pool, the member volumes are left intact so that
they can be reused. The is_dynamic option must be set to n before
shrinking system-defined storage pools.
Note: Shrinking of a system-defined storage pool by default deletes
member volumes automatically. Specifying the -deep option on the
system-defined storage pool shrink does not make any difference.
SEE ALSO
Managing Volumes and File Systems with VNX Automatic Volume
Management, Managing Volumes and File Systems for VNX Manually,
Controlling Access to System Objects on VNX, Using TimeFinder/FS,
NearCopy, and FarCopy on VNX for File, fs_timefinder, nas_fs,
nas_volume, and nas_slice.
SYSTEM OUTPUT
VNX for block supports the following traditional system-defined
storage pools: clar_r1, clar_r5_performance, clar_r5_economy,
clar_r6, clarata_r3, clarata_r6, clarata_r10, clarata_archive, cm_r1,
cm_r5_performance, cm_r5_economy, cm_r6, cmata_r3,
cmata_archive, cmata_r6, cmata_r10, clarsas_archive, clarsas_r6,
clarsas_r10, clarefd_r5, clarefd_r10, cmsas_archive, cmsas_r6,
cmsas_r10, and cmefd_r5.
A mapped pool was formerly called a thin or virtual pool.
Disk types when using VNX for block are CLSTD, CLEFD, CLATA,
MIXED (indicates that tiers used in the pool contain multiple disk
types), Performance, Capacity, and Extreme_performance and for
VNX for block involving mirrored disks are: CMEFD, CMSTD,
CMATA, Mirrored_mixed, Mirrored_performance,
Mirrored_capacity, and Mirrored_extreme_performance.
Disk types when using VNX for block are CLSTD, CLEFD, and
CLATA, and for VNX for block involving mirrored disks are CMEFD,
CMSTD, and CMATA.
VNX with a Symmetrix system supports the following
system-defined storage pools: symm_std, symm_std_rdf_src,
nas_pool
299
The nas Commands
symm_ata, symm_ata_rdf_src, symm_ata_rdf_tgt,
symm_std_rdf_tgt, symm_efd, symm_fts, symm_fts_rdf_tgt and
symm_fts_rdf_src.
For user-defined storage pools, the difference in output is in the disk
type. Disk types when using a Symmetrix are STD, R1STD, R2STD,
BCV, R1BCV, R2BCV, ATA, R1ATA, R2ATA, BCVA, R1BCA,
R2BCA, EFD, FTS, R1FTS, R2FTS, R1BCF, R2BCF, BCVF,
BCVMIXED, R1MIXED, R2MIXED, R1BCVMIXED, and
R2BCVMIXED.
EXAMPLE #1
id
name
description
acl
in_use
clients
members
storage_system(s)
default_slice_flag
is_user_defined
thin
disk_type
server_visibility
is_greedy
template_pool
num_stripe_members
stripe_size
To create a storage pool with the name, marketing, with a
description, with the following disk members, d12, d13, and with the
default slice flag set to y, type:
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
$ nas_pool -create -name marketing -description 'Storage
Pool' -volumes d12,d13 -default_slice_flag y
20
marketing
Storage Pool
0
False
d12,d13
FNM00105000212
True
True
False
CLSTD
server_2,server_3,server_4,server_5
False
N/A
N/A
N/A
Where:
Value
id
name
description
acl
in_use
clients
members
storage_systems(s)
default_slice_flag
is_user_defined
300
Definition
ID of the storage pool.
Name of the storage pool.
Comment assigned to the storage pool.
Access control level value assigned to the storage pool.
Whether the storage pool is being used by a file system.
File systems using the storage pool.
Volumes used by the storage pool.
Storage systems used by the storage pool.
Allows slices from the storage pool.
User-defined as opposed to system-defined.
Command Line Interface Reference for File 8.1
The nas Commands
Indicates whether thin provisioning is enabled or disabled.
Type of disk contingent on the system attached. CLSTD, CLATA, CMSTD, CLEFD,
CMEFD, CMATA, MIXED (indicates tiers used in the pool contain multiple disk types),
Performance, Capacity, Extreme_performance, Mirrored_mixed, Mirrored_performance,
Mirrored_capacity, and Mirrored_extreme_performance are for VNX for block, and STD,
BCV, R1BCV, R2BCV, R1STD, R2STD, ATA, R1ATA, R2ATA, BCVA, R1BCA, R2BCA,
EFD, BCVMIXED, R1MIXED, R2MIXED, R1BCVMIXED, and R2BCVMIXED are for
Symmetrix.
Storage pool is visible to the physical Data Movers specified.
Indicates whether the system-defined storage pool will use new member volumes as
needed.
System pool template used to create the user pool. Only applicable to user pools created
by size or if the last member volume is a stripe or both.
Number of stripe members used to create the user pool. Applicable to system pools and
user pools created by size or if the last member volume is a stripe or both.
Stripe size used to create the user pool. Applicable to system pools and user pools
created by size or if the last member volume is a stripe or both.
thin
disk_type
server_visibility
is_greedy
template_pool
num_stripe_members
stripe_size
EXAMPLE #2
id
name
description
acl
in_use
clients
members
storage_system(s)
default_slice_flag
is_user_defined
thin
disk_type
server_visibility
is_greedy
template_pool
num_stripe_members
stripe_size
To change the description for the marketing storage pool to include a
descriptive comment, type:
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
$ nas_pool -modify marketing -description 'Marketing
Storage Pool'
20
marketing
Marketing Storage Pool
0
False
d12,d13
FNM00105000212
True
True
False
CLSTD
server_2,server_3,server_4,server_5
False
N/A
N/A
N/A
EXAMPLE #1 provides a description of command output.
EXAMPLE #3
To view the size information for the FP1 mapped pool, type:
$ nas_pool -size FP1
id
= 40
name
= FP1
used_mb
= 0
avail_mb
= 0
nas_pool
301
The nas Commands
total_mb
= 0
potential_mb = 2047
Where:
Value
used_mb
avail_mb
total_mb
potential_mb
Definition
Space in use by the storage pool specified.
Unused space still available in the storage pool.
Total space in the storage pool (total of used and unused).
Available space that can be added to the storage pool.
Note: Each of the options used with the command nas_pool - size filters for
the output of the command. For example, if you specify a Data Mover, the
output will reflect only the space to which the specified Data Mover has
visibility. Physical used_mb, Physical avail_mb, and Physical total_mb are
applicable for system-defined virtual AVM pools only.
EXAMPLE #4
To view the size information for the TP1 mapped pool that contains
only virtual LUNs, type:
$ nas_pool -size TP1
id
= 40
name
= TP1
used_mb
= 0
avail_mb
= 0
total_mb
= 0
potential_mb = 2047
Physical storage usage in tp1 on FCNTR074200038:
used_mb = 0
avail_mb = 20470
Where:
Value
Physical used_mb
Physical avail_mb
Definition
Used physical size of a system mapped pool in MB (some may be used by non-VNX hosts).
Available physical size of a system mapped pool in MB.
Note: Physical used_mb and Physical avail_mb are applicable for
system-defined AVM pools that contain virtual LUNs only.
EXAMPLE #5
id
name
description
acl
302
For a VNX system, to change the -is_greedy and -is_dynamic options
for the system defined, clar_r5_performance storage pool, type:
=
=
=
=
$ nas_pool -modify clar_r5_performance -is_dynamic n
-is_greedy y
3
clar_r5_performance
CLARiiON RAID5 4plus1
421
Command Line Interface Reference for File 8.1
The nas Commands
in_use
clients
members
storage_system(s)
default_slice_flag
is_user_defined
thin
disk_type
server_visibility
volume_profile
is_dynamic
is_greedy
num_stripe_members
stripe_size
=
=
=
=
=
=
=
=
=
=
=
=
=
=
False
v120
True
False
False
CLSTD
server_2,server_3,server_4,server_5
clar_r5_performance_vp
False
True
4
32768
EXAMPLE #1 provides a description of command output.
EXAMPLE #6
For a VNX for file with a Symmetrix system, to change the -is_greedy
and -is_dynamic options for the system-defined, symm_std storage
pool, type:
$ nas_pool -modify symm_std -is_dynamic y -is_greedy y
id
name
description
acl
in_use
clients
members
default_slice_flag
is_user_defined
thin
disk_type
compressed
server_visibility
volume_profile
is_dynamic
is_greedy
num_stripe_members
stripe_size
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
1
symm_std
Symmetrix STD
1421, owner=nasadmin, ID=201
True
ufs3
v169,v171
False
False
False
STD
True
server_2,server_3,server_4,server_5
symm_std_vp
True
True
8
32768
Where:
Value
id
name
description
acl
in_use
Definition
ID of the storage pool.
Name of the storage pool.
Comment assigned to the storage pool.
Access control level value assigned to the storage pool.
Whether the storage pool is being used by a file system.
nas_pool
303
The nas Commands
clients
members
default_slice_flag
is_user_defined
thin
disk_type
compressed
server_visibility
volume_profile
is_dynamic
is_greedy
template_pool
num_stripe_members
stripe_size
EXAMPLE #7
File systems using the storage pool.
Disks used by the storage pool.
Allows slices from the storage pool.
User-defined as opposed to system-defined.
Indicates whether thin provisioning is enabled or disabled.
Contingent on the system attached.
For VNX with Symmetrix backend, indicates whether data is compressed. Values are:
True, False, Mixed (indicates some of the LUNs, but not all, are compressed).
Storage pool is visible to the physical Data Movers specified.
Volume profile used.
Whether the system-defined storage pool can add or remove volumes.
Indicates whether the system-defined storage pool will use new member volumes as
needed.
System pool template used to create the user pool. Only applicable to user pools created
by size or if the last member volume is a stripe or both.
Number of stripe members used to create the user pool. Applicable to system pools and
user pools created by size or if the last member volume is a stripe or both.
Stripe size used to create the user pool. Applicable to system pools and user pools created
by size or if the last member volume is a stripe or both.
To change the -is_greedy option for the user-defined, user_pool
storage pool, type:
$ nas_pool -modify user_pool -is_greedy y
id
= 58
name
= user_pool
description
=
acl
= 0
in_use
= False
clients
=
members
= d21,d22,d23,d24
storage_system(s)
= FNM00105000212
default_slice_flag
= True
is_user_defined
= True
thin
= False
disk_type
= CLSTD
server_visibility
= server_2
is_greedy
= True
template_pool
= N/A
num_stripe_members
= N/A
stripe_size
= N/A
EXAMPLE #1 provides a description of command output.
EXAMPLE #8
id
304
To add the volumes, d7 and d8, to the marketing storage pool, type:
$ nas_pool -xtend marketing -volumes d7,d8
= 20
Command Line Interface Reference for File 8.1
The nas Commands
name
description
acl
in_use
clients
members
default_slice_flag
is_user_defined
thin
disk_type
server_visibility
template_pool
num_stripe_members
stripe_size
=
=
=
=
=
=
=
=
=
=
=
=
=
=
marketing
Marketing Storage Pool
0
False
d12,d13,d7,d8
True
True
True
CLSTD
server_2,server_3,server_4,server_5
N/A
N/A
N/A
EXAMPLE #1 provides a description of command output.
EXAMPLE #9
id
name
description
acl
in_use
clients
members
default_slice_flag
is_user_defined
thin
disk_type
server_visibility
volume_profile
is_dynamic
is_greedy
num_stripe_members
stripe_size
For a VNX system, to extend the system-defined storage pool by a
specified size with a specified system, type:
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
$ nas_pool -xtend clar_r5_performance -size 128M -storage
APM00042000818
3
clar_r5_performance
CLARiiON RAID5 4plus1
1421, owner=nasadmin, ID=201
False
v120
True
False
False
CLSTD
server_2,server_3,server_4,server_5
clar_r5_performance_vp
False
True
4
32768
EXAMPLE #1 provides a description of command output.
EXAMPLE #10
id
name
description
acl
in_use
clients
For VNX system, to remove d7 and d8 from the marketing storage
pool, type:
=
=
=
=
=
=
$ nas_pool -shrink marketing -volumes d7,d8
20
marketing
Marketing Storage Pool
0
False
nas_pool
305
The nas Commands
members
default_slice_flag
is_user_defined
thin
disk_type
server_visibility
template_pool
num_stripe_members
stripe_size
=
=
=
=
=
=
=
=
=
d12,d13
True
True
True
CLSTD
server_2,server_3,server_4,server_5
N/A
N/A
N/A
EXAMPLE #1 provides a description of command output.
id
2
3
4
10
11
20
40
41
EXAMPLE #11
To list the storage pools, type:
inuse
n
n
n
n
n
n
y
y
$ nas_pool -list
name
clar_r1
clar_r5_performance
clar_r5_economy
clarata_archive
clarata_r3
marketing
TP1
FP1
acl
421
421
421
421
421
0
0
0
storage_system
N/A
FCNTR074200038
N/A
FCNTR074200038
N/A
FCNTR074200038
FCNTR074200038
FCNTR074200038
Where:
Value
id
inuse
acl
name
storage_system
Definition
ID of the storage pool.
Whether the storage pool is being used by a file system.
Access control level value assigned to the storage pool.
Name of the storage pool.
Name of the system where the storage pool resides.
EXAMPLE #12
id
name
description
acl
in_use
clients
members
storage_system(s)
default_slice_flag
is_user_defined
thin
disk_type
server_visibility
306
To display information about the user-defined storage pool called
marketing, type:
=
=
=
=
=
=
=
=
=
=
=
=
=
$ nas_pool -info marketing
20
marketing
Marketing Storage Pool
0
False
d12,d13
True
True
True
CLSTD
server_2,server_3,server_4,server_5
Command Line Interface Reference for File 8.1
The nas Commands
is_greedy
template_pool
num_stripe_members
stripe_size
=
=
=
=
False
N/A
N/A
N/A
EXAMPLE #1 provides a description of command output.
EXAMPLE #13
id
name
description
acl
in_use
clients
members
default_slice_flag
is_user_defined
thin
disk_type
server_visibility
volume_profile
is_dynamic
is_greedy
num_stripe_members
stripe_size
To display information about the system-defined
clar_r5_performance storage pool, type:
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
$ nas_pool -info clar_r5_performance
3
clar_r5_performance
CLARiiON RAID5 4plus1
1421, owner=nasadmin, ID=201
False
v120
True
False
False
CLSTD
server_2,server_3,server_4,server_5
clar_r5_performance_vp
False
True
4
32768
EXAMPLE #1 provides a description of command output.
EXAMPLE #14
id
name
description
acl
in_use
clients
members
default_slice_flag
is_user_defined
thin
disk_type
server_visibility
volume_profile
is_dynamic
is_greedy
num_stripe_members
stripe_size
To display information about the system-defined engineer virtual
pool, type:
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
$ nas_pool -info engineer
40
engineer
Mapped Pool engineer on APM00084401666
0
True
DA_BE_VIRT_FS,vp_test,vp_test1,vp_test12,cvpfs1,cvpfs3
v363
True
False
True
CLSTD
server_2,server_3
engineer_APM00084401666_vp
True
True
N/A
N/A
nas_pool
307
The nas Commands
EXAMPLE #1 provides a description of command output.
EXAMPLE #15
id
name
description
acl
in_use
clients
members
default_slice_flag
is_user_defined
thin
tiering_policy
compressed
mirrored
disk_type
volume_profile
is_dynamic
is_greedy
To display information about the mapped storage pool called FP1
from a VNX for block, type:
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
$ nas_pool -info FP1
40
FP1
Mapped Pool on FCNTR074200038
0
False
True
False
True
Auto-tier
False
False
Mixed
FP1
True
True
Where:
Value
Definition
tiering_policy Indicates the tiering policy in effect. If the initial tier and the tiering policy are the same, the values are:
Auto-Tier, Highest Available Tier, Lowest Available Tier. If the initial tier and the tiering policy are not the
same, the values are: Auto-Tier/No Data Movement, Highest Available Tier/No Data Movement, Lowest
Available Tier/No Data Movement.
compressed
For VNX for block, indicates whether data is compressed. Values are: True, False, Mixed (indicates that
some of the LUNs, but not all, are compressed).
mirrored
Indicates whether the disk is mirrored.
EXAMPLE #16
id
name
description
acl
in_use
clients
members
default_slice_flag
is_user_defined
thin
308
To display information about the mapped storage pool called SG0
from a Symmetrix system, type:
=
=
=
=
=
=
=
=
=
=
$ nas_pool -info SG0
40
SG0
Symmetrix Mapped Pool on 000192601245
0
False
True
False
True
Command Line Interface Reference for File 8.1
The nas Commands
tiering_policy
compressed
frontend_io_quota
disk_type
volume_profile
is_dynamic
is_greedy
=
=
=
=
=
=
=
symm_policy_1
True
maxiopersec 500,maxmbpersec 500
Mixed
True
True
N/A
Where:
Value
id
name
description
acl
in_use
clients
members
default_slice_flag
is_user_defined
thin
tiering_policy
compressed
frontend_io_quota
disk_type
volume_profile
is_dynamic
is_greedy
EXAMPLE #17
Definition
ID of the storage pool.
Name of the storage pool.
Comment assigned to the storage pool.
Access control level value assigned to the storage pool.
Whether the storage pool is being used by a file system.
File systems using the storage pool.
Volumes used by the storage pool.
Allows slices from the storage pool.
User-defined as opposed to system-defined.
Indicates whether thin provisioning is enabled or disabled.
Indicates the tiering policy in effect. If the initial tier and the tiering policy are the same,
the values are: Auto-Tier, Highest Available Tier, Lowest Available Tier. If the initial tier
and the tiering policy are not the same, the values are: Auto-Tier/No Data Movement,
Highest Available Tier/No Data Movement, Lowest Available Tier/No Data Movement.
For VNX with Symmetrix backend, indicates whether data is compressed. Values are:
True, False, Mixed (indicates some of the LUNs, but not all, are compressed).
For VNX with Symmetrix backend, indicates if Frotend IO Quota is configured on this
mapped pool, could also have value as False (indicates Frontend IO Quota is not
configured on mapped SG in Symmetrix backend).
Type of disk contingent on the system attached. CLSTD, CLATA, CMSTD, CLEFD,
CMEFD, CMATA, MIXED (indicates tiers used in the pool contain multiple disk types),
Performance, Capacity, Extreme_performance, Mirrored_mixed, Mirrored_performance,
Mirrored_capacity, and Mirrored_extreme_performance are for VNX for block, and STD,
BCV, R1BCV, R2BCV, R1STD, R2STD, ATA, R1ATA, R2ATA, BCVA, R1BCA, R2BCA,
EFD, BCVMIXED, R1MIXED, R2MIXED, R1BCVMIXED, and R2BCVMIXED are for
Symmetrix.
Volume profile used.
Whether the system-defined storage pool can add or remove volumes.
Indicates whether the system-defined storage pool will use new member volumes as
needed.
To delete the storage pool, marketing, and each of the storage pool
member volumes recursively, type:
$ nas_pool -delete marketing -deep
nas_pool
309
The nas Commands
id
name
description
acl
in_use
clients
members
storage_system(s)
default_slice_flag
is_user_defined
is_greedy
thin
template_pool
num_stripe_members
stripe_size
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
20
marketing
Marketing Storage Pool
0
False
True
True
True
True
N/A
N/A
N/A
EXAMPLE #1 provides a description of command output.
310
Command Line Interface Reference for File 8.1
The nas Commands
nas_quotas
Manages quotas for mounted file systems.
SYNOPSIS
nas_quotas
-edit [-user|-group] {-mover <movername>|-fs <fs_name>} [-path
<pathname>]} [[-proto <proto_id>]|[-block <hard_limit>[:<soft_limit>]]
[-inode <hard_limit>[:<soft_limit>]]] <id> [<id>...]
| -edit -config {-mover <movername>|-fs <fs_name>} [-path <pathname>]}
[-option <options>]
| -edit -tree -fs <fs_name>
[[-proto <proto_id>]|[-block <hard_limit>[:<soft_limit>]]
[-inode <hard_limit>[:<soft_limit>]]]
[-comment <comment>] <id> [<id>...]
| -report [-user|-group] {-mover <movername>|-fs <fs_name>}
[-path <pathname>]} [<id> <id>...]
| -report -config {-mover <movername>|-fs <fs_name>} [-path <pathname>]}
| -report -tree -fs <fs_name> [<id> <id>...]
| {-on|-off|-clear} [-user|-group|-both]
{-mover <movername>|-fs <fs_name>|[-path <pathname>] -all}
| -on -tree -fs <fs_name> -path <pathname> [-comment <comment>]
| -off -tree -fs <fs_name> -path <pathname>
| {-list|-clear} -tree -fs <fs_name>
| -check -start [-mode online|offline] [-tree] -fs <fs_name> [-path
<pathname>]
| -check {-stop|-status} -fs <fs_name> [-path <pathname>]
| -quotadb {-info|-upgrade [-Force]} {-mover <movername>|-fs <fs_name>}
DESCRIPTION
nas_quotas edits quotas for mounted file systems, and displays a
listing of quotas and disk usage at the file system level (by the user,
group, or tree), or at the quota-tree level (by the user or group).
nas_quotas also turns quotas on and off, and clears quotas records
for a file system, quota tree, or a Data Mover. When a Data Mover is
specified, the action applies to all mounted file systems on the Data
Mover.
nas_quotas also starts and stops quota database checks either online
or offline for quota trees and file systems, and allows you to upgrade
the quota database limits to the maximum limit value for a file
system. When a Data Mover is specified, the action applies to all
mounted file systems on the Data Mover.
nas_quotas
311
The nas Commands
!
CAUTION
Quotas should be turned on (enabled) before file systems go into a
production environment. Enabling (or disabling, or clearing)
quotas in a production environment is time consuming and the
process may disrupt file system operation. CIFS clients are
disconnected during these events and NFS clients receive a
message that the server is not responding. However, once enabled,
quotas can be changed at any time without impact.
OPTIONS
-edit [-user|-group] {-mover <movername>|-fs <fs_name>
[-path <pathname>]} [<id> [<id>...]
Sets the quota limits for users or groups on a specified Data Mover,
mounted file system, or directory tree.
For a user, the ID can be a user ID or UID, however, if NIS or the local
password file on the Data Mover is available, a user name can also be
used.
For a group, the ID can be a group ID or GID, however, if NIS or the
local password file is available, a group name can also be used.
Upon execution, a vi session (unless the EDITOR environment
variable specifies otherwise) is opened to edit the quota configuration
file. Changes to the file are applied when the vi session is saved and
exited.
[-proto <proto_id>]|[-block <hard_limit>
[:<soft_limit>]]
Applies the quota configuration defined for the prototype user
for each specified ID, and sets a hard and soft limit for storage
(block) usage in kilobytes.
[-inode <hard_limit>[:<soft_limit>]][<id>
[<id>...][-block <hard_limit>[:<soft_limit>]]
Edits the inode (file count) limits and the block (storage in KBs)
limits directly into the quota configuration file without opening
an editing session.
-edit -config {-mover <movername>|-fs <fs_name>} [-path
<pathname>]}
Edits the default quota configuration for all users/groups currently
without quotas or subsequently added to the specified Data Mover or
file system or quota tree. Also edits the grace periods for soft quotas,
and the conditions upon which to generate a quotas-event message to
the system log.
312
Command Line Interface Reference for File 8.1
The nas Commands
[-option <options>]
Specifies the following comma-separated options:
BGP=<integer>
Sets the block grace period in seconds.
IGP=<integer>
Sets the inode grace period in seconds.
DUBSL=<integer>
Sets the default user block soft limit in KB.
DUBHL=<integer>
Sets the default user block hard limit in KB.
DUISL=<integer>
Sets the default user inode soft limit.
DUIHL=<integer>
Sets the default user inode hard limit.
DGBSL=<integer>
Sets the default group block soft limit in KB.
DGBHL=<integer>
Sets the default group block hard limit in KB.
DGISL=<integer>
Sets the default group inode soft limit.
DGIHL=<integer>
Sets the default group inode hard limit.
HLE={True|False}
Specifies whether the hard limit is enforced.
ESFCS={True|False}
Specifies the event for check start has been sent.
ESFCE={True|False}
Specifies the event for check end has been sent.
ESFBSL={True|False}
Specifies that the event for block soft limits has been sent.
ESFBHL={True|False}
Specifies that the event for block hard limits has been sent.
nas_quotas
313
The nas Commands
-edit -tree -fs <fs_name> [[-proto <proto_id>]| [-block
<hard_limit> [:<soft_limit>]][-inode
<hard_limit>[:<soft_limit>]]][-comment <comment>] <id>
[<id>...]
Edits the quota limits for trees (inodes or blocks used by a tree
directory) where the <id> is the tree ID. This option can only be
applied on each file system basis. The -list option to display the tree
IDs.
The -proto option applies the quota configuration of the prototype
tree for each specified tree ID, or sets a hard and soft limit for blocks.
The <proto_id> must be a tree ID.
The -inode and -block options edit the inode/block limits for the tree
directly in the quota configuration file without opening an editing
session.
The -comment option associates a comment with the quota tree. The
comment is delimited by single quotes. Comment length is limited to
256 bytes (represented as 256 ASCII characters or a variable number
of Unicode multibyte characters) and cannot include single quotes (’
’), double quotes (" "), semicolons (;), NL (New Line), or FF (Form
Feed).
-report [-user|-group] {-mover <movername>|-fs <fs_name>}
[-path <pathname>]} [<id> <id> ...]
Displays a summary of disk usage and quotas for the user or group,
including the number of files and space in kilobytes for the specified
<fs_name>, or all file systems mounted on the specified
<movername>, or for the specified quota tree. The -edit option
provides more information for the usage of UIDs and GIDs.
Note: The nas_quotas can show report for maximum 1024 IDs at a time.
-report -config {-mover <movername>|-fs <fs_name>} [-path
<pathname>]}
Displays quota configuration information as viewed from the
specified Data Mover, file system, or quota-tree level, including:
◆
◆
◆
◆
◆
◆
314
Active quota policy
Quota status (user/group quotas enabled or disabled)
Grace period
Default limits currently set for users/groups
Hard-quota enforcement option setting (deny disk space enabled
or disabled)
Quota conditions that trigger event-logging
Command Line Interface Reference for File 8.1
The nas Commands
-report -tree -fs <fs_name>[<id> <id>...]
Displays the quota limits for a specified quota tree in a file system.
The <id> is a tree ID.
Note: The <id> is either a user ID, a group ID, or a tree ID. If the quota type is
not specified, the default is set to the ’-user’ ID.
{-on|-off|-clear} [-user|-group|-both] {-mover
<movername>|-fs <fs_name>|[-path <pathname>]|-all}
Turns quotas on, off, and clears quotas for the user, group, or both
(users and groups at once) on the <movername>, <fs_name>,
<pathname>, for all users, or groups on all file systems on all Data
Movers in the cabinet.
The -clear option permanently removes all quota records, deletes the
quota configuration file, and turns quotas off.
!
CAUTION
While quotas are being turned on, off, or cleared, other operations
to a file system may be disrupted. CIFS clients are disconnected
during this execution.
-on -tree -fs <fs_name> -path <pathname>
Turns on (enables) tree quotas so that quota tracking and hard-limit
enforcement (if enabled) can occur. When enabling tree quotas, the
directory must not exist; it is created in this tree-quota-enabling
process.
Note: The quota path length (which VNX for file calculates as including the
file system mountpoint) must be less than 1024 bytes. If Unicode is enabled
on the selected Data Mover, -path accepts any characters defined by the
Unicode 3.0 standard. Otherwise, it accepts only ASCII characters.
[-comment <comment>]
The -comment option associates a comment with the quota tree.
The comment is delimited by single quotes. Comment length is
limited to 256 bytes (represented as 256 ASCII characters or a
variable number of Unicode multibyte characters) and cannot
include single quotes (’ ’), double quotes (" "), semicolons (;), NL
(New Line), or FF (Form Feed).
nas_quotas
315
The nas Commands
-off -tree -fs <fs_name> -path <pathname>
Turns tree quotas off. When turning tree quotas off, the tree directory
must be empty.
{-list|-clear} -tree -fs <fs_name>
The -list option displays all active quota trees and their respective
tree IDs used by -edit and -report with the specified file system.
Use the -tree -clear option to clear all the information from the
database after you disable (turn off) quotas for all trees within a file
system. Once cleared, the database information is not recoverable.
!
CAUTION
The -clear option deletes the usage and the limit information for
tree quotas. The limits cannot be recovered.
-check -start [-mode online|offline] [-tree] -fs
<fs_name> [-path <pathname>]
Starts a check of a quota database in online or offline mode for a tree
quota or a file system quota. The default mode is online if the -mode
option is not specified, and a quota check is run while the file system
remains online.
-check {-stop|-status} -fs <fs_name> [-path <pathname>]
Stops or provides status of a file system quota database check that is
in progress.
-quotadb {-info|-upgrade [-Force]} {-mover
<movername>|-fs <fs_name>}
Either displays status related to the quota database upgrade or starts
an upgrade of the quota database for a specific file system or all file
systems on a Data Mover.
The -info option displays the status related to the quota database
limits upgrade.
Use the -upgrade option to perform an upgrade of the quota
database. If the -Force option is not specified, you are in interactive
mode while upgrading the quota database. If the -Force option is
specified, you are in non-interactive mode while upgrading the quota
database.
Use -mover <movername> to upgrade all mounted file systems’
quota databases on a Data Mover.
Use -fs <fs_name> to upgrade a specific file system’s quota database.
316
Command Line Interface Reference for File 8.1
The nas Commands
Note: Before the upgrade process runs, the Control Station displays the
estimated upgrade time on the file system whose quota database will be
upgraded, and also displays a warning message to notify users that the file
system will be unavailable during the upgrade process. If users are in
interactive mode, a dialog box displays letting users choose whether they
want to continue. If users are in non-interactive mode, after displaying the
estimated upgrade time message and warning message, the upgrade process
starts immediately.
SEE ALSO
EXAMPLE #1
Using Quotas on VNX.
To enable quotas on for users and groups of a file system, type:
$ nas_quotas -on -both -fs ufs1
done
EXAMPLE #2
To open a vi session to edit file system quotas on ufs1 for the
specified user, 1000, type:
$ nas_quotas -edit -user -fs ufs1 1000
Userid : 1000
fs ufs1 blocks (soft = 2000, hard = 3000) inodes (soft = 0, hard = 0)
~
~
~
~
"/tmp/EdP.agGQuIz" 2L, 84C written
done
EXAMPLE #3
To change the block limit and inode limit for a file system without
opening up a vi session, type:
$ nas_quotas -edit -user -fs ufs1 -block 7000:6000 -inode
700:600 2000
done
EXAMPLE #4
To view a report of user quotas for ufs1, type:
$ nas_quotas -report -user -fs ufs1
Report for user quotas on filesystem ufs1 mounted on /ufs1
+-----------+--------------------------------+------------------------------+
|User
|
Bytes Used (1K)
|
Files
|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
|
| Used | Soft | Hard |Timeleft| Used | Soft | Hard |Timeleft|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
|#1000
|
1328|
2000|
3000|
|
54|
0|
0|
|
|#2000
|
6992|
6000|
7000| 7.0days|
66|
600|
700|
|
|#5000
| 141592|
0|
0|
|
516|
0|
0|
|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
done
nas_quotas
317
The nas Commands
EXAMPLE #5
To select user 300 as prototype user for ufs1, and assign other users
the same limits, type:
$ nas_quotas -group -edit -fs ufs1 -proto 300 301 302 303
done
EXAMPLE #6
To display the group quotas information for ufs1, type:
$ nas_quotas -report -group -fs ufs1
Report for group quotas on filesystem ufs1 mounted on /ufs1
+-----------+--------------------------------+------------------------------+
| Group
|
Bytes Used (1K)
|
Files
|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
|
| Used | Soft | Hard |Timeleft| Used | Soft | Hard |Timeleft|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
|#1
|
296|
0|
0|
|
12|
0|
0|
|
|#300
|
6992|
6000|
7000| 7.0days|
67|
600|
700|
|
|#301
|
0|
6000|
7000|
|
0|
600|
700|
|
|#302
|
0|
6000|
7000|
|
0|
600|
700|
|
|#303
|
0|
6000|
7000|
|
0|
600|
700|
|
|#32772
| 22296|
0|
0|
|
228|
0|
0|
|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
done
EXAMPLE #7
To edit the default quota configuration for server_2, type:
$ nas_quotas -edit -config -mover server_2
File System Quota Parameters:
fs "ufs1"
Block Grace: (1.0 weeks)
Inode Grace: (1.0 weeks)
* Default Quota Limits:
User: block (soft = 5000, hard = 8000) inodes (soft = 100, hard= 200)
Group: block (soft = 6000, hard = 9000) inodes (soft = 200, hard= 400)
Deny disk space to users exceeding quotas: (yes)
* Generate Events when:
Quota check starts:
(no)
Quota check ends:
(no)
soft quota crossed:
(no)
hard quota crossed:
(no)
fs "ufs2"
Block Grace: (1.0 weeks)
Inode Grace: (1.0 weeks)
* Default Quota Limits:
User: block (soft = 0, hard = 0) inodes (soft = 0, hard= 0)
Group: block (soft = 0, hard = 0) inodes (soft = 0, hard= 0)
Deny disk space to users exceeding quotas: (yes)
* Generate Events when:
Quota check starts:
(no)
Quota check ends:
(no)
soft quota crossed:
(no)
hard quota crossed:
(no)
318
Command Line Interface Reference for File 8.1
The nas Commands
~
~
"/tmp/EdP.ahCPdAB" 25L, 948C written
done
EXAMPLE #8
To open a vi session and edit the quotas for a file system, type:
$ nas_quotas -edit -config -fs ufs1
File System Quota Parameters:
fs "ufs1"
Block Grace: (1.0 weeks)
Inode Grace: (1.0 weeks)
* Default Quota Limits:
User: block (soft = 5000, hard = 8000) inodes (soft = 100, hard= 200)
Group: block (soft = 6000, hard = 9000) inodes (soft = 200, hard= 400)
Deny disk space to users exceeding quotas: (yes)
* Generate Events when:
Quota check starts:
(no)
Quota check ends:
(no)
soft quota crossed:
(yes)
hard quota crossed:
(yes)
~
~
"/tmp/EdP.a4slhyg" 13L, 499C written
done
EXAMPLE #9
To view the quota configuration for the file system, ufs1, type:
$ nas_quotas -report -config -fs ufs1
+------------------------------------------------------------------------------+
| Quota parameters for filesystem ufs1 mounted on /ufs1:
+------------------------------------------------------------------------------+
| Quota Policy: blocks
| User Quota: ON
| Group Quota: ON
| Block grace period: (1.0 weeks)
| Inode grace period: (1.0 weeks)
| Default USER quota limits:
|
Block Soft: (
5000), Block Hard: (
8000)
|
Inode Soft: (
100), Inode Hard: (
200)
| Default GROUP quota limits:
|
Block Soft: (
6000), Block Hard: (
9000)
|
Inode Soft: (
200), Inode Hard: (
400)
| Deny Disk Space to users exceeding quotas:
YES
| Log an event when ...
|
Block hard limit reached/exceeded:
YES
|
Block soft limit (warning level) crossed:
YES
|
Quota check starts:
NO
|
Quota Check ends:
NO
+------------------------------------------------------------------------------+
done
nas_quotas
319
The nas Commands
EXAMPLE #10
To enable tree quotas on for ufs1, type:
$ nas_quotas -on -tree -fs ufs1 -path /tree1 -comment
'Tree #1'
done
EXAMPLE #11
To create a tree quota in a language that uses multibyte characters
(such as Japanese), type:
$ nas_quotas -on -tree -fs fs_22 -path
/<path_in_local_language_text> -comment
<comment_in_local_language_text>
done
EXAMPLE #12
To list the tree quotas for ufs1, type:
$ nas_quotas -list -tree -fs ufs1
+----------------------------------------------------------------------------+
| Quota trees for filesystem ufs1 mounted on /ufs1:
+------+---------------------------------------------------------------------+
|TreeId| Quota tree path (Comment)
|
+------+---------------------------------------------------------------------+
|
1 | /tree1 (Tree #1)
|
|
2 | /tree2 (Tree #2)
|
|
3 | /<tree_path_in_local_language_text> (Tree #3)
|
+------+---------------------------------------------------------------------+
done
EXAMPLE #13
To edit or add a comment for a tree quota for ufs1, type:
$ nas_quotas -edit -tree -fs ufs1 -comment 'Quota for
Tree1' 1
done
EXAMPLE #14
To edit tree quotas for ufs1, type:
$ nas_quotas -edit -tree -fs ufs1 1
treeid : 1
fs ufs1 blocks (soft = 6000, hard = 8000) inodes (soft = 200, hard = 300)
~
~
"/tmp/EdP.aiHKgh5" 2L, 85C written
done
EXAMPLE #15
To edit tree quotas for ufs1 and change the block and inodes, type:
$ nas_quotas -edit -tree -fs ufs1 -block 8000:6000 -inode
900:800 1
done
EXAMPLE #16
320
To edit tree quotas for ufs1 and apply the quota configuration of the
prototype tree, type:
Command Line Interface Reference for File 8.1
The nas Commands
$ nas_quotas -edit -tree -fs ufs1 -proto 1 2
done
EXAMPLE #17
To display any currently active trees on a file system, type:
$ nas_quotas -report -tree -fs ufs1
Report for tree quotas on filesystem ufs1 mounted on /ufs1
+-----------+--------------------------------+------------------------------+
| Tree
|
Bytes Used (1K)
|
Files
|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
|
| Used | Soft | Hard |Timeleft| Used | Soft | Hard |Timeleft|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
|#1
|
384|
6000|
8000|
|
3|
800|
900|
|
|#2
|
7856|
6000|
8000| 7.0days|
60|
800|
900|
|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
done
EXAMPLE #18
To enable tree quotas off, type:
$ nas_quotas -tree -off -fs ufs1 -path /tree1
done
EXAMPLE #19
To enable quotas on for users and groups on tree quota, /tree3, of a
file system, ufs1, type:
$ nas_quotas -on -both -fs ufs1 -path /tree3
done
EXAMPLE #20
To open a vi session to edit file system quotas on quota tree, /tree3, on
ufs1 for the specified user, 1000, type:
$ nas_quotas -edit -user -fs ufs1 -path /tree3 1000
Userid : 1000
fs ufs1 tree "/tree3" blocks (soft = 4000, hard = 6000) inodes (soft = 30, hard =
50)
~
~
~
~
"/tmp/EdP.aMdtIQR" 2L, 100C written
done
EXAMPLE #21
To change the block limit and inode limit on quota tree, /tree3, on
ufs1 for the specified user, 1000, without opening up a vi session,
type:
$ nas_quotas -edit -user -fs ufs1 -path /tree3 -block
6000:4000 -inode 300:200 1000
done
nas_quotas
321
The nas Commands
EXAMPLE #22
To view a report of user quotas on tree quota, /tree3, for ufs1, type:
$ nas_quotas -report -user -fs ufs1 -path /tree3
Report for user quotas on quota tree /tree3 on filesystem ufs1 mounted on /ufs1
+-----------+--------------------------------+------------------------------+
|User
|
Bytes Used (1K)
|
Files
|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
|
| Used | Soft | Hard |Timeleft| Used | Soft | Hard |Timeleft|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
|#1000
|
2992|
4000|
6000|
|
34|
200|
300|
|
|#32768
|
9824|
0|
0|
|
28|
0|
0|
|
+-----------+-------+-------+-------+--------+-------+------+------+--------+
done
EXAMPLE #23
To open a vi session and edit the quota configuration for tree quota,
/tree3, on a file system, ufs1, type:
$ nas_quotas -edit -config -fs ufs1 -path /tree3
Tree Quota Parameters:
fs "ufs1"
tree "/tree3"
Block Grace: (1.0 weeks)
Inode Grace: (1.0 weeks)
* Default Quota Limits:
User: block (soft = 8000, hard = 9000) inodes (soft = 200, hard= 300)
Group: block (soft = 8000, hard = 9000) inodes (soft = 300, hard= 400)
Deny disk space to users exceeding quotas: (yes)
* Generate Events when:
Quota check starts:
(no)
Quota check ends:
(no)
soft quota crossed:
(yes)
hard quota crossed:
(yes)
~
~
~
~
"/tmp/EdP.aDTOKeU" 14L, 508C written
done
EXAMPLE #24
To view the quota configuration for tree quota, /tree3, on file system,
ufs1, type:
$ nas_quotas -report -config -fs ufs1 -path /tree3
+------------------------------------------------------------------------------+
| Quota parameters for tree quota /tree3 on filesystem ufs1 mounted on /ufs1:
+------------------------------------------------------------------------------+
| Quota Policy: blocks
| User Quota: ON
| Group Quota: ON
| Block grace period: (1.0 weeks)
| Inode grace period: (1.0 weeks)
| Default USER quota limits:
322
Command Line Interface Reference for File 8.1
The nas Commands
|
Block Soft: (
8000), Block Hard: (
9000)
|
Inode Soft: (
200), Inode Hard: (
300)
| Default GROUP quota limits:
|
Block Soft: (
8000), Block Hard: (
9000)
|
Inode Soft: (
300), Inode Hard: (
400)
| Deny Disk Space to users exceeding quotas:
YES
| Log an event when ...
|
Block hard limit reached/exceeded:
YES
|
Block soft limit (warning level) crossed:
YES
|
Quota check starts:
NO
|
Quota Check ends:
NO
+------------------------------------------------------------------------------+
done
EXAMPLE #25
To disable user quota and group quota on tree quota, /tree3, type:
$ nas_quotas -off -both -fs ufs1 -path /tree3
done
EXAMPLE #26
To disable group quotas for ufs1, type:
$ nas_quotas -off -group -fs ufs1
done
EXAMPLE #27
To clear all tree quotas for ufs1, type:
$ nas_quotas -clear -tree -fs ufs1
done
EXAMPLE #28
To clear quotas for users and groups of a Data Mover, type:
$ nas_quotas -clear -both -mover server_2
done
EXAMPLE #29
To start a tree quota check in quota tree /mktg-a/dir1 in file system
ufs1 with the file system online, type:
$ nas_quotas -check -start -mode online -tree -fs ufs1
/mktg-a/dir1
done
EXAMPLE #30
To stop a tree quota check in file system ufs1, type:
$ nas_quotas -check -stop -fs ufs1
done
EXAMPLE #31
To view the status of a tree quota check in quota tree /mktg-a/dir1 in
file system ufs1, type:
$ nas_quotas -check -status -tree -fs ufs1 -path
/mktg-a/dir1
nas_quotas
323
The nas Commands
Tree quota check on filesystem ufs1 and path /mktg-a/dir is running and is 60%
complete.
Done
EXAMPLE #32
To list quota database limits for all file systems on a Data Mover, type:
$ nas_quotas -quotadb -info -mover server_2
Info 13421850365 : The quota limit on ufs0 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Info 13421850365 : The quota limit on ufs1 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Info 13421850365 : The quota limit on ufs2 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Info 13421850365 : The quota limit on ufs3 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Info 13421850366 : The quota limit on ufs4 is at 256 TB
EXAMPLE #33
To list quota database limits for file system ufs4, type:
$ nas_quotas -quotadb -info -fs ufs4
Info 13421850366 : The quota limit on ufs4 is at 256 TB
EXAMPLE #34
To upgrade all file systems on a Data Mover, in interactive mode,
type:
$ nas_quotas -quotadb -upgrade -mover server_2
Info 13421850365 : The quota limit on ufs0 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Info 13421850365 : The quota limit on ufs1 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
324
Command Line Interface Reference for File 8.1
The nas Commands
Info 13421850365 : The quota limit on ufs2 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Info 13421850365 : The quota limit on ufs3 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Info 13421850366 : The quota limit on ufs4 is at 256 TB
Warning 17716861297: The file systems specified in the list above will not be
accessible during the quota database upgrade, and a file system's CIFS share and
NFS export also will not be accessible during the upgrade. The file systems shown
above are listed in the order that the quota database conversion is performed, one
by one sequentially. The estimated time ( shown above ) needed to upgrade the quota
database may change based on the file system's quota configuration and I/O
performance when the conversion is running.
Do you really want to upgrade the file system quota database now[Y/N]: Y
Info 13421850367 : quota db upgraded on ufs0
Info 13421850367 : quota db upgraded on ufs1
Info 13421850367 : quota db upgraded on ufs2
Error 13421850368 : Timeout occurred when upgrading quota db on ufs3. The Quota
db upgrade may still be in progress. Use the "-info" option to check status.
Info 13421850369 : quota db already upgraded on ufs4
EXAMPLE #35
To list quota database limits for file system ufs3 after an upgrade has
timed out, type:
$ nas_quotas -quotadb -info -fs ufs3
Info 13421850370 : The quota limit on ufs3 is at 4TB. Upgrade is 48% complete.
EXAMPLE #36
To upgrade all file systems on a Data Mover, in non-interactive mode,
type:
$ nas_quotas -quotadb -upgrade -Force -mover server_2
Info 13421850365 : The quota limit on ufs0 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Info 13421850365 : The quota limit on ufs1 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Info 13421850365 : The quota limit on ufs2 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
nas_quotas
325
The nas Commands
Info 13421850365 : The quota limit on ufs3 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Info 13421850366 : The quota limit on ufs4 is at 256 TB
Warning 17716861297: The file systems specified in the list above will not be
accessible during the quota database upgrade, and a file system's CIFS share and
NFS export also will not be accessible during the upgrade. The file systems shown
above are listed in the order that the quota database conversion is performed, one
by one sequentially. The estimated time ( shown above ) needed to upgrade the quota
database may change based on the file system's quota configuration and I/O
performance when the conversion is running.
Info 13421850367 : quota db upgraded on ufs0
Info 13421850367 : quota db upgraded on ufs1
Info 13421850367 : quota db upgraded on ufs2
Error 13421850368 : Timeout occurred when upgrading quota db on ufs3. The Quota
db upgrade may still be in progress. Use the "-info" option to check status.
Info 13421850369 : quota db already upgraded on ufs4
EXAMPLE #37
To upgrade file system ufs3, in interactive mode, type:
$ nas_quotas -quotadb -upgrade -fs ufs3
Info 13421850365 : The quota limit on ufs3 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
Warning 17716861297: The file systems specified in the list above will not be
accessible during the quota database upgrade, and a file system's CIFS share and
NFS export also will not be accessible during the upgrade. The file systems shown
above are listed in the order that the quota database conversion is performed, one
by one sequentially. The estimated time ( shown above ) needed to upgrade the quota
database may change based on the file system's quota configuration and I/O
performance when the conversion is running.
Do you really want to upgrade the file system quota database now[Y/N]: Y
Info 13421850367 : quota db upgraded on ufs3
done
EXAMPLE #38
To upgrade file system ufs3, in non-interactive mode, type:
$ nas_quotas -quotadb -upgrade -Force -fs ufs3
Info 13421850365 : The quota limit on ufs3 is at 4TB. The upgrade to 256 TB is
estimated to take 5 seconds.
A total number of 1500 data blocks in the quota database will be converted at a
speed of 300 blocks per second.
326
Command Line Interface Reference for File 8.1
The nas Commands
Warning 17716861297: The file systems specified in the list above will not be
accessible during the quota database upgrade, and a file system's CIFS share and
NFS export also will not be accessible during the upgrade. The file systems shown
above are listed in the order that the quota database conversion is performed, one
by one sequentially. The estimated time ( shown above ) needed to upgrade the quota
database may change based on the file system's quota configuration and I/O
performance when the conversion is running.
Info 13421850367 : quota db upgraded on ufs3
done
nas_quotas
327
The nas Commands
nas_rdf
Facilitates communication between two VNX systems. Its primary
use is to manage VNX for file systems and define the relationships
needed for disaster recovery in a SRDF environment.
SYNPOSIS
nas_rdf
-init
| -activate [-reverse|-skip_rdf_operations][-skip_SiteA_shutdown][-nocheck]
| -restore [-skip_rdf_operations [-skip_SiteA_shutdown]][-nocheck]
-check {-all|<test>,...}
DESCRIPTION
nas_rdf establishes and manages relationships for Control Stations
and Data Movers that physically reside in separate VNX for file
cabinets.
For SRDF, nas_rdf initializes the VNX, activates a failover to a
destination VNX for file, or restores a source VNX. For Dynamic
SRDF, nas_rdf activates a failover and reverses the system from a
destination volume (R2) to a source volume (R1). Configuration
details depend on the type of SRDF: active/passive or active/active’
SRDF/S for synchronous replication with disaster recovery, or
active/passive SRDF/A for extended-distance, or asynchronous
replication with a point-in-time replica.
Note: The apostrophe in active/active’ indicates that both sites have a source
volume mirrored at the other site.
SRDF is supported only on a VNX attached to a Symmetrix system.
Also, this command must be run from a primary Control Station in
slot 0; it will report an error if run from a Control Station in slot 1.
Note: This command must be executed from the /nas/sbin directory, unless
otherwise directed. Log in with your administrative username and password,
and execute this functionality from root.
OPTIONS
-init
Initializes a source or destination (target) VNX for SRDF/S or
SRDF/A.
-activate [-reverse]
Initiates an SRDF failover from the source VNX for file to the
destination. The -activate option is executed on the destination VNX
328
Command Line Interface Reference for File 8.1
The nas Commands
at the discretion of the user. The -activate option sets each
SRDF-protected volume on the source VNX as read-only, and each
mirrored volume on the destination VNX is set as read-write. The
SRDF standby Data Movers acquire the IP and MAC addresses, file
systems, and export tables of their source counterparts. The -reverse
option reverses SRDF direction by converting R2 volumes at
destination site to R1 and synchronizing the destination and source
sites. The -reverse option adds SYMCLI swap and establishes
operations on the system after the normal activate operation is
performed. When the -activate option is executed, an automatic,
internal SRDF health check is performed before activating a failover.
The -nocheck option allows you to skip this health check.
-activate -skip_rdf_operations
Skips RDF backend operations like symrdf failover. The backend
operations must be done using Solution Enablers or Mainframe host
component prior to this command. SiteA shutdown (Data Mover
shutdown and reboot Control Station) will be skipped all the time
when this option is specified. However Control Station reboot is sent
to SiteA at the end of the activate operation when the backend RDF
status is not “Split” to clean up old processes. (The “Split” status
means SiteA is read write, and the production site is up and running).
For failover from SiteB to SiteC or SiteC to SiteB, the Control Station
reboot is sent to SiteB or SiteC. SiteB/SiteC must be read write before
starting this operation. The -activate -skip_rdf_operations
-skip_SiteA_shutdown will do the same operation.
-activate -skip_SiteA_shutdown
Skips SiteA shutdown (Data Mover shutdown and reboot Control
Station) operation. However the SiteA shutdown is sent to SiteA at
the end of the activate operation. This option is mainly used to
minimize the failover time.
-restore –skip_rdf_operations
Skips RDF backend operations like symrdf failback. This option also
completes only SiteB/SiteC restore operations and skip SiteA restore
operation. The SiteA restore operation must be done separately at
SiteA after the SiteB/SiteC restore operation completes. SiteB/SiteC
must be read write before starting this operation.
-restore –skip_rdf_operations –skip_SiteA_shutdown
Skips RDF backend operations like symrdf failback and also skip
SiteA shutdown operation. This is mainly used to failover from SiteB
to SiteC or from SiteC to SiteB.
nas_rdf
329
The nas Commands
-restore
Restores a source VNX after a failover. The -restore option is initially
executed on the destination VNX. The data on each destination
volume is copied to the corresponding volume on the source VNX.
On the destination VNX, services on each SRDF standby Data Mover
are stopped. (NFS clients connected to these Data Movers see a
“server unavailable” message; CIFS client connections time out.)
Each volume on the source VNX is set as read-write, and each
mirrored volume on the destination VNX is set as read-only.
Finally, nas_rdf -restore can be remotely executed on the source VNX
to restore the original configuration. Each primary Data Mover
reacquires its IP and MAC addresses, file systems, and export tables.
When the -restore option is executed, an automatic, internal SRDF
health check is performed before restoring source and destination
VNX systems. The -nocheck option allows you to skip this health
check.
-check { -all|<test>,…}
Runs SRDF health checks on the VNX. The -check option can be
executed either by using the -all option or by specifying one or more
of the following individual checks: SRDF standby Data Mover
configuration check (r1_dm_config, r2_dm_config), SRDF session
state check (r1_session, r2_session), Device group configuration
check (r1_dev_group, r2_dev_group), Data Mover mirrored device
accessibility check (r1_dev_avail, r2_dev_avail), Symmetrix device
state check (dev_not_normal), and SRDF restored state check
(restored). In these checks, r1 represents the source side and r2
represents the destination side.
When the -all option is used, all the checks are performed
automatically. If the -check option detects invalid configurations or
state issues, it prints relevant warning messages with recommended
actions so that the issues can be resolved before running the activate
or restore options. You can use the -check option to perform health
checks at any time.
Note: To run the -check option, you must log in to the VNX either as
nasadmin and then switch (su) to root, or as rdfadmin and then switch (su)
to root.
SEE ALSO
EXAMPLE #1
330
Using SRDF/S with VNX for Disaster Recovery, Using SRDF/S with
VNX, and nas_cel.
To start the initialization process on a destination VNX in an
Command Line Interface Reference for File 8.1
The nas Commands
active/passive SRDF/S configuration, as a nasadmin su to root user,
type:
# /nas/sbin/nas_rdf -init
Discover local storage devices ...
Discovering storage on eng564168 (may take several minutes)
done
Start R2 dos client ...
done
Start R2 nas client ...
done
Contact CS_A ... is alive
Create a new login account to manage the RDF site CELERRA
Caution: For an active-active configuration, avoid using the same UID
that was used for the rdfadmin account on the other side.
New login username and UID (example: rdfadmin:500): rdfadmin:600
done
New UNIX password:
BAD PASSWORD: it is based on a dictionary word
Retype new UNIX password:
Changing password for user rdfadmin.
passwd: all authentication tokens updated successfully.
done
operation in progress (not interruptible)...
id = 1
name = CS_A
owner = 600
device = /dev/ndj1
channel = rdev=/dev/ndg, off_MB=391; wdev=/dev/nda, off_MB=391
net_path = 10.245.64.169
celerra_id = 0001949004310028
passphrase = nasadmin
Discover remote storage devices ...done
The following servers have been detected on the system (CS_B):
id type acl slot groupID state name
1 4 2000 2 0 server_2
2 1 0 3 0 server_3
Please enter the id(s) of the server(s) you wish to reserve
(separated by spaces) or "none" for no servers.
Select server(s) to use as standby: 1
operation in progress (not interruptible)...
id = 1
name = CS_A
owner = 600
device = /dev/ndj1
channel = rdev=/dev/ndg, off_MB=391; wdev=/dev/nda, off_MB=391
net_path = 10.245.64.169
celerra_id = 0001949004310028
passphrase = nasadmin
nas_rdf
331
The nas Commands
EXAMPLE #2
To initiate an SRDF failover from the source VNX to the destination,
as a rdfadmin su to root, type:
# /nas/sbin/nas_rdf -activate
Is remote site CELERRA completely shut down (power OFF)?
Do you wish to continue? [yes or no]: yes
Successfully pinged (Remotely) Symmetrix ID: 000187430809
Successfully pinged (Remotely) Symmetrix ID: 000190100559
Successfully pinged (Remotely) Symmetrix ID: 000190100582
Write Disable devices on SA at source (R1)..............Done.
Suspend RDF links.......................................Done.
Read/Write Enable devices on RA at target (R2)..........Done.
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
fsck 1.35 (28-Feb-2004)
/dev/ndj1: recovering journal
/dev/ndj1: clean, 13780/231360 files, 233674/461860 blocks
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
id
type acl slot groupID state name
1
1
1000 2
0
server_2
2
4
1000 3
0
server_3
3
1
1000 4
0
server_4
4
4
1000 5
0
server_5
server_2 :
server_2 : going offline
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
server_3 :
server_3 : going offline
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
server_4 :
Error 4003: server_4 : standby is not configured
server_5 :
Error 4003: server_5 : standby is not configured
Suspend RDF links.......................................Done.
Merge device track tables between source and target.......Started.
Device: 045A in (0557,005)............................... Merged.
Merge device track tables between source and target.......Done.
Resume RDF links........................................Started.
Resume RDF links........................................Done.
332
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #3
To initiate an SRDF failover from the source VNX to the destination,
without the SRDF health check, as rdfadmin su to root user, type:
# /nas/sbin/nas_rdf -activate -nocheck
Skipping SRDF health check ....
Is remote site CELERRA completely shut down (power OFF)?
Do you wish to continue? [yes or no]: yes
Successfully pinged (Remotely) Symmetrix ID: 000187430809
Successfully pinged (Remotely) Symmetrix ID: 000190100559
Successfully pinged (Remotely) Symmetrix ID: 000190100582
Write Disable devices on SA at source (R1)..............Done.
Suspend RDF links.......................................Done.
Read/Write Enable devices on RA at target (R2)..........Done.
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
fsck 1.35 (28-Feb-2004)
/dev/ndj1: recovering journal
/dev/ndj1: clean, 13780/231360 files, 233674/461860 blocks
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
id type acl slot groupID state name
1 1 1000 2 0 server_2
2 4 1000 3 0 server_3
3 1 1000 4 0 server_4
4 4 1000 5 0 server_5
server_2 :
server_2 : going offline
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
server_3 :
server_3 : going offline
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
server_4 :
Error 4003: server_4 : standby is not configured
server_5 :
Error 4003: server_5 : standby is not configured
Suspend RDF links.......................................Done.
Merge device track tables between source and target.......Started.
Device: 045A in (0557,005)............................... Merged.
Merge device track tables between source and target.......Done.
Resume RDF links........................................Started.
Resume RDF links........................................Done.
nas_rdf
333
The nas Commands
EXAMPLE #4
To initiate a Dynamic SRDF failover from the source VNX to the
destination, as rdfadmin su to root user, type:
# /nas/sbin/nas_rdf -activate -reverse
Is remote site CELERRA completely shut down (power OFF)?
Do you wish to continue? [yes or no]: yes
Successfully pinged (Remotely) Symmetrix ID: 000280600118
Write Disable devices on SA at source (R1)..............Done.
Suspend RDF links.......................................Done.
Read/Write Enable devices on RA at target (R2)..........Done.
fsck 1.35 (28-Feb-2004)
/dev/sdj1: recovering journal
Clearing orphaned inode 37188 (uid=0, gid=0, mode=0100644, size=0)
/dev/sdj1: clean, 12860/219968 files, 194793/439797 blocks
id
type acl slot groupID state name
1
1
1000 2
0
server_2
2
4
1000 3
0
server_3
3
4
2000 4
0
server_4
4
4
2000 5
0
server_5
server_2 :
server_2 : going offline
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
server_3 :
server_3 : going offline
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
An RDF 'Swap Personality' operation execution is
in progress for device group '1R2_500_1'. Please wait...
Swap RDF Personality......................................Started.
Swap RDF Personality......................................Done.
The RDF 'Swap Personality' operation successfully executed for
device group '1R2_500_1'.
An RDF 'Incremental Establish' operation execution is
in progress for device group '1R2_500_1'. Please wait...
Suspend RDF links.......................................Done.
Resume RDF links........................................Started.
Merge device track tables between source and target.......Started.
Devices: 0009-000B ...................................... Merged.
Devices: 0032-0034 ...................................... Merged.
334
Command Line Interface Reference for File 8.1
The nas Commands
Devices:
Devices:
Devices:
Devices:
Devices:
Devices:
Devices:
0035-0037
0038-003A
003B-003D
003E-0040
0041-0043
0044-0046
0047-0049
......................................
......................................
......................................
......................................
......................................
......................................
......................................
Merged.
Merged.
Merged.
Merged.
Merged.
Merged.
Merged.
.
Merge device track tables between source and target.......Done.
Resume RDF links........................................Done.
The RDF 'Incremental Establish' operation successfully initiated for
device group '1R2_500_1'.
nas_rdf
335
The nas Commands
EXAMPLE #5
To restore a source VNX after failover, as rdfadmin su to root user,
type:
# /nas/sbin/nas_rdf -restore
Is remote site CELERRA ready for Storage restoration?
Do you wish to continue? [yes or no]: yes
Contact Joker_R1_CS0 ... is alive
Restore will now reboot the source site control station.
Do you wish to continue? [yes or no]: yes
Device Group (DG) Name
DG's Type
DG's Symmetrix ID
336
: 1R2_500_5
: RDF2
: 000190100557
Target (R2) View
Source (R1) View
------------------------------------------------------ST
LI
ST
Standard
A
N
A
Logical
T R1 Inv
R2 Inv K
T R1 Inv
R2 Inv
Device Dev
E Tracks
Tracks S Dev
E Tracks
Tracks
-------------------------------- -- ------------------------
MODES
----- ------------
DEV001
DEV002
DEV003
DEV004
DEV005
DEV006
DEV007
DEV008
DEV009
DEV010
DEV011
DEV012
DEV013
DEV014
DEV015
DEV016
DEV017
DEV018
DEV019
DEV020
DEV021
DEV022
DEV023
DEV024
DEV025
DEV026
DEV027
DEV028
DEV029
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
045A
045B
045C
045D
045E
045F
0467
0468
0469
046A
046B
046C
046D
046E
046F
0470
0471
0472
0473
0474
0475
0476
0477
0478
0479
047A
047B
047C
047D
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
RW
10
2054
0
0
1284
0
0
2
0
0
2
0
0
0
2
0
2
0
0
0
0
0
2
2
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
RW
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
NR
045A
045B
045C
045D
045E
045F
0467
0468
0469
046A
046B
046C
046D
046E
046F
0470
0471
0472
0473
0474
0475
0476
0477
0478
0479
047A
047B
047C
047D
Command Line Interface Reference for File 8.1
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
WD
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
RDF Pair
MDA
STATE
----- -----------R1 Updated
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
Failed Over
The nas Commands
DEV030 047E RW
0
0 NR 047E WD
0
0 S..
Failed Over
DEV031 047F RW
0
0 NR 047F WD
0
0 S..
Failed Over
DEV032 0480 RW
0
0 NR 0480 WD
0
0 S..
Failed Over
DEV033 0481 RW
0
0 NR 0481 WD
0
0 S..
Failed Over
DEV034 0482 RW
0
0 NR 0482 WD
0
0 S..
Failed Over
DEV035 0483 RW
0
0 NR 0483 WD
0
0 S..
Failed Over
DEV036 0484 RW
0
0 NR 0484 WD
0
0 S..
Failed Over
DEV037 0485 RW
0
0 NR 0485 WD
0
0 S..
Failed Over
DEV038 0486 RW
0
0 NR 0486 WD
0
0 S..
Failed Over
DEV039 0487 RW
0
0 NR 0487 WD
0
0 S..
Failed Over
DEV040 0488 RW
0
0 NR 0488 WD
0
0 S..
Failed Over
DEV041 0489 RW
0
0 NR 0489 WD
0
0 S..
Failed Over
DEV042 048A RW
0
0 NR 048A WD
0
0 S..
Failed Over
DEV043 048B RW
0
0 NR 048B WD
0
0 S..
Failed Over
DEV044 048C RW
0
0 NR 048C WD
0
0 S..
Failed Over
DEV045 048D RW
0
0 NR 048D WD
0
0 S..
Failed Over
DEV046 048E RW
0
0 NR 048E WD
0
0 S..
Failed Over
DEV047 048F RW
2
0 NR 048F WD
0
0 S..
Failed Over
DEV048 0490 RW
0
0 NR 0490 WD
0
0 S..
Failed Over
DEV049 0491 RW
0
0 NR 0491 WD
0
0 S..
Failed Over
DEV050 0492 RW
0
0 NR 0492 WD
0
0 S..
Failed Over
DEV051 0493 RW
0
0 NR 0493 WD
0
0 S..
Failed Over
DEV052 0494 RW
0
0 NR 0494 WD
0
0 S..
Failed Over
DEV053 0495 RW
0
0 NR 0495 WD
0
0 S..
Failed Over
DEV054 0496 RW
0
0 NR 0496 WD
0
0 S..
Failed Over
DEV055 0497 RW
2
0 NR 0497 WD
0
0 S..
Failed Over
DEV056 0498 RW
2
0 NR 0498 WD
0
0 S..
Failed Over
DEV057 0499 RW
0
0 NR 0499 WD
0
0 S..
Failed Over
DEV058 049A RW
0
0 NR 049A WD
0
0 S..
Failed Over
DEV059 049B RW
0
0 NR 049B WD
0
0 S..
Failed Over
DEV060 049C RW
0
0 NR 049C WD
0
0 S..
Failed Over
DEV061 049D RW
0
0 NR 049D WD
0
0 S..
Failed Over
DEV062 049E RW
0
0 NR 049E WD
0
0 S..
Failed Over
DEV063 049F RW
0
0 NR 049F WD
0
0 S..
Failed Over
DEV064 04A0 RW
0
0 NR 04A0 WD
0
0 S..
Failed Over
DEV065 04A1 RW
0
0 NR 04A1 WD
0
0 S..
Failed Over
DEV066 04A2 RW
0
0 NR 04A2 WD
0
0 S..
Failed Over
DEV067 04A3 RW
0
0 NR 04A3 WD
0
0 S..
Failed Over
DEV068 04A4 RW
0
0 NR 04A4 WD
0
0 S..
Failed Over
DEV069 04A5 RW
0
0 NR 04A5 WD
0
0 S..
Failed Over
DEV070 04A6 RW
0
0 NR 04A6 WD
0
0 S..
Failed Over
Total
-------- --------------- -------Tracks
3366
0
0
0
MBs
105.2
0.0
0.0
0.0
Legend for MODES:
M(ode of Operation): A = Async, S = Sync, E = Semi-sync, C = Adaptive Copy
D(omino)
: X = Enabled, . = Disabled
A(daptive Copy)
: D = Disk Mode, W = WP Mode, . = ACp off
Suspend RDF link(s).......................................Done.
Merge device track tables between source and target.......Started.
Devices: 045A-045F, 0467-0477 in (0557,005).............. Merged.
Devices: 0478-0489 in (0557,005)......................... Merged.
nas_rdf
337
The nas Commands
Devices: 048A-049B in (0557,005)......................... Merged.
Devices: 049C-04A6 in (0557,005)......................... Merged.
Merge device track tables between source and target.......Done.
Resume RDF link(s)........................................Started.
Resume RDF link(s)........................................Done.
Is remote site CELERRA ready for Network restoration?
Do you wish to continue? [yes or no]: yes
server_2 : done
server_3 : done
server_4 :
Error 4003: server_4 : standby is not configured
server_5 :
Error 4003: server_5 : standby is not configured
fsck 1.35 (28-Feb-2004)
/dev/ndj1: clean, 13836/231360 files, 233729/461860 blocks
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
Waiting for 1R2_500_5 access ...done
Write Disable device(s) on RA at target (R2)..............Done.
Suspend RDF link(s).......................................Done.
Merge device track tables between source and target.......Started.
Devices: 045A-045F, 0467-0477 in (0557,005).............. Merged.
Devices: 0478-0489 in (0557,005)......................... Merged.
Devices: 048A-049B in (0557,005)......................... Merged.
Devices: 049C-04A6 in (0557,005)......................... Merged.
Merge device track tables between source and target.......Done.
Resume RDF link(s)........................................Started.
Resume RDF link(s)........................................Done.
Read/Write Enable device(s) on SA at source (R1)..........Done.
Waiting for 1R2_500_5 sync ...done
Starting restore on remote site CELERRA ...
Waiting for nbs clients to start ... done
Waiting for nbs clients to start ... done
Suspend RDF link(s).......................................Done.
server_2 :
server_2 : going standby
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
server_3 :
server_3 : going standby
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
server_4 :
Error 4003: server_4 : standby is not configured
338
Command Line Interface Reference for File 8.1
The nas Commands
server_5 :
Error 4003: server_5 : standby is not configured
Resume RDF link(s)........................................Started.
Resume RDF link(s)........................................Done.
If the RDF device groups were setup to operate in ASYNCHRONOUS ( SRDF/A ) mode,
now would be a good time to set it back to that mode.
Would you like to set device group 1R2_500_5 to ASYNC Mode ? [yes or no]: no
done
EXAMPLE #6
To restore a source VNX after failover, without the SRDF health
check, as rdfadmin su to root user, type:
# /nas/sbin/nas_rdf -restore -nocheck
Skipping SRDF health check ....
Is remote site CELERRA ready for Storage restoration?
Do you wish to continue? [yes or no]: yes
Contact Joker_R1_CS0 ... is alive
Restore will now reboot the source site control station.
Do you wish to continue? [yes or no]: yes
Device Group (DG) Name : 1R2_500_5
DG's Type : RDF2
DG's Symmetrix ID : 000190100557
Target (R2) View Source (R1) View MODES
-------------------------------- ------------------------ ----- -----------ST
LI
ST
Standard
A
N
A
Logical
T R1 Inv
R2 Inv K
T R1 Inv
R2 Inv
RDF Pair
Device Dev
E Tracks
Tracks S Dev
E Tracks
Tracks MDA
STATE
-------------------------------- -- ------------------------ ----- -----------DEV001 045A RW
10
0 RW 045A WD
0
0 S..
R1 Updated
DEV002 045B RW
2054
0 NR 045B WD
0
0 S..
Failed Over
DEV003 045C RW
0
0 NR 045C WD
0
0 S..
Failed Over
DEV004 045D RW
0
0 NR 045D WD
0
0 S..
Failed Over
DEV005 045E RW
1284
0 NR 045E WD
0
0 S..
Failed Over
DEV006 045F RW
0
0 NR 045F WD
0
0 S..
Failed Over
DEV007 0467 RW
0
0 NR 0467 WD
0
0 S..
Failed Over
DEV008 0468 RW
2
0 NR 0468 WD
0
0 S..
Failed Over
DEV009 0469 RW
0
0 NR 0469 WD
0
0 S..
Failed Over
DEV010 046A RW
0
0 NR 046A WD
0
0 S..
Failed Over
DEV011 046B RW
2
0 NR 046B WD
0
0 S..
Failed Over
DEV012 046C RW
0
0 NR 046C WD
0
0 S..
Failed Over
DEV013 046D RW
0
0 NR 046D WD
0
0 S..
Failed Over
DEV014 046E RW
0
0 NR 046E WD
0
0 S..
Failed Over
DEV015 046F RW
2
0 NR 046F WD
0
0 S..
Failed Over
DEV016 0470 RW
0
0 NR 0470 WD
0
0 S..
Failed Over
DEV017 0471 RW
2
0 NR 0471 WD
0
0 S..
Failed Over
DEV018 0472 RW
0
0 NR 0472 WD
0
0 S..
Failed Over
DEV019 0473 RW
0
0 NR 0473 WD
0
0 S..
Failed Over
DEV020 0474 RW
0
0 NR 0474 WD
0
0 S..
Failed Over
DEV021 0475 RW
0
0 NR 0475 WD
0
0 S..
Failed Over
DEV022 0476 RW
0
0 NR 0476 WD
0
0 S..
Failed Over
nas_rdf
339
The nas Commands
DEV022 0476 RW
0
0 NR 0476 WD
DEV023 0477 RW
2
0 NR 0477 WD
DEV024 0478 RW
2
0 NR 0478 WD
DEV025 0479 RW
0
0 NR 0479 WD
DEV026 047A RW
0
0 NR 047A WD
DEV027 047B RW
0
0 NR 047B WD
DEV028 047C RW
0
0 NR 047C WD
DEV029 047D RW
0
0 NR 047D WD
DEV030 047E RW
0
0 NR 047E WD
DEV031 047F RW
0
0 NR 047F WD
DEV032 0480 RW
0
0 NR 0480 WD
DEV033 0481 RW
0
0 NR 0481 WD
DEV034 0482 RW
0
0 NR 0482 WD
DEV035 0483 RW
0
0 NR 0483 WD
DEV036 0484 RW
0
0 NR 0484 WD
DEV037 0485 RW
0
0 NR 0485 WD
DEV038 0486 RW
0
0 NR 0486 WD
DEV039 0487 RW
0
0 NR 0487 WD
DEV040 0488 RW
0
0 NR 0488 WD
DEV041 0489 RW
0
0 NR 0489 WD
DEV042 048A RW
0
0 NR 048A WD
DEV043 048B RW
0
0 NR 048B WD
DEV044 048C RW
0
0 NR 048C WD
DEV045 048D RW
0
0 NR 048D WD
DEV046 048E RW
0
0 NR 048E WD
DEV047 048F RW
2
0 NR 048F WD
DEV048 0490 RW
0
0 NR 0490 WD
DEV049 0491 RW
0
0 NR 0491 WD
DEV050 0492 RW
0
0 NR 0492 WD
DEV051 0493 RW
0
0 NR 0493 WD
DEV052 0494 RW
0
0 NR 0494 WD
DEV053 0495 RW
0
0 NR 0495 WD
DEV054 0496 RW
0
0 NR 0496 WD
DEV055 0497 RW
2
0 NR 0497 WD
DEV056 0498 RW
2
0 NR 0498 WD
DEV057 0499 RW
0
0 NR 0499 WD
DEV058 049A RW
0
0 NR 049A WD
DEV059 049B RW
0
0 NR 049B WD
DEV060 049C RW
0
0 NR 049C WD
DEV061 049D RW
0
0 NR 049D WD
DEV062 049E RW
0
0 NR 049E WD
DEV063 049F RW
0
0 NR 049F WD
DEV064 04A0 RW
0
0 NR 04A0 WD
DEV065 04A1 RW
0
0 NR 04A1 WD
DEV066 04A2 RW
0
0 NR 04A2 WD
DEV067 04A3 RW
0
0 NR 04A3 WD
DEV068 04A4 RW
0
0 NR 04A4 WD
DEV069 04A5 RW
0
0 NR 04A5 WD
DEV070 04A6 RW
0
0 NR 04A6 WD
Total -------- -------- -------- -------Track(s) 3366 0 0 0he nas Commands
MB(s) 105.2 0.0 0.0 0.0
340
Command Line Interface Reference for File 8.1
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
0
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
S..
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Failed
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
Over
The nas Commands
Legend for MODES:
M(ode of Operation): A = Async, S = Sync, E = Semi-sync, C = Adaptive Copy
D(omino) : X = Enabled, . = Disabled
A(daptive Copy) : D = Disk Mode, W = WP Mode, . = ACp off
Suspend RDF link(s).......................................Done.
Merge device track tables between source and target.......Started.
Devices: 045A-045F, 0467-0477 in (0557,005).............. Merged.
Devices: 0478-0489 in (0557,005)......................... Merged.
Devices: 048A-049B in (0557,005)......................... Merged.
Devices: 049C-04A6 in (0557,005)......................... Merged.
Merge device track tables between source and target.......Done.
Resume RDF link(s)........................................Started.
Resume RDF link(s)........................................Done.
Is remote site CELERRA ready for Network restoration?
Do you wish to continue? [yes or no]: yes
server_2 : done
server_3 : done
server_4 :
Error 4003: server_4 : standby is not configured
server_5 :
Error 4003: server_5 : standby is not configured
fsck 1.35 (28-Feb-2004)
/dev/ndj1: clean, 13836/231360 files, 233729/461860 blocks
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
Waiting for 1R2_500_5 access ...done
Write Disable device(s) on RA at target (R2)..............Done.
Suspend RDF link(s).......................................Done.
Merge device track tables between source and target.......Started.
Devices: 045A-045F, 0467-0477 in (0557,005).............. Merged.
Devices: 0478-0489 in (0557,005)......................... Merged.
Devices: 048A-049B in (0557,005)......................... Merged.
Devices: 049C-04A6 in (0557,005)......................... Merged.
Merge device track tables between source and target.......Done.
Resume RDF link(s)........................................Started.
Resume RDF link(s)........................................Done.
Read/Write Enable device(s) on SA at source (R1)..........Done.
Waiting for 1R2_500_5 sync ...done
Starting restore on remote site CELERRA ...
Waiting for nbs clients to start ... done
Waiting for nbs clients to start ... done
Suspend RDF link(s).......................................Done.
server_2 :
server_2 : going standby
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
doneThe nas Commands
server_3 :
nas_rdf
341
The nas Commands
server_3 : going standby
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
server_4 :
Error 4003: server_4 : standby is not configured
server_5 :
Error 4003: server_5 : standby is not configured
Resume RDF link(s)........................................Started.
Resume RDF link(s)........................................Done.
If the RDF device groups were setup to operate in ASYNCHRONOUS ( SRDF/A ) mode,
now would be a good time to set it back to that mode.
Would you like to set device group 1R2_500_5 to ASYNC Mode ? [yes or no]: no
done
EXAMPLE #7
To restore a source VNX after failover, when using Dynamic SRDF, as
rdfadmin su to root user, type:
# /nas/sbin/nas_rdf -restore
Is remote site CELERRA ready for Storage restoration?
Do you wish to continue? [yes or no]: yes
Contact eng17335 ... is alive
Restore will now reboot the source site control station.
Do you wish to continue? [yes or no]: yes
Device Group (DG) Name
DG's Type
DG's Symmetrix ID
Remote Symmetrix ID
RDF (RA) Group Number
:
:
:
:
:
1R2_500_1
RDF1
000280600187
000280600118
1 (00)
(Microcode Version: 5568)
(Microcode Version: 5568)
Source (R1) View
Target (R2) View MODES
------------------------------------------------ ---- -----------ST
LI
ST
Standard
A
N
A
Logical
T R1 Inv R2 Inv K
T R1 Inv R2 Inv
RDF Pair
Device Dev
E Tracks Tracks S Dev
E Tracks Tracks MDA STATE
------------------------------- -- ------------------- --- -----------DEV001 0056 RW
DEV002 0057 RW
DEV003 0032 RW
...............
BCV008 0069 RW
BCV009 006A RW
BCV010 006B RW
Total
Track(s)
342
0
0
0
0 RW 0030 WD
0 RW 0031 WD
0 RW 000C WD
0
0
0
0 S.. Synchronized
0 S.. Synchronized
0 S.. Synchronized
0
0
0
0 RW 005F WD
0 RW 0060 WD
0 RW 0061 WD
0
0
0
0 S.. Synchronized
0 S.. Synchronized
0 S.. Synchronized
------ -----0
0
Command Line Interface Reference for File 8.1
------ -----0
0
The nas Commands
MB(s)
0.0
0.0
0.0
0.0
Legend for MODES: M(ode of Operation): A = Async, S = Sync, E = Semi-sync, C =
Adaptive Copy
D(omino)
: X = Enabled, . = Disabled
A(daptive Copy)
: D = Disk Mode, W = WP Mode, . = ACp off
Is remote site CELERRA ready for Network restoration?
Do you wish to continue? [yes or no]: yes
server_2 : done
server_3 : done
server_4 :
Error 4003: server_4 : standby is not configured
server_5 :
Error 4003: server_5 : standby is not configured
fsck 1.35 (28-Feb-2004)
/dev/sdj1: clean, 12956/219968 files, 188765/439797 blocks
An RDF 'Failover' operation execution is
in progress for device group '1R2_500_1'. Please wait...
Write Disable device(s) on SA at source (R1)..............Done.
Suspend RDF link(s).......................................Done.
Swap RDF Personality......................................Started.
Swap RDF Personality......................................Done.
Suspend RDF link(s).......................................Done.
Read/Write Enable device(s) on SA at source (R1)..........Done.
Resume RDF link(s)........................................Started.
Resume RDF link(s)........................................Done.
Read/Write Enable device(s) on SA at target (R2)..........Done.
The RDF 'Failover' operation successfully executed for
device group '1R2_500_1'.
Waiting for 1R2_500_1 sync ...done
Starting restore on remote site CELERRA ...
Suspend RDF link(s).......................................Done.
server_2 :
server_2 : going standby
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
server_3 :
server_3 : going standby
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
nas_rdf
343
The nas Commands
server_4 :
Error 4003: server_4: standby is not configured
server_5 :
Error 4003: server_5: standby is not configured
done
EXAMPLE #8
To run all available checks on a source VNX, as a nasadmin su to root
user, type:
# /nas/sbin/nas_rdf -check -all
--------------------- SRDF Health Checks --------------------SRDF: Checking device is normal.......................... Pass
SRDF: Checking system is restored........................ Pass
SRDF: Checking R1 SRDF session is Synch or Consistent.... Pass
SRDF: Checking R1 Data Mover configuration is valid...... Pass
SRDF: Checking R1 devices are available.................. Pass
SRDF: Checking R1 device group has all devices........... Pass
SRDF: Checking R2 SRDF session is Synch or Consistent.... Pass
SRDF: Checking R2 Data Mover configuration is valid...... Pass
SRDF: Checking R2 devices are available.................. Pass
SRDF: Checking R2 device group has all devices........... Pass
EXAMPLE #9
To run one or more specific available checks on a source VNX, as a
nasadmin su to root user, type:
# /nas/sbin/nas_rdf -check r1_dev_group,r2_dev_group
--------------------- SRDF Health Checks --------------------SRDF: Checking R1 device group has all devices........... Pass
SRDF: Checking R2 device group has all devices........... Pass
EXAMPLE #10
To initiate an SRDF failover from the source VNX to the destination,
without the SRDF health check for the following use cases, a
rdfadmin su to root user, type:
# /nas/sbin/nas_rdf -activate -skip_rdf_operations
-nocheck
◆
SRDF STAR concurrent or cascaded
◆
SRDF concurrent or cascaded
◆
SRDF R2 enable (Split)
SiteA to SiteB/SiteC failover case
Skipping SRDF health check ....
Skipping Site A shutdown process for the skip_rdf_opertaions option ....
Successfully pinged (Remotely) Symmetrix ID: 000194900462
Successfully pinged (Remotely) Symmetrix ID: 000194900546
344
Command Line Interface Reference for File 8.1
The nas Commands
Skipping symrdf failover process ....
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
fsck 1.39 (29-May-2006)
/dev/ndj1: recovering journal
/dev/ndj1: clean, 15012/252928 files, 271838/516080 blocks
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
id
type acl slot groupID state name
1
1
0
2
0
server_2
2
1
0
3
0
server_3
server_2 :
server_2 : going offline
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
Skipping symrdf update process ....
A reboot Control Station request was sent to Site A to clean up old processes ....
SiteB to SiteC failover case
[root@CS_C rdfadmin]# /nas/sbin/nas_rdf -activate
-skip_rdf_operations -nocheck
Skipping Site A shutdown process ....
For Site B to Site C failover or Site C to Site B failover, nas_rdf -restore
-skip_rdf_operations -skip_SiteA_shutdown and reboot -f -n operations must be
done on the source side Control Station (with read write backend) to clean up
old processes before continue this activate operation unless the source side
is not reachable or destroyed.
Do you wish to continue? [yes or no]: yes
Successfully pinged (Remotely) Symmetrix ID: 000194900431
Successfully pinged (Remotely) Symmetrix ID: 000194900546
Successfully pinged (Remotely) Symmetrix ID: 000194900673
Skipping symrdf failover process ....
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
fsck 1.39 (29-May-2006)
/dev/ndj1: clean, 14717/252928 files, 279439/516080 blocks
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
server_2 :
server_2 : going standby
rdf : going active
replace in progress ...done
nas_rdf
345
The nas Commands
failover activity complete
commit in progress (not interruptible)...done
done
Skipping symrdf update process ....
A reboot Control Station request was sent to 10.245.64.168 to clean up old
processes ....
EXAMPLE #11
To initiate an SRDF failover from the source VNX to the destination,
without the SRDF health check for the case SiteA Data Movers are
already shutdown and the Control Station is already rebooted, type:
# /nas/sbin/nas_rdf -activate -skip_SiteA_shutdown
-nocheck
Skipping SRDF health check ....
Skipping Site A shutdown process ....
This skip_SiteA_shutdown option is only for the case the Site A Data Movers have
been already shutdown and the Site A Control Station has been already rebooted
to clean up old processes.
Do you wish to continue? [yes or no]: yes
Successfully pinged (Remotely) Symmetrix ID: 000194900431
Successfully pinged (Remotely) Symmetrix ID: 000194900462
Successfully pinged (Remotely) Symmetrix ID: 000194900673
Write Disable device(s) on SA at source (R1)..............Done.
Suspend RDF link(s).......................................Done.
Read/Write Enable device(s) on RA at target (R2)..........Done.
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
fsck 1.39 (29-May-2006)
/dev/ndj1: recovering journal
/dev/ndj1: clean, 14237/252928 files, 297432/516080 blocks
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
id
type acl slot groupID state name
1
4
2000 2
0
server_2
2
1
1000 3
0
server_3
server_3 :
server_3 : going offline
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
Suspend RDF link(s).......................................Done.
Merge device track tables between source and target.......Started.
346
Command Line Interface Reference for File 8.1
The nas Commands
Devices: 0078-0078 in (0546,011)..........................Merged.
Merge device track tables between source and target.......Done.
Resume RDF link(s)........................................Started.
Resume RDF link(s)........................................Done.
A shutdown request was sent to Site A to clean up old processes ....
EXAMPLE #12
To restore a source VNX after failover for the following use cases, as a
nasadmin su to root user, type:
# /nas/sbin/nas_rdf -restore -skip_rdf_operations
◆
SRDF STAR concurrent or cascaded
◆
SRDF concurrent or cascaded
◆
SRDF R2 enable (Split)
Restore on SiteB/SiteC
Skipping session check ....
Is remote site CELERRA ready for Storage restoration?
Do you wish to continue? [yes or no]: yes
Contact eng564169 ... is alive
Restore will now reboot the source site control station. This process may take
several minutes.
Do you wish to continue? [yes or no]: yes
Halting SiteA Data Movers and rebooting SiteA Control Station ....
Checking SiteA Data Mover halt status ....
Skipping symrdf update operation ....
Is remote site CELERRA ready for Network restoration?
Do you wish to continue? [yes or no]: yes
server_2 : done
server_3 :
Error 4003: server_3 : standby is not configured
fsck 1.39 (29-May-2006)
/dev/ndj1: clean, 14716/252928 files, 279441/516080 blocks
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
Skipping symrdf failback operation & Site A restore ....
Restore on SiteA
To restore on siteA as a nasadmin su to root user, type:
[root@CS_A nasadmin]# /nasmcd/sbin/nas_rdf -restore
-skip_rdf_operations
Waiting for NAS services to finish starting......................... Done
Ensure that SiteA is currently write-enabled to continue this restore operation.
nas_rdf
347
The nas Commands
Do you wish to continue? [yes or no]: yes
Waiting for nbs clients to start ... done
Waiting for nbs clients to start ... done
server_2 :
server_2 : going standby
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
done
server_3 :
Error 4003: server_3 : standby is not configured
Skipping symrdf set async operation ....
Run 'nas_diskmark -mark -all' on all Control Stations in the SRDF configuration
to make sure the SRDF configuration and nasdb are restored completely.
Starting Services ...done
Example # 13
To disable SiteB for failover from SiteB to SiteC, as a rdfadmin su to
root user, type:
# /nas/sbin/nas_rdf -restore -skip_rdf_operations
-skip_SiteA_shutdown
Skipping session check ....
Skipping Site A shutdown process ....
Skipping symrdf update operation ....
Is remote site CELERRA ready for Network restoration?
Do you wish to continue? [yes or no]: yes
server_2 : done
server_3 :
Error 4003: server_3 : standby is not configured
fsck 1.39 (29-May-2006)
/dev/ndj1: clean, 14717/252928 files, 279439/516080 blocks
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
Waiting for nbs clients to die ... done
Waiting for nbs clients to start ... done
Skipping symrdf failback operation & Site A restore ....
348
Command Line Interface Reference for File 8.1
The nas Commands
nas_replicate
Manages loopback, local, and remote VNX Replicator sessions.
SYNOPSIS
nas_replicate
-list [-id]
| -info {-all|id=<sessionId>|<name>}
| -create <name>
-source -fs {<fsName>|id=<fsId>}
[-sav {<srcSavVolStoragePool>|id=<srcSavVolStoragePoolId>}
[-storageSystem <srcSavStorageSerialNumber>]]
-destination {-fs {id=<dstFsId>|<existing_dstFsName>}
| -pool {id=<dstStoragePoolId>|<dstStoragePool>}
[-storageSystem <dstStorageSerialNumber> ] }
[-vdm <dstVdmName>]}
[-sav {id=<dstSavVolStoragePoolId>|<dstSavVolStoragePool>}
[-storageSystem <dstSavStorageSerialNumber> ] ]
-interconnect {<name>|id=<interConnectId>}
[-source_interface {ip=<ipAddr>|<nameServiceInterfaceName>}]
[-destination_interface
{ip=<ipAddr>|<nameServiceInterfaceName>}]
[{-max_time_out_of_sync <maxTimeOutOfSync>|-manual_refresh}]
[-overwrite_destination][-tape_copy][-background]
| -create <name>
-source -vdm <vdmName>
-destination {-vdm <existing_dstVdmName>|-pool
{id=<dstStoragePoolId>|<dstStoragePool>}[-storageSystem
<dstStorageSerialNumber> ]}
-interconnect {<name>|id=<interConnectId>}
[-source_interface {ip=<ipAddr>|<nameServiceInterfaceName>}]
[-destination_interface
{ip=<ipAddr>|<nameServiceInterfaceName>}]
[{-max_time_out_of_sync <maxTimeOutOfSync>|-manual_refresh}]
[-overwrite_destination][-background]
| -start {<name>|id=<sessionId>}
[-interconnect {<name>|id=<interConnectId>}]
[-source_interface {ip=<ipAddr>|<nameServiceInterfaceName>}]
[-destination_interface
{ip=<ipAddr>|<nameServiceInterfaceName>}]
[{-max_time_out_of_sync <maxTimeOutOfSync>|-manual_refresh}]
[-overwrite_destination][-reverse][-full_copy][-background]
| -modify {<name>|id=<sessionId>} [-name <new name>]
[-source_interface {ip=<ipAddr>|<nameServiceInterfaceName>}]
[-destination_interface
{ip=<ipAddr>|<nameServiceInterfaceName>}]
[{-max_time_out_of_sync <maxTimeOutOfSync>|-manual_refresh}]
| -stop {<name>|id=<sessionId>} [-mode {source|destination|both}]
[-background]
| -delete {<name>|id=<sessionId>} [-mode {source|destination|both}]
[-background]
| -failover {<name>|id=<sessionId>} [-background]
| -switchover {<name>|id=<sessionId>}
| -reverse {<name>|id=<sessionId>} [-background]
| -refresh {<name>|id=<sessionId>} [-source {<ckptName>|id=<ckptId>}
nas_replicate
349
The nas Commands
-destination {<ckptName>|id=<ckptId>}] [-background]
Note: This command manages replication sessions using VNX Replicator. For
a one-time file system copy using VNX Replicator, use the nas_copy
command. For ongoing file system copy, use the nas_replicate command.
DESCRIPTION
nas_replicate creates, manages, or displays session information for
ongoing VNX Replicator replication of a file system, Virtual Data
Mover (VDM) at a destination using an existing Data Mover
interconnect. Each session handles a single source object and
destination, and is assigned a globally unique ID, fixed for the life of
the session.
In response to a potential disaster scenario, use nas_replicate to
perform a failover of a specified replication session with possible data
loss. The -switchover option switches over a replication relationship
and performs synchronization of the source and destination without
data loss. Use nas_replicate to also reverse the direction of a
replication session or refresh the destination side with updates to the
source based on a time-out of synchronization value or on demand.
OPTIONS
-list [-id]
Displays all configured (or stopped) replication sessions on each Data
Mover in the VNX for file cabinet. Each session is represented by
either a name or a session ID that is generated automatically
whenever a session is configured and is globally unique.
Use this option to obtain the session ID needed for another
command. Since session IDs are lengthy, the session ID obtained from
this command can be copied and pasted into the command.
-info {-all|id=<sessionId>|<name>}
Displays the status of a specific configured (or stopped) replication
session or copy session, or the status of all replication sessions.
CREATING FILE SYSTEM
REPLICATION
-create <name>
Assigns a name to the file system replication session. The name must
be unique for each Data Mover pair, which is defined by the
interconnect.
-source -fs {<fsName>|id=<fsId>}
-sav<srcSavVolStoragePool>|
id=<srcSavVolStoragePoolId>}[-storageSystem<srcSavSto
rageSerialNumber>]]
350
Command Line Interface Reference for File 8.1
The nas Commands
Specifies the name or ID of the existing source file system to
replicate. The source file system must be mounted as read-only or
read and write.
Note: If the source file system is mounted to a VDM and the goal is to
replicate a CIFS environment for disaster recovery (that is, replicate a
VDM and the file systems mounted to the VDM), create a session to
replicate the VDM first, before replicating a file system mounted to the
VDM.
The -sav option allocates a storage pool for all subsequent
checkpoints for the file system. By default, if checkpoint storage
(the checkpoint SavVol) needs to be allocated for checkpoints of
the file system, the command uses the same storage pool used to
create the source file system.
The -storageSystem option identifies the system on which all
subsequent checkpoints for the source file system reside. For
RAID group-based pools, specifies the system when there are
multiple systems attached. For mapped pools, specify the pool ID
or the pool ID and system serial number to uniquely identify a
pool.
-destination {-fs {<existing_dstFsName>|
id=<dstFsId>}| -pool {<dstStoragePool>|
id=<dstStoragePoolId>}[-storageSystem
<dstStorageSerialNumber>]}
Specifies an existing destination file system or the storage needed
to create the destination file system. An existing destination file
system must be mounted as read-only and the same size as the
source. Specifying a storage pool or ID creates the destination file
system automatically, as read-only, using the same name and size
as the source file system.
The -storageSystem option identifies the system on which the
destination file system will reside. This is necessary when there
are multiple back-end systems attached. Use nas_storage -list to
obtain attached system serial numbers.
-vdm <dstVdmName>]} [-sav {id=<dstSavVolStoragePoolId>
|<dstSavVolStoragePool>}[-storageSystem
<dstStorageSerialNumber>] Specifying a pool with the -vdm
option mounts the destination file system to an existing VDM as
part of replication in a CIFS environment.
nas_replicate
351
The nas Commands
The -sav option allocates a storage pool for all subsequent
checkpoints of the destination file system. By default, if
destination checkpoint storage needs to be allocated for
checkpoints, the command uses the same storage pool used to
create the destination file system. The -storageSystem option
identifies the system on which the destination checkpoint will
reside. This is necessary when there are multiple back-end
systems attached. Use nas_storage -list to obtain attached system
serial numbers.
By default, the destination file system name will be the same as
the source file system name. If a file system with the same name
as the source file system already exists on the destination, the
naming convention <source_fs_name>_replica<#> will be used.
A number 1–4 is assigned according to how many replicas of that
file system already exist.
-interconnect {<name>|id=<interConnectId>}
Specifies the local (source) side of an established Data Mover
interconnect to use for this replication session.
Use the nas_cel -interconnect -list command on the source VNX
for file to list the interconnects available to the replication
sessions.
[-source_interface
{<nameServiceInterfaceName>|ip=<ipAddr>}]
Instructs the replication session to use a specific local interface
defined for the interconnect on the source VNX instead of
selecting the local interface supporting the lowest number of
sessions (the default). If this local interface was defined for the
interconnect using a name service interface name, specify the
name service interface name; if it was defined using an IP
address, specify the IP address. If you define an interface
using an IP address, ensure that the destination interface uses
the same IPv4/IPv6 protocol. An IPv4 interface cannot
connect to an IPv6 interface and vice versa. Both sides of the
connection must use the same protocol.
The source_interfaces field of the output from the nas_cel
-interconnect -info command shows how the source interface
was defined. This option does not apply to a loopback
interconnect, which always uses 127.0.0.1.
352
Command Line Interface Reference for File 8.1
The nas Commands
If no source interface is specified, the system will select an
interface. This ensures that the interface selected can
communicate with the destination interface.
[-destination_interface{<nameServiceInterfaceName>|ip
=<ipaddr>}]
Instructs the replication session to use a specific peer interface
defined for the interconnect on the destination VNX instead of
selecting the peer interface supporting the lowest number of
sessions (the default). If this peer interface was defined for the
interconnect using a name service interface name, specify the
name service interface name; if it was defined using an IP
address, specify the IP address. If you define an interface using an
IP address, ensure that the source interface uses the same
IPv4/IPv6 protocol. An IPv4 interface cannot connect to an IPv6
interface and vice versa. Both sides of the connection must use the
same protocol.
The destination_interfaces field of the output from the nas_cel
-interconnect -info command shows how the peer interface
was defined. This option does not apply to a loopback
interconnect, which always uses 127.0.0.1.
If no destination interface is specified, the system will select an
interface. This ensures that the interface selected can
communicate with the source interface.
[{-max_time_out_of_sync <maxTimeOutOfSync>|
-manual_refresh}]
Specifies the time, in 1.1440 minutes (up to 24 hours), that the
source and destination can be out of synchronization before an
update occurs. If you do not specify a max_time_out_of_sync
value, use the -manual_refresh option to indicate that the
destination will be updated on demand using the
nas_replicate -refresh command. If no option is selected, the
refresh default time for a file system replication is 10 minutes.
[-overwrite_destination]
For an existing destination object, discards any changes made
to the destination object and restores it from the established
common base, thereby starting the replication session from a
differential copy. If this option is not specified and the
destination object contains different content than the
established common base, an error is returned.
nas_replicate
353
The nas Commands
[-tape_copy]
For file system replication only, creates and stops the
replication session to enable an initial copy using the physical
tape backup and process instead of an initial copy over the
network. Using VNX Replicator describes the procedures for
performing a tape copy, which involves a manually issued
backup to tape from the source file system, a restore from tape
to the destination file system, and a start of the replication
session.
[-background]
Executes the command in an asynchronous mode. Use the
nas_task command to check the status of the command.
CREATING VDM
REPLICATION
-create <name>
Assigns a name to the VDM replication session. The name must be
unique for each Data Mover pair, which is defined by the
interconnect.
-source -vdm <vdmName>|[id=<VdmId>]
Specifies the name or ID of an existing VDM to replicate. This
replicates the CIFS working environment information contained
in the root file system of the VDM. The source VDM must be in a
loaded read/write or mounted read-only state. The source VDM
can be the source or destination VDM of another replication
session.
Note: Any file system mounted to a VDM must be replicated using file
system replication. VDM replication affects the VDM only.
-destination {-vdm {<existing_dstVdmName>|
id=<dstVdmId>|-pool {id=<dstStoragePoolId>|
<dstStoragePool>}[-storageSystem
<dstStorageSerialNumber>]}
Specifies either an existing destination VDM or the storage
needed to create the destination VDM. An existing destination
VDM must be mounted as read-only, the same size as the source,
and not loaded. The destination VDM can be the source of
another replication but cannot be the destination of another
replication. Specifying a storage pool creates the destination
VDM automatically, as read-only, using the same name and size
as the source VDM.
354
Command Line Interface Reference for File 8.1
The nas Commands
The -storageSystem option identifies the system on which the
destination VDM will reside. This is necessary when there are
multiple back-end systems attached. Use nas_storage -list to
obtain attached system serial numbers.
-interconnect {<name>|id=<interConnectId>}
Specifies the local (source) side of an established Data Mover
interconnect to use for this replication session.
Use the nas_cel -interconnect -list command on the source VNX
to list the interconnects available to replication sessions. The
nas_cel -interconnect -create command is executed twice, one
from each side, to create an interconnect between a pair of Data
Movers (two local Data Movers for local replication, or one local
and one remote, for remote replication). Loopback interconnects
are created for each Data Mover and are named automatically.
[-source_interface {<nameServiceInterfaceName>
|ip=<ipAddr>}]
Instructs the replication session to use a specific local interface
defined for the interconnect on the source VNX instead of
selecting the local interface supporting the lowest number of
sessions (the default). If this local interface was defined for the
interconnect using a name service interface name, specify the
name service interface name; if it was defined using an IP
address, specify the IP address. If you define an interface using an
IP address, ensure that the destination interface uses the same
IPv4/IPv6 protocol. An IPv4 interface cannot connect to an IPv6
interface and vice versa. Both sides of the connection must use the
same protocol.
The source_interfaces field of the output from the nas_cel
-interconnect -info command shows how the source interface
was defined. This option does not apply to a loopback
interconnect, which always uses 127.0.0.1.
If no source interface is specified, the system will select an
interface. This ensures that the interface selected can
communicate with the destination interface.
[-destination_interface{<nameServiceInterfaceName>|ip
=<ipaddr>}]
Instructs the replication session to use a specific peer interface
defined for the interconnect on the destination VNX instead of
selecting the peer interface supporting the lowest number of
sessions (the default). If this peer interface was defined for the
nas_replicate
355
The nas Commands
interconnect using a name service interface name, specify the
name service interface name; if it was defined using an IP
address, specify the IP address. If you define an interface using an
IP address, ensure that the source interface uses the same
IPv4/IPv6 protocol. An IPv4 interface cannot connect to an IPv6
interface and vice versa. Both sides of the connection must use the
same protocol.
The destination_interfaces field of the output from the nas_cel
-interconnect -info command shows how the peer interface was
defined. This option does not apply to a loopback interconnect,
which always uses 127.0.0.1.
If no destination interface is specified, the system will select an
interface. This ensures that the interface selected can
communicate with the source interface.
[{-max_time_out_of_sync <maxTimeOutOfSync>
|-manual_refresh}]
Specifies the time, in 1.1440 minutes (up to 24 hours), that the
source and destination can be out of synchronization before an
update occurs. If you do not specify a max_time_out_of_sync
value, use the -manual_refresh option to indicate that the
destination will be updated on demand using the nas_replicate
-refresh command. If no option is selected, the refresh default
time for a VDM replication is 5 minutes.
[-overwrite_destination]
For an existing destination object, discards any changes made to
the destination object and restores it from the established
common base, thereby starting the replication session from a
differential copy. If this option is not specified, and the
destination object contains different content than the established
common base, an error is returned.
[-background]
Executes the command in an asynchronous mode. Use the
nas_task command to check the status of the command.
START OPTIONS
-start {<name>|id=<sessionId>}
From the source side only, specifies the name or session ID needed to
start the replication session. A replication name is unique for each
Data Mover pair; if a duplicate name is detected on the system, the
session ID is required. To get the session ID, use nas_replicate -list.
356
Command Line Interface Reference for File 8.1
The nas Commands
[-interconnect {<name>|id=<interConnectId>}]
Specifies an established source-side (local) Data Mover
interconnect to use for the replication session. Use the nas_cel
-interconnect -list command to list the interconnects available to
replication sessions. The nas_cel -interconnect -create command
creates an interconnect between a pair of Data Movers (two local
Data Movers for local replication, or one local and one remote, for
remote replication). Loopback interconnects are created and
named automatically, and always use IP address 127.0.0.1.
[-source_interface {<nameServiceInterfaceName>
|ip=<ipaddr>}]
As the source interface for the replication session, uses a specific
local interface defined for the interconnect instead of any local
interface defined for the interconnect (the default, which enables
the software to select the interface supporting the lowest number
of sessions). If this interface was defined for the interconnect
using a name service interface name, specify the name service
interface name; if it was defined using an IP address, specify the
IP address (IPv4 or IPv6). If you define an interface using an IP
address, ensure that the destination interface uses the same
IPv4/IPv6 protocol. An IPv4 interface cannot connect to an IPv6
interface and vice versa. Both sides of the connection must use the
same protocol.
[-destination_interface{<nameServiceInterfaceName>|ip
=<ipaddr>}]
As the destination interface for the replication session, uses a
specific peer interface defined for the interconnect instead of any
peer interface defined for the interconnect (the default, which
enables the software to select the interface supporting the lowest
number of sessions). If this interface was defined for the
interconnect using a name service interface name, specify the
name service interface name; if it was defined using an IP
address, specify the IP address (IPv4 or IPv6). If you define an
interface using an IP address, ensure that the source interface uses
the same IPv4/IPv6 protocol. An IPv4 interface cannot connect to
an IPv6 interface and vice versa. Both sides of the connection
must use the same protocol.
[{-max_time_out_of_sync <maxtimeOutOfSync>
|-manual_refresh}]
Specifies the time, in 1.1440 minutes (up to 24 hours), that the
source and destination can be out of synchronization before an
update occurs. If you do not specify a max_time_out_of_sync
nas_replicate
357
The nas Commands
value, use the -manual_refresh option to indicate that the
destination will be updated on demand using the nas_replicate
-refresh command. If no option is selected, the refresh default
time for file system replication is 10 minutes, and 5 minutes for
VDM replication sessions.
[-overwrite_destination]
For an existing destination object, discards any changes made to
the destination object and restores the destination object from the
established, internal common base checkpoint, thereby starting
the replication session from a differential copy. If this option is not
specified and the destination object has different content than the
established common base, an error is returned.
[-reverse]
Reverses the direction of the replication session when invoked
from the new source side (the original destination). A reverse
operation continues to use the established replication name or
replication session ID. Use this option to restart replication after a
failover or switchover.
[-full_copy]
For an existing destination object that contains content changes,
performs a full copy of the source object to the destination object.
If replication cannot be started from a differential copy using the
-overwrite_destination option, omitting this option causes the
command to return an error.
[-background]
Executes the command in an asynchronous mode. Use the
nas_task command to check the status of the command.
MODIFY OPTIONS
-modify {<name>|id=<sessionId>}
From the source side only, specifies the name or session ID of the
replication session to be modified. If a duplicate name is detected on
the system, the session ID (fixed for the life of the session) is required.
To get the session ID, use nas_replicate -list.
Note: A session cannot be modified if a -stop, -delete, -reverse, -failover,
-switchover, -create, or -start operation is running. However, once a -modify
operation is running, any other operation is permitted.
358
Command Line Interface Reference for File 8.1
The nas Commands
[-name <newName>]
Renames the replication session to the new name specified. When
renaming a session, note that the name must be unique for each
Data Mover pair.
[-source_interface {<nameServiceInterfaceName>
|ip=<ipaddr>}]
Changes the source interface used for the session to another local
interface from the list defined for the interconnect. If this interface
was defined for the interconnect using a name service interface
name, specify the name service interface name; if it was defined
using an IP address, specify the IP address (IPv4 or IPv6). If you
change an IP address, ensure that the destination interface uses
the same IPv4/IPv6 protocol. An IPv4 interface cannot connect to
an IPv6 interface and vice versa. Both sides of the connection
must use the same protocol.
[-destination_interface{<nameServiceInterfaceName>|ip
=<ipaddr>}]
Changes the destination interface used for the session to another
peer interface from the list defined for the interconnect. If this
interface was defined for the interconnect using a name service
interface name, specify the name service interface name; if it was
defined using an IP address, specify the IP addres (IPv4 or IPv6).
If you change an IP address, ensure that the source interface uses
the same IPv4/IPv6 protocol. An IPv4 interface cannot connect to
an IPv6 interface and vice versa. Both sides of the connection
must use the same protocol..
[-max_time_out_of_sync <maxTimeOutOfSync>
|-manual_refresh]
Specifies the time, from 1.1440 minutes (up to 24 hours), that the
source and destination can be out of synchronization before an
update occurs. If you do not specify a max_time_out_of_sync
value, use the -manual_refresh option to indicate that the
destination will be updated on demand using the nas_replicate
-refresh command. If no option is selected, the refresh default
time for file system replication is 10 minutes, and 5 minutes for
VDM replication sessions.
STOP OPTIONS
-stop {<name>|id=<session_id>}
Executed from the Control Station on the source VNX, stops the
specified replication session but retains the session’s configuration
nas_replicate
359
The nas Commands
information. Any data transfer in progress is terminated immediately
and the destination object is restored to a consistent state.
Note: A session cannot be stopped if the -delete option is already running for
the session. Once a stop operation is in progress, only the options -list, -info,
and the nas_task command are permitted.
[-mode {source|destination|both}]
When stopping a session handling a local or remote replication
from the source side, the -mode both option immediately stops
both sides of the replication session. The -mode source option
stops only the replication session on the source and ignores the
other side of the replication relationship. If the destination side is
not operational, the -mode source option is required to stop the
session. From the destination side, only the -mode destination
option can be issued. When stopping a session handling a
loopback replication, you can specify any -mode option to stop
the session.
[-background]
Executes the command in an asynchronous mode. Use the
nas_task command to check progress.
DELETE OPTIONS
-delete {<name>|id=<session_id>}
Executed from the Control Station on the source VNX, cancels
replication data transfer if it is in progress, performs an internal
checkpoint restore of the latest destination checkpoint to bring the file
system back to a consistent state and then deletes the replication
session specified by the -mode options.
[-mode {source|destination|both}]
When deleting a local or remote replication session from the
source side, the -mode both option deletes both sides of the
replication session. The -mode source option immediately aborts
only the replication session on the source and ignores the other
side of the replication relationship. If the destination side is not
operational, the -mode source option is required to delete the
session. From the destination side, only the -mode destination
option can be issued. When deleting a loopback replication, you
can specify any -mode option to stop the session.
[-background]
Executes the command in an asynchronous mode. Use the
nas_task command to check progress.
360
Command Line Interface Reference for File 8.1
The nas Commands
The execution of the -delete option is asynchronous and can be
delayed if there is a network problem. During the delete process,
other operations on the replication session are not allowed.
FAILOVER OPTIONS
-failover {<name>|id=<session_id>}
In response to a potential disaster scenario, performs a failover of the
specified replication session with possible data loss. Execute this
command from the Control Station on the destination VNX only. This
command cancels any data transfer that is in process and marks the
destination object as read-write so that it can serve as the new source
object. When the original source Data Mover becomes reachable, the
source object is changed to read-only.
!
CAUTION
The execution of the failover operation is asynchronous and results
in data loss if all the data was not transferred to the destination site
prior to issuing the failover.
If there are multiple sessions using the same source object, only one
replication session can be failed over. After the selected session is
failed over, the other sessions become inactive until the session is
restarted or failed back.
[-background]
Executes the command in an asynchronous mode. Use the
nas_task command to check progress.
SWITCHOVER OPTIONS
-switchover {<name>|id=<session_id>}
For test or migration purposes, switches over the specified replication
relationship and performs synchronization of the source and
destination without data loss. Execute this command from the
Control Station on the source VNX only. This command stops
replication, mounts the source object as read-only, and marks the
destination object as read-write so that it can act as the new source
object.
Unlike a reverse operation, a switchover operation does not restart
replication.
[-background]
Executes the command in an asynchronous mode. Use the
nas_task command to check progress.
REVERSE OPTIONS
-reverse {<name>|id=<session_id>}
If executed from the source side of a replication session, reverses the
nas_replicate
361
The nas Commands
direction of the specified replication session without data loss. A
reverse synchronizes the destination with the source, mounts the
source object as read-only, stops replication, marks the destination
object as read-write so that it can act as the new source object, then
restarts replication in the reverse direction from a differential copy
(using the same configuration parameters established originally for
the session).
[-background]
Executes the command in an asynchronous mode. Use the
nas_task command to check progress.
REFRESH OPTIONS
-refresh{<name>|id=<session_id>}
Updates the destination side of the specified replication session based
on changes to the source side. Execute this command from the
Control Station on the source side only. A refresh operation handles
updates on demand; as an alternative, the -max_time_out_of_sync
option performs an update automatically after a specified number of
minutes.
If the data changes on the source are large, this command can take a
long time to complete. Consider running this command in
background mode.
[-source{<ckptName>|id=<ckptId>}
-destination{<ckptName>|id=<ckptId>}]
Instructs the replication -refresh option to use a specific checkpoint
on the source side and a specific checkpoint on the destination side.
Specifying source and destination checkpoints for the -refresh option
is optional. However, if you specify a source checkpoint, you must
also specify a destination checkpoint. Replication transfers the
contents of the user-specified source checkpoint to the destination file
system. This transfer can be either a full copy or a differential copy
depending on the existing replication semantics. After the transfer,
the replication internally refreshes the user-specified destination
checkpoint and marks the two checkpoints as common bases.
After the replication refresh operation completes successfully, both
the source and destination checkpoints have the same view of their
file systems. The replication continues to use these checkpoints as
common bases until the next transfer is completed. After a user
checkpoint is marked with a common base property, the property is
retained until the checkpoint is refreshed or deleted. A checkpoint
that is already paired as a common base with another checkpoint
propagates its common base property when it is specified as the
362
Command Line Interface Reference for File 8.1
The nas Commands
source in a replication refresh operation. This propagation makes it
possible for file systems without a direct replication relationship to
have common base checkpoints.
[-background]
Executes the command in an asynchronous mode. Use the nas_task
command to check progress.
SYSTEM OUTPUT
The number associated with the storage device is dependent on the
attached system. VNX for block displays a prefix of APM before a set
of integers, for example, APM00033900124-0019. For example,
Symmetrix systems appear as 002804000190-003C. The outputs
displayed in the examples use a VNX system.
EXAMPLE #1
To list all the VNX Replicator sessions, type:
$ nas_replicate -list
Type
Local Mover Interconnect
filesystem server_3 -->NYs3_LAs2
vdm
server_3 -->NYs3_LAs2
Name
ufs1_rep1
vdm1_rep1
Celerra
cs110
cs110
Status
OK
OK
Where:
Value
Name
Type
Source Mover
Interconnect
Celerra
Status
EXAMPLE #2
Definition
Either the name of the session or the globally unique session ID for the session, if there are
duplicate names on the system.
The type of replication session (ongoing file system (fs), copy, or VDM).
The source Data Mover for the session.
The name of the source-side interconnect used for the session.
The name of the VNX system.
The status of the session (OK, Active, Idle, Stopped, Error, Waiting) Info, Critical.
To create a file system replication session ufs1_rep1 on the source file
system ufs1 and destination pool clar_r5_performance on the
interconnect NYs3_LAs2 using the specified source and destination
IP addresses to be updated automatically every 5 minutes, type:
$ nas_replicate -create ufs1_rep1 -source -fs ufs1
-destination -pool clar_r5_performance -interconnect
NYs3_LAs2 -source_interface ip=10.6.3.190
-destination_interface ip=10.6.3.173
-max_time_out_of_sync 5
OK
EXAMPLE #3
ID
To display information for a replication session ufs1_rep1, type:
$ nas_replicate -info ufs1_rep1
= 184_APM00064600086_0000_173_APM00072901601_0000
nas_replicate
363
The nas Commands
Name
Source Status
Network Status
Destination Status
Last Sync Time
Type
Celerra Network Server
Dart Interconnect
Peer Dart Interconnect
Replication Role
Source Filesystem
Source Data Mover
Source Interface
Source Control Port
Source Current Data Port
Destination Filesystem
Destination Data Mover
Destination Interface
Destination Control Port
Destination Data Port
Max Out of Sync Time (minutes)
Next Transfer Size (Kb)
Latest Snap on Source
Latest Snap on Destination
Current Transfer Size (KB)
Current Transfer Remain (KB)
Estimated Completion Time
Current Transfer is Full Copy
Current Transfer Rate (KB/s)
Current Read Rate (KB/s)
Current Write Rate (KB/s)
Previous Transfer Rate (KB/s)
Previous Read Rate (KB/s)
Previous Write Rate (KB/s)
Average Transfer Rate (KB/s)
Average Read Rate (KB/s)
Average Write Rate (KB/s)
EXAMPLE #4
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
ufs1_rep1
OK
OK
OK
Thu Dec 13 14:47:16 EST 2007
filesystem
cs110
NYs3_LAs2
20004
source
ufs1
server_3
10.6.3.190
0
0
ufs1_replica3
server_2
10.6.3.173
5081
8888
5
0
0
0
No
76
11538
580
0
0
0
6277
0
0
To create a VDM replication session vdm_rep1 on source VDM vdm1
and destination pool clar_r5_performance on the interconnect
NYs3_LAs2 with the given source and destination IP addresses to be
updated automatically every 5 minutes, type:
$ nas_replicate -create vdm1_rep1 -source -vdm vdm1
-destination -pool clar_r5_performance -interconnect
NYs3_LAs2 -source_interface ip=10.6.3.190
-destination_interface ip=10.6.3.173
-max_time_out_of_sync 5
OK
EXAMPLE #5
364
To list existing replication sessions, type:
Command Line Interface Reference for File 8.1
The nas Commands
Name
ufs1_rep1
vdm1_rep1
$ nas_replicate
Type Local Mover
filesystem
server_3
vdm
server_3
EXAMPLE #6
-list
Interconnect
-->NYs3_LAs2
-->NYs3_LAs2
Celerra Status
cs110
OK
cs110
OK
To manually synchronize source and destination for the replication
session ufs1_rep1, type:
$ nas_replicate -refresh ufs1_rep1
OK
EXAMPLE #7
To manually synchronize source and destination for the replication
session ufs1_rep1 by using user checkpoints on the source and the
destination, type:
$ nas_replicate -refresh ufs1_rep1 -source id=101
-destination id=102
OK
EXAMPLE #8
To stop replication on both source and destination for the replication
session ufs1_rep1, type:
$ nas_replicate -stop ufs1_rep1 -mode both
OK
EXAMPLE #9
To start stopped replication session ufs1_rep1 on interconnect
NYs3_LAs2, specify manual refresh and to overwrite the destination
LUN performing a full copy, type:
$ nas_replicate -start ufs1_rep1 -interconnect NYs3_LAs2
-manual_refresh -overwrite_destination -full_copy
OK
EXAMPLE #10
ID
Name
Source Status
Network Status
Destination Status
Last Sync Time
Type
Celerra Network Server
Dart Interconnect
Peer Dart Interconnect
Replication Role
Source VDM
Source Data Mover
Source Interface
Source Control Port
To display information for the VDM replication session vdm_rep1,
type:
$ nas_replicate -info vdm1_rep1
= 278_APM00064600086_0000_180_APM00072901601_0000
= vdm1_rep1
= OK
= OK
= OK
= Fri Dec 14 16:49:54 EST 2007
= vdm
= cs110
= NYs3_LAs2
= 20004
= source
= vdm1
= server_3
= 10.6.3.190
= 0
nas_replicate
365
The nas Commands
Source Current Data Port
Destination VDM
Destination Data Mover
Destination Interface
Destination Control Port
Destination Data Port
Max Out of Sync Time (minutes)
Next Transfer Size (Kb)
Latest Snap on Source
Latest Snap on Destination
Current Transfer Size (KB)
Current Transfer Remain (KB)
Estimated Completion Time
Current Transfer is Full Copy
Current Transfer Rate (KB/s)
Current Read Rate (KB/s)
Current Write Rate (KB/s)
Previous Transfer Rate (KB/s)
Previous Read Rate (KB/s)
Previous Write Rate (KB/s)
Average Transfer Rate (KB/s)
Average Read Rate (KB/s)
Average Write Rate (KB/s)
EXAMPLE #11
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
0
vdm1
server_2
10.6.3.173
5081
8888
5
0
0
0
No
313
19297
469
0
0
0
155
0
0
To change the session name vdm1_rep1 to vdm1_rep2, and to change
max time out of sync value to 90, type:
$ nas_replicate -modify vdm1_rep1 -name vdm1_rep2
-max_time_out_of_sync 90
OK
EXAMPLE #12
To failover the replication session ufs1_rep1, type on destination:
$ nas_replicate -failover ufs1_rep1
OK
EXAMPLE #13
To start failed over replication in the reverse direction, type:
$ nas_replicate -start ufs1_rep1 -interconnect LAs2_NYs3
-reverse -overwrite_destination
OK
EXAMPLE #14
To reverse direction of the replication session ufs1_rep1, type:
$ nas_replicate -reverse ufs1_rep1
OK
EXAMPLE #15
To switch over the replication session ufs1_rep1 using the
-background option, type:
$ nas_replicate -switchover ufs1_rep1 -background
366
Command Line Interface Reference for File 8.1
The nas Commands
Info 26843676673: In Progress: Operation is still running.
the Task Status screen for results.
Check task id 4058 on
*** Comment: Use nas_task -info command to find out the status of background task.
EXAMPLE #16
To delete the replication session fs1_rep1 on both source and
destination, type:
$ nas_replicate -delete fs1_rep1 -mode both
OK
nas_replicate
367
The nas Commands
nas_rp
Manages VNX systems that use RecoverPoint and defines the
relationships needed for cabinet level disaster recovery.
SYNOPSIS
nas_rp
-cabinetdr {
| -info
| -init <cel_name>
| -failover
| -failback }
|-cg {
| -list
| -info {<cg_name>|id=<cg_id>}
| -suspend {<cg_name>|id=<cg_id>}
| -resume {<cg_name>|id=<cg_id>}
| -modify {<cg_name>|id=<cg_id>} -rpo <time in seconds> }
|-rpa{
| -add <rpa_name> -local_ip <local_ip> -remote_ip <remote_ip> -admin
<rpa_adminname> [-password <password>]
| -delete {<rpa_name>|id=<rpa_Id>}
| -update {<rpa_name>|id=<rpa_Id>}
| -update {<rpa_name>|id=<rpa_Id>}-local_ip <local_ip> -remote_ip
<remote_ip> -admin <rpa_adminname> [-password <password>]
| -info {<rpa_name>|id=<rpa_Id>} [-version] [-verbose]
| -list }
DESCRIPTION
OPTIONS
nas_rp manages VNX systems and cabinet level disaster recovery
using RecoverPoint.
-cabinetdr
Manages the activities needed for cabinet level disaster recovery.
-info
Displays the CG to be failed over, the state and transfer
information of the CG, servers that have been configured as
standby, and the state of these servers.
-init <cel_name>
Initializes a source or destination (target) Culham system.
-failover
Initiates a RecoverPoint cabinet failover from the source VNX to
the destination VNX. The -failover option is executed on the
destination Control Station at the discretion of the user. The
368
Command Line Interface Reference for File 8.1
The nas Commands
-failover option fails over the RecoverPoint NAS CGs. It sets each
RecoverPoint-protected volume on the source VNX as read-only
and each mirrored volume on the destination VNX as read-write.
-failback
Restores a source VNX after a failover. The -failback option is
initially executed on the destination Control Station. The data on
each destination volume is copied to the corresponding volume
on the source VNX. On the destination VNX, services on each
standby Data Mover are stopped. (NFS clients connected to these
Data Movers see a “server unavailable” message; CIFS client
connections time out.) Each volume on the source VNX is set as
read-write, and each mirrored volume on the destination VNX is
set as read-only.
-cg
Manages the NAS Consistency Groups or CG.
-list
Lists all the NAS CGs discovered on the RPA.
-info {<cg_name>|id=<cg_id>}
Displays the properties of a specified NAS CG, including the
devices or LUNs that are being replicated.
-suspend {<cg_name>|id=<cg_id>}
Suspends the data transfer for a specified NAS CG.
-resume {<cg_name>|id=<cg_id}
Resumes the suspended data transfer for a specified NAS CG.
-modify {<cg_name>|id=<cg_id>} –rpo <time in seconds>
Modifies the RPO settings on a specified NAS CG to a
user-specified value. The RPO value must be greater than or
equal to 1 minute.
-rpa
Manages the configurations of RecoverPoint Appliances or RPA.
-add <rpa_name> -local_ip <local_ip> -remote_ip
<remote_ip> -admin <rpa_adminname> [-password
<password>]
Adds an RPA entry to the NAS database with the RPA’s local IP
address, remote IP address and administrator login credentials.
The -add option can be run in the non-interactive mode with the
-password option specified.
nas_rp
369
The nas Commands
–delete {<rpa_name>|id=<rpa_Id>}
Deletes the RPA entry from the NAS database.
–update {<rpa_name>|id=<rpa_Id>}
Repairs all SSH RSA key issues to make the CS-RPA
communication go through based on SSH Key authentication.
–update {<rpa_name>|id=<rpa_Id>} -local_ip <local_ip>
-remote_ip <remote_ip> -admin <rpa_adminname>
[-password <password>]
Updates the local IP and remote IP of RPA entry.
-info {<rpa_name>|id=<rpa_id>} [-version] [-verbose]
Displays the properties of a specified RPA and the license
settings.
-list
Lists all the RPAs configured in the system.
EXAMPLE #1
To add an RPA to the NAS database, type:
$ /nas/sbin/nas_rp -rpa -add rpa1 -ip 10.245.64.16 -admin
admin -password admin
done
EXAMPLE #2
To fix all the SSH key related problems for an RPA, type:
$ /nas/sbin/nas_rp -rpa -update rpa1
done
EXAMPLE #3
To get detailed information about an RPA, type:
$ /nas/sbin/nas_rp -rpa -info rpa1
Name = rpa1
ID = 1
IP = 172.24.173.9
Administrator = admin
Installation ID = f442ea06c265ac910b153fcf855d98a8b86fe1f0
License =
Rts6q8aDCHxl175rYkKxuQs/NqZtosXmSWwXd0ZIojIVaXw6Ot9TLuMXkTD7/rAnVnvEUL2wfqF+5eBM
w0RS0wAA
Activation Code =
13dfacBpSohMgUIzHwgRJpYhboBvKlL56RnfU6CELkDoWXNa/PJB7whLptoiXfGsASdTgaqtHs57UFP7
M7wECwAA
Version = 3.2.SP2.P2(h.26)
EXAMPLE #4
To list all the RPAs configured in the system, type:
$ nas_rp -rpa -list
Id name
1 rpa1
370
ipaddress
172.24.173.9
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #5
ID
13
Name
NASCG_dev10_dev13
EXAMPLE #6
id
=
name
=
rpa
=
source copy
=
remote copy
=
source clar id
=
remote clar id
=
contains control luns =
transfer state
=
replication direction =
role
=
transfer mode
=
rpo
=
Replication sets
Id Name
Src
277 RSet 1
4
278 RSet 10
119
279 RSet 11
120
280 RSet 12
101
281 RSet 13
102
282 RSet 14
103
283 RSet 15
104
284 RSet 16
105
285 RSet 17
106
286 RSet 18
107
287 RSet 19
108
288 RSet 2
0
289 RSet 20
109
290 RSet 21
110
291 RSet 22
111
292 RSet 23
112
293 RSet 3
1
294 RSet 4
113
295 RSet 5
114
296 RSet 6
115
297 RSet 7
116
298 RSet 8
117
299 RSet 9
118
EXAMPLE #7
To list all the NAS CGs discovered on the RPA, type:
$ nas_rp -cg -list
RPA ID
Prod Copy
4
Src_dev10
Remote Copy
DR_dev13
Control LUN CG
True
To display the properties of a specified NAS CG, including the
devices or LUNs that are being replicated, type:
$ nas_rp -cg -info id=13
13
NASCG_dev10_dev13
rpa1
Src_dev10
DR_dev13
APM00102102333
APM00102400657
True
ACTIVE
remote -> local
Secondary
Async
SYSTEM
LUN
Dst LUN
25
119
120
101
102
103
104
105
106
107
108
18
109
110
111
112
19
113
114
115
116
117
118
Size
2147483648
214748364800
214748364800
214748364800
214748364800
214748364800
214748364800
214748364800
214748364800
214748364800
214748364800
11811160064
214748364800
214748364800
214748364800
214748364800
11811160064
214748364800
214748364800
214748364800
214748364800
214748364800
214748364800
To suspend data transfer for a specified NAS CG, type:
$ nas_rp -cg -suspend id=13
nas_rp
371
The nas Commands
done
EXAMPLE #8
To resume the suspended data transfer for a specified NAS CG, type:
$ nas_rp -cg -resume id=13
done
EXAMPLE #9
To modify the RPO settings on a specified NAS CG to a user-specified
value, type:
$ nas_rp -cg -modify id=13 -rpo 300
done
EXAMPLE #10
To display the CG to be failed over, the state and transfer information
of the CG, servers that have been configured as standby, and the state
of these servers, type:
# nas_rp -cabinetdr -info
***** Consistency Group Configuration *****
name
= NASCG_dev10_dev13
description
=
uid
= 50:6:1:60:B0:60:26:BC:0:0:0:0:0:0:0:0
state
= Consistent
role
= Primary
condition
= Active
recovery policy
= Automatic
number of mirrors
= 16
mode
= SYNC
owner
= 0
mirrored disks
=
root_disk,root_ldisk,d5,d8,d10,d11,d24,d25,d26,d27,d29,d30,d31,d32,d33,d39,
local clarid
= APM00053001552
remote clarid
= APM00053001549
mirror direction
= local -> remote
***** Servers configured with RPstandby *****
id
=
name
=
acl
=
type
=
slot
=
member_of =
standby
=
RDFstandby=
status
:
defined =
actual =
372
1
server_2
2000, owner=rpadmin, ID=500
nas
2
server_3, policy=auto
slot=2
enabled
online, active
Command Line Interface Reference for File 8.1
The nas Commands
id
=
name
=
acl
=
type
=
slot
=
member_of =
standbyfor=
RDFstandby=
status
:
defined =
actual =
2
server_3
1000, owner=nasadmin, ID=201
standby
3
server_2
slot=3
enabled
online, ready
***** Servers configured as standby *****
id
=
name
=
acl
=
type
=
slot
=
member_of =
standbyfor=
status
:
defined =
actual =
1
server_2
2000, owner=rpadmin, ID=500
standby
2
id
=
name
=
acl
=
type
=
slot
=
member_of =
standbyfor=
status
:
defined =
actual =
2
server_3
2000, owner=rpadmin, ID=500
standby
3
enabled
online, ready
enabled
online, ready
EXAMPLE #11
To initialize a source or destination (target) VNX, type:
# /nas/sbin/nas_rp -cabinetdr -init new_york
Culham with RecoverPoint Disaster Recovery
Initializing new_york --> new_jersey
Contacting new_york for remote storage info
Local storage system: FNM00093600019
Remote storage system: FNM00094700042
Discovering storage on new_york (may take several minutes)
Setting security information for FNM00093600019
Discovering storage at 172.24.173.26 (may take several minutes)
Contacting new_york for remote storage info
Contacting new_york for server capabilities...
Analyzing server information...
Source servers available to be configured for remote DR
nas_rp
373
The nas Commands
------------------------------------------------------1. server_2:new_york
2. server_3:new_york [ local standby ]
v. Verify standby server configuration
q. Quit initialization process
c. Continue initialization
Select a new_york server: 1
Destination servers available to act as remote standby
-----------------------------------------------------1. server_2:new_jersey
server_3:new_jersey [ local standby ]
b. Back
Select a new_jersey server: 1
Source servers available to be configured for remote DR
------------------------------------------------------1. server_2:new_york [ remote standby is server_2:new_jersey
]
2. server_3:new_york [ local standby ]
v. Verify standby server configuration
q. Quit initialization process
c. Continue initialization
Select a new_york server: c
Standby configuration validated OK
Enter user information for managing remote site new_york
Username: dradmin
Password: ********* Retype your response to validate
Password: *********
Setting up server_2 on new_york
Rebooting server_2 on new_jersey as standby ... done
Setting acl for server_2 on new_jersey
Updating the Culham domain information
done
EXAMPLE #12
To activate a failover, log in to destination Control Station by using
the dradmin account, su to root, and type:
# /nas/sbin/nas_rp -cabinetdr -failover
Sync with CLARiiON backend ...... done
Validating consistency group configuration ...... done
Is source site new_york ready for complete shut down (power OFF)? [yes or no] yes
Contacting source site source_cs, please wait... done
Shutting down remote site source_cs ...................................... done
Sync with CLARiiON backend ...... done
STARTING an RP 'FAILOVER' operation.
Device group: cg_new_york ............ done
The RP 'FAILOVER' operation SUCCEEDED.
Failing over Devices ... done
Adding NBS access for server_2 ........ done
Adding NBS access for server_3 ........ done
Activating the target environment ... done
server_2 : going offline
374
Command Line Interface Reference for File 8.1
The nas Commands
rdf : going active
replace in progress ...done
failover activity complete
server_3 : going offline
rdf : going active
replace in progress ...done
failover activity complete
commit in progress (not interruptible)...done
commit in progress (not interruptible)...done
commit in progress (not interruptible)...done
commit in progress (not interruptible)...done
done
EXAMPLE #13
To restore, log in to the destination Control Station by using the
dradmin account, su to root, and type:
# /nas/sbin/nas_rp -cabinetdr -failback
Sync with CLARiiON backend ...... done
Validating consistency group configuration ...... done
Contacting source site new_york, please wait... done
Running restore requires shutting down source site new_york.
Do you wish to continue? [yes or no] yes
Shutting down remote site new_york ....... done
Is source site new_york ready for storage restoration ? [yes or no] yes
Sync with CLARiiON backend ...... done
STARTING an RP 'RESUME' operation.
Consistency group: cg_new_york ............ done
The RP 'RESUME' operation SUCCEEDED.
Percent synchronized: 100
Updating consistency group ... done
Is source site ready for network restoration ? [yes or no] yes
Restoring servers ...... done
Waiting for servers to reboot ...... done
Removing NBS access for server_2 .. done
Removing NBS access for server_3 .. done
Waiting for consistency group ready to failback .... done
Sync with CLARiiON backend ...... done
STARTING an RP 'FAILBACK' operation.
Consistency group: cg_new_york ............ done
The RP 'FAILBACK' operation SUCCEEDED.
Restoring remote site cg_new_york, please wait... done
done
nas_rp
375
The nas Commands
nas_server
Manages the Data Mover (server) table.
SYNOPSIS
nas_server
-list [-all|-vdm]
| -delete <movername>
| -info {-all|<movername>|<slot_number>|id=<mover_id>
| -vdm {-all|<vdm_name>|id=<vdm_id>}}
| -rename <old_movername> <new_movername>
| -acl <acl_value> <movername>
| [-name <name>][-type <type>] -create <movername> [-setstate <state>]
[-fs <fs_name>|pool=<pool>][-option <options>]
| -vdm <vdm_name> -attach <interface> [,<interface2>...]
| -vdm <vdm_name> -detach <interface> [,<interface2>...]
| -vdm <vdm_name> -setstate <state> [<movername>][-ConvertI18N]
| -vdm <vdm_name> -move <movername> [-ConvertI18N]
DESCRIPTION
OPTIONS
nas_server manages the server tables for both physical and virtual
Data Movers (VDMs), creates a VDM, sets an access control value for
a physical Data Mover or VDM, renames a Data Mover and displays
attributes for a specified Data Mover or all Data Movers, deletes a
physical Data Mover entry from the server table, and deletes the
VDM configuration for a Data Mover, and attaches or detaches the
network interface to a VDM.
-list
Lists the Data Mover server table. The server table displays the ID,
type, access control level value, slot number, group ID, state, and
name of a Data Mover. VDMs have their own server table and do not
have a numeric reference in the general server table.
Note: The ID of the object is an integer and is assigned automatically. The
name of the Data Mover might be truncated if it is too long for the display. To
display the full name, use the -info option with the Data Mover ID.
Valid Data Mover types are:
• 1=nas
• 2=not used
• 3=not used
• 4=standby
• 5=not used
376
Command Line Interface Reference for File 8.1
The nas Commands
• 6=rdf
Note: The nas type is set automatically, vdm is set using nas_rp, rdf and
standby are set up using server_ssh.
[-all|-vdm]
The -all option displays the physical Data Mover and VDMs. The
-vdm option lists the VDMs only.
-delete <movername>
Deletes the specified physical Data Mover entry from the server table
or deletes the VDM configuration. A Data Mover that is being deleted
cannot contain mounted file systems.
Deleting a physical Data Mover requires the root command. Use
/nas/sbin/rootnas_server to execute a delete.
-info {-all|<movername>|<slot_number>|id =<mover_id>}
Displays attributes for all physical Data Movers, or a Data Mover
specified by its <movername>, <slot_number>, or <mover_id>.
-info -vdm {-all|<vdm_name>|id=<vdm_id>}
Displays attributes for all VDMs, or a specified VDM, including the
network interfaces that are being used by the CIFS servers.
-rename <old_movername> <new_movername>
Changes the name of the physical Data Mover or the VDM to the
specified name. The -create option provides more information for
rules applicable to naming a Data Mover.
-acl <acl_value> <movername>
Sets an access control level value that defines the owner of the
physical Data Mover or the Virtual Data Mover, and the level of
access allowed for users and groups defined in the access control
level table. The nas_acl command provides more information.
[-name <name>][-type vdm] -create <movername>
Creates a VDM with an optional name for the specified physical or
VDM. The movername is case-sensitive and supports the following
characters: a through z, A through Z, 0 through 9, _(underscore), (hyphen) though names may not start with a hyphen. The default
type is nas.
[-setstate <state>]
Sets the Data Mover to loaded or mounted.
nas_server
377
The nas Commands
The loaded option installs the image of the VDM onto the
physical Data Mover, but does not mount the non-root file
systems. The mounted option mounts the root_fs as read-only,
but the VDM image is not installed. The -setstate option is for use
with replication.
Note: Before a VDM image is loaded, the image must be unloaded from
the previous Data Mover, and the CIFS server must be joined using
server_cifs. The network interfaces used by the CIFS servers on the VDM
must be available on the destination Data Mover.
[-fs <fs_name>|pool=<pool>]
Specifies a file system or assigns a rule set known as a storage
pool for the VDM root file system.
For the -fs option, the file system must be unmounted, clean
(nas_fsck provides more information), and be either of type uxfs
or rawfs. For a loaded state VDM, only an uxfs file system type
can be used, but for mounted state VDM, both uxfs and rawfs can
be used. The file system to be provided as the VDM root file
system is renamed to root_fs_vdm_<vdm_name>. This is deleted
when the VDM is deleted.
[storage=<system_name>]
The storage pool option assigns a rule set for the root file system
of the VDM that contains automatically created volumes and
defines the type of disk volumes used and how they are
aggregated. Storage pools are system defined (storage pool
description provides more information) or user defined. nas_pool
provides a description of pool types.
[-option <options>]
Specifies the following comma-separated list of options:
fstype={rawfs|uxfs}
Specifies the file system type of the root file system for the server.
It can be either rawfs or uxfs type.
log_type={common|split}
Specifies the type of log file associated with the file system. Log
files can be either shared (common) or uniquely assigned to
individual file systems(split). For SRDF Async or STAR feature,
split option is strongly recommended to avoid fsck before
mounting a BCV file system on SiteB or SiteC.
378
Command Line Interface Reference for File 8.1
The nas Commands
-vdm <vdm_name> -attach <interface> [,<interface2>...]
Allows the user to manage the network interface(s) for a VDM. The
interfaces are attached to a VDM when the VDM state is loaded.
When an interface is attached to a VDM, the NFS clients connecting
the Data Mover through this interface have access to the file system
exported by the VDM configuration.
-vdm <vdm_name> -detach <interface> [,<interface2>...]
An attempt to delete an interface attached to the VDM with the
server_ifconfig command fails with an error message. It indicates
that the interface is currently used by the VDM <vdm_name>. The
user must detach the interface from the VDM before deleting it.
Note: If the user wants to share a VDM interface for both CIFS and NFSv3 or
NFSv4 clients, the administrator must create a CIFS server and assign it to the
interface by using the server_cifs command.
-vdm <vdm_name> -setstate <state>
Sets the state of the VDM to loaded, mounted, tempunloaded, or
permunloaded.
The loaded option installs the image of the VDM onto the physical
Data Mover, but does not mount the non-root file systems. The
mounted option mounts the root_fs read-only, but the VDM image is
not installed.
The tempunloaded option temporarily unloads the VDM image,
while the permunloaded option permanently unloads the image.
[<movername>]
Specifies a physical Data Mover for the VDM.
[-ConvertI18N]
When loading the VDM image, forces the conversion of the I18N
mode of the VDM’s root file system from ASCII to UNICODE.
The I18N mode of the Data Mover can be either ASCII or
UNICODE. The mode of the VDM must be the same as the
physical Data Mover, for example, when performing the -move
option, or when replicating.
This mode is used when the mode of the VDM root file system is
different from that of the physical Data Mover.
Default states are
nas_server
379
The nas Commands
-vdm <vdm_name> -move <movername>
Moves the image of the VDM onto the physical Data Mover, and
mounts the non-root file systems.
Note: Before a VDM image is loaded, the image must be unloaded from the
previous Data Mover, and the CIFS server must be joined using server_cifs.
The network interfaces used by the CIFS servers on the VDM must be
available on the destination Data Mover.
[-ConvertI18N]
When loading the VDM image, forces the conversion of the I18N
mode of the VDM’s root file system from ASCII to UNICODE.
The I18N mode of the Data Mover can be either ASCII or
UNICODE. The mode of the VDM must be the same as the
physical Data Mover, for example, when performing the -move
option, or when replicating.
This mode is used when the mode of the VDM root file system is
different from that of the physical Data Mover.
SEE ALSO
380
Configuring Virtual Data Mover on VNX, Using International Character
Sets for File, nas_fs, nas_volume, and server_cifs.
Command Line Interface Reference for File 8.1
The nas Commands
SYSTEM OUTPUT
VNX systems support the following system-defined storage pools:
clar_r1, clar_r5_performance, clar_r5_economy, clar_r6, clarata_r3,
clarata_r6, clarata_r10, clarata_archive, cm_r1, cm_r5_performance,
cm_r5_economy, cm_r6, cmata_r3, cmata_archive, cmata_r6,
cmata_r10, clarsas_archive, clarsas_r6, clarsas_r10, clarefd_r5,
clarefd_r10, cmsas_archive, cmsas_r6, cmsas_r10, and cmefd_r5.
Disk types when using VNX for block are CLSTD, CLEFD, and
CLATA, and for VNX for block involving mirrored disks are CMEFD,
CMSTD, and CMATA.
VNX with a Symmetrix system supports the following
system-defined storage pools: symm_std, symm_std_rdf_src,
symm_ata, symm_ata_rdf_src, symm_ata_rdf_tgt,
symm_std_rdf_tgt, symm_ata_rdf_tgt, symm_std_rdf_tgt, and
symm_efd.
For user-defined storage pools, the difference in output is in the disk
type. Disk types when using a Symmetrix are STD, R1STD, R2STD,
BCV, R1BCV, R2BCV, ATA, R1ATA, R2ATA, BCVA, R1BCA,
R2BCA, and EFD.
EXAMPLE #1
id
1
2
3
4
type
1
1
1
4
acl
1000
1000
1000
1000
To list the physical Data Mover table, type:
$ nas_server -list
slot groupID state
name
2
0 server_2
3
0 server_3
4
0 server_4
5
0 server_5
Where:
Value
id
type
acl
slot
groupID
state
name
Definition
ID of the Data Mover.
Type assigned to Data Mover.
Access control level value assigned to the Data Mover or VDM.
Physical slot in the cabinet where the Data Mover resides.
ID of the Data Mover group.
Whether the Data Mover is enabled=0, disabled=1, failed over=2.
Name given to the Data Mover.
nas_server
381
The nas Commands
EXAMPLE #2
id
1
2
3
4
type
1
1
1
4
acl
1000
1000
1000
1000
id
3
acl
0
server
1
To list the physical Data Mover and VDM table, type:
$ nas_server -list
slot groupID
state
2
0
3
0
4
0
5
0
mountedfs
rootfs
31
-all
name
server_2
server_3
server_4
server_5
name
vdm_1
EXAMPLE #1 provides a description of outputs for the physical Data
Movers. The following table provides a description of the command
output for the VDM table.
Where:
Value
id
acl
server
mountedfs
rootfs
name
Definition
ID of the Data Mover.
Access control level value assigned to the Data Mover or VDM.
Server on which the VDM is loaded on.
File systems that are mounted on this VDM.
ID number of the root file system.
Name given to the Data Mover or VDM.
To list the VDM server table, type:
acl
0
$ nas_server -list -vdm
mountedfs
rootfs
31
id
3
EXAMPLE #3
server
1
EXAMPLE #4
id
name
acl
type
slot
member_of
standby
status
defined
actual
=
=
=
=
=
=
=
:
=
=
To list information for a Data Mover, type:
$ nas_server -info server_2
1
server_2
1000, owner=nasadmin, ID=201
nas
2
server_5, policy=auto
enabled
online, ready
Where:
Value
id
name
382
name
vdm_1
Definition
ID of the Data Mover.
Name given to the Data Mover.
Command Line Interface Reference for File 8.1
The nas Commands
Access control level value assigned to the Data Mover or VDM.
Type assigned to Data Mover.
Physical slot in the cabinet where the Data Mover resides.
Group to which the Data Mover is a member.
If the Data Mover has a local standby associated with it.
Whether the Data Mover is enabled or disabled, and whether it is active.
acl
type
slot
member_of
standby
status
EXAMPLE #5
To display detailed information for all servers, type:
$ nas_server -info -all
1
server_2
1000, owner=nasadmin, ID=201
nas
2
id
name
acl
type
slot
member_of
standby
status
defined
actual
=
=
=
=
=
=
=
:
=
=
id
name
acl
type
slot
member_of
standby
status
defined
actual
=
=
=
=
=
=
=
:
=
=
2
server_3
1000, owner=nasadmin, ID=201
nas
3
id
name
acl
type
slot
member_of
standby
status
defined
actual
=
=
=
=
=
=
=
:
=
=
3
server_4
1000, owner=nasadmin, ID=201
nas
4
id
=
name
=
acl
=
type
=
slot
=
member_of =
standbyfor=
4
server_5
1000, owner=nasadmin, ID=201
standby
5
server_5, policy=auto
enabled
online, active
server_5, policy=auto
enabled
online, ready
server_5, policy=auto
enabled
online, ready
server_4,server_2,server_3
nas_server
383
The nas Commands
status
:
defined = enabled
actual = online, ready
EXAMPLE #4 provides a description of command outputs.
EXAMPLE #6
To display information for all VDMs, type:
$ nas_server -info -vdm -all
id
= 3
name
= vdm_1
acl
= 0
type
= vdm
server
= server_2
rootfs
= root_fs_vdm_1
I18N mode = UNICODE
mountedfs =
member_of =
status
:
defined = enabled
actual = mounted
Interfaces to services mapping:
Where:
Value
id
name
acl
type
server
rootfs
I18N mode
mountedfs
member_of
status
Interfaces to
services mapping
384
Definition
ID of the Data Mover.
Name of the Data Mover.
Access control level value assigned to the VDM.
For VDM server, the type is always VDM.
Server on which the VDM loaded.
Root file system of the VDM.
L18N mode of the VDM. I18N mode is either ASCII or UNICODE.
File systems that are mounted on this VDM.
If it is a member of a cluster, then this field will show the cluster name.
Whether the VDM is enabled or disabled, and whether it can be loaded ready, loaded active,
mounted, temporarily unloaded and permanently unloaded.
List of interfaces that are used for the services configured on this VDM. Currently, only CIFS
service is provided, so this field lists all the interfaces used in the CIFS servers configured on
this VDM.
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #7
To create a mounted VDM named vdm_1 on server_2 using the
storage pool, clar_r5_performance with a rawfs, type:
$ nas_server -name vdm_1 -type vdm -create server_2
-setstate mounted pool=clar_r5_performance -option
fstype=uxfs
id
= 3
name
= vdm_1
acl
= 0
type
= vdm
server
= server_2
rootfs
= root_fs_vdm_1
I18N mode = UNICODE
mountedfs =
member_of =
status
:
defined = enabled
actual = mounted
Interfaces to services mapping:
EXAMPLE #6 provides a description of command outputs.
EXAMPLE #8
To set the state of a vdm_1 to mounted, type:
$ nas_server -vdm vdm_1 -setstate mounted
id
= 3
name
= vdm_1
acl
= 0
type
= vdm
server
= server_2
rootfs
= root_fs_vdm_1
I18N mode = UNICODE
mountedfs =
member_of =
status
:
defined = enabled
actual = mounted
Interfaces to services mapping:
EXAMPLE #6 provides a description of command outputs.
EXAMPLE #9
id
name
acl
type
server
rootfs
I18N mode
mountedfs
=
=
=
=
=
=
=
=
To move the image of vdm_1 onto server_4, type:
$ nas_server -vdm vdm_1 -move server_4
3
vdm_1
0
vdm
server_4
root_fs_vdm_1
UNICODE
nas_server
385
The nas Commands
member_of =
status
:
defined = enabled
actual = loaded, ready
Interfaces to services mapping:
EXAMPLE #6 provides a description of command outputs.
EXAMPLE #10
id
name
acl
type
slot
member_of
standby
status
defined
actual
=
=
=
=
=
=
=
:
=
=
To rename a Data Mover entry from server_2 to dm2, type:
$ nas_server -rename server_2 dm2
1
dm2
1000, owner=nasadmin, ID=201
nas
2
server_5, policy=auto
enabled
online, active
EXAMPLE #4 provides a description of command outputs.
EXAMPLE #11
id
name
acl
type
slot
member_of
standby
status
defined
actual
=
=
=
=
=
=
=
:
=
=
To set the access control level for server_2, type:
$ nas_server -acl 1432 server_2
1
server_2
1432, owner=nasadmin, ID=201
nas
2
server_5, policy=auto
enabled
online, ready
Note: The value 1432 specifies nasadmin as the owner, gives users with an
access level of at least observer read-only access, users with an access level of
at least operator read/write access, and users with an access level of at least
admin read/write/delete access.
EXAMPLE #4 provides a description of command outputs.
386
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #12
To delete vdm_1, type:
$ nas_server -delete vdm_1
id
= 3
name
= vdm_1
acl
= 0
type
= vdm
server
=
rootfs
= root_fs_vdm_1
I18N mode = UNICODE
mountedfs =
member_of =
status
:
defined = enabled
actual = permanently unloaded
Interfaces to services mapping:
EXAMPLE #6 provides a description of command outputs.
EXAMPLE #13
To delete a physical Data Mover using root command, type:
$ /nas/sbin/rootnas_server -delete server_3
id
name
acl
type
slot
member_of
standby
status
defined
actual
=
=
=
=
=
=
=
:
=
=
2
server_3
0
nas
3
server_5, policy=auto
disabled
boot_level=0
EXAMPLE #6 provides a description of command outputs.
EXAMPLE #14
To create a VDM named vdm1 on the server 3, type:
$ nas_server -name vdm1 -type vdm -create server_3
id
= 43
name
= vdm1
acl
= 0
type
= vdm
server
= server_3
rootfs
= root_fs_vdm_vdm1
I18N mode = UNICODE
mountedfs =
member_of =
status
:
defined = enabled
actual = loaded, ready
nas_server
387
The nas Commands
Interfaces to services mapping:
EXAMPLE #15
To assign the network interface to vdm1, assuming vdm1if1 and
vdm1if2 exist and are not attached to another vdm, type:
$ nas_server -vdm vdm1 -attach vdm1if1, vdm1if2
id
= 43
name
= vdm1
acl
= 0
type
= vdm
server
= server_2
rootfs
= root_fs_vdm_vdm1
I18N mode = UNICODE
mountedfs =
member_of =
status
:
defined = enabled
actual = loaded, ready
Interfaces to services mapping:
interface=vdm1if1 :vdm
interface=vdm1if2 :vdm
EXAMPLE #16
To query the vdm1 state, type:
$ nas_server -info -vdm vdm1
id
= 43
name
= vdm1
acl
= 0
type
= vdm
server
= server_2
rootfs
= root_fs_vdm_vdm1
I18N mode = UNICODE
mountedfs =
member_of =
status
:
defined = enabled
actual = loaded, ready
Interfaces to services mapping:
interface=vdm1if2 :cifs vdm
interface=vdm1if1 :vdm
EXAMPLE #17
To create a VDM named vdm2 on the server_3 using split ufs log
type, type:
$ nas_server -name vdm2 -type vdm -create server_3
-setstate loaded pool=symm_std_rdf_src -o log_type=split
388
Command Line Interface Reference for File 8.1
The nas Commands
d
= 2
name
= vdm2
acl
= 0
type
= vdm
server
= server_3
rootfs
= root_fs_vdm_vdm2
I18N mode = ASCII
mountedfs =
member_of =
status
:
defined = enabled
actual = loaded, ready
Interfaces to services mapping:
To confirm a VDM ufs log type, type:
/nas/sbin/rootnas_fs -i root_fs_vdm_vdm2
id
= 49
name
= root_fs_vdm_vdm2
acl
= 0
in_use
= True
type
= uxfs
worm
= off
volume
= v1260
pool
= symm_std_rdf_src
member_of = root_avm_fs_group_8
rw_servers= server_3
ro_servers=
rw_vdms
=
ro_vdms
=
auto_ext = no,thin=no
log_type = split
fast_clone_level = 2
deduplication
= Off
stor_devs =
000194900462-10C6,000194900462-10CE,000194900462-10D6,000194900462-10DE,00019
4900462-10E6,000194900462-10EE,000194900462-10F6,000194900462-10FE
disks
= d1102,d1103,d1104,d1105,d1106,d1107,d1108,d1109
disk=d1102 stor_dev=000194900462-10C6
addr=c4t3l4-72-0
server=server_3
disk=d1102 stor_dev=000194900462-10C6
addr=c20t3l4-71-0
server=server_3
disk=d1102 stor_dev=000194900462-10C6
addr=c36t3l4-71-0
server=server_3
disk=d1102 stor_dev=000194900462-10C6
addr=c52t3l4-72-0
server=server_3
disk=d1103 stor_dev=000194900462-10CE
addr=c4t3l5-72-0
server=server_3
disk=d1103 stor_dev=000194900462-10CE
addr=c20t3l5-71-0
server=server_3
disk=d1103 stor_dev=000194900462-10CE
addr=c36t3l5-71-0
server=server_3
disk=d1103 stor_dev=000194900462-10CE
addr=c52t3l5-72-0
server=server_3
disk=d1104 stor_dev=000194900462-10D6
addr=c4t3l6-72-0
server=server_3
disk=d1104 stor_dev=000194900462-10D6
addr=c20t3l6-71-0
server=server_3
disk=d1104 stor_dev=000194900462-10D6
addr=c36t3l6-71-0
server=server_3
disk=d1104 stor_dev=000194900462-10D6
addr=c52t3l6-72-0
server=server_3
disk=d1105 stor_dev=000194900462-10DE
addr=c4t3l7-72-0
server=server_3
disk=d1105 stor_dev=000194900462-10DE
addr=c20t3l7-71-0
server=server_3
nas_server
389
The nas Commands
disk=d1105
disk=d1105
disk=d1106
disk=d1106
disk=d1106
disk=d1106
disk=d1107
disk=d1107
disk=d1107
disk=d1107
disk=d1108
disk=d1108
disk=d1108
disk=d1108
disk=d1109
disk=d1109
disk=d1109
disk=d1109
390
stor_dev=000194900462-10DE
stor_dev=000194900462-10DE
stor_dev=000194900462-10E6
stor_dev=000194900462-10E6
stor_dev=000194900462-10E6
stor_dev=000194900462-10E6
stor_dev=000194900462-10EE
stor_dev=000194900462-10EE
stor_dev=000194900462-10EE
stor_dev=000194900462-10EE
stor_dev=000194900462-10F6
stor_dev=000194900462-10F6
stor_dev=000194900462-10F6
stor_dev=000194900462-10F6
stor_dev=000194900462-10FE
stor_dev=000194900462-10FE
stor_dev=000194900462-10FE
stor_dev=000194900462-10FE
Command Line Interface Reference for File 8.1
addr=c36t3l7-71-0
addr=c52t3l7-72-0
addr=c4t3l8-72-0
addr=c20t3l8-71-0
addr=c36t3l8-71-0
addr=c52t3l8-72-0
addr=c4t3l9-72-0
addr=c20t3l9-71-0
addr=c36t3l9-71-0
addr=c52t3l9-72-0
addr=c4t3l10-72-0
addr=c20t3l10-71-0
addr=c36t3l10-71-0
addr=c52t3l10-72-0
addr=c4t3l11-72-0
addr=c20t3l11-71-0
addr=c36t3l11-71-0
addr=c52t3l11-72-0
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
server=server_3
The nas Commands
nas_slice
Manages the slice table.
SYNOPSIS
nas_slice
-list
| -delete <slice_name>
| -info {<slice_name>|id=<slice_id>}
| -rename <old_name> <new_name>
| [-name <name>] -create <volume_name> <size>[<offset>]
DESCRIPTION
OPTIONS
nas_slice creates, deletes, and renames a slice. nas_slice also displays
the slice table and displays the attributes of a specified slice.
-list
Lists the slice table.
Note: The ID of the slice is an integer and is assigned automatically. The
name of the slice may be truncated if it is too long for the display. To display
the full name, use the -info option with the slice ID.
-delete <slice_name>
Deletes an unused slice. Slices cannot be used by a storage pool.
-info {<slice_name>|id=<slice_id>}
Displays attributes for the slice as specified by the <slice_name> or
<slice_id>.
-rename <old_name> <new_name>
Changes the current name of a slice to the new name.
[-name <name>] -create <volume_name> <size>
Creates a slice with the specified size in megabytes. Also assigns an
optional name for the slice. Names are case-sensitive.
[<offset>]
Specifies the offset which is the distance in megabytes from the
beginning of the volume to the start of a slice. If an offset is not
specified, the system defaults to the next best-fit size.
SEE ALSO
Managing Volumes and File Systems with VNX Automatic Volume
Management, Managing Volumes and File Systems for VNX Manually,
nas_volume to manage metavolumes, stripe volumes, and disk
volumes and nas_cel.
nas_slice
391
The nas Commands
EXAMPLE #1
id
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
inuse
n
n
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
y
n
n
n
n
n
n
n
n
n
n
n
n
n
n
slice_of
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
1
To display the slice table, type:
$ nas_slice -list
offsetMB sizeMB
0
134
134
1
135
16
151
16
167
16
183
16
199
16
215
16
231
16
247
16
263
16
279
16
295
16
311
16
327
16
343
16
359
16
375
16
391
1
392
1
393
1
394
1
395
1
396
1
397
1
398
1
399
1
400
1
401
1
402
1
403
1
404
1
name
root_dos
root_layout
root_slice_1
root_slice_2
root_slice_3
root_slice_4
root_slice_5
root_slice_6
root_slice_7
root_slice_8
root_slice_9
root_slice_10
root_slice_11
root_slice_12
root_slice_13
root_slice_14
root_slice_15
root_slice_16
root_log_1
root_log_2
root_log_3
root_log_4
root_log_5
root_log_6
root_log_7
root_log_8
root_log_9
root_log_10
root_log_11
root_log_12
root_log_13
root_log_14
Note: This is a partial listing of the slice table.
Where:
Value
id
inuse
slice_of
offsetMB
sizeMB
name
392
Definition
ID of the slice.
Whether the slice is used.
Volume the slice is sliced from.
Distance in megabytes from the beginning of the disk to the start of the slice.
Slice size in megabytes.
Name of the slice.
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #2
To create a slice named slv1 on volume d8 of 1024 MB in size, type:
$ nas_slice -name slv1 -create d8 1024
id
=
name
=
acl
=
in_use
=
slice_of =
offset(MB)=
size (MB)=
volume_name
76
slv1
0
False
d8
0
1024
= slv1
Where:
Value
id
name
acl
inuse
slice_of
offset(MB)
size (MB)
volume_name
Definition
ID of the slice.
Name of the slice.
Access control level value assigned to the slice.
Whether the slice is used.
Volume the slice is sliced from.
Distance in Megabytes from the beginning of the disk to the start of the slice.
Slice size in Megabytes.
Volume used to create the slice.
EXAMPLE #3
To create a slice volume of 1 GB (1024 MB) with an offset of 1 GB
called slv2 from volume d8, type:
$ nas_slice -name slv2 -create d8 1024 1024
id
=
name
=
acl
=
in_use
=
slice_of =
offset(MB)=
size (MB)=
volume_name
77
slv2
0
False
d8
1024
1024
= slv2
EXAMPLE #2 provides a description of command outputs.
EXAMPLE #4
To display attributes of slv1, type:
$ nas_slice -info slv1
id
=
name
=
acl
=
in_use
=
slice_of =
offset(MB)=
size (MB)=
volume_name
76
slv1
0
False
d8
0
1024
= slv1
nas_slice
393
The nas Commands
EXAMPLE #2 provides a description of command outputs.
EXAMPLE #5
To rename a slice from slv to slv1, type:
$ nas_slice -rename slv slv1
id
=
name
=
acl
=
in_use
=
slice_of =
offset(MB)=
size (MB)=
volume_name
76
slv1
0
False
d8
0
1024
= slv1
EXAMPLE #2 provides a description of command outputs.
EXAMPLE #6
To delete slv1, type:
$ nas_slice -delete slv1
id
=
name
=
acl
=
in_use
=
slice_of =
offset(MB)=
size (MB)=
76
slv1
0
False
d8
0
1024
EXAMPLE #2 provides a description of command outputs.
394
Command Line Interface Reference for File 8.1
The nas Commands
nas_stats
Manages Statistics Groups.
SYNOPSIS
nas_stats
-groups
{ -list
| -info [-all|<statgroup_name>[,...]]
| -create <statgroup_name>
[-description "<description_line>"]
{<statpath_name>|<statgroup_name>}[,...]
| -modify <statgroup_name>
{[-rename <new_statgroup_name>]
[-description "<description_line>"]
[{<statpath_name>|<statgroup_name>}[,...]]}
| -add <statgroup_name>
{<statpath_name>|<statgroup_name>}[,...]
| -remove <statgroup_name>
{<statpath_name>|<statgroup_name>}[,...]
| -delete <statgroup_name> [-Force]
| -database
{ -recover [-Force]
|-verify }
DESCRIPTION
nas_stats allows the user to manage Statistics Groups. There are two
types of Statistics Groups: system-defined and user-defined groups.
System-defined statistics groups are created as part of the installation
(or upgrade) process and include the following statistics, which
correspond to the summary and table collections used by
server_stats:
System-defined statistics group name
basic-std
caches-std
cifs-std
nfs-std
cifsOps-std
diskVolumes-std
metaVolumes-std
netDevices-std
nfsOps-std
server_stats collection name
-summary basic
-summary caches
-summary cifs
-summary nfs
-table cifs
-table dvol
-table fsvol
-table net
-table nfs
nas_stats
395
The nas Commands
Note: server_stats collection names are deprecated and will not be supported
in future releases.
Statistics Groups can be created to include any combination of
statpath names, displayed through the server_stats command, as
well as other existing statgroup names.
Any Statistics Group name can be used with server_stats -monitor to
collect statistics as defined in its member_stats list.
OPTIONS
-list
Lists system and user-defined Statistics Groups.
-info
Provides detailed information on all (or specified) Statistics Groups.
-create
Creates a statistics group and specifies what statpath names it
includes. It also allows the nesting of statgroups by adding existing
statgroups to new statgroups.
Statgroup names can be used with the -info request. A statgroup
name is limited to 255 characters. Space, slash, back slash, quote,
double quote, and comma are the illegal characters in it.
[-description]
The -description option is optional and defaults to the statgroup
name. If the -description option is used, its argument must be
enclosed in quotation marks.
-modify
Allows you to modify a statgroup’s member_stats list by specifying
the new member statistics of the group, overriding the previous
contents.
-add
Allows you to add statpath and existing statgroup names to a
statgroup by specifying additional items to be appended to the
statgroup’s list member_stats list.
-remove
Allows you to remove member statpath and statgroup names from a
statgroup by specifying the items to remove from the statgroup’s
member_stats list.
396
Command Line Interface Reference for File 8.1
The nas Commands
-delete
Allows you to delete a statgroup. However, this option does not
delete any statsgroups that are members of the statgroup.
-recover
Attempts to recover the latest uncorrupted copy of the Statistics
Groups database from the NAS database backups. nas_stats searches
through the available backups and restores the latest copy. In this
event, NAS database backups do not contain a healthy version of the
Statistics Groups; a new Statistics Groups database is installed. In the
case of a new Statistics Groups database, all user-defined information
is lost. NAS database backups run hourly and VNX maintains the last
12 backups.
[-Force]
Use the -Force option with the -recover option to skip the
warning prompt.
-verify
Checks the health status of the Statistics Groups database.
SEE ALSO
EXAMPLE #1
server_stats
To list the system-defined and user-defined Statistics Groups, type:
$ nas_stats -groups -list
Type
Name
System
System
...
User
User
...
basic-std
basicCifs-std
...
basic
nfsNet
...
EXAMPLE #2
To provide detailed information on all (or specified) Statistics
Groups, type:
$ nas_stats -groups -info
name
= basic-std
description
= The basic system-defined group.
type
= System-defined
member_stats
=
kernel.cpu.utilization.cpuUtil,net.basic.inBytes,net.basic.outBytes,store.readBy
tes,store.writeBytes
member_elements =
member_of
=
nas_stats
397
The nas Commands
name
description
type
member_stats
member_elements
member_of
=
=
=
=
=
=
basic3
CPU and Memory
User-defined
kernel.cpu.utilization.cpuUtil,kernel.memory.freeBytes
name
= caches-std
description
= The caches system-defined group.
type
= System-defined
member_stats
=
fs.dnlc.hitRatio,fs.ofCache.hitRatio,kernel.memory.bufferCache.hitRatio
member_elements =
member_of
=
name
= cifs-std
description
= The cifs system-defined group.
type
= System-defined
member_stats
=
cifs.global.basic.totalCalls,cifs.global.basic.reads,cifs.global.basic.readBytes
,cifs.global.basic.readAvgSize,cifs.global.basic.writes,cifs.global.basic.writeB
ytes,cifs.global.basic.writeAvgSize,cifs.global.usage.currentConnections,cifs.gl
obal.usage.currentOpenFiles
member_elements =
member_of
= newSG
398
name
description
type
member_stats
member_elements
member_of
=
=
=
=
=
=
cifsOps-std
The cifs table system-defined group.
System-defined
cifs.smb1.op,cifs.smb2.op
name
description
type
member_stats
member_elements
=
=
=
=
=
diskVolumes-std
The disk volume table system-defined group.
System-defined
store.diskVolume
name
description
type
member_stats
member_elements
member_of
=
=
=
=
=
=
metaVolumes-std
The meta volume table system-defined group.
System-defined
store.logicalVolume.metaVolume
name
description
type
member_stats
member_elements
member_of
=
=
=
=
=
=
netDevices-std
The net table system-defined group.
System-defined
net.device
Command Line Interface Reference for File 8.1
The nas Commands
name
description
type
member_stats
member_elements
member_of
=
=
=
=
=
=
newSG
newSG
User-defined
cifs-std,nfs.v3.op,nfs.v4.op
name
= nfs-std
Description
= The nfs system-defined group.
type
= System-defined
member_stats
=
nfs.totalCalls,nfs.basic.reads,nfs.basic.readBytes,nfs.basic.readAvgSize,nfs.bas
ic.writes,nfs.basic.writeBytes,nfs.basic.writeAvgSize,nfs.currentThreads
member_elements =
member_of
=
name
description
type
member_stats
member_elements
member_of
=
=
=
=
=
=
nfsOps-std
The nfs table system-defined group.
System-defined
nfs.v2.op,nfs.v3.op,nfs.v4.op
name
= statgroup1
description
= My first group
type
= User-defined
member_stats
=
net.basic.inBytes,net.basic.outBytes,store.readBytes,store.writeBytes
member_elements =
member_of
= statgroup2
name
= statgroup2
description
= My first group
type
= User-defined
member_stats
=
net.basic.inBytes,net.basic.outBytes,store.readBytes,store.writeBytes,kernel.cpu
.utilization.cpuUtil,statgroup1
member_elements =
member_of
=
EXAMPLE #3
name
description
type
member_stats
member_elements
member_of
=
=
=
=
=
=
To provide detailed information on all (or specified) Statistics
Groups, type:
$ nas_stats -groups -info statsA
statsA
My group # 2
user-defined
statpath1, statpath2, statpath3, statsC
statsB
nas_stats
399
The nas Commands
EXAMPLE #4
To create a statistics group called basic3, type:
$ nas_stats -groups -create basic3 -description "CPU and
Memory" kernel.cpu.utilization.cpuUtil,kernel.memory.
freeBytes
'basic3' created successfully.
EXAMPLE #5
To create a statistics group called statgroup2, type:
$ nas_stats -groups -create statgroup2 statgroup1,nfs,net
’statgroup2’ created successfully.
EXAMPLE #6
To use an existing statgroup, type:
$ nas_stats -groups -create statgroup1 -description "My
first group" kernel.cpu.utilization.cpuUtil,
net.basic.inBytes,net.basic.outBytes,store.readBytes,
store.writeBytes
ERROR (13421969439): ’statgroup1’ already exists.
EXAMPLE #7
To modify a statgroup by specifying the new contents of the group,
overriding the previous contents, type:
$ nas_stats -groups -modify statgroup2 cifs,nfs-std
’statgroup2’ modified successfully.
EXAMPLE #8
To modify the description of a statgroup, type:
$ nas_stats -groups -modify basic1 -description "My basic
group"
’basic1’ modified successfully.
EXAMPLE #9
To rename a user-defined statgroup, type:
$ nas_stats -groups -modify statgroup2 -rename basic2
’statgroup2’ modified successfully.
EXAMPLE #10
To add to the member_stats list of a statgroup, type:
$ nas_stats -groups -add statgroup2
kernel.cpu.utilization.cpuUtil,statgroup1
Adding the following statistics:
... kernel.cpu.utilization.cpuUtil
... statgroup1
Statistics added to ’statgroup2’ successfully.
EXAMPLE #11
To remove from the member_stats list of a statgroup, type:
$ nas_stats -groups -remove statgroup1
kernel.cpu.utilization.cpuUtil
Removing the following statistics:
... kernel.cpu.utilization.cpuUtil
Statistics removed from ’statgroup1’ successfully.
400
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #12
To delete a statgroup, type:
$ nas_stats -groups -delete statgroup1
’statgroup1’ deleted successfully.
EXAMPLE #13
To delete reference from other groups using statgroupA, type:
$ nas_stats -groups -delete statgroupA
’statgroupA’ is used in group (s): mystats1, mystats2.
Clear ’statgroupA’ from other groups? [Y/N] Y
’statgroupA’ deleted successfully.
EXAMPLE #14
To delete reference from other groups using statgroupA and the
-Force option to skip the warning prompt, type:
$ nas_stats -groups -delete statgroupA -F
’statgroupA’ is used in group (s): mystats1, mystats2.
’statgroupA’ deleted successfully.
EXAMPLE #15
To recover the latest healthy (uncorrupted) copy of a statgroup
database from the NAS database backups, type:
$ nas_stats -groups -database -recover
Latest healthy database modified last on Tue Apr 7 17:29:06 EDT 2009.
Any updates performed after the latest backup will be lost. Continue? [Y/N] Y
The nas_stats command recover operation is completed successfully.
EXAMPLE #16
To recover the latest healthy (uncorrupted) copy of the statgroup
database from the NAS database backups using the -Force option to
skip the warning prompt, type:
$ nas_stats -groups -database -recover -Force
Latest healthy database modified last on Tue Apr 7 17:29:06 EDT 2009.
The nas_stats command recover operation is completed successfully.
EXAMPLE #17
To check the health status of the Statistics Groups database, type:
$ nas_stats -groups -database -verify
Database is healthy.
nas_stats
401
The nas Commands
nas_storage
Controls system access and performs some management tasks.
SYNOPSIS
nas_storage
-list
| -info {-all|<name>|id=<storage_id>} [-option <options>]
| -rename <old_name> <new_name>
| -acl <acl_value> <name>
| -delete {<name>|id=<storage_id>} [-spare <spindle-id>|-group
<diskgroup-id>]
| -failback {<name>|id=<storage_id>}
| -sync {-all|<name>|id=<storage_id>}
| -check {-all|<name>|id=<storage_id>}
| -modify {<name>|id=<storage_id>} -network
{-spa|-spb} <IP>
| -modify {<name>|id=<storage_id>}
-security [-username <username>][-password <password>]
[-newpassword <new_password>]]
Note: Output from this command is determined by the type of system
attached to the VNX.
DESCRIPTION
OPTIONS
nas_storage sets the name for a system, assigns an access control
value, displays attributes, performs a health check, synchronizes the
system with the Control Station, and performs a failback for VNX for
block systems.
-list
Displays a list of all attached systems available for the VNX.
Note: The ID of the object is an integer and is assigned automatically. The
name of the system may be truncated if it is too long for the display. To
display the full name, use the -info option with the system ID.
-info {-all|<name>|id=<storage_id>}
Displays the configuration of the attached system.
[-option <options>]
Specifies a comma-separated list of options.
sync={yes|no}
Synchronizes the Control Station’s view with that of the system
before displaying configuration information. Default=yes.
402
Command Line Interface Reference for File 8.1
The nas Commands
-rename <old_name> <new_name>
Renames the current system name to a new name. By default, the
system name is its serial number.
-acl <acl_value> <name>
Sets an access control level value that defines the owner of the
system, and the level of access allowed for users and groups defined
in the access control level table (nas_acl provides information).
-delete {<name>|id=<storage_id> [-spare
<spindle-id>|-group <diskgroup-id>]
Deletes an entry from the system table. The system can only be
deleted after all disks on the system have been deleted using
nas_disk. The system and disks can be rediscovered using the
server_devconfig command. The -spare option deletes the hot spare
disk from the hot spare pool on the VNX for block storage used by
NAS. The -group option deletes the disk group specified. This deletes
and unbinds the LUNs in the RAID groups used by VNX for file. If
there are other LUNs in the RAID group not allocated to the VNX, the
RAID group is not unbound. After removing the VNX LUNs, the
RAID group is empty and it will be destroyed.
-sync {-all|<name>|id=<storage_id>}
Synchronizes the Control Station's view with that of the system.
-check {-all|<name>|id=<storage_id>}
Performs a health check on the system to verify if it is configured for,
and in a state to provide the level of high availability that is required.
Use this option after making any management changes to your
system (for example, changes to VNX for block array properties, such
as enabling/disabling statistics polling).
Note: This option does not support remote storage. For example, for
recoverpoint configurations where remote storage is listed, the check will
only run on first listed system.
For VNX for Block only
-failback {<name>|id=<storage_id>}
Returns the system’s normal operating state by returning ownership
of all disk volumes to their default storage processor.
To verify that the system failed over, type the -info option. If the
value appears as failed_over=True, then the system has failed over.
nas_storage
403
The nas Commands
-modify {<name>|id=<storage_id>} -network {-spa|-spb}
<IP>
Modifies the IP address of the VNX for block in the VNX database.
-modify {<name>|id=<storage_id>} -security [-username
<username>][-password <password>]
Updates the login information the VNX for file uses to authenticate
with the VNX and changes the VNX username, or password if the
VNX account is changed or the following error is reported:
Error 5010: APM00055105668: Storage API code=4651:
SYMAPI_C_CLAR_NOT_PRIVILEGED
Operation denied by Clariion array - you are not
privileged to perform the requested operation
[-newpassword <new_password]
Assigns a new password to the username on the VNX for block.
Note: This operation is not supported for Symmetrix systems.
SEE ALSO
SYSTEM OUTPUT
EXAMPLE #1
id
1
acl
0
VNX System Operations, nas_rdf, nas_disk, and server_devconfig.
The number associated with the storage device is dependent on the
attached system. VNX for block displays a prefix of APM before a set
of integers, for example, APM00033900124-0019. For example,
Symmetrix systems appear as 002804000190-003C.
For the VNX system, to list all attached systems, type:
$ nas_storage -list
name
serial_number
APM00042000818
APM00042000818
For the VNX with a Symmetrix system, to list all attached systems,
type:
id
1
$ nas_storage -list
name
serial_number
000187940260
000187940260
acl
0
Where:
Value
id
acl
name
serial_number
404
Definition
ID number of the attached system.
Access control level value assigned to the attached system.
Name assigned to the attached system.
Serial number of the attached system.
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #2
id
arrayname
name
type
model_type
model_num
db_sync_time
API_version
num_disks
num_devs
num_pdevs
num_storage_grps
num_raid_grps
cache_page_size
wr_cache_mirror
low_watermark
high_watermark
unassigned_cache
is_local
failed_over
captive_storage
For the VNX system, to display information for the attached system,
type:
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
$ nas_storage -info APM00042000818
1
APM00042000818
APM00042000818
Clariion
RACKMOUNT
700
1131986667 == Mon Nov 14 11:44:27 EST 2005
V6.0-629
60
34
8
1
16
8
True
60
80
0
True
False
False
Active Software
-AccessLogix
= FLARE-Operating-Environment= 02.16.700.5.004
-NavisphereManager
= -
Storage Processors
SP Identifier
signature
microcode_version
serial_num
prom_rev
agent_rev
phys_memory
sys_buffer
read_cache
write_cache
free_memory
raid3_mem_size
failed_over
hidden
network_name
ip_address
subnet_mask
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
A
1057303
2.16.700.5.004
LKE00040201171
3.30.00
6.16.0 (4.80)
3967
773
122
3072
0
0
False
False
spa
172.24.102.5
255.255.255.0
nas_storage
405
The nas Commands
gateway_address
= 172.24.102.254
num_disk_volumes
= 20 - root_disk root_ldisk d3 d4 d5 d6 d7 d8 d9 d10 d11 d12
d13 d14 d15 d16 d17 d18 d19 d20
Port Information
Port 1
uid
link_status
port_status
switch_present
switch_uid
sp_source_id
=
=
=
=
=
=
50:6:1:60:B0:60:1:CC:50:6:1:61:30:60:1:CC
UP
ONLINE
True
10:0:8:0:88:A0:36:F3:20:42:8:0:88:A0:36:F3
6373907
=
=
=
=
=
=
50:6:1:60:B0:60:1:CC:50:6:1:62:30:60:1:CC
UP
ONLINE
True
10:0:8:0:88:A0:36:F3:20:41:8:0:88:A0:36:F3
6373651
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
B
1118484
2.16.700.5.004
LKE00041700812
3.30.00
6.16.0 (4.80)
3967
773
122
3072
0
0
False
False
spb
172.24.102.6
255.255.255.0
172.24.102.254
0
<...removed...>
Port 2
uid
link_status
port_status
switch_present
switch_uid
sp_source_id
SP Identifier
signature
microcode_version
serial_num
prom_rev
agent_rev
phys_memory
sys_buffer
read_cache
write_cache
free_memory
raid3_mem_size
failed_over
hidden
network_name
ip_address
subnet_mask
gateway_address
num_disk_volumes
Port Information
Port 1
uid
link_status
406
= 50:6:1:60:B0:60:1:CC:50:6:1:69:30:60:1:CC
= UP
Command Line Interface Reference for File 8.1
The nas Commands
port_status
switch_present
switch_uid
sp_source_id
=
=
=
=
ONLINE
True
10:0:8:0:88:A0:36:F3:20:3E:8:0:88:A0:36:F3
6372883
=
=
=
=
=
=
50:6:1:60:B0:60:1:CC:50:6:1:6A:30:60:1:CC
UP
ONLINE
True
10:0:8:0:88:A0:36:F3:20:3D:8:0:88:A0:36:F3
6372627
<...removed...>
Port 2
uid
link_status
port_status
switch_present
switch_uid
sp_source_id
Storage Groups
id
name
num_hbas
num_devices
shareable
hidden
=
=
=
=
=
=
A4:74:8D:50:6E:A1:D9:11:96:E1:8:0:1B:43:5E:4F
ns704g-cs100
18
24
True
False
uid
storage_processor
port
server
=
=
=
=
50:6:1:60:90:60:3:49:50:6:1:60:10:60:3:49
B
1
server_4
uid
storage_processor
port
server
=
=
=
=
50:6:1:60:90:60:3:49:50:6:1:60:10:60:3:49
A
0
server_4
uid
storage_processor
port
server
=
=
=
=
50:6:1:60:80:60:4:F0:50:6:1:61:0:60:4:F0
B
0
server_2
=
=
=
=
50:6:1:60:80:60:4:F0:50:6:1:68:0:60:4:F0
B
1
server_3
Hosts
<...removed...>
uid
storage_processor
port
server
uid
= 20:0:0:0:C9:2B:98:77:10:0:0:0:C9:2B:98:77
storage_processor = B
port
= 0
nas_storage
407
The nas Commands
uid
= 20:0:0:0:C9:2B:98:77:10:0:0:0:C9:2B:98:77
storage_processor = A
port
= 0
ALU
HLU
-----------0000 -> 0000
0001 -> 0001
0002 -> 0002
0003 -> 0003
0004 -> 0004
0005 -> 0005
0018 -> 0018
0019 -> 0019
0020 -> 0020
0021 -> 0021
0022 -> 0022
0023 -> 0023
0024 -> 0024
0025 -> 0025
0026 -> 0026
0027 -> 0027
0028 -> 0028
0029 -> 0029
0030 -> 0030
0031 -> 0031
0032 -> 0032
0033 -> 0033
0034 -> 0034
0035 -> 0035
Disk Groups
id
storage profiles
raid_type
logical_capacity
num_spindles
num_luns
num_disk_volumes
spindle_type
bus
raw_capacity
used_capacity
free_capacity
hidden
=
=
=
=
=
=
=
=
=
=
=
=
=
0000
2 - clar_r5_performance,cm_r5_performance
RAID5
1068997528
5 - 0_0_0 0_0_1 0_0_2 0_0_3 0_0_4
6 - 0000 0001 0002 0003 0004 0005
6 - root_disk root_ldisk d3 d4 d5 d6
FC
0
1336246910
62914560
1006082968
False
<...removed...>
id
408
= 2_0_14
Command Line Interface Reference for File 8.1
The nas Commands
product
revision
serial
capacity
used_capacity
disk_group
hidden
type
bus
enclosure
slot
vendor
remapped_blocks
state
=
=
=
=
=
=
=
=
=
=
=
=
=
=
ST314670 CLAR146
6A06
3KS02RHM
280346624
224222822
0014
False
FC
2
0
14
SEAGATE
-1
ENABLED
For the VNX with a Symmetrix system, to display information for the
attached system, type:
id
serial_number
name
type
ident
model
microcode_version
microcode_version_num
microcode_date
microcode_patch_level
microcode_patch_date
symmetrix_pwron_time
db_sync_time
db_sync_bcv_time
db_sync_rdf_time
last_ipl_time
last_fast_ipl_time
API_version
cache_size
cache_slot_count
max_wr_pend_slots
max_da_wr_pend_slots
max_dev_wr_pend_slots
permacache_slot_count
num_disks
num_symdevs
num_pdevs
sddf_configuration
config_checksum
num_powerpath_devs
config_crc
is_local
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
$ nas_storage -info 000187940260
1
000187940260
000187940260
Symmetrix
Symm6
800-M2
5670
16260000
03012004
69
03012004
1130260200 == Tue Oct 25 13:10:00
1133215405 == Mon Nov 28 17:03:25
1133215405 == Mon Nov 28 17:03:25
1133215405 == Mon Nov 28 17:03:25
1128707062 == Fri Oct 7 13:44:22
1130260200 == Tue Oct 25 13:10:00
V6.0-629
32768
860268
180000
90000
6513
0
60
378
10
ENABLED
0x01ca544
0
0x07e0ba1e6
True
EDT
EST
EST
EST
EDT
EDT
2005
2005
2005
2005
2005
2005
nas_storage
409
The nas Commands
Physical Devices
/nas/dev/c0t0l15s2
/nas/dev/c0t0l15s3
/nas/dev/c0t0l15s4
/nas/dev/c0t0l15s6
/nas/dev/c0t0l15s7
/nas/dev/c0t0l15s8
/nas/dev/c16t0l15s2
/nas/dev/c16t0l15s3
/nas/dev/c16t0l15s4
/nas/dev/c16t0l15s8
type
DA
DA
DA
DA
DA
DA
DA
DA
FA
FA
FA
FA
FA
num
1
2
15
16
17
18
31
32
33
34
47
48
49
slot
1
2
15
16
1
2
15
16
1
2
15
16
1
ident
DF-1A
DF-2A
DF-15A
DF-16A
DF-1B
DF-2B
DF-15B
DF-16B
FA-1C
FA-2C
FA-15C
FA-16C
FA-1D
stat
On
On
On
On
On
On
On
On
On
On
On
On
On
Director Table
vols ports p0_stat
21
2
On
8
2
On
21
2
On
8
2
On
8
2
On
21
2
On
152 2
On
165 2
On
0
2
On
0
2
On
0
2
On
0
2
On
0
2
On
scsi
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
p1_stat
On
On
On
On
On
On
On
On
On
On
On
On
On
p2_stat
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
p3_stat
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
NA
Note: This is a partial listing due to the length of the outputs.
EXAMPLE #3
To rename a system, type:
$ nas_storage -rename APM00042000818 cx700_1
id
= 1
serial_number = APM00042000818
name
= cx700_1
acl
= 0
EXAMPLE #4
To set the access control level for the system cx700_1, type:
$ nas_storage -acl 1000 cx700_1
id
= 1
serial_number = APM00042000818
name
= cx700_1
acl
= 1000, owner=nasadmin, ID=201
Note: The value 1000 specifies nasadmin as the owner and gives read, write,
and delete access only to nasadmin.
410
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #5
To change the existing password on the VNX for block, type:
$ nas_storage -modify APM00070204288 -security -username
nasadmin -password nasadmin -newpassword abc
Changing password on APM00070204288
EXAMPLE #6
To avoid specifying passwords in clear text on the command line,
type:
$ nas_storage -modify APM00070204288 -security
-newpassword
Enter the Global CLARiiON account information
Username: nasadmin
Password: *** Retype your response to validate
Password: ***
New Password
Password: ******** Retype your response to validate
Password: ********
Changing password on APM00070204288
Done
EXAMPLE #7
To failback a VNX for block, type:
$ nas_storage -failback cx700_1
id
= 1
serial_number = APM00042000818
name
= cx700_1
acl
= 1000, owner=nasadmin, ID=201
EXAMPLE #8
id
arrayname
name
type
model_type
model_num
db_sync_time
API_version
num_disks
num_devs
num_pdevs
num_storage_grps
num_raid_grps
cache_page_size
wr_cache_mirror
low_watermark
high_watermark
unassigned_cache
is_local
To display information for a VNX for block and turn synchronization
off, type:
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
$ nas_storage -info cx700_1 -option sync=no
1
APM00042000818
cx700_1
Clariion
RACKMOUNT
700
1131986667 == Mon Nov 14 11:44:27 EST 2005
V6.0-629
60
34
8
1
16
8
True
60
80
0
True
nas_storage
411
The nas Commands
failed_over
captive_storage
= False
= False
Active Software
-AccessLogix
= FLARE-Operating-Environment= 02.16.700.5.004
-NavisphereManager
= -
Storage Processors
SP Identifier
=
signature
=
microcode_version
=
serial_num
=
prom_rev
=
agent_rev
=
phys_memory
=
sys_buffer
=
read_cache
=
write_cache
=
free_memory
=
raid3_mem_size
=
failed_over
=
hidden
=
network_name
=
ip_address
=
subnet_mask
=
gateway_address
=
num_disk_volumes
=
d11 d12 d13 d14 d15 d16
A
1057303
2.16.700.5.004
LKE00040201171
3.30.00
6.16.0 (4.80)
3967
773
122
3072
0
0
False
False
spa
172.24.102.5
255.255.255.0
172.24.102.254
20 - root_disk root_ldisk d3 d4 d5 d6 d7 d8 d9 d10
d17 d18 d19 d20
Port Information
Port 1
uid
link_status
port_status
switch_present
switch_uid
sp_source_id
=
=
=
=
=
=
50:6:1:60:B0:60:1:CC:50:6:1:61:30:60:1:CC
UP
ONLINE
True
10:0:8:0:88:A0:36:F3:20:42:8:0:88:A0:36:F3
6373907
=
=
=
=
50:6:1:60:B0:60:1:CC:50:6:1:62:30:60:1:CC
UP
ONLINE
True
<...removed...>
Port 2
uid
link_status
port_status
switch_present
412
Command Line Interface Reference for File 8.1
The nas Commands
switch_uid
sp_source_id
SP Identifier
signature
microcode_version
serial_num
prom_rev
agent_rev
phys_memory
sys_buffer
read_cache
write_cache
free_memory
raid3_mem_size
failed_over
hidden
network_name
ip_address
subnet_mask
gateway_address
num_disk_volumes
= 10:0:8:0:88:A0:36:F3:20:41:8:0:88:A0:36:F3
= 6373651
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
B
1118484
2.16.700.5.004
LKE00041700812
3.30.00
6.16.0 (4.80)
3967
773
122
3072
0
0
False
False
spb
172.24.102.6
255.255.255.0
172.24.102.254
0
=
=
=
=
=
=
50:6:1:60:B0:60:1:CC:50:6:1:69:30:60:1:CC
UP
ONLINE
True
10:0:8:0:88:A0:36:F3:20:3E:8:0:88:A0:36:F3
6372883
=
=
=
=
=
=
50:6:1:60:B0:60:1:CC:50:6:1:6A:30:60:1:CC
UP
ONLINE
True
10:0:8:0:88:A0:36:F3:20:3D:8:0:88:A0:36:F3
6372627
Port Information
Port 1
uid
link_status
port_status
switch_present
switch_uid
sp_source_id
<...removed...>
Port 2
uid
link_status
port_status
switch_present
switch_uid
sp_source_id
Storage Groups
id
name
num_hbas
num_devices
shareable
=
=
=
=
=
A4:74:8D:50:6E:A1:D9:11:96:E1:8:0:1B:43:5E:4F
ns704g-cs100
18
24
True
nas_storage
413
The nas Commands
hidden
= False
Hosts
uid
storage_processor
port
server
=
=
=
=
50:6:1:60:90:60:3:49:50:6:1:60:10:60:3:49
B
1
server_4
uid
storage_processor
port
server
=
=
=
=
50:6:1:60:90:60:3:49:50:6:1:60:10:60:3:49
A
0
server_4
uid
storage_processor
port
server
=
=
=
=
50:6:1:60:80:60:4:F0:50:6:1:61:0:60:4:F0
B
0
server_2
=
=
=
=
50:6:1:60:80:60:4:F0:50:6:1:68:0:60:4:F0
B
1
server_3
<...removed...>
uid
storage_processor
port
server
uid
= 20:0:0:0:C9:2B:98:77:10:0:0:0:C9:2B:98:77
storage_processor = B
port
= 0
uid
= 20:0:0:0:C9:2B:98:77:10:0:0:0:C9:2B:98:77
storage_processor = A
port
= 0
ALU
HLU
-----------0000 -> 0000
0001 -> 0001
0002 -> 0002
0003 -> 0003
0004 -> 0004
0005 -> 0005
0018 -> 0018
0019 -> 0019
0020 -> 0020
0021 -> 0021
0022 -> 0022
0023 -> 0023
0024 -> 0024
0025 -> 0025
0026 -> 0026
0027 -> 0027
414
Command Line Interface Reference for File 8.1
The nas Commands
0028
0029
0030
0031
0032
0033
0034
0035
->
->
->
->
->
->
->
->
0028
0029
0030
0031
0032
0033
0034
0035
Disk Groups
id
storage profiles
raid_type
logical_capacity
num_spindles
num_luns
num_disk_volumes
spindle_type
bus
raw_capacity
used_capacity
free_capacity
hidden
=
=
=
=
=
=
=
=
=
=
=
=
=
0000
2 - clar_r5_performance,cm_r5_performance
RAID5
1068997528
5 - 0_0_0 0_0_1 0_0_2 0_0_3 0_0_4
6 - 0000 0001 0002 0003 0004 0005
6 - root_disk root_ldisk d3 d4 d5 d6
FC
0
1336246910
62914560
1006082968
False
=
=
=
=
=
=
=
=
=
=
=
=
=
0205
0
SPARE
622868992
1 - 0_1_0
1 - 0205
0
ATA
0
622868992
622868992
0
False
=
=
=
=
=
=
=
=
0_0_0
ST314670 CLAR146
6A06
3KS088SQ
280346624
12582912
0000
False
<...removed...>
id
storage profiles
raid_type
logical_capacity
num_spindles
num_luns
num_disk_volumes
spindle_type
bus
raw_capacity
used_capacity
free_capacity
hidden
Spindles
id
product
revision
serial
capacity
used_capacity
disk_group
hidden
nas_storage
415
The nas Commands
type
bus
enclosure
slot
vendor
remapped_blocks
state
=
=
=
=
=
=
=
FC
0
0
0
SEAGATE
-1
ENABLED
=
=
=
=
=
=
=
=
=
=
=
=
=
=
=
2_0_14
ST314670 CLAR146
6A06
3KS02RHM
280346624
224222822
0014
False
FC
2
0
14
SEAGATE
-1
ENABLED
<...removed...>
id
product
revision
serial
capacity
used_capacity
disk_group
hidden
type
bus
enclosure
slot
vendor
remapped_blocks
state
Note: This is a partial display due to the length of the outputs.
EXAMPLE #9
To delete a system with no attached disks, type:
$ nas_storage -delete APM00035101740
id = 0
serial_number = APM00035101740
name = APM00035101740
acl = 0
EXAMPLE #10
To turn synchronization on for all systems, type:
$ nas_storage -sync -all
done
416
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #11
To perform a health check on the system, type:
$ nas_storage -check -all
Discovering storage (may take several minutes)
done
EXAMPLE #12
To set the access control level for the system APM00042000818, type:
$ nas_storage -acl 1432 APM00042000818
id
= 1
serial_number = APM00042000818
name
= APM00042000818
acl
= 1432, owner=nasadmin, ID=201
Note: The value 1432 specifies nasadmin as the owner and gives users with
an access level of at least observer read access only, users with an access level
of at least operator read/write access, and users with an access level of at
least admin read/write/delete access.
EXAMPLE #13
To modify the IP address of the VNX for block, type:
$ nas_storage -modify APM00072303347 -network -spa
10.6.4.225
Changing IP address for APM00072303347
Discovering storage (may take several minutes)
done
nas_storage
417
The nas Commands
nas_task
Manages in-progress or completed tasks.
SYNOPSIS
nas_task
-list [-remote_system {<remoteSystemName>|id=<id>}]
| -info {-all|<taskId>}
[-remote_system {<remoteSystemName>|id=<id>}]
| -abort <taskId>
[-mover <moverName>][-remote_system {<remoteSystemName>|id=<id>}]
| -delete <taskId>
-remote_system {<remoteSystemName>|id=<id>}]
DESCRIPTION
nas_task lists the tasks associated with commands currently in
progress or completed, reports information about a particular task,
aborts a task, or deletes a task. Each task can be uniquely identified
by its task ID and the remote VNX system name or ID.
Use the nas_task command to monitor, abort, and delete long
running tasks and tasks started in asynchronous mode.
OPTIONS
-list
Lists all local tasks that are in progress, or completed tasks that have
not been deleted. For each task, lists the task ID, remote system name,
a description of the task, and the task state (running, recovering,
succeeded, or failed).
-remote_system {<remoteSystemName>|id=<id>}
Lists local tasks initiated by the specified remote VNX system.
Specify the remote system name or ID.
-info {-all|<taskId>}
Provides more detailed status information for all tasks or for a
particular task. Displays the run time status, estimated completion
time, and percent complete for running tasks. Displays the
completion status and actual end time for completed tasks.
The taskID is the ID returned from a command run in the background
mode or from the nas_task -list command.
Note: The ID of the task is an integer and is assigned automatically. The task
ID is unique to the VNX.
418
Command Line Interface Reference for File 8.1
The nas Commands
[-remote_system {<remoteSystemName>|id=<id>}]
Provides more detailed status information of local tasks initiated
by the specified remote VNX system. Specify the remote system
name or remote system ID. The remote system name is returned
from the nas_task -list command.
-abort <taskId>
Aborts the specified task leaving the system in a consistent state. For
example, it aborts a one-time copy in progress. This might take a long
time to complete because a remote system may be unavailable or the
network may be down. You should check the status of the task to
verify that the task was aborted. This command can be executed from
the source only.
!
CAUTION
This option might leave the system in an inconsistent state. Use
caution when using this option.
[-mover <moverName>]
Aborts a task running locally on the specified Data Mover. This
command can be executed from the source or destination side.
Use this command when the source and destination VNX systems
cannot communicate. You should run this command on both
sides.
[-remote_system {<remoteSystemName>|id=<id>]
Aborts a task that was initiated on a remote VNX leaving the
source side intact. Specify the Data Mover to abort a task from the
destination side. Specify the Data Mover and remote system
name or remote system id along with the task id.
-delete id <taskId>
Based on the task ID, deletes a completed task from the database on
the Control Station.
[-remote_system {<remoteSystemName>|id=<id>]
Deletes a task that was initiated on a remote VNX. Specify the
remote system name or remote system id along with the task id.
SEE ALSO
Using VNX Replicator, nas_copy, nas_replicate, and nas_cel.
nas_task
419
The nas Commands
EXAMPLE #1
Task Id
Celerra Network Server
Task State
Percent Complete
Description
Originator
Start Time
Estimated End Time
Schedule
To display detailed information about the task with task ID 4241,
type:
$
=
=
=
=
=
=
=
=
=
nas_task -info 4241
4241
cs100
Running
95
Create Replication ufs1_replica1.
nasadmin@cli.localhost
Mon Dec 17 14:21:35 EST 2007
Mon Dec 17 19:24:21 EST 2007
n/a
Where:
Value
Task Id
VNX
Remote Task Id
State
Current Activity
Percent Completed
Description
Originator
Start Time/End Time
Estimated End Time
Schedule
Response Statuses
EXAMPLE #2
Definition
Globally unique character string used as the identifier of the task.
When set, local.
When set, identifies a remote task.
Running, Recovering, Completed, or Failed. Running could be a combination of completed
and failed.
Displays state property when available.
Appears only when set and not complete.
Appears if details are set.
User or host that initiated the task.
The starting time and ending time (or status) for the task.
Appears instead of previous line when available and task is incomplete.
The schedule in effect, or n/a for a task that is not a scheduled checkpoint refresh.
Displayed list of messages, if any. A completed task should always have one.
To display the list of all tasks, type:
$ nas_task -list
ID
Task State
Originator
Start
Time
escription Schedule
Remote System
4241
Running nasadmin@cli+ Mon Dec 17 14:21:35 EST 2007 Create Replication
ufs1_r+ cs100
4228 Succeeded nasadmin@cli+ Mon Dec 17 14:04:02 EST 2007 Delete task NONE: 4214.
cs100
4177
Failed nasadmin@cli+ Mon Dec 17 13:59:26 EST 2007 Create Replication
ufs1_r+ cs100
4150 Succeeded nasadmin@cli+ Mon Dec 17 13:55:39 EST 2007 Delete task NONE: 4136.
cs100
4127 Succeeded nasadmin@cli+ Mon Dec 17 11:38:32 EST 2007 Delete task NONE: 4113.
cs100
4103 Succeeded nasadmin@cli+ Mon Dec 17 11:21:00 EST 2007 Delete task NONE: 4098.
cs100
420
Command Line Interface Reference for File 8.1
The nas Commands
4058 Succeeded nasadmin@cli+ Fri Dec 14 16:43:23 EST 2007 Switchover Replication
NONE. cs100
2277 Succeeded nasadmin@cli+ Fri Dec 14 16:42:08 EST 2007 Reverse Replication
NONE.
cs110
2270 Succeeded nasadmin@cli+ Fri Dec 14 16:40:29 EST 2007 Start Replication NONE.
cs110
2265
Failed nasadmin@cli+ Fri Dec 14 16:40:11 EST 2007 Start Replication NONE.
cs110
EXAMPLE #1 provides a description of the outputs.
EXAMPLE #3
To abort task 4267 running locally on server_3, type:
$ nas_task -abort 4267 -mover server_3
OK
EXAMPLE #4
To delete the existing task 4267, type:
$ nas_task -delete 4267
OK
nas_task
421
The nas Commands
nas_version
Displays the software version running on the Control Station.
SYNOPSIS
nas_version
[-h|-l]
DESCRIPTION
OPTIONS
nas_version displays the Control Station version in long form or
short form. When used during a software upgrade, informs the user
about the upgrade in progress.
No arguments
Displays the software version running on the Control Station.
-h
Displays command usage.
-l
Displays detailed software version information for the Control
Station.
EXAMPLE #1
To display the software version running on the Control Station
during a software upgrade, type:
$ nas_version
5.6.25-0
EXAMPLE #2
To display the system output during a software upgrade, type:
$ nas_version
5.6.19-0
Warning!!Upgrade is in progress from 5.6.19-0 to 5.6.20-0
Warning!!Please log off IMMEDIATELY if you are not upgrading the Celerra
EXAMPLE #3
To display the usage for nas_version, type:
$ nas_version -h
usage: /nas/bin/nas_version [-h|-l]
-h
help
-l
long_format
EXAMPLE #4
Name
Version
Release
Size
422
:
:
:
:
emcnas
5.6.19
0
454239545
To display detailed software version information for the Control
Station, type:
$ nas_version -l
Relocations: /nas
Vendor: EMC
Build Date: Tue 19 Dec 2006 08:53:31 PM EST
License: EMC Copyright
Command Line Interface Reference for File 8.1
The nas Commands
Signature
Packager
URL
Summary
Description
:
:
:
:
:
(none)
EMC Corporation
http://www.emc.com
EMC nfs base install
EMC nfs base install
EXAMPLE #5
To display detailed software version information for the Control
Station during a software upgrade, type:
$ nas_version -l
Name
: emcnas
Relocations: /nas
Version
: 5.6.19
Vendor: EMC
Release
: 0
Build Date: Wed 14 Mar 2007 12:36:55 PM EDT
Size
: 500815102
License: EMC Copyright
Signature
: (none)
Packager
: EMC Corporation
URL
: http://www.emc.com
Summary
: EMC nfs base install
Description : EMC nfs base install
Warning!!Upgrade is in progress from 5.6.19-0 to 5.6.20-0
Warning!!Please log off IMMEDIATELY if you are not upgrading the Celerra
nas_version
423
The nas Commands
nas_volume
Manages the volume table.
SYNOPSIS
nas_volume
-list
| -delete <volume_name>
| -info [-size] {-all|<volume_name>} [-tree]
| -rename <old_name> <new_name>
| -size <volume_name>
| -acl <acl_value> <volume_name>
| -xtend <volume_name> {<volume_name>,...}
| [-name <name>] -create [-Stripe [<stripe_size>]|-Meta]
[-Force] {<volume_name>,...}
| -Clone <volume_name> [{<svol>:<dvol>,...}][-option <options>]
DESCRIPTION
OPTIONS
nas_volume creates metavolumes and stripe volumes and lists,
renames, extends, clones, and deletes metavolumes, stripe, and slice
volumes. nas_volume sets an access control value for a volume, and
displays detailed volume attributes, including the total size of the
volume configuration.
-list
Displays the volume table.
Note: The ID of the object is an integer and is assigned automatically. The
name of the volume may be truncated if it is more than 17 characters. To
display the full name, use the -info option with the volume ID.
-delete <volume_name>
Deletes the specified volume.
-info [-size] {-all|<volume_name>} [-tree]
Displays attributes and the size for all volumes, or the specified
<volume_name>. The -tree option recursively displays the volume
set, that is, the list of component volumes for the specified volume or
all volumes.
-rename <old_name> <new_name>
Changes the current name of a volume to a new name.
-size <volume_name>
Displays the total size in MB of the <volume_name>, including used
and available space.
424
Command Line Interface Reference for File 8.1
The nas Commands
-acl <acl_value> <volume_name>
Sets an access control level value that defines the owner of the
volume, and the level of access allowed for users and groups defined
in the access control level table. The nas_acl command provides
information.
-xtend <volume_name> {<volume_name>,...}
Extends the specified metavolume by adding volumes to the
configuration. The total size of the metavolume increases by the sum
of all the volumes added.
Note: Only metavolumes can be extended. The volume that was added
remains in use until the original metavolume is deleted. Volumes containing
mounted file systems cannot be extended using this option. The nas_fs
command provides information to extend a volume that is hosting a
mounted file system.
-create {<volume_name>,...}
Creates a volume configuration from the specified volumes. Unless
otherwise specified, volumes are automatically created as
metavolumes.
[-name <name>]
Assigns a <name> to volume. If a name is not specified, one is
assigned automatically. The name of a volume is case-sensitive.
[-Stripe <stripe_size>|-Meta]
Sets the type for the volume to be either a stripe volume or
metavolume (default). If -Stripe is specified, type a stripe size in
multiples of 8192 bytes with a recommended size of 262,144 bytes
(256 KB) for all environments and drive types. If a stripe size is
not specified, the system creates a 256 KB stripe by default.
nas_slice provides information to create a slice volume.
[-Force] {<volume_name>,...}
Forces the creation of a volume on a mixed system.
-Clone <volume_name>
Creates an exact clone of the specified <volume_name>. Volumes can
be cloned from slice, stripe, or metavolumes. The name automatically
assigned to the clone is derived from the ID of the volume.
[{<svol>:<dvol>,...}]
Sets a specific disk volume set for the source volume and the
destination volume. The size of the destination volume must be
the same as the source volume.
nas_volume
425
The nas Commands
-option disktype=<type>
Specifies the type of disk to be created.
Disk types when using VNX for block are CLSTD, CLEFD, and
CLATA, and for VNX for block involving mirrored disks are
CMEFD, CMSTD, and CMATA.
Disk types when using a Symmetrix are STD, R1STD, R2STD,
BCV, R1BCV, R2BCV, ATA, R1ATA, R2ATA, BCVA, R1BCA,
R2BCA, and EFD.
SEE ALSO
id
1
Managing Volumes and File Systems with VNX Automatic Volume
Management, Managing Volumes and File Systems for VNX Manually,
Using TimeFinder/FS, NearCopy, and FarCopy on VNX for File,
Controlling Access to System Objects on VNX, nas_slice, nas_disk,
nas_acl, and nas_fs.
EXAMPLE #1
To list all volumes, type:
inuse type acl
y
4
0
$ nas_volume -list
name cltype
root_disk
0
2
y
4
0
root_ldisk
0
3
4
5
6
7
8
9
10
11
12
13
...
y
y
y
y
n
n
y
y
y
y
y
4
4
4
4
1
1
1
3
1
3
1
0
0
0
0
0
0
0
0
0
0
0
d3
d4
d5
d6
root_dos
root_layout
root_slice_1
root_volume_1
root_slice_2
root_volume_2
root_slice_3
1
1
1
1
0
0
1
2
1
2
1
clid
1,2,3,4,5,6,7,8,9,10,11,
12,13,14,15,16,17,18,19,20,
21,22,23,24,25,26,27,28,29,
30,31,32,33,34,51
35,36,37,38,39,40,41,42,
43,44,45,46,47,48,49,50,52
76
77
78
79
10
1
12
2
14
Note: This is a partial listing due to the length of the outputs.
Where:
Value
id
inuse
type
acl
426
Definition
ID of the volume.
Whether the volume is used.
Type assigned to the volume. Available types are: 1=slice, 2=stripe, 3=meta, 4=disk, and 100=pool.
Access control level assigned tothe volume.
Command Line Interface Reference for File 8.1
The nas Commands
Name assigned to the volume.
The client type of the volume. Available values are:
• 0 — If the clid field is not empty then the client is a slice.
• 1 — The client is another volume (meta, stripe, volume_pool).
• 2 — The client is a file system.
ID of the client.
name
cltype
clid
EXAMPLE #2
To create a metavolume named, mtv1, on disk volume, d7, type:
$ nas_volume -name mtv1 -create d7
id
name
acl
in_use
type
volume_set
disks
=
=
=
=
=
=
=
146
mtv1
0
False
meta
d7
d7
Where:
Value
id
name
acl
in_use
type
volume_set
disks
Definition
ID of the volume.
Name assigned to the volume.
Access control level value assigned to the volume.
Whether the volume is used.
Type assigned to the volume. Types are meta, stripe, slice, disk, and pool.
Name assigned to the volume.
Disks used to build a file system.
EXAMPLE #3
To display configuration information for mtv1, type:
$ nas_volume -info mtv1
id
name
acl
in_use
type
volume_set
disks
=
=
=
=
=
=
=
146
mtv1
0
False
meta
d7
d7
nas_volume
427
The nas Commands
EXAMPLE #4
To rename a mtv1 to mtv2, type:
$ nas_volume -rename mtv1 mtv2
id
name
acl
in_use
type
volume_set
disks
=
=
=
=
=
=
=
146
mtv2
0
False
meta
d7
d7
EXAMPLE #5
To create a stripe volume named, stv1, with a size of 32768 bytes on
disk volumes d10, d12, d13, and d15, type:
$ nas_volume -name stv1 -create -Stripe 32768
d10,d12,d13,d15
id
name
acl
in_use
type
stripe_size
volume_set
disks
=
=
=
=
=
=
=
=
147
stv1
0
False
stripe
32768
d10,d12,d13,d15
d10,d12,d13,d15
Where:
Value
stripe_size
Definition
Specified size of the stripe volume.
EXAMPLE #6
To clone mtv1, type:
$ nas_volume -Clone mtv1
428
id
name
acl
in_use
type
volume_set
disks
=
=
=
=
=
=
=
146
mtv1
0
False
meta
d7
d7
id
name
acl
in_use
type
volume_set
disks
=
=
=
=
=
=
=
148
v148
0
False
meta
d8
d8
Command Line Interface Reference for File 8.1
The nas Commands
EXAMPLE #7
To clone the volume mtv1 and set the disk type to BCV, type:
$ /nas/sbin/rootnas_volume -Clone mtv1 -option
disktype=BCV
id
name
acl
in_use
type
volume_set
disks
=
=
=
=
=
=
=
322
mtv1
0
False
meta
d87
d87
id
name
acl
in_use
type
volume_set
disks
=
=
=
=
=
=
=
323
v323
0
False
meta
rootd99
rootd99
EXAMPLE #8
To extend mtv1 with mtv2, type:
$ nas_volume -xtend mtv1 mtv2
id
name
acl
in_use
type
volume_set
disks
=
=
=
=
=
=
=
146
mtv1
0
False
meta
d7,mtv2
d7,d8
EXAMPLE #9
To display the size of mtv1, type:
$ nas_volume -size mtv1
total = 547418 avail = 547418 used = 0 ( 0% ) (sizes in MB)
Where:
Value
total
avail
used
Definition
Total size of the volume.
Amount of unused space on the volume.
Amount of space used on the volume.
nas_volume
429
The nas Commands
EXAMPLE #10
id
name
acl
in_use
type
volume_set
disks
=
=
=
=
=
=
=
To set the access control level for the metavolume mtv1, type:
$ nas_volume -acl 1432 mtv1
125
mtv1
1432, owner=nasadmin, ID=201
False
meta
d7,mtv2
d7,d8
Note: The value 1432 specifies nasadmin as the owner and gives users with
an access level of at least observer read access only, users with an access level
of at least operator read/write access, and users with an access level of at
least admin read/write/delete access.
EXAMPLE #11
id
name
acl
in_use
type
volume_set
disks
430
=
=
=
=
=
=
=
To delete mtv2, type:
$ nas_volume -delete mtv1
146
mtv1
1432, owner=nasadmin, ID=201
False
meta
d7,mtv2
d7,d8
Command Line Interface Reference for File 8.1
6
The server Commands
This chapter lists the VNX Command Set provided for managing,
configuring, and monitoring Data Movers. The commands are
prefixed with server and appear alphabetically. The command line
syntax (Synopsis), a description of the options, and an example of
usage are provided for each command. Commands included are:
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
server_archive...................................................................................
server_arp..........................................................................................
server_certificate ..............................................................................
server_cdms ......................................................................................
server_cepp .......................................................................................
server_checkup.................................................................................
server_cifs..........................................................................................
server_cifssupport............................................................................
server_cpu .........................................................................................
server_date........................................................................................
server_dbms......................................................................................
server_devconfig ..............................................................................
server_df............................................................................................
server_dns .........................................................................................
server_export ....................................................................................
server_file ..........................................................................................
server_fileresolve .............................................................................
server_http ........................................................................................
server_ifconfig ..................................................................................
server_ip ............................................................................................
server_kerberos ................................................................................
server_kerberos ................................................................................
server_ldap........................................................................................
The server Commands
433
447
449
455
461
464
481
509
523
525
530
537
545
547
549
563
564
584
590
597
602
602
609
431
The server Commands
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
◆
432
server_log..........................................................................................
server_mount....................................................................................
server_mountpoint ..........................................................................
server_mpfs ......................................................................................
server_mt...........................................................................................
server_name......................................................................................
server_netstat ...................................................................................
server_nfs ..........................................................................................
server_nis ..........................................................................................
server_nsdomains ............................................................................
server_param....................................................................................
server_pax.........................................................................................
server_ping .......................................................................................
server_ping6 .....................................................................................
server_rip ..........................................................................................
server_route ......................................................................................
server_security .................................................................................
server_setup......................................................................................
server_snmpd ...................................................................................
server_ssh..........................................................................................
server_standby .................................................................................
server_stats .......................................................................................
server_sysconfig...............................................................................
server_sysstat ...................................................................................
server_tftp .........................................................................................
server_umount .................................................................................
server_uptime...................................................................................
server_user........................................................................................
server_usermapper ..........................................................................
server_version ..................................................................................
server_viruschk ................................................................................
server_vtlu ........................................................................................
Command Line Interface Reference for File 8.1
621
628
637
639
642
644
645
649
658
662
665
673
677
679
681
682
685
688
691
695
707
710
758
767
769
772
774
775
778
782
783
787
The server Commands
server_archive
Reads and writes file archives, and copies directory hierarchies.
SYNOPSIS
server_archive <movername> [-cdnvN] -f <archive_file> [-J [p][w|d|u]]
[-I <client_dialect>]
[-e <archive_name>][-s <replstr>] ...
[-T [<from_date>][,<to_date>]][<pattern>] ...
server_archive <movername> -r [-cdiknuvDNYZ][-E <limit>]
[-J [w|d|u]][-C d|i|m][-I <client_dialect>]
[-f <file_name>][-e <archive_name>]
[-p <string>] ... [-s <replstr>] ...
[-T [<from_date>][,<to_date>]] ... [<pattern> ...]
server_archive <movername> -w [-dituvLNPX]
[-J [w|d|u]][-I <client_dialect>]
[-b <block_size>][-f <file_name>][-e <archive_name>]
[-x <format>][-B bytes][-s <replstr>] ...
[-T [<from_date>][,<to_date>][/[c][m]]] ...
[[-0]|[-1]][<file> ...]
server_archive -r -w [-diklntuvDLPXYZ]
[-J [w|d|u]][-C d|i|m]
[-p <string>] ... [-s <replstr>] ...
[-T [<from_date>][,<to_date>][/[c][m]]] ...
[<file> ...] <directory>
DESCRIPTION
server_archive reads, writes, and lists the members of an archive file,
and copies directory hierarchies. The server_archive operation is
independent of the specific archive format, and supports a variety of
different archive formats.
Note: A list of supported archive formats can be found under the description
of the -x option.
The presence of the -r and the -w options specifies the following
functional modes: list, read, write, and copy.
-List (no arguments)
server_archive writes to standard output a table of contents of the
members of the archive file read from archive, whose pathnames
match the specified patterns.
Note: If no options are specified, server_archive lists the contents of the
archive.
server_archive
433
The server Commands
Read (-r)
server_archive extracts the members of the archive file read from the
archive, with pathnames matching the specified patterns. The archive
format and blocking are automatically determined on input. When an
extracted file is a directory, the entire file hierarchy rooted at that
directory is extracted.
Note: Ownership, access, and modification times, and file mode of the
extracted files are discussed in more detail under the -p option.
Write (-w)
server_archive writes an archive containing the file operands to
archive using the specified archive format. When a file operand is
also a directory, the entire file hierarchy rooted at that directory is
included.
Copy (-r -w)
server_archive copies the file operands to the destination directory.
When a file operand is also a directory, the entire file hierarchy rooted
at that directory is included. The effect of the copy is as if the copied
files were written to an archive file and then subsequently extracted,
except that there may be hard links between the original and the
copied files. The -l option provides more information.
!
CAUTION
The destination directory must exist and must not be one of the file
operands or a member of a file hierarchy rooted at one of the file
operands. The result of a copy under these conditions is
unpredictable.
While processing a damaged archive during a read or list operation,
server_archive attempts to recover from media defects and searches
through the archive to locate and process the largest number of
archive members possible (the -E option provides more details on
error handling).
OPERANDS
The directory operand specifies a destination directory pathname. If
the directory operand does not exist, or it is not writable by the user,
or it is not a directory name, server_archive exits with a non-zero exit
status.
The pattern operand is used to select one or more pathnames of
archive members. Archive members are selected using the pattern
matching notation described by fnmatch 3. When the pattern
434
Command Line Interface Reference for File 8.1
The server Commands
operand is not supplied, all members of the archive are selected.
When a pattern matches a directory, the entire file hierarchy rooted at
that directory is selected. When a pattern operand does not select at
least one archive member, server_archive writes these pattern
operands in a diagnostic message to standard error and then exits
with a non-zero exit status.
The file operand specifies the pathname of a file to be copied or
archived. When a file operand does not select at least one archive
member, server_archive writes these file operand pathnames in a
diagnostic message to standard error and then exits with a non-zero
exit status.
The archive_file operand is the name of a file where the data is stored
(write) or read (read/list). The archive_name is the name of the
streamer on which the data will be stored (write) or read (read/list).
Note: To obtain the device name, you can use server_devconfig -scsi.
OPTIONS
The following options are supported:
-r
Reads an archive file from archive and extracts the specified files. If
any intermediate directories are needed to extract an archive member,
these directories will be created as if mkdir 2 was called with the
bit-wise inclusive OR of S_IRWXU, S_IRWXG, and S_IRWXO, as the
mode argument. When the selected archive format supports the
specification of linked files and these files cannot be linked while the
archive is being extracted, server_archive writes a diagnostic
message to standard error and exits with a non-zero exit status at the
completion of operation.
-w
Writes files to the archive in the specified archive format.
-0 (zero)
With this option, a full referenced backup is performed with the time
and date of launching put in a reference file. This reference file is an
ASCII file and is located in /.etc/BackupDates. The backup is
referenced by the pathname of the files to back up and the time and
date when the backup was created. This file is updated only if the
backup is successful.
Backup files can be copied using the server_file command.
server_archive
435
The server Commands
-<x>
Level x (x=1–9) indicates a backup of all files in a file system that have
been modified since the last backup of a level smaller than the
previous backup.
For example, a backup is performed for:
• Monday: level 0 = full backup
• Tuesday: level 3 = files modified since Monday
• Friday: level 5 = files modified since Tuesday
• Saturday: level 4 = files modified since Tuesday
• Sunday: level 4 = files modified since Tuesday
Note: If the backup type is not indicated, a full backup is performed
automatically.
-b <block_size>
When writing an archive, blocks the output at a positive decimal
integer number of bytes per write to the archive file. The
<block_size> must be a multiple of 512 bytes with a maximum size of
40 kilobytes.
Note: To remain POSIX-compatible, do not exceed 32256 Bytes.
A <block_size> can end with k or b to specify multiplication by 1024
(1K) or 512, respectively. A pair of <block_size> can be separated by x
to indicate a product. A specific archive device may impose
additional restrictions on the size of blocking it will support. When
blocking is not specified, the default for <block_size> is dependent
on the specific archive format being used. The -x option provides
more information.
-c
Matches all file or archive members except those specified by the
pattern and file operands.
-d
Causes files of type directory being copied or archived, or archive
members of type directory being extracted, to match only the
directory file or archive member, and not the file hierarchy rooted at
the directory.
-e <archive_name>
Specifies the archive name when it is streamed.
436
Command Line Interface Reference for File 8.1
The server Commands
Note: To prevent the tape from rewinding at the end of command execution,
use the -N option with the -e <archive_name> option.
-f <archive_file>
Specifies the archive name when it is a file.
Note: A single archive may span multiple files and different archive devices.
When required, server_archive prompts for the pathname of the file or
device of the next volume in the archive.
-i
Interactively renames files or archive members. For each archive
member matching a pattern operand, or each file matching a file
operand, server_archive prompts to /dev/tty giving the name of the
file, its file mode, and its modification time. Then server_archive
reads a line from /dev/tty. If this line is blank, the file or archive
member is skipped. If this line consists of a single period, the file or
archive member is processed with no modification to its name.
Otherwise, its name is replaced with the contents of the line. Then
server_archive immediately exits with a non-zero exit status if
<EOF> is encountered when reading a response, or if /dev/tty cannot
be opened for reading and writing.
-k
Does not allow overwriting existing files.
-l
Links files. In the copy mode (-r, -w), hard links are made between the
source and destination file hierarchies whenever possible.
-I <client_dialect>
Allows filename information recovered from an archive to be
translated into UTF-8.
-n
Selects the first archive member that matches each pattern operand.
No more than one archive member is matched for each pattern. When
members of type directory are matched, the file hierarchy rooted at
that directory is also matched (unless -d is also specified).
-p <string>
Specifies one or more file characteristic options (privileges). The
<string> option-argument is a string specifying file characteristics to
be retained or discarded on extraction. The string consists of the
specification characters a, e, m, o, and p. Multiple characteristics can
server_archive
437
The server Commands
be concatenated within the same string and multiple -p options can
be specified. The meaning of the specification characters is as follows:
a
Do not preserve file access times. By default, file access times are
preserved whenever possible.
e
Preserve everything (default mode), the user ID, group ID, file
mode bits, file access time, and file modification time.
Note: The e flag is the sum of the o and p flags.
m
Do not preserve file modification times. By default, file
modification times are preserved whenever possible.
o
Preserve the user ID and group ID.
p
Preserve the file mode bits. This specification character is
intended for a user with regular privileges who wants to preserve
all aspects of the file other than the ownership. The file times are
preserved by default, but two other flags are offered to disable
this and use the time of extraction instead.
In the preceding list, preserve indicates that an attribute stored in
the archive is given to the extracted file, subject to the permissions
of the invoking process. Otherwise, the attribute of the extracted
file is determined as part of the normal file creation action. If
neither the e nor the o specification character is specified, or the
user ID and group ID are not preserved for any reason,
server_archive will not set the S_ISUID (setuid) and S_ISGID
(setgid) bits of the file mode. If the preservation of any of these
items fails for any reason, server_archive writes a diagnostic
message to standard error.
Note: Failure to preserve these items will affect the final exit status, but
will not cause the extracted file to be deleted.
If the file characteristic letters in any of the string
option-arguments are duplicated, or in conflict with one another,
the ones given last will take precedence. For example, if you
specify -p eme, file modification times are still preserved.
438
Command Line Interface Reference for File 8.1
The server Commands
-s <replstr>
Modifies the file or archive member names specified by the pattern or
<file> operand according to the substitution expression <replstr>
using the syntax of the ed utility regular expressions.
Note: The ed 1 manual page provides information.
Multiple -s expressions can be specified. The expressions are applied
in the order they are specified on the command line, terminating with
the first successful substitution. The optional trailing g continues to
apply the substitution expression to the pathname substring, which
starts with the first character following the end of the last successful
substitution.
The optional trailing p causes the final result of a successful
substitution to be written to standard error in the following format:
<original pathname> >> <new pathname>
File or archive member names that substitute the empty string are not
selected and are skipped.
-t
Resets the access times of any file or directory read or accessed by
server_archive to be the same as they were before being read or
accessed by server_archive.
-u
Ignores files that are older (having a less recent file modification time)
than a pre-existing file, or archive member with the same name.
During read, an archive member with the same name as a file in a file
system is extracted if the archive member is newer than the file.
During copy, the file in the destination hierarchy is replaced by the
file in the source hierarchy, or by a link to the file in the source
hierarchy if the file in the source hierarchy is newer.
server_archive
439
The server Commands
-v
During a list operation, produces a verbose table of contents using
the format of the ls 1 utility with the -l option. For pathnames
representing a hard link to a previous member of the archive, the
output has the format:
<ls -l listing> == <link name>
For pathnames representing a symbolic link, the output has the
format:
<ls -l listing> => <link name>
where <ls -l listing> is the output format specified by the ls 1 utility
when used with the -l option. Otherwise, for all the other operational
modes (read, write, and copy), pathnames are written and flushed a
standard error without a trailing <newline> as soon as processing
begins on that file or archive member. The trailing <newline> is not
buffered, and is written only after the file has been read or written.
-x format
Specifies the output archive format, with the default format being
ustar. The server_archive command currently supports the following
formats:
cpio
The extended cpio interchange format specified in the -p1003.2
standard. The default blocksize for this format is 5120 bytes.
Inode and device information about a file (used for detecting file
hard links by this format) which may be truncated by this format
is detected by server_archive and is repaired.
Note: To be readable by server_archive, the archive must be built on
another machine with the option -c (write header information in ASCII).
bcpio
The old binary cpio format. The default blocksize for this format
is 5120 bytes.
Note: This format is not very portable and should not be used when
other formats are available.
Inode and device information about a file (used for detecting file
hard links by this format) which may be truncated by this format
is detected by server_archive and is repaired.
440
Command Line Interface Reference for File 8.1
The server Commands
sv4cpio
The System V release 4 cpio. The default blocksize for this format
is 5120 bytes. Inode and device information about a file (used for
detecting file hard links by this format) which may be truncated
by this format is detected by server_archive and is repaired.
sv4crc
The System V release 4 cpio with file crc checksums. The default
blocksize for this format is 5120 bytes. Inode and device
information about a file (used for detecting file hard links by this
format) which may be truncated by this format is detected by
server_archive and is repaired.
tar
The old BSD tar format as found in BSD4.3. The default blocksize
for this format is 10240 bytes. Pathnames stored by this format
must be 100 characters or less in length. Only regular files, hard
links, soft links, and directories will be archived (other file system
types are not supported).
ustar
The extended tar interchange format specified in the -p1003.2
standard. The default blocksize for this format is 10240 bytes.
Note: Pathnames stored by this format must be 250 characters or less in
length (150 for basename and 100 for <file_name>).
emctar
This format is not compatible with -p1003.2 standard. It allows
archiving to a file greater than 8 GB. Pathnames stored by this
format are limited to 3070 characters. The other features of this
format are the same as ustar.
server_archive detects and reports any file that it is unable to store or
extract as the result of any specific archive format restrictions. The
individual archive formats may impose additional restrictions on use.
Note: Typical archive format restrictions include (but are not limited to) file
pathname length, file size, link pathname length, and the type of the file.
-B bytes
Limits the number of bytes written to a single archive volume to
bytes. The bytes limit can end with m, k, or b to specify multiplication
by 1048576 (1M), 1024 (1K) or 512, respectively. A pair of bytes limits
can be separated by x to indicate a product.
server_archive
441
The server Commands
Note: The limit size will be rounded up to the nearest block size.
-C [d|i|m]
When performing a restore, this allows you to choose PAX behaviors
on CIFS collision names.
d: delete
i: ignore
m: mangle
-D
Ignores files that have a less recent file inode change time than a
pre-existing file, or archive member with the same name. The -u
option provides information.
Note: This option is the same as the -u option, except that the file inode
change time is checked instead of the file modification time. The file inode
change time can be used to select files whose inode information (such as uid,
gid, and so on) is newer than a copy of the file in the destination directory.
-E limit
Has the following two goals:
442
◆
In case of medium error, to limit the number of consecutive read
faults while trying to read a flawed archive to limit. With a
positive limit, server_archive attempts to recover from an archive
read error and will continue processing starting with the next file
stored in the archive. A limit of 0 (zero) will cause server_archive
to stop operation after the first read error is detected on an
archive volume. A limit of "NONE" will cause server_archive to
attempt to recover from read errors forever.
◆
In case of no medium error, to limit the number of consecutive
valid header searches when an invalid format detection occurs.
With a positive value, server_archive will attempt to recover from
an invalid format detection and will continue processing starting
with the next file stored in the archive. A limit of 0 (zero) will
cause server_archive to stop operation after the first invalid
header is detected on an archive volume. A limit of "NONE" will
cause server_archive to attempt to recover from invalid format
errors forever. The default limit is 10 retries.
Command Line Interface Reference for File 8.1
The server Commands
!
CAUTION
Using this option with NONE requires extreme caution as
server_archive may get stuck in an infinite loop on a badly flawed
archive.
-J
Backs up, restores, or displays CIFS extended attributes.
p: Displays the full pathname for alternate names (for listing and
archive only)
u: Specifies UNIX name for pattern search
w: Specifies M256 name for pattern search
d: Specifies M83 name for pattern search
-L
Follows all symbolic links to perform a logical file system traversal.
-N
Used with the -e archive_name option, prevents the tape from
rewinding at the end of command execution.
-P
Does not follow symbolic links.
Note: Performs a physical file system traversal. This is the default mode.
-T [from_date][,to_date][/[c][m]]
Allows files to be selected based on a file modification or inode
change time falling within a specified time range of from_date to
to_date (the dates are inclusive). If only a from_date is supplied, all
files with a modification or inode change time equal to or less than
are selected. If only a to_date is supplied, all files with a modification
or inode change time equal to or greater than will be selected. When
the from_date is equal to the to_date, only files with a modification or
inode change time of exactly that time will be selected.
When server_archive is in the write or copy mode, the optional
trailing field [c][m] can be used to determine which file time (inode
change, file modification or both) is used in the comparison. If neither
is specified, the default is to use file modification time only. The m
specifies the comparison of file modification time (the time when the
file was last written). The c specifies the comparison of inode change
time (the time when the file inode was last changed; for example, a
change of owner, group, mode, and so on). When c and m are both
server_archive
443
The server Commands
specified, then the modification and inode change times are both
compared. The inode change time comparison is useful in selecting
files whose attributes were recently changed, or selecting files which
were recently created and had their modification time reset to an
older time (as what happens when a file is extracted from an archive
and the modification time is preserved). Time comparisons using both
file times are useful when server_archive is used to create a
time-based incremental archive (only files that were changed during
a specified time range will be archived).
A time range is made up of six different fields and each field must
contain two digits. The format is:
[yy[mm[dd[hh]]]]mm[ss]
where yy is the last two digits of the year, the first mm is the month
(from 01 to 12), dd is the day of the month (from 01 to 31), hh is the
hour of the day (from 00 to 23), the second mm is the minute (from 00
to 59), and ss is seconds (from 00 to 59). The minute field mm is
required, while the other fields are optional, and must be added in
the following order: hh, dd, mm, yy. The ss field may be added
independently of the other fields. Time ranges are relative to the
current time, so -T 1234/cm selects all files with a modification or
inode change time of 12:34 p.m. today or later. Multiple -T time range
can be supplied, and checking stops with the first match.
-X
When traversing the file hierarchy specified by a pathname does not
allow descending into directories that have a different device ID. The
st_dev field as described in stat 2 provides more information about
device IDs.
-Y
Ignores files that have a less recent file inode change time than a
pre-existing file, or archive member with the same name.
Note: This option is the same as the -D option, except that the inode change
time is checked using the pathname created after all the filename
modifications have completed.
-Z
Ignores files that are older (having a less recent file modification time)
than a pre-existing file, or archive member with the same name.
444
Command Line Interface Reference for File 8.1
The server Commands
Note: This option is the same as the -u option, except that the modification
time is checked using the pathname created after all the filename
modifications have completed.
The options that operate on the names of files or archive members (-c,
-i, -n, -s, -u, -v, -D, -T, -Y, and -Z) interact as follows.
When extracting files during a read operation, archive members are
selected, based only on the user-specified pattern operands as
modified by the -c, -n, -u, -D, and -T options. Then any -s and -i
options will modify, in that order, the names of those selected files.
Then the -Y and -Z options will be applied based on the final
pathname. Finally, the -v option will write the names resulting from
these modifications.
When archiving files during a write operation, or copying files
during a copy operation, archive members are selected, based only on
the user-specified pathnames as modified by the -n, -u, -D, and -T
options (the -D option applies only during a copy operation). Then
any -s and -i options will modify, in that order, the names of these
selected files. Then during a copy operation, the -Y and -Z options
will be applied based on the final pathname. Finally, the -v option
will write the names resulting from these modifications.
When one or both of the -u or -D options are specified along with the
-n option, a file is not considered selected unless it is newer than the
file to which it is compared.
SEE ALSO
EXAMPLE #1
Using the server_archive Utility on VNX.
To archive the contents of the root directory to the device rst0, type:
$ server_archive <movername> -w -e rst0
EXAMPLE #2
To display the verbose table of contents for an archive stored in
<file_name>, type:
$ server_archive <movername> -v -f <file_name>
EXAMPLE #3
To copy the entire olddir directory hierarchy to newdir, type:
$ server_archive <movername> -rw <olddir newdir>
EXAMPLE #4
To interactively select the files to copy from the current directory to
dest_dir, type:
$ server_archive <movername> -rw -i <olddir dest_dir>
server_archive
445
The server Commands
EXAMPLE #5
To extract all files from the archive stored in <file_name>, type:
$ server_archive <movername> -r -f <file_name>
EXAMPLE #6
To update (and list) only those files in the destination directory
/backup that are older (less recent inode change or file modification
times) than files with the same name found in the source file tree
home, type:
$ server_archive <movername> -r -w -v -Y -Z home /backup
STANDARDS
The server_archive utility is a superset of the -p1003.2 standard.
Note: The archive formats bcpio, sv4cpio, sv4crc, and tar, and the flawed
archive handling during list and read operations are extensions to the POSIX
standard.
ERRORS
The server_archive command exits with one of the following system
messages:
All files were processed successfully.
or
An error occurred.
Whenever server_archive cannot create a file or a link when reading
an archive, or cannot find a file when writing an archive, or cannot
preserve the user ID, group ID, or file mode when the -p option is
specified, a diagnostic message is written to standard error, and a
non-zero exit status is returned. However, processing continues.
In the case where server_archive cannot create a link to a file, this
command will not create a second copy of the file.
If the extraction of a file from an archive is prematurely terminated by
a signal or error, server_archive may have only partially extracted a
file the user wanted. Additionally, the file modes of extracted files
and directories may have incorrect file bits, and the modification and
access times may be wrong.
If the creation of an archive is prematurely terminated by a signal or
error, server_archive may have only partially created the archive
which may violate the specific archive format specification.
If while doing a copy, server_archive detects a file is about to
overwrite itself, the file is not copied, a diagnostic message is written
to standard error and when server_archive completes, it exits with a
non-zero exit status.
446
Command Line Interface Reference for File 8.1
The server Commands
server_arp
Manages the Address Resolution Protocol (ARP) table for Data
Movers.
SYNOPSIS
server_arp {<movername>|ALL}
<ip_addr>
| -all
| -delete <ip_addr>
| -set <ip_addr> <physaddr>
DESCRIPTION
server_arp displays and modifies the IP-to-MAC address translation
tables used by the ARP for the specified Data Mover.
The ALL option executes the command for all Data Movers.
OPTIONS
<ip_addr>
Displays the ARP entry for the specified IP address.
-all
Displays the first 64 of the current ARP entries.
-delete <ip_addr>
Deletes an ARP entry.
-set <ip_addr> <physaddr>
Creates an ARP entry with an IP address and physical address.
EXAMPLE #1
To create an ARP entry, type:
$ server_arp server_2 -set 172.24.102.20
00:D0:B7:82:98:E0
server_2 : added: 172.24.102.20 at 0:d0:b7:82:98:e0
EXAMPLE #2
To display all ARP entries for a specified Data Mover, type:
$ server_arp server_2 -all
server_2 :
172.24.102.254 at 0:d0:3:f9:37:fc
172.24.102.20 at 0:d0:b7:82:98:e0
172.24.102.24 at 0:50:56:8e:1d:5
128.221.253.100 at 0:4:23:a7:b1:35
EXAMPLE #3
To display an ARP entry specified by IP address, type:
$ server_arp server_2 172.24.102.20
server_2 : 172.24.102.20 at 0:d0:b7:82:98:e0
server_arp
447
The server Commands
EXAMPLE #4
To delete an ARP entry, type:
$ server_arp server_2 -delete 172.24.102.24
server_2 : deleted: 172.24.102.24 at 0:50:56:8e:1d:5
448
Command Line Interface Reference for File 8.1
The server Commands
server_certificate
Manages VNX for file system’s Public Key Infrastructure (PKI) for the
specified Data Movers.
SYNOPSIS
server_certificate {<movername>|ALL}
-ca_certificate
[-list]
| -info {-all|<certificate_id>}
| -import [-filename <path>]
| -delete {-all|<certificate_id>}}
-persona
[-list]
| -info {-all|<persona_name>|id=<persona_id>}
| -generate {<persona_name>|id=<persona_id>} -key_size {2048|4096}
[-cs_sign_duration <# of months>]
{-cn|-common_name} <common_name>[;<common_name>]
[-ou <org_unit>[;<org_unit>]]
[-organization <organization>]
[-location <location>]
[-state <state>]
[-country <country>]
[-filename <output_path>]
| -clear {<persona_name>|id=<persona_id>}{-next|-current|-both}
| -import {<persona_name>|id=<persona_id>} [-filename <path>]
DESCRIPTION
OPTIONS
server_certificate manages the use of public key certificates between
Data Movers acting as either clients or servers. server_certificate
-ca_certificate manages the Certificate Authority (CA) certificates the
VNX uses to confirm a server’s identity when the Data Mover is
acting as a client. server_certificate -persona manages the certificates
presented by the Data Mover to a client application when the Data
Mover is acting as a server as well as the certificates presented by the
Data Mover to a server configured to require client authentication.
-ca_certificate
Lists the CA certificates currently available on the VNX. The ouput
from this command is identical to the output from the -list option.
-ca_certificate -list
Lists the CA certificates currently available on the VNX.
-ca_certificate -info {-all|<certificate_id>}
Displays the properties of a specified CA certificate or all CA
certificates.
server_certificate
449
The server Commands
-ca_certificate -import [-filename <path>]
Imports a CA certificate. You can only paste text in PEM format at the
command prompt. Specify -filename and provide a path to import a
CA certificate in either DER or PEM format.
-ca_certificate -delete {-all|<certificate_id>}
Deletes a specified CA certificate or all CA certificates.
-persona
Lists the key sets and associated certificates currently available on the
VNX. The ouput from this command is identical to the output from
the -list option.
-persona -list
Lists the key sets and associated certificates currently available on the
VNX.
-persona -info {-all|<persona_name>|id= <persona_id>}
Displays the properties of the key sets and associated certificates,
including the text of a pending certificate request, of a specified
persona or all personas.
-persona -generate {<persona_name>|id=<persona_id>}
-key_size <bits> {-cn|-common_name} <common_name>
[;<common_name>]
Generates a public/private key set along with a request to sign the
certificate. Specify either the persona name or ID. The ID is
automatically generated when the persona is created. You can
determine the ID using the -list or -info options. The key size can be
either 2048 or 4096 bits. Use either -cn or -common_name to specify
the commonly used name. The common name is typically a hostname
that describes the Data Mover with which the persona is associated.
Multiple common names are allowed but must be separated by
semicolon.
[-cs_sign_duration <# of months>]
Specifies the number of months the certificate is valid. A month is
defined as 30 days. This option is valid only if the certificate will
be signed by the Control Station. If this option is specified, you
cannot save the request to a file using the -filename option.
[-ou <org_unit>[;<org_unit>]]
Identifies the organizational unit. Multiple organizational units
are allowed but must be separated by semicolon.
[-organization <organization>]
Identifies the organization.
450
Command Line Interface Reference for File 8.1
The server Commands
[-location <location>]
Identifies the physical location of the organizational unit.
[-state <state>]
Identifies the state where the organizational unit is located.
[-country <country>]
Identifies the country where the organization unit is located. This
value is limited to two characters.
[-filename <output_path>]
Provides a path to where the request should be saved to a file.
This option is valid only if the certificate will be signed by an
external CA. If this option is specified, you cannot specify the
number of months the certificate is valid using the
-cs_sign_duration option.
-persona -clear {<persona_name>|id=<persona_id>}
{-next|-current|-both}
Deletes a key set and the associated certificate. You can delete the
current key set and certificate, the next key set and certificate, or both.
-persona -import {<persona_name>|id=<persona_id>}
[-filename <path>]
Imports a CA-signed certificate. You can only paste text in PEM
format at the command prompt. Specify -filename and provide a
path to import a CA-signed certificate in either DER or PEM format.
SEE ALSO
EXAMPLE #1
nas_ca_certificate.
To import a CA certificate, specifying a filename and path, type:
$ server_certificate server_2 -ca_certificate -import
-filename "/tmp/ca_cert.pem"
done
server_certificate
451
The server Commands
EXAMPLE #2
To list all the CA certificates currently available on the VNX, type:
$ server_certificate ALL -ca_certificate -list
server_2 :
id=1
subject=O=VNX Certificate Authority;CN=sorento
issuer=O=VNX Certificate Authority;CN=sorento
expire=20120318032639Z
id=2
subject=C=US;O=VeriSign, Inc.;OU=Class 3 Public Primary Certification Author
issuer=C=US;O=VeriSign, Inc.;OU=Class 3 Public Primary Certification Author
expire=20280801235959Z
server_3 :
id=1
subject=O=VNX Certificate Authority;CN=zeus-cs
issuer=O=VNX Certificate Authority;CN=zeus-cs
expire=20120606181215Z
EXAMPLE #3
To list the properties of the CA certificate identified by certificate ID
2, type:
$ server_certificate server_2 -ca_certificate -info 2
server_2 :
id=2
subject = C=US;O=VeriSign, Inc.;OU=Class 3 Public Primary Certification Authority
issuer = C=US;O=VeriSign, Inc.;OU=Class 3 Public Primary Certification Authority
start
= 19960129000000Z
expire = 20280801235959Z
signature alg. = md2WithRSAEncryption
public key alg. = rsaEncryption
public key size = 1024 bits
serial number
= 70ba e41d 10d9 2934 b638 ca7b 03cc babf
version
= 1
EXAMPLE #4
To generate a key set and certificate request to be sent to an external
CA for the persona identified by the persona name default, type:
$ server_certificate server_2 -persona -generate default
-key_size 2048 -common_name division.xyz.com
server_2 :
Starting key generation. This could take a long time ...
done
EXAMPLE #5
To list all the key sets and associated certificates currently available
on the VNX, type:
$ server_certificate ALL -persona -list
server_2 :
id=1
name=default
next state=Request Pending
request subject=CN=name;CN=1.2.3.4
452
Command Line Interface Reference for File 8.1
The server Commands
server_3 :
id=1
name=default
next state=Not Available
CURRENT CERTIFICATE:
id=1
subject=CN=test;CN=1.2.3.4
expire=20070706183824Z
issuer=O=VNX Certificate Authority;CN=eng173100
EXAMPLE #6
To list the properties of the key set and certificate identified by
persona ID 1, type:
$ server_certificate server_2 -persona -info id=1
server_2 :
id=1
name=default
next state=Request Pending
request subject=CN=name;CN=1.2.3.4
Request:
-----BEGIN CERTIFICATE REQUEST----MIIEZjCCAk4CAQAwITENMAsGA1UEAxMEbmFtZTEQMA4GA1UEAxMHMS4yLjMuNDCC
AiIwDQYJKoZIhvcNAQEBBQADggIPADCCAgoCggIBANKW3Q/F6eMqIxrCO5IeXLET
bWkm5RzrbI5lHxLNuhobR5S9G2o+k47X0QZFkGzq/2F7kR06vVIH7CPH9X2gGAzV
7GmZaFO0wPcktPJYzjQE8guNhcL1qZpPl4IZrbnSGEAWcAAE0nvNwLp9aN0WSC+N
TDJZY4A9yTURiUc+Bs8plhQh16wLLL0zjUKIvKjAqiTE0F3RApVJEE/9y6N+Idsb
Vwf/rvzP6/z0wZW5Hl84HKXInJaHTBDK59G+e/Y2JgvUY1UNBZ5SODunOakHabex
k6COFYjDu7Vd+yHpvcyTalHJ2RcIavpQuM02o+VVpxgUyX7M1+VXJXTJm0yb4j4g
tZITOSVZ2FqEpOkoIpzqoAL7A9B69WpFbbpIX8danhReafDh4oj4yWocvSwMKYv1
33nLak3+wpMQNrwJ2L9FIHP2fXClnvThBgupm7uqqHP3TfNBbBPTYY3qkNPZ78wx
/njUrZKbfWd81Cc+ngUi33hbMuBR3FFsQNASYZUzgl5+JexALH5jhBahd2aRXBag
itQLhvxYK0dEqIEwDfdDedx7i+yro2gbNxhLLdtkuBtKrmOnuT5g2WWXNKzNa/H7
KWv8JSwCv1mW1N/w7V9aEbDizBBfer+ZdMPkGLbyb/EVXZnHABeWH3iKC6/ecnRd
4Kn7KO9F9qXVHlzzTeYVAgMBAAGgADANBgkqhkiG9w0BAQUFAAOCAgEAzSS4ffYf
2WN0vmZ0LgsSBcVHPVEVg+rP/aU9iNM9KDJ4P4OK41UDU8tOGy09Kc8EvklBUm59
fyjt2T/3RqSgvvkCEHJsVW3ZMnSsyjDo6Ruc0HmuY4q+tuLl+dilSQnZGUxt8asw
dhEpdEzXA6o9cfmVZMSt5QicfAmmBNr4BaO96+VAlg59fu/chU1pvKWWMGXz4I2s
7z+UdMBYO4pEfyG1i34Qof/z4K0SVNICn3CEkW5TIsSt8qA/E2JXXlLhbMYWKYuY
9ur/gspHuWzkIXZFx4SmTK9/RsE1Vy7fBztIoN8myFN0nma84D9pyqls/yhvXZ/D
iDF6Tgk4RbNzuanRBSYiJFu4Tip/nJlK8uv3ZyFJ+3DK0c8ozlBLuQdadxHcJglt
m/T4FsHa3JS+D8CdA3uDPfIvvVNcwP+4RBK+Dk6EyQe8uKrVL7ShbacQCUXn0AAd
Ol+DQYFQ7Mczcm84L98srhov3JnIEKcjaPseB7S9KtHvHvvs4q1lQ5U2RjQppykZ
qpSFnCbYDGjOcqOrsqNehV9F4h9fTszEdUY1UuLgvtRj+FTT2Ik7nMK641wfVtSO
LCial6kuYsZg16SFxncnH5gKHtQMWxd9nv+UyJ5VwX3aN12N0ZQbaIDcQp75Em2E
aKjd28cZ6FEavimn69sz0B8PHQV+6dPwywM=
-----END CERTIFICATE REQUEST-----
server_certificate
453
The server Commands
EXAMPLE #7
To generate a key set and certificate request that is automatically
received by the Control Station for the persona identified by the
persona name default, type:
$ server_certificate server_2 -persona -generate default
-key_size 2048 -cs_sign_duration 12 -common_name
division.xyz.com
server_2 :
Starting key generation. This could take a long time ...
done
EXAMPLE #8
To generate a key set and certificate request to be sent to an external
CA specifying subject information, type:
$ server_certificate server_2 -persona -generate default
-key_size 2048 -common_name division.xyz.com -ou QA
-organization XYZ -location Bethesda -state Maryland
-country US -filename /tmp/server_2.1.request.pem
server_2 :
Starting key generation. This could take a long time ...
done
EXAMPLE #9
To import a signed certificate and paste the certificate text, type:
$ server_certificate server_2 -persona -import default
server_2 : Please paste certificate data. Enter a carriage return and on the new
line type ‘end of file’ or ‘eof’ followed by another carriage return.
454
Command Line Interface Reference for File 8.1
The server Commands
server_cdms
Provides File Migration Service for VNX functionality for the
specified Data Movers.
SYNOPSIS
server_cdms {<movername>|ALL}
-connect <mgfs> -type {nfsv2|nfsv3} -path <localpath>
-source <srcName>:/<srcPath>[-option <options>]
| -connect <mgfs> -type cifs -path <localpath> -netbios <netbios> -source
\\<srcServer>[.<domain>]\<srcShare>[\<srcPath>] -admin
[<domain>\]<admin_name> [-wins <wins>]
| -disconnect <mgfs> {-path <localpath>|-path <cid>|-all}
| -verify <mgfs> [-path {<localpath>|<cid>}]
| -Convert <mgfs>
| -start <mgfs> -path <localpath> [-Force] -log <logpath>
[-include <include_path>][-exclude <exclude_path>]
| -halt <mgfs> -path <localpath>
| -info [<mgfs>][-state {START|STOP|ON_GOING|ERROR|SUCCEED|FAIL}]
DESCRIPTION
server_cdms establishes and removes connections to remote systems,
and allows users to start on-access migration.
server_cdms creates an auto-migration process on the Data Mover to
ensure that all data has been migrated from the remote system.
server_cdms also checks the state of the migrated file system (MGFS),
all auto-migration processes, and the connection, and reports if all
data has been migrated successfully.
CDMS supports NFSv2 and NFSv3 only.
The ALL option executes the command for all Data Movers.
OPTIONS
-connect <mgfs> -type {nfsv2|nfsv3} -path <localpath>
-source <srcName>:/<srcPath>
Provides a connection for the VNX with the remote NFS server. The
-type option specifies the protocol type to be used for communication
with the remote NFS server. The directory <localpath> in the file
system must be unique for that file system.
The -source option specifies the source file server name or IP address
of the remote server as the <srcName> and the export path for
migration. For example, nfs_server:/export/path
Note: After the -connect command completes, the file system must be
exported.
server_cdms
455
The server Commands
[-option <options>]
Specifies the following comma-separated options:
[useRootCred={true|false}]
When the file system is mounted, true ensures that the MGFS
reads from the source file server using root access UID=0, GID=0.
This assumes that the source file server path is exported to allow
root access from the specified Data Mover. When false (default),
the MGFS uses the owner’s UID and GID to access data.
[proto={TCP|UDP}]
Sets the connection protocol type. The default is TCP.
[nfsPort=<port>]
Sets a remote NFS port number in case the Portmapper or RPC
bind is not running, and the port is not the default of 2049.
[mntPort=<port>]
Sets a remote mount port number in case Portmapper or RPC
bind is not running.
[mntVer={1|2|3}]
Sets the version used for mount protocol. By default, NFSv2 uses
mount version 2, unless user specified version 1; NFSv3 uses
mount version 3.
[localPort=<port>]
Sets the port number used for NFS services, if it needs to be
different from the default. The default port number is always
greater than 1024.
-connect <mgfs> -type cifs -path <localpath> -netbios
<netbios> -source \\<srcServer>[.<domain>]
\<srcShare>[\<srcPath>] -admin [<domain>\]
<admin_name>[-wins <wins>]
Provides a connection for the VNX with the remote CIFS server as
specified by its NetBIOS name. The directory <localpath> in the file
system must be unique for that file system. The -source option
specifies the source file server name of the remote server as the
<srcName> and the share path for migration that is not at the root of
the share. For example, \\share\dir1...
The -source and -admin option strings must be enclosed by quotes
when issued in a Linux shell.
The -admin option specifies an administrator for the file system. A
password is asked interactively when the command is issued. The
-wins option specifies an IP address for the WINS server.
456
Command Line Interface Reference for File 8.1
The server Commands
Note: This is required only for Windows NT 4.0.
-disconnect <mgfs> {-path <localpath>|-path <cid>|-all}
Removes a connection without migrating the data. The <localpath> is
not removed nor is any partially migrated data.
The administrator should manually remove this data before
attempting a -verify or -Convert command. It may require the
administrator to handle a partial migration of old data as well as
potentially new data created by users.
It is recommended not to use the -disconnect option if the
administrator has exported this directory for user access.
-verify <mgfs>
Checks that all data has completed the migration for the <mgfs>.
[-path {<localpath>|<cid>}]
If the -path option is provided, it can check on a communication
basis. If no path is provided, the system defaults to checking all
connections on the file system.
-Convert <mgfs>
Performs a verify check on the entire file system, then changes the file
system type from MGFS to UxFS. After the -Convert option succeeds,
no data migration can be done on that file system.
-start <mgfs> -path <localpath> [-Force] -log <logpath>
Directs the Data Mover to migrate all files from the source file server
to the VNX. The -log option provides detailed information on the
state of the migration, and any failures that might occur. The
<localpath> is the path where the migration thread is started. The
-Force option is used if you need to start a migration thread a second
time on the same <localpath> where a previous migration thread had
already finished. For example, -Force would be needed to start a
thread which had no include file (that is, to migrate all remaining
files) on <localpath> where a thread with an include file had already
been run.
[-include <include_path>]
Starts the thread in the <include_path> which is the path of the
file containing the specified directories.
[-exclude <exclude_path>]
Excludes files or directories from migration. The <include_path>
is the path of the file containing the specified directories.
server_cdms
457
The server Commands
-halt <mgfs> -path <localpath>
Stops a running thread, and halts its execution on the Data Mover.
The <mgfs> is the name of the migration file system and the
<localpath> is the full path where the migration thread was started.
The -start option resumes thread execution.
-info
Displays a status on the migration file system and the threads.
[<mgfs>]
Specifies the migration file system.
[-state {START|STOP|ON_GOING|ERROR|SUCCEED|
FAIL}]
Displays only the threads that are in the state that is specified.
SEE ALSO
EXAMPLE #1
VNX CDMS Version 2.0 for NFS and CIFS, server_export,
server_mount, and server_setup.
To provide a connection for the migration file system to communicate
with the remote NFS server, type:
$ server_cdms server_2 -connect ufs1 -type nfsv3 -path
/nfsdir -source 172.24.102.144:/srcdir -option proto=TCP
server_2 : done
EXAMPLE #2
To provide a connection for the migration file system to communicate
with the remote CIFS server, type:
$ server_cdms server_2 -connect ufs1 -type cifs -path
/dstdir -netbios dm112-cge0 -source
"\\\winserver1.nasdocs.emc.com\srcdir" -admin
"nasdocs.emc.com\administrator" -wins 172.24.102.25
server_2 : Enter Password:*******
done
EXAMPLE #3
To display a status on the migration file system, type:
$ server_cdms server_2
server_2 :
CDMS enabled with 32 threads.
ufs1:
path
=
cid
=
type
=
source =
options=
path
=
cid
=
458
/nfsdir
0
NFSV3
172.24.102.144:/srcdir
proto=TCP
/dstdir
1
Command Line Interface Reference for File 8.1
The server Commands
type
=
source =
netbios=
admin =
CIFS
\\winserver1.nasdocs.emc.com\srcdir\
DM112-CGE0.NASDOCS.EMC.COM
nasdocs.emc.com\administrator
When migration is started:
$ server_cdms server_2
server_2 :
CDMS enabled with 32 threads.
ufs1:
path
= /nfsdir
cid
= 0
type
= NFSV3
source = 172.24.102.144:/srcdir
options= proto=TCP
path
= /dstdir
cid
= 1
type
= CIFS
source = \\winserver1.nasdocs.emc.com\srcdir\
netbios= DM112-CGE0.NASDOCS.EMC.COM
admin = nasdocs.emc.com\administrator
threads:
path
= /dstdir
state
= ON_GOING
log
= /
cid
= NONE
Where:
Value
ufs1
path
cid
type
source
options
netbios
admin
threads
state
log
Definition
Migration file system.
Directory in the local file system.
Connection ID (0 through 1023).
Protocol type to be used to communicate with the remote server.
Source file server name or IP address of the remote server and the export path for migration.
Connection protocol type.
NetBIOS name of the remote CIFS server.
Administrator for the file system.
Currently existing migration threads.
Current status of migration threads.
Location of the log file that provides detailed information.
EXAMPLE #4
To direct server_2 to migrate all files from the source file server to the
VNX, type:
$ server_cdms server_2 -start ufs1 -path /dstdir -log /
server_2 : done
server_cdms
459
The server Commands
EXAMPLE #5
To display information about migration with the specified status,
type:
$ server_cdms server_2 -info ufs1 -state ON_GOING
server_2 :
ufs1:
path
= /nfsdir
cid
= 0
type
= NFSV3
source = 172.24.102.144:/srcdir
options= proto=TCP
path
= /dstdir
cid
= 1
type
= CIFS
source = \\winserver1.nasdocs.emc.com\srcdir\
netbios= DM112-CGE0.NASDOCS.EMC.COM
admin = nasdocs.emc.com\administrator
threads:
path
= /dstdir
state
= ON_GOING
log
= /
cid
= NONE
EXAMPLE #6
To stop data migration on server_2 for ufs1, type:
$ server_cdms server_2 -halt ufs1 -path /dstdir
server_2 : done
EXAMPLE #7
To check that all data has completed the migration, type:
$ server_cdms server_2 -verify ufs1 -path /dstdir
server_2 : done
EXAMPLE #8
To disconnect the path on server_2 for data migration, type:
$ server_cdms server_2 -disconnect ufs1 -path /nfsdir
server_2 : done
EXAMPLE #9
To disconnect all paths for data migration, type:
$ server_cdms server_2 -disconnect ufs1 -all
server_2 : done
EXAMPLE #10
To perform a verify check on ufs1, and then convert it to a uxfs, type:
$ server_cdms server_2 -Convert ufs1
server_2 : done
460
Command Line Interface Reference for File 8.1
The server Commands
server_cepp
Manages the Common Event Publishing Agent (CEPA) service on the
specified Data Mover.
SYNOPSIS
server_cepp {<movername>|ALL}
-service {-start|-stop|-status|-info}
| -pool {-info|-stats}
DESCRIPTION
server_cepp starts or stops the CEPA service on the specified Data
Mover or all Data Movers and displays information on the status,
configuration, and statistics for the service and the pool. The CEPA
service is set up in the cepp.conf configuration file. The CEPA
configuration is displayed using -service -status, but changes can
only be made by directly editing the file with a text editor.
ALL executes the command for all Data Movers.
OPTIONS
-service {-start|-stop|-status|-info}
The -start option starts the CEPA service on the specified Data Mover.
-stop stops the CEPA service, -status returns a message indicating
whether the CEPA service has started or been stopped, and -info
displays information about the CEPA service including key
properties of the configured pool.
-pool {-info|-stats}
Displays properties or statistics for the CEPA pool on the specified
Data Mover.
SEE ALSO
EXAMPLE #1
Using VNX Event Enabler
To start the CEPA service on a Data Mover, type:
$ server_cepp server_2 -service -start
server_2 : done
EXAMPLE #2
To display the status of the CEPA service, type:
$ server_cepp server_2 -service -status
server_2 : CEPP Stopped
server_cepp
461
The server Commands
EXAMPLE #3
To display the configuration of the CEPA service, type:
$ server_cepp server_2 -service -info
server_2 :
CIFS share name = \\DVBL\CHECK$
cifs_server = DVBL
heartbeat_interval = 15 seconds
ft level = 1
ft size = 1048576
ft location = /.etc/cepp
msrpc user = OMEGA13$
msrpc client name = OMEGA13.CEE.LAB.COM
pool_name
server_required access_checks_ignored req_timeout retry_timeout
pool_1
no
0
5000
25000
Where:
Value
CIFS share name
cifs_server
heartbeat_interval
ft level
Definition
The name of the shared directory and CIFS server used to access files in the Data Movers.
CIFS server to access files.
The time taken to scan each CEPA server.
Fault tolerance level assigned. This option is required. 0 (continue and tolerate lost events;
default setting), 1 (continue and use a persistence file as a circular event buffer for lost
events), 2 (continue and use a persistence file as a circular event buffer for lost events until
the buffer is filled and then stop CIFS), or 3 (upon heartbeat loss of connectivity, stop CIFS).
ft location
Directory where the persistence buffer file resides relative to the root of a file system. If a
location is not specified, the default location is the root of the file system.
ft size
Maximum size in MB of the persistence buffer file. The default is 1 MB and the range is 1 MB
to 100 MB.
msrpc user
Name assigned to the user account that the CEPA service is running under on the CEE
machine. For example, ceeuser.
msrpc client name
Domain name assigned if the msrpc user is a member of a domain. For example,
domain.ceeuser.
pool_name
Name assigned to the pool that will use the specified CEPA options.
server_required
Displays availability of the CEPA server. If a CEPA server is not available and this option is
yes, an error is returned to the requestor that access is denied. If a CEPA server is not
available and this option is no, an error is not returned to the requestor and access is
allowed.
access_checks_ignored The number of CIFS requests processed when a CEPA server is not available and the
server_required option is set to “no.” This option is reset when the CEPA server becomes
available.
req_timeout
Time out in ms to send a request that allows access to the CEPA server.
retry_timeout
Time out in ms to retry the access request sent to the CEPA server.
462
Command Line Interface Reference for File 8.1
The server Commands
EXAMPLE #4
To display information about the CEPA pool, type:
$ server_cepp server_2 -pool -info
server_2 :
pool_name = pool1
server_required = yes
access_checks_ignored = 0
req_timeout = 5000 ms
retry_timeout = 25000 ms
pre_events = OpenFileNoAccess, OpenFileRead
post_events = CreateFile,DeleteFile
post_err_events = CreateFile,DeleteFile
CEPP Servers:
IP = 10.171.10.115, state = ONLINE, vendor = Unknown
...
Where:
Value
pre_events
post_events
post_err_events
CEPP Servers
Definition
Sends notification before selected event occurs. An empty list indicates that no pre-event
messages are generated.
Sends notification after selected event occurs. An empty list indicates that no post-event
messages are generated.
Sends notification if selected event generates an error. An empty list indicates that no
post-error-event messages are generated.
IP addresses of the CEPA servers; state of the CEPA servers; vendor software installed on
CEPA servers.
EXAMPLE #5
To display statistics for the CEPA pool, type:
$ server_cepp server_2 -pool -stats
server_2 :
pool_name = pool1
Event Name
OpenFileWrite
CloseModified
Total Requests = 4
Min(us) = 604
Max(us) = 758
Average(us) = 664
Requests
2
2
Min(us)
659
604
Max(us)
758
635
Average(us)
709
620
server_cepp
463
The server Commands
server_checkup
Checks the configuration parameters, and state of a Data Mover and
its dependencies.
SYNOPSIS
server_checkup {<movername>|ALL}
{[-test <component> [-subtest <dependency>]][-quiet][-full]}
| -list
| -info {<component>|all}
DESCRIPTION
server_checkup performs a sanity check of a specific Data Mover
component and its dependencies by checking configuration
parameters, and the current state of the component and
dependencies.
A component is any basic feature that is available on the Data Mover,
for example, CIFS. A dependency is a configuration component of a
Data Mover that the proper operation of a Data Mover functionality
(like CIFS) is depending upon. This configuration component can be
owned by multiple Data Mover components. For example, proper
operation of a CIFS service depends on correctly specified DNS,
WINS, Antivirus, and so on.
server_checkup displays a report of errors and warnings detected in
the specified Data Mover component and its dependencies.
OPTIONS
No arguments
Performs a sanity check of all the components and all their
dependencies on the specified Data Mover or all Data Movers.
-test <component>
Performs a sanity check of a specific component and all of its
dependencies.
[-subtest <dependency>]
Performs a sanity check of a specific component and its specified
dependency only. If the dependency is not defined, executes the
command for all the dependencies of the component.
-quiet
Displays only the number of errors and warnings for the sanity
check.
-full
Provides a full sanity check of the specified Data Movers.
464
Command Line Interface Reference for File 8.1
The server Commands
-list
Lists all available components that can be checked on a Data Mover.
-info <component>
Lists all dependencies of the specified component, with details of
checks that can be performed on each dependency.
EXAMPLE #1
To list the available component in the Data Mover, type:
$ server_checkup server_2 -list
server_2 : done
REPV2
HTTPS
CIFS
FTPDS
EXAMPLE #2
To execute the check of the CIFS component, type:
$ server_checkup server_2 -test CIFS
server_2 :
------------------------------------Checks-------------------------------------Component CIFS :
ACL
:
Connection:
Credential:
DC
:
DFS
:
DNS
:
EventLog :
FS_Type
:
GPO
:
HomeDir
:
I18N
:
Kerberos :
LocalGrp :
NIS
:
NTP
:
Ntxmap
:
Security :
Server
:
Share
:
SmbList
:
Threads
:
UM_Client :
UM_Server :
UnsupOS
:
UnsupProto:
VC
:
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
the number of ACL per file system.....................*Pass
the load of TCP connections of CIFS................... Pass
the validity of credentials........................... Pass
the connectivity and configuration of the DCs.........*Fail
the DFS configuration files and DFS registry.......... Pass
the DNS configuration and connectivity to DNS servers. Pass
the configuration of Windows Event Logs............... Pass
if all file systems are all DIR3 type................. Pass
the GPO configuration................................. Pass
the configuration of home directory share............. Pass
the I18N mode and the Unicode/UTF8 translation tables. Pass
machine password update for Kerberos.................. Fail
the local groups database configuration............... Fail
the connectivity to the NIS servers, if defined....... Pass
the connectivity to theNTP servers, if defined........ Pass
the ntxmap configuration file......................... Pass
the CIFS security settings............................ Pass
the CIFS files servers configuration.................. Pass
the network shares database........................... Pass
the range availability of SMB ID......................*Pass
for CIFS blocked threads.............................. Pass
for the connectivity to usermapper servers, if any.... Pass
the consistency of usermapper database, if primary....*Pass
for unsupported client network OS..................... Pass
for unsupported client network protocols.............. Pass
the configuration to Virus Checker servers............ Pass
server_checkup
465
The server Commands
WINS
: Checking for the connectivity to WINS servers, if defined...... Pass
NB: a result with a '*' means that some tests were not executed. use -full to run
them
----------------------------------------------------------------------------------------------------------CIFS : Kerberos Warnings----------------------------Warning 17451974742: server_2 : No update of the machine password of server
'DM102-CGE1'. hold.
--> Check the log events to find out the reason of this issue.
Warning 17451974742: server_2 : No update of the machine password of server
'DM102-CGE0'. hold.
--> Check the log events to find out the reason of this issue.
---------------------------CIFS : LocalGrp Warnings----------------------------Warning 17451974726: server_2 : The local group 'Guests' of server 'DM102-CGE1'
contains an unmapped member: S-1-5-15-60415a8a-335a7a0d-6b635f23-202.The access
to some network resources may be refused.
--> According the configured resolver of your system (NIS, etc config files,
usermapper, LDAP...),add the missing members.
--------------------------------------------------------------------------------------------------------------CIFS : DC Errors--------------------------------Error 13160939577: server_2 : pingdc failed due to NT error ACCESS_DENIED at step
SAMR lookups
--> check server configuration and/or DC policies according to reported error.
Error 13160939577: server_2 : pingdc failed due to NT error ACCESS_DENIED at step
SAMR lookups
--> check server configuration and/or DC policies according to reported error.
--------------------------------------------------------------------------------
EXAMPLE #3
To execute only the check of the DNS dependency of the CIFS
component, type:
$ server_checkup server_2 -test CIFS -subtest DNS
server_2 :
------------------------------------Checks-------------------------------------Component CIFS :
DNS
: Checking the DNS configuration and connectivity to DNS servers. Pass
--------------------------------------------------------------------------------
466
Command Line Interface Reference for File 8.1
The server Commands
EXAMPLE #4
To list the available dependencies of the CIFS component, type:
$ server_checkup server_2 -info CIFS
server_2 :
done
COMPONENT
: CIFS
DEPENDENCY
: ACL
DESCRIPTION : Number of ACL per file system.
TESTS
:
In full mode, check if the number of ACL per file system doesn't exceed 90% of the
maximum limit.
COMPONENT
DEPENDENCY
DESCRIPTION
TESTS
Check if the
number.
: CIFS
: Connection
: TCP connection number
:
number of CIFS TCP connections doesn't exceed 80% of the maximum
COMPONENT
DEPENDENCY
DESCRIPTION
TESTS
Check if all
: CIFS
: Credential
: Users and groups not mapped
:
credentials in memory are mapped to a valid SID.
COMPONENT
: CIFS
DEPENDENCY
: DC
DESCRIPTION : Connectivity to the domain controllers
TESTS
:
Check the connectivity to the favorite DC (DCPing),
In full mode, check the connectivity to all DC of the domain,
Check if DNS site information are defined for each computer name,
Check if the site of each computer name has an available DC,
Check if trusted domain of each computer name can be reached,
Check the ds.useDCLdapPing parameter is enabled,
Check the ds.useADSite parameter is enabled.
COMPONENT
: CIFS
DEPENDENCY
: DFS
DESCRIPTION : DFS service configuration on computer names
TESTS
:
Check the DFS service is enabled in registry if DFS metadata exists,
Check the DFS metadata of each share with DFS flag are correct,
Check if share names in DFS metadata are valid and have the DFS flag,
Check if each DFS link is valid and loaded,
Check in the registry if the WideLink key is enabled and corresponds to a valid
share name.
COMPONENT
DEPENDENCY
DESCRIPTION
: CIFS
: DNS
: DNS domain configuration
server_checkup
467
The server Commands
TESTS
Check
Check
Check
Check
Check
:
if each DNS domain has at least 2 defined servers,
the connectivity to each DNS server of each DNS domain,
if each DNS server of each DNS domain supports really the DNS service,
the ds.useDSFile parameter (automatic discovery of DC),
the ds.useDSFile parameter is enabled if the directoryservice file exists.
COMPONENT
: CIFS
DEPENDENCY
: EventLog
DESCRIPTION : Event Logs parameters on servers
TESTS
:
Check if the pathnames of each event logs files are valid (application, system and
security),
Check if the maximum file size of each event logs file doesn't exceed 1GB,
Check if the retention time of each event logs file doesn't exceed 1 month.
COMPONENT
: CIFS
DEPENDENCY
: FS_Type
DESCRIPTION : DIR3 mode of filesystems
TESTS
:
Check if each file system is configured in the DIR3 mode.
COMPONENT
: CIFS
DEPENDENCY
: GPO
DESCRIPTION : GPO configuration on Win2K servers
TESTS
:
Check if the size of the GPO cache file doesn't exceed 10% of the total size of
the root file system,
Check the last modification date of the GPO cache file is up-to-date,
Check the cifs.gpo and cifs.gpoCache parameters have not been changed,
COMPONENT
: CIFS
DEPENDENCY
: HomeDir
DESCRIPTION : Home directory shares configuration
TESTS
:
Check if the home directory shares configuration file exists, the feature is
enabled,
Check if the home directory shares configuration file is optimized (40 lines
maximum),
Check the syntax of the home directory shares configuration file.
COMPONENT
: CIFS
DEPENDENCY
: I18N
DESCRIPTION : Internationalization and translation tables
TESTS
:
Check if computer name exists, the I18N mode is enabled,
Check the .etc_common file system is correctly mounted,
Check the syntax of the definition file of the Unicode characters,
Check the uppercase/lowercase conversion table of Unicode character is valid.
COMPONENT
DEPENDENCY
468
: CIFS
: Kerberos
Command Line Interface Reference for File 8.1
The server Commands
DESCRIPTION : Kerberos configuration
TESTS
:
Check the machine password update is enabled and up-to-date.
COMPONENT
: CIFS
DEPENDENCY
: LocalGrp
DESCRIPTION : Local groups and local users
TESTS
:
Check the local group database doesn't contain more than 80% of the maximum number
of servers,
Check if the servers in the local group database are all valid servers,
Check the state of the local group database (initialized and writable),
Check if the members of built-in local groups are all resolved in the domain,
Check the number of built-in local groups and built-in local users,
Check if the number of defined local users doesn't exceed 90% of the maximum
number.
COMPONENT
: CIFS
DEPENDENCY
: NIS
DESCRIPTION : Network Information System (NIS) configuration
TESTS
:
If NIS is configured, check at least 2 NIS servers are defined (redundancy check),
Check if each NIS server can be contacted on the network,
Check if each NIS server really supports the NIS service.
COMPONENT
: CIFS
DEPENDENCY
: NTP
DESCRIPTION : Network Time Protocol (NTP) configuration
TESTS
:
If NTP is configured, check at least 2 NTP servers are defined (redundancy check),
Check if each NIS server can be contacted on the network,
If computer names exist, check if NTP is configured and is running.
COMPONENT
:
DEPENDENCY
:
DESCRIPTION :
TESTS
:
Check the data
CIFS
Ntxmap
Checking the ntxmap.conf file.
consistency of the ntxmap configuration file.
COMPONENT
: CIFS
DEPENDENCY
: Security
DESCRIPTION : Security settings
TESTS
:
If the I18N mode is enabled, check the share/unix security setting is not in use,
Discourage to use the share/unix security setting,
Check the cifs.checkAcl parameter is enabled if the security setting is set to NT.
COMPONENT
: CIFS
DEPENDENCY
: Server
DESCRIPTION : Files servers
TESTS
:
Check if each CIFS server is configured with a valid IP interface,
server_checkup
469
The server Commands
Check
Check
Check
Check
if
if
if
if
each computer name has joined its domain,
each computer name is correctly registered in their DNS servers,
the DNS servers have the valid IP addresses of each computer name,
a DNS domain exists if at least one computer name exists,
COMPONENT
: CIFS
DEPENDENCY
: Share
DESCRIPTION : Network shares
TESTS
:
Check the available size and i-nodes on the root file system are at least 10% of
the total size,
Check the size of the share database doesn't exceed 30% of the total size of the
root file system,
Check if the pathname of each share is valid and is available,
Check if each server in the share database really exists,
Check if the I18N mode is enabled, all the share names are UTF-8 compatible,
Check the list of ACL of each share contains some ACE,
Check the length of each share name doesn't exceed 80 Unicode characters.
COMPONENT
: CIFS
DEPENDENCY
: SmbList
DESCRIPTION : 64k UID, TID and FID limits
TESTS
:
In full mode, check the 3 SMB ID lists (UID, FID and TID) don't exceed 90% of the
maximum ID number.
COMPONENT
: CIFS
DEPENDENCY
: Threads
DESCRIPTION : Blocked threads and overload
TESTS
:
Check CIFS threads blocked more than 5 and 30 seconds,
Check the maximum number of CIFS threads in use in the later 5 minutes doesn't
exceed 90% of the total number,
Check the number of threads reserved for Virus Checker doesn't exceed 20% of the
total number of CIFS threads.
COMPONENT
: CIFS
DEPENDENCY
: UM_Client
DESCRIPTION : Connectivity to the usermapper server
TESTS
:
If usermapper servers are defined, check each server can be contacted,
Check if usermapper servers are defined, NIS is not simultaneously activated.
COMPONENT
: CIFS
DEPENDENCY
: UM_Server
DESCRIPTION : Primary usermapper
TESTS
:
If a primary usermapper is defined
30% of the total size,
Check if configuration file is in
exceed 90%,
Check if configuration file is in
470
server
locally, check its database size doesn't exceed
use, the filling rate of the ranges doesn't
use, 2 ranges do not overlap,
Command Line Interface Reference for File 8.1
The server Commands
Check if secmap is enabled,
In full mode, check the SID/UID and SID/GID mappings and reverses are correct and
coherent.
COMPONENT
: CIFS
DEPENDENCY
: UnsupOS
DESCRIPTION : Client OS not supported
TESTS
:
Check for unsupported client network OS.
COMPONENT
: CIFS
DEPENDENCY
: UnsupProto
DESCRIPTION : Unsupported protocol commands detected
TESTS
:
Check for unsupported client network protocol commands.
COMPONENT
: CIFS
DEPENDENCY
: VC
DESCRIPTION : Virus checker configuration
TESTS
:
If VC is enabled, check the syntax of the VC configuration file,
Check if the VC 'enable' file and the VC configuration are compatible,
Check the number of VC servers. Make sure at least 2 servers are defined, for
redundancy,
Check if there are offline VC servers,
Check if the VC high watermark has not been reached,
Check the connection of VC servers to the Data Mover.
COMPONENT
: CIFS
DEPENDENCY
: WINS
DESCRIPTION : WINS servers.
TESTS
:
If NetBIOS names are defined, check if at least one WINS server is defined,
Check the number of WINS servers. check if two servers are defined for redundancy,
Check if each WINS server can be contacted on the network,
Check these servers are really WINS servers,
Check if the NetBIOS are correctly registered on the servers.
EXAMPLE #5
To execute additional tests, type:
$ server_checkup server_2 -full
server_2 :
------------------------------------Checks-------------------------------------Component REPV2 :
F_RDE_CHEC: Checking the F-RDE compatibilty of Repv2 sessions.............. Fail
Component HTTPS :
HTTP
: Checking the configuration of HTTP applications................ Pass
server_checkup
471
The server Commands
SSL
: Checking the configuration of SSL applications................. Fail
Component CIFS :
ACL
:
Connection:
Credential:
DC
:
DFS
:
DNS
:
EventLog :
FS_Type
:
GPO
:
HomeDir
:
I18N
:
Kerberos :
LDAP
:
LocalGrp :
NIS
:
NS
:
NTP
:
Ntxmap
:
Secmap
:
Security :
Server
:
Share
:
SmbList
:
Threads
:
UM_Client :
UM_Server :
UnsupOS
:
UnsupProto:
VC
:
WINS
:
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
the number of ACLs per file system....................
the load of CIFS TCP connections......................
the validity of credentials...........................
the connectivity and configuration of Domain Controlle
the DFS configuration files and DFS registry..........
the DNS configuration and connectivity to DNS servers.
the configuration of Windows Event Logs...............
if all file systems are in the DIR3 format............
the GPO configuration.................................
the configuration of home directory shares............
the I18N mode and the Unicode/UTF8 translation tables.
password updates for Kerberos.........................
the LDAP configuration................................
the database configuration of local groups............
the connectivity to the NIS servers...................
the naming services configuration.....................
the connectivity to the NTP servers...................
the ntxmap configuration file.........................
the SECMAP database...................................
the CIFS security settings............................
the CIFS file servers configuration...................
the network shares database...........................
the range availability of SMB IDs.....................
for CIFS blocked threads..............................
the connectivity to usermapper servers................
the usermapper server database........................
for unsupported client network operating systems......
for unsupported client network protocols..............
the configuration of Virus Checker servers............
the connectivity to WINS servers......................
Pass
Pass
Fail
Fail
Fail
Fail
Fail
Pass
Pass
Fail
Fail
Fail
Pass
Fail
Pass
Fail
Fail
Pass
Fail
Pass
Fail
Fail
Fail
Pass
Pass
Pass
Pass
Pass
Pass
Fail
if all file systems are in the DIR3 format............
the configuration of FTPD.............................
the connectivity to the NIS servers...................
the naming services configuration.....................
the connectivity to the NTP servers...................
the configuration of SSL applications.................
Pass
Fail
Pass
Fail
Fail
Fail
Component FTPDS :
FS_Type
FTPD
NIS
NS
NTP
SSL
:
:
:
:
:
:
Checking
Checking
Checking
Checking
Checking
Checking
------------------------------------------------------------------------------------------------------------HTTPS : SSL Warnings------------------------------Warning 17456169084: server_2 : The SSL feature 'DHSM' can not get certificate
from the persona default. Because this feature needs a certificate and a private
key, it can not start,
472
Command Line Interface Reference for File 8.1
The server Commands
--> Run the server_certificate command to generate a new key set and certificate
for this persona. Or run the appropriate command (like server_http for instance)
to set a correct persona for this SSL feature.
Warning 17456169084: server_2 : The SSL feature 'DIC' can not get certificate
from the persona default. Because this feature needs a certificate and a private
key, it can not start,
--> Run the server_certificate command to generate a new key set and certificate
for this persona. Or run the appropriate command (like server_http for instance)
to set a correct persona for this SSL feature.
Warning 17456169084: server_2 : The SSL feature 'DIC_S' can not get certificate
from the persona default. Because this feature needs a certificate and a private
key, it can not start,
--> Run the server_certificate command to generate a new key set and certificate
for this persona. Or run the appropriate command (like server_http for instance)
to set a correct persona for this SSL feature.
Warning 17456169084: server_2 : The SSL feature 'DIC_L' can not get certificate
from the persona default. Because this feature needs a certificate and a private
key, it can not start,
--> Run the server_certificate command to generate a new key set and certificate
for this persona. Or run the appropriate command (like server_http for instance)
to set a correct persona for this SSL feature.
Warning 17456169084: server_2 : The SSL feature 'DBMS_FILE_TRANSFER' can not get
certificate from the persona default. Because this feature needs a certificate and
a private key, it can not start,
--> Run the server_certificate command to generate a new key set and certificate
for this persona. Or run the appropriate command (like server_http for instance)
to set a correct persona for this SSL feature.
--------------------------CIFS : Credential Warnings---------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
------------------------------CIFS : DC Warnings-------------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
-----------------------------CIFS : DFS Warnings-------------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
server_checkup
473
The server Commands
---------------------------CIFS : EventLog Warnings----------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
---------------------------CIFS : HomeDir Warnings-----------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
-----------------------------CIFS : I18N Warnings------------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
---------------------------CIFS : Kerberos Warnings----------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
---------------------------CIFS : LocalGrp Warnings----------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
-----------------------------CIFS : NTP Warnings-------------------------------Warning 17456169044: server_2 : The Network Time Protocol subsystem (NTP) has
been stopped or is not connected to its server. It may cause potential errors
during Kerberos authentication (timeskew).
--> If the NTP service is not running, start it using the server_date command. If
it is not connected, check the IP address of the NTP server and make sure the NTP
service is up and running on the server. If needed, add another NTP server in the
configuration of the Data Mover. Use the server_date command to manage the NTP
service and the parameters on the Data Mover.
----------------------------CIFS : Secmap Warnings-----------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
474
Command Line Interface Reference for File 8.1
The server Commands
--> Start the CIFS server by executing the 'server_setup' command, and try again.
----------------------------CIFS : Server Warnings-----------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
----------------------------CIFS : Share Warnings------------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
---------------------------CIFS : SmbList Warnings-----------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
-----------------------------CIFS : WINS Warnings------------------------------Warning 17456168968: server_2 : The CIFS service is currently stopped. Many CIFS
sanity check tests cannot be done as all CIFS servers are currently disabled on
this Data Mover.
--> Start the CIFS server by executing the 'server_setup' command, and try again.
-----------------------------FTPDS : NTP Warnings------------------------------Warning 17456169044: server_2 : The Network Time Protocol subsystem (NTP) has
been stopped or is not connected to its server. It may cause potential errors
during Kerberos authentication (timeskew).
--> If the NTP service is not running, start it using the server_date command. If
it is not connected, check the IP address of the NTP server and make sure the NTP
service is up and running on the server. If needed, add another NTP server in the
configuration of the Data Mover. Use the server_date command to manage the NTP
service and the parameters on the Data Mover.
-----------------------------FTPDS : SSL Warnings------------------------------Warning 17456169084: server_2 : The SSL feature 'DHSM' can not get certificate
from the persona default. Because this feature needs a certificate and a private
key, it can not start,
--> Run the server_certificate command to generate a new key set and certificate
for this persona. Or run the appropriate command (like server_http for instance)
to set a correct persona for this SSL feature.
server_checkup
475
The server Commands
Warning 17456169084: server_2 : The SSL feature 'DIC' can not get certificate
from the persona default. Because this feature needs a certificate and a private
key, it can not start,
--> Run the server_certificate command to generate a new key set and certificate
for this persona. Or run the appropriate command (like server_http for instance)
to set a correct persona for this SSL feature.
Warning 17456169084: server_2 : The SSL feature 'DIC_S' can not get certificate
from the persona default. Because this feature needs a certificate and a private
key, it can not start,
--> Run the server_certificate command to generate a new key set and certificate
for this persona. Or run the appropriate command (like server_http for instance)
to set a correct persona for this SSL feature.
Warning 17456169084: server_2 : The SSL feature 'DIC_L' can not get certificate
from the persona default. Because this feature needs a certificate and a private
key, it can not start,
--> Run the server_certificate command to generate a new key set and certificate
for this persona. Or run the appropriate command (like server_http for instance)
to set a correct persona for this SSL feature.
Warning 17456169084: server_2 : The SSL feature 'DBMS_FILE_TRANSFER' can not get
certificate from the persona default. Because this feature needs a certificate and
a private key, it can not start,
--> Run the server_certificate command to generate a new key set and certificate
for this persona. Or run the appropriate command (like server_http for instance)
to set a correct persona for this SSL feature.
---------------------------------------------------------------------------------------------------------REPV2 : F_RDE_CHECK Errors---------------------------Error 13160415855: server_2 : For the Replication session: rep1,
Data Mover version on the source fs: 5.6.47
Data Mover version on the destination fs: 5.5.5
Minimum required Data Mover version on the destination fs: 5.6.46
The Data Mover version on the destination file system is incompatible with the
Data Mover version on the source file system. After data transfer, the data in the
destination file system may appear to be corrupt, even though the data is in fact
intact.
Upgrade the Data Mover where the destination file system resides to at least
5.6.46.
Error 13160415855: server_2 : For the Replication session:rsd1,
F-RDE version on the source fs: 5.6.46
F-RDE version on the destination fs: 5.5.5
Minimum required F-RDE version on the destination fs: 5.6.46
The F-RDE versions are incompatible.
After data transfer, the data in the dst FS may appear to be corrupt.
476
Command Line Interface Reference for File 8.1
The server Commands
--> Upgrade the DataMover where the dst fs resides to atleast the version on the
source.
Error 13160415855: server_2 : For the Replication session:rsd2,
F-RDE version on the source fs: 5.6.46
F-RDE version on the destination fs: 5.5.5
Minimum required F-RDE version on the destination fs: 5.6.46
The F-RDE versions are incompatible.
After data transfer, the data in the dst FS may appear to be corrupt.
--> Upgrade the DataMover where the dst fs resides to atleast the version on the
source.
Error 13160415855: server_2 : For the Replication session:rsd3, F-RDE version on
the source fs: 5.6.46
F-RDE version on the destination fs: 5.5.5 Minimum required
F-RDE version on the destination fs: 5.6.46
The F-RDE versions are incompatible.
After data transfer, the data in the dst FS may appear to be corrupt.
--> Upgrade the DataMover where the dst fs resides to atleast the version on the
source.
------------------------------HTTPS : SSL Errors-------------------------------Error 13156876314: server_2 : The persona 'default' contains nor certificate
neither private keys sets. So, this persona can not be used by a SSL feature on
the Data Mover.
--> Run the server_certificate command to generate a new key set and certificate
for this persona.
------------------------------CIFS : DNS Errors--------------------------------Error 13161070637: server_2 : The DNS service is currently stopped and does not
contact any DNS server. The CIFS clients may not be able to access the Data Mover
on the network.
--> Start the DNS service on the Data Mover, using the 'server_dns' command.
-------------------------------CIFS : NS Errors--------------------------------Error 13156352011: server_2 : None of the naming services defined for the entity
'host' in nsswitch.conf is configured.
--> Make sure each entity (e.g. host, passwd..) in the nsswitch.conf file contains
naming services, (e.g. local files, NIS or usermapper), and make sure these
services are configured. Use the corresponding commands like server_nis,
server_dns or server_ldap to make sure they are configured.
Error 13156352011: server_2 : None of the naming services defined for the entity
'group' in nsswitch.conf is configured.
--> Make sure each entity (e.g. host, passwd..) in the nsswitch.conf file contains
naming services, (e.g. local files, NIS or usermapper), and make sure these
services are configured. Use the corresponding commands like server_nis,
server_dns or server_ldap to make sure they are configured.
server_checkup
477
The server Commands
Error 13156352011: server_2 : None of the naming services defined for the entity
'netgroup' in nsswitch.conf is configured.
--> Make sure each entity (e.g. host, passwd..) in the nsswitch.conf file contains
naming services, (e.g. local files, NIS or usermapper), and make sure these
services are configured. Use the corresponding commands like server_nis,
server_dns or server_ldap to make sure they are configured.
-----------------------------FTPDS : FTPD Errors-------------------------------Error 13156876314: server_2 : The persona 'default' contains nor certificate
neither private keys sets. So, this persona can not be used by a SSL feature on
the Data Mover.
--> Run the server_certificate command to generate a new key set and certificate
for this persona.
------------------------------FTPDS : NS Errors--------------------------------Error 13156352011: server_2 : None of the naming services defined for the entity
'host' in nsswitch.conf is configured.
--> Make sure each entity (e.g. host, passwd..) in the nsswitch.conf file contains
naming services, (e.g. local files, NIS or usermapper), and make sure these
services are configured. Use the corresponding commands like server_nis,
server_dns or server_ldap to make sure they are configured.
Error 13156352011: server_2 : None of the naming services defined for the entity
'group' in nsswitch.conf is configured.
--> Make sure each entity (e.g. host, passwd..) in the nsswitch.conf file contains
naming services, (e.g. local files, NIS or usermapper), and make sure these
services are configured. Use the corresponding commands like server_nis,
server_dns or server_ldap to make sure they are configured.
Error 13156352011: server_2 : None of the naming services defined for the entity
'netgroup' in nsswitch.conf is configured.
--> Make sure each entity (e.g. host, passwd..) in the nsswitch.conf file contains
naming services, (e.g. local files, NIS or usermapper), and make sure these
services are configured. Use the corresponding commands like server_nis,
server_dns or server_ldap to make sure they are configured.
------------------------------FTPDS : SSL Errors-------------------------------Error 13156876314: server_2 : The persona 'default' contains nor certificate
neither private keys sets. So, this persona can not be used by a SSL feature on
the Data Mover.
--> Run the server_certificate command to generate a new key set and certificate
for this persona.
-------------------------------------------------------------------------------Total :
14 errors, 25 warnings
478
Command Line Interface Reference for File 8.1
The server Commands
EXAMPLE #6
To display only the number of errors and warnings for a Data Mover
and dependency, type:
$ server_checkup server_2 -quiet
server_2 :
------------------------------------Checks-------------------------------------Component REPV2 :
F_RDE_CHEC: Checking the F-RDE compatibilty of Repv2 sessions.............. Fail
Component HTTPS :
HTTP
SSL
: Checking the configuration of HTTP applications................ Pass
: Checking the configuration of SSL applications................. Pass
Component CIFS :
ACL
:
Connection:
Credential:
DC
:
DFS
:
DNS
:
EventLog :
FS_Type
:
GPO
:
HomeDir
:
I18N
:
Kerberos :
LDAP
:
LocalGrp :
NIS
:
NS
:
NTP
:
Ntxmap
:
Secmap
:
Security :
Server
:
Share
:
SmbList
:
Threads
:
UM_Client :
UM_Server :
UnsupOS
:
UnsupProto:
VC
:
WINS
:
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
Checking
the number of ACLs per file system....................*Pass
the load of CIFS TCP connections...................... Pass
the validity of credentials........................... Fail
the connectivity and configuration of Domain Controlle Fail
the DFS configuration files and DFS registry.......... Fail
the DNS configuration and connectivity to DNS servers. Fail
the configuration of Windows Event Logs............... Fail
if all file systems are in the DIR3 format............ Pass
the GPO configuration................................. Pass
the configuration of home directory shares............ Fail
the I18N mode and the Unicode/UTF8 translation tables. Fail
password updates for Kerberos......................... Fail
the LDAP configuration................................ Pass
the database configuration of local groups............ Fail
the connectivity to the NIS servers................... Pass
the naming services configuration..................... Fail
the connectivity to the NTP servers................... Fail
the ntxmap configuration file......................... Pass
the SECMAP database................................... Fail
the CIFS security settings............................ Pass
the CIFS file servers configuration................... Fail
the network shares database........................... Fail
the range availability of SMB IDs.....................*Pass
for CIFS blocked threads.............................. Pass
the connectivity to usermapper servers................ Pass
the usermapper server database........................*Pass
for unsupported client network operating systems...... Pass
for unsupported client network protocols.............. Pass
the configuration of Virus Checker servers............ Pass
the connectivity to WINS servers...................... Fail
Component FTPDS :
server_checkup
479
The server Commands
FS_Type
FTPD
NIS
NS
NTP
SSL
:
:
:
:
:
:
Checking
Checking
Checking
Checking
Checking
Checking
if all file systems are in the DIR3 format............
the configuration of FTPD.............................
the connectivity to the NIS servers...................
the naming services configuration.....................
the connectivity to the NTP servers...................
the configuration of SSL applications.................
Pass
Fail
Pass
Fail
Fail
Pass
NB: a result with a '*' means that some tests were not executed. use -full to run
them
-------------------------------------------------------------------------------Total :
12 errors, 14 warnings
------------------------------------Checks--------------------------------------
480
Command Line Interface Reference for File 8.1
The server Commands
server_cifs
Manages the CIFS configuration for the specified Data Movers or
Virtual Data Movers (VDMs).
SYNOPSIS
server_cifs {<movername>|ALL} [<options>]
'options' can be one of the following:
| -option {{audit [,user=<user_name>][,client=<client_name>][,full]}
| homedir[=NO]}
| -add netbios=<netbios_name>,domain=<domain_name>[,alias=<alias_name>...]
[,hidden={y|n} [[,interface=<if_name>[,wins=<ip>[:<ip>]]]...]
[,local_users][-comment <comment>]
| -add
compname=<comp_name>,domain=<full_domain_name>[,alias=<alias_name>...]
[,hidden={y|n}][,authentication={kerberos|all}]
[,netbios=<netbios_name>][[,interface=<if_name>[,wins=<ip>[:<ip>]]
[,dns=<if_suffix>]]...][,local_users][-comment <comment>]
|-add standalone=<netbios_name>,workgroup=<workgroup_name>
[,alias=<alias_name>...][,hidden={y|n}]
[[,interface=<if_name>[,wins=<ip>[:<ip>]]...][,local_users]
[-comment <comment>]
| -rename -netbios <old_name> <new_name>
| -Join compname=<comp_name>,domain=<full_domain_name>,admin=<admin_name>
[,ou=<organizational_unit>]
[-option {reuse|resetserverpasswd|addservice=nfs}]
| -Unjoin compname=<comp_name>,domain=<full_domain_name>,admin=<admin_name>
| -add security={NT|UNIX|SHARE} [,dialect=<dialect_name>]
| -add wins=<ip_addr>[,wins=<ip_addr>...]
| -add usrmapper=<ip_addr>[,usrmapper=<ip_addr>...]
| -Disable <interface>[,<interface>...]
| -Enable <interface>[,<interface>...]
| -delete netbios=<netbios_name> [-remove_localgroup]
[,alias=<alias_name>...][,interface=<if_name>]
| -delete compname=<comp_name> [-remove_localgroup]
[,alias=<alias_name>...][,interface=<if_name>]
| -delete wins=<ip_addr>[,wins=<ip_addr>...]
| -delete usrmapper=<ip_addr>[,usrmapper=<ip_addr>...]
| -delete standalone=<netbios_name> [-remove_localgroup]
[,alias=<alias_name>...][,interface=<if_name>]
| -update {<share_name>|<path>} [mindirsize=<size>][force]
| -Migrate {<fs_name> -acl|<netbios_servername> -localgroup}
<src_domain>{:nb=<netbios>|:if=<interface>}
<dst_domain>{:nb=<netbios>|:if=<interface>}
| -Replace {<fs_name> -acl|<netbios_servername> -localgroup}
{:nb=<netbios>|:if=<interface>}
| -smbhash
{-hashgen <path> [-recursive] [-minsize <size>]
| -hashdel <path> [-recursive]
server_cifs
481
The server Commands
| -abort <id>
| -info
| -fsusage <fs_name>
| -exclusionfilter <filter>
| -audit {enable|disable} [-task] [-service] [-access]
| -service {enable|disable}
| -cleanup <fs_name> [-all |-unusedfor <days>|-unusedsince
<date>}}
| -setspn {-list [server=<full_comp_name>]
| -add <SPN> compname=<comp_name>,domain=<full_domain_name>,
admin=<admin_name>
| -delete <SPN> compname=<comp_name>,domain=<full_domain_name>,
admin=<admin_name>}
}
}
DESCRIPTION
server_cifs manages the CIFS configuration for the specified
<movername> which can be the physical Data Mover or VDMs.
Most command options are used with both VDMs and physical Data
Movers, whereas others are only used with physical Data Movers.
Options available for physical Data Movers only are:
-add security/dialect…
-add/delete usrmapper
-enable/disable interface
The ALL option executes the command for all Data Movers.
OPTIONS
No arguments
Displays the CIFS protocol configuration. Certain inputs are not casesensitive; however, variables may be automatically converted to
uppercase.
<options>
CIFS options include:
-option audit
Audits the CIFS configuration by testing for live connections to a
Data Mover.
[,user=<user_name>][,client=<client_name>] [,full]
Audits the live connections created when the session is initiated
by the specified <client_name> or audits the live connections for
those owned by the specified <user_name>. The full option can
be used to identify open files. The <client_name> can be a string
or an IPV4 address and the <user_name> can be a string of
maximum 20 characters.
482
Command Line Interface Reference for File 8.1
The server Commands
-option homedir[=NO]
Enables and disables (default) the home directory feature. The Data
Mover reads information from the homedir map file.
-add netbios=<netbios_name>, domain=<domain_name>
Configures a Windows NT 4.0-like CIFS server on a Data Mover,
assigning the specified <netbios_name> and <domain_name> to the
server. The domain name is limited to 15 bytes.
!
CAUTION
Each NetBIOS name must be unique to the domain and the Data
Mover.
[,alias=<alias_name>...]
Assigns a NetBIOS alias to the <netbios_name> associated with the
NetBIOS name. The <alias_name> must:
• Be unique on a Data Mover
• Be limited to 15 bytes
• Not begin with an @ (at sign) or - (dash) character
• Not include spaces, tab characters, or the following symbols: /
\ : ; , = * +|[] ? < > "
[,hidden={y|n}]
By default, the <netbios_name> is displayed in the Network
Neighborhood. If hidden=y is specified, the <netbios_name>
does not appear.
[[,interface=<if_name>[,wins=<ip>[:<ip>]]]...]
Specifies a logical IP interface for the CIFS server in the Windows
NT 4.0 domain and associates up to two WINS IP addresses with
each interface. The interface name is case-sensitive.
Note: When configuring a CIFS server without any interfaces for a Data
Mover, it becomes the default CIFS server and is available on all
interfaces not used by other CIFS servers. The default CIFS server can be
deleted at any time. It is recommended that IP interfaces should always
be specified. VDMs do not have default CIFS servers.
[,local_users]
Enables local user support that allows the creation of a limited
number of local user accounts on the CIFS server. When this
command executes, type and confirm a password that is assigned
to the local Administrator account on the CIFS server. In addition
server_cifs
483
The server Commands
to the Administrator account, a Guest account is also created. The
Guest account is disabled by default. The Administrator account
password must be changed before the Administrator can log in to
the CIFS server.
After initial creation of the stand-alone server, the local_users
option resets the local Administrator account password. The
password can only be reset if it has not been changed through a
Windows client. If the password has already been changed
through Windows, the reset will be refused.
[-comment <comment>]
Assigns a comment to the configuration. The comment is
delimited by quotes. Comment length is limited to 48 bytes
(represented as 48 ASCII characters or a variable number of
Unicode multibyte characters) and cannot include colons since
they are recognized as delimiters.
-add compname=<comp_name>,domain= <full_domain_name>
Configures a CIFS server as the <comp_name> in the specified
Windows Active Directory workgroup. A default NetBIOS name is
automatically assigned to the <comp_name>. Since the default for
<netbios_name> is derived from the <comp_name>, the
<comp_name> must not contain any characters that are invalid for a
<netbios_name>.
In the case of disjointed namespaces, you must use the fully qualified
domain name for the <comp_name>. For example, for a disjointed
namespace, you must always specify the fully qualified domain
name (FQDN) with the computer name when joining a CIFS server to
a domain, that is, dm112-cge0.emc.com, not just dm112-cge0.
The <comp_name> is limited to 63 bytes. The fully qualified domain
name is limited to 155 bytes. The <full_domain_name> must contain
a dot (.). There cannot be a @ (at sign) or - (dash) character. The name
also cannot include spaces, tab characters, or the symbols: / \ : ; , = *
+|[] ? < > "
!
CAUTION
Each computer name must be unique to the domain and the Data
Mover.
Note: Using International Character Sets for File provides details. Only
Windows NT security mode can be configured when UTF-8 is enabled.
484
Command Line Interface Reference for File 8.1
The server Commands
[,alias=<alias_name>...]
Assigns an alias to the NetBIOS name. The <alias_name> must:
•
•
•
•
Be unique on a Data Mover
Be limited to 15 bytes
Not begin with an @ (at sign) or - (dash) character
Not include spaces, tab characters, or the following symbols: /
\ : ; , = * +|[] ? < > "
[,hidden={y|n}]
By default, the computer name appears in the Network
Neighborhood. If hidden=y is specified, then the computer name
does not appear.
[,authentication={kerberos|all}]
Specifies the type of user authentication. The kerberos option
limits the server usage to Kerberos authentication; the all option
(default) allows both Kerberos and NTLM authentication.
[,netbios=<netbios_name>]
Specifies a <netbios_name> for the <comp_name> in place of the
default. The default for <netbios_name> is assigned
automatically and is derived from the first 15 bytes of the
<comp_name>. The <netbios_name> cannot begin with an @ (at
sign) or - (dash) character. The name also cannot include spaces,
tab characters, or the symbols: / \ : ; , = * +|[] ? < > "
[[,interface=<if_name>[,wins=<ip>[:<ip>]]]...]
Specifies a logical IP interface for the CIFS server in the Active
Directory domain and associates up to two WINS IP addresses
with each interface. The interface name is case-sensitive.
Note: When configuring a CIFS server without any interfaces for a Data
Mover, it becomes the default CIFS server and is available on all
interfaces not used by other CIFS servers. The default CIFS server can be
deleted at any time. It is recommended that IP interfaces should always
be specified. VDMs do not have default CIFS servers.
[,dns=<if_suffix>]
Specifies a different DNS suffix for the interface for DNS updates.
By default, the DNS suffix is derived from the domain. This DNS
option does not have any impact on the DNS settings of the Data
Mover.
server_cifs
485
The server Commands
[,local_users]
Enables local user support that allows the creation of a limited
number of local user accounts on the CIFS server. When this
command executes, type and confirm a password that is assigned
to the local Administrator account on the CIFS server. In addition
to the Administrator account, a Guest account is also created. The
Guest account is disabled by default. The Administrator account
password must be changed before the Administrator account can
log in to the CIFS server.
After initial creation of the stand-alone server, the local_users
option resets the local Administrator account password. The
password can only be reset if it has not been changed through a
Windows client. If the password has already been changed
through Windows, the reset will be refused.
[-comment <comment>]
Assigns a comment to the configuration. The comment is
delimited by quotes. Comment length is limited to 48 bytes
(represented as 48 ASCII characters or a variable number of
Unicode multibyte characters) and cannot include colons since
they are recognized as delimiters.
-add standalone=<netbios_name>, workgroup=
<workgroup_name>
Creates or modifies a stand-alone CIFS server on a Data Mover,
assigning the specified <netbios_name> and <workgroup_name>
to the server. The NetBIOS and workgroup names are limited to
15 bytes. When creating a stand-alone CIFS server for the first
time, the local_users option must be typed, or the command will
fail. It is not required when modifying the CIFS server. A
stand-alone CIFS server does not require any Windows domain
infrastructure. A stand-alone server has local user accounts on the
Data Mover and NTLM is used to authenticate users against the
local accounts database.
!
CAUTION
Each NetBIOS name must be unique to the workgroup and the
Data Mover.
[,alias=<alias_name>...]
Assigns an alias to the NetBIOS name. The <alias_name> must:
• Be unique on a Data Mover
• Be limited to 15 bytes
486
Command Line Interface Reference for File 8.1
The server Commands
• Not begin with an @ (at sign) or - (dash) character
• Not include spaces, tab characters, or the following symbols: /
\ : ; , = * +|[] ? < > "
[,hidden={y|n}]
By default, the <netbios_name> is displayed in the Network
Neighborhood. If hidden=y is specified, the <netbios_name>
does not appear.
[[,interface=<if_name>[,wins=<ip>[:<ip>]]]...]
Specifies a logical IP interface for the CIFS server and associates
up to two WINS IP addresses with each interface. The interface
name is case-sensitive.
Note: When configuring a CIFS server without any interfaces for a Data
Mover, it becomes the default CIFS server and is available on all
interfaces not used by other CIFS servers. The default CIFS server can be
deleted at any time. It is recommended that IP interfaces should always
be specified. VDMs do not have default CIFS servers.
[,local_users]
Enables local user support that allows the creation of a limited
number of local user accounts on the CIFS server. When this
command executes, type and confirm a password that is assigned
to the local Administrator account on the CIFS server. In addition
to the Administrator account, a Guest account is also created. The
Guest account is disabled by default. The Administrator account
password must be changed before the Administrator can log in to
the CIFS server.
After initial creation of the stand-alone server, the local_users
option resets the local Administrator account password. The
password can only be reset if it has not been changed through a
Windows client. If the password has already been changed
through Windows, the reset will be refused.
[-comment <comment>]
Assigns a comment to the configuration. The comment is
delimited by quotes. Comment length is limited to 48 bytes
(represented as 48 ASCII characters or a variable number of
Unicode multibyte characters) and cannot include colons since
they are recognized as delimiters.
server_cifs
487
The server Commands
-rename -netbios <old_name> <new_name>
Renames a NetBIOS name. For Windows Server, renames a
Compname after the CIFS server is unjoined from the domain.
Note: Before performing a rename, the new NetBIOS name must be added to
the domain using the Windows Server Users and Computers MMC snap-in.
-Join compname=<comp_name>,domain=
<full_domain_name>,admin=<admin_name>
Creates an account for the CIFS server in the Active Directory. By
default, the account is created under the domain root as
ou=Computers,ou=EMC VNX.
!
CAUTION
Before performing a -Join, CIFS service must be started using
server_setup.
The <comp_name> is limited to 63 bytes and represents the name of
the server to be registered in DNS. The <full_domain_name> is the
full domain name to which the server belongs. This means the name
must contain at least one period (.). The <admin_name> is the logon
name of the user with the right to create and manage computer
accounts in the Organizational Unit that the CIFS server is being
joined to. If a domain is given as part of the admin username it
should be of the form: admin@FQDN. If no domain is given the
admin user account is assumed to be part of the domain the CIFS
Server is being joined to. The user is prompted to type a password for
the admin account.
An Active Directory and a DNS can have the same domain name, or a
different domain name (disjoint namespace). For each type of Active
Directory and DNS domain relationship, specific VNX parameters
and command values must be used. For example, for a disjoint
namespace, you must always specify the fully qualified domain
name (FQDN) with the computer name when joining a CIFS server to
a domain, that is, dm112-cge0.emc.com, not just dm112-cge0.
!
CAUTION
Time services must be synchronized using server_date.
[,ou=<organizational_unit>]
Specifies the organizational unit or container where computer
accounts are created in the Active Directory. By default, computer
488
Command Line Interface Reference for File 8.1
The server Commands
accounts are created in an organizational unit called Computers.
The name must be in a valid distinguished name format, for
example, ou="cn=My_mover". The name may contain multiple
nested elements, such as ou="cn=comp:ou=mach". The colon (:)
must be used as a separator for multiple elements. By default,
ou=Computers,ou=EMC VNX is used. The organizational unit
name is limited to 256 bytes.
[-option {reuse|resetserverpasswd|addservice= nfs}]
The reuse option reuses the existing computer account with the
original principal or joins a CIFS server to the domain where the
computer account has been created manually.
The resetserverpasswd option resets the CIFS server password
and encryption keys on a domain controller. This option could be
used for security reasons, such as changing the server password
in the Kerberos Domain Controller.
The addservice option adds the NFS service to the CIFS server,
making it possible for NFS users to access the Windows Kerberos
Domain Controller. Before adding NFS service, the
<comp_name> must already be joined to the domain, otherwise
the command will fail.
-Unjoin compname=<comp_name>,domain=
<full_domain_name>,admin=<admin_name>
Deletes the account for the CIFS server as specified by its
<comp_name> from the Active Directory database. The user is
prompted to type a password for the admin account.
-add security={NT|UNIX|SHARE}
Defines the user authentication mechanism used by the Data Mover
for CIFS services. NT (default) security mode uses standard
Windows domain based user authentication. The local password and
group files, NIS, EMC Active Directory UNIX users and groups
extension, or UserMapper are required to translate Windows user
and group names into UNIX UIDs and GIDs. NT security mode is
required for the Data Mover to run Windows 2000 or later native
environments. Unicode should be enabled for NT security mode.
!
CAUTION
EMC does not recommend the use of UNIX or SHARE security
modes.
server_cifs
489
The server Commands
For UNIX security mode, the client supplies a username and a
plain-text password to the server. The server uses the local (password
or group) file or NIS to authenticate the user. To use UNIX security
mode, CIFS client machines must be configured to send user
passwords to the Data Mover unencrypted in plain text. This requires
a registry or security policy change on every CIFS client machine.
For VDM, UNIX and SHARE security modes are global to the Data
Mover and cannot be set for each VDM. Unicode must not be
enabled.
For SHARE security mode, clients supply a read-only or read-write
password for the share. No user authentication is performed using
SHARE security. Since this password is sent through the network in
clear text, you must modify the Client Registry to allow for clear text
passwords.
!
CAUTION
Before adding or changing a security mode, CIFS service must be
stopped using server_setup, then restarted once options have been
set.
[,dialect=<dialect_name>]
Specifies a dialect. Optimum dialects are assigned by default.
Options include CORE, COREPLUS, LANMAN1 (default for
UNIX and SHARE security modes), LANMAN2, and NT1 (which
represents SMB1 and is the default for NT security mode), SMB2
and SMB3.
• SMB1 dialect is NT1 dialect.
• SMB2 dialect means max dialect in SMB2 which is SMB2.1.
SMB2.0 or SMB2.1 can be specified explicitly to refine the
dialect revision.
• SMB3 dialect means max dialect in SMB3 which is SMB3.0.
SMB3.0 can be specified explicitly.
Note: SMB3 is enabled by default.
-add wins=<ip_addr>[,wins=<ip_addr>...]
Adds the WINS servers to the CIFS configuration. The list of WINS
servers is processed in the order in which they are added. The first
one is the preferred WINS server. If after 1500 milliseconds, the first
WINS server times out, the next WINS server on the list is used.
490
Command Line Interface Reference for File 8.1
The server Commands
-add usrmapper=<ip_addr>[,usrmapper=<ip_addr>...]
Adds the IP address(es) of a secondary Usermapper hosts to the CIFS
configuration. A single IP address can point to a primary or
secondary Usermapper host. If you are using distributed
Usermappers, up to eight subsequent IP addresses can point to
secondary Usermapper hosts.
-Disable <interface> [<interface>,...]
Disables the specified IP interfaces for CIFS service. Interface names
are case-sensitive. All unused interfaces should be disabled.
-Enable <interface> [<interface>,...]
Enables the specified IP interfaces for CIFS service. Interface names
are case-sensitive.
-delete standalone=<netbios_name>
[-remove_localgroup][,alias=<alias_name>...][,interface=
<if_name>]
Deletes the stand-alone CIFS server as identified by its NetBIOS
name from the CIFS configuration of the Data Mover.
-delete netbios=<netbios_name>
[-remove_localgroup][,alias=<alias_name>...][,interface=
<if_name>]
Deletes the CIFS server as identified by its NetBIOS name from the
CIFS configuration of the Data Mover.
-delete compname=<comp_name> [-remove_localgroup]
[,alias=<alias_name>...][,interface=<if_name>]
Deletes the CIFS server as identified by its compname from the CIFS
configuration of the Data Mover. This does not remove the account
from the Active Directory. It is recommended that an -Unjoin be
executed prior to deleting the computer name.
!
CAUTION
The -remove_localgroup option permanently deletes the local
group information of the CIFS server from the permanent storage
of the Data Mover. The alias and interface options delete the alias
and the interface only, however, the CIFS server exists. The alias
and interface options can be combined in the same delete
command.
-delete wins=<ip_addr>[,wins=<ip_addr>...]
Deletes the WINS servers from the CIFS configuration.
server_cifs
491
The server Commands
-delete usrmapper=<ip_addr>[,usrmapper= <ip_addr>...]
Deletes the IP addresses of a secondary Usermapper hosts from the
CIFS configuration.
-update {<share_name>|<path>}
Updates the attributes and their CIFS names for COMPAT file
systems. For every file system, CIFS maintains certain attributes for
which there are no NFS equivalents. Updating CIFS attributes
updates file attributes and CIFS names by searching the
subdirectories of the defined share or path, generating a listing of
Microsoft clients filenames (M8.3 and M256), and converting them to
a format that CIFS supports. It is not necessary to use this command
for DIR3 file systems. Options include:
[mindirsize=<size>]
Updates the directories with the minimum size specified. Size
must be typed in multiples of 512 bytes. A value of 0 ensures that
all directories are rebuilt.
[force]
Forces a previous update to be overwritten.
!
CAUTION
The initial conversion of a directory can take considerable time
when the directory contains a large number of files. Although the
process is designed to take place in the background, an update
should be run only during periods of light system usage.
-Migrate {<fs_name> -acl|<netbios_servername>
-localgroup}<src_domain>{:nb=<netbios>|:if=<interface>}<
dst_domain>{:nb=<netbios>|:if=<interface>}
Updates all security IDs (SIDs) from a <src_domain> to the SIDs of a
<dst_domain> by matching the user and group account names in the
source domain to the user and group account names in the
destination domain. The interface that is specified in this option
queries the local server, then its corresponding source and target
Domain Controllers to look up each object’s SID.
If -acl is specified, all secure IDs in the ACL database are migrated for
the specified file system.
The -localgroup option must be used to migrate the SID members of
local group defined for the specified NetBIOS name.
On the source domain, an interface specified to issue a lookup of the
SID is defined by either the NetBIOS name or the interface name. On
492
Command Line Interface Reference for File 8.1
The server Commands
the destination domain, an interface specified to issue a lookup of the
SID is defined by either the NetBIOS name or the interface name.
-Replace {<fs_name> -acl|<netbios_servername>
-localgroup}{:nb=<netbios>|:if=<interface>}
Replaces the history SIDs from the old domain with the new SIDS in
the new domain. An interface that can be specified to issue a lookup
of the SIDs is defined by the interface name or the NetBIOS name.
The -localgroup option must be used to migrate the SID members of
the local group defined for the specified NetBIOS name. When the
-Replace option is used, the user or group migrated in the new
domain keeps their old SID in addition to the new SID created in the
new domain.
The -localgroup option does the same kind of migration for a
specified NetBIOS name in the local groups (instead of the ACL in a
file system for the history argument).
-smbhash -hashgen <path> [-recursive][-minsize <size>]
Triggers the generation of all SMB Hash Files for this path. Both
BranchCache V1 and BranchCache V2 hash files are generated. This
path is an absolute path from the root of the VDM.
If the path is a file, only the SMB Hash File for this file will be
generated.
If the path is a directory, then SMB Hash File for all files will be
generated in this directory. Additionally, if the -recursive option is
specified, then the SMB Hash File for all files will be generated
recursively inside the sub-directories.
By default, only files greater than 64KB are considered. If -minsize
option is specified, then all files greater or equal to the specified size
in KB will be considered. Any size specified smaller than 64 KB will
be ignored. SMB Hash Files are generated only if they are missing or
obsolete.
The hash file generation is asynchronous, so the command will reply
immediately. Use -info or check the system event log to monitor if the
request has been completed.
server_cifs
493
The server Commands
-smbhash -hashdel <path> [-recursive]
Triggers the deletion of all SMB Hash Files for this path. Both
BranchCache V1 and BranchCache V2 hash files are deleted. This
path is an absolute path from the root of the VDM.
If the path is a file, only the SMB Hash File for this file will be deleted.
If the path is a directory, then SMB Hash File will be deleted for all
files in this directory. Additionally, if the -recursive option is
specified, then the SMB Hash File will be deleted recursively for all
files inside the sub-directories.
The hash file deletion is asynchronous, so the command will reply
immediately. Use -info or check the system event log to monitor if the
request has been completed.
-smbhash -abort <id>
Cancels the pending or ongoing request (generation or deletion)
provided its ID is given. Request ID is received from the output of the
command -info.
-smbhash -info
Get all kinds of information relative to the hash generation service:
◆
The list of pending requests with their ID.
◆
The list of under processing requests with their ID.
◆
Values of the parameters which are actually in use.
◆
Value of the GPO setting taken into account for each server.
◆
Statistics
-smbhash -fsusage <fs_name>
Displays the SMB Hash File disk usage of the specified file system.
The return values are:
◆
Total size in bytes of the file system
◆
Usage in bytes of the SMB Hash Files of the file system
◆
Usage in percentage of the file system of the SMB Hash Files
-smbhash -exclusionfilter <filter>
Files which match the exclusion filter will not have a SMB Hash File
generated. This is to avoid waste of resources spent on files that
frequently change like temporary files.
494
Command Line Interface Reference for File 8.1
The server Commands
This command directly modifies the parameter ExclusionFilter as
defined with the specified format:
Type: REG_STRING
Meaning: Hash files are not generated for files which match one
of the specified filters. The comparison between this parameter
and the filename is done case less. Any change is taken into
account immediately.
Values: Default is no filter. A filter is a list of items separated by a
character “:”. Each item is made of:
• Any valid character for a filename
• *: means any string
• ?:means any character
-smbhash -audit { enable | disable } [-service] [-task]
[-access]
Enables the generation of audits in the smbhash event log. By default,
it is not validated. The parameters are one of the following:
◆
enable: Enables generation of specified event. If no event is
specified in the optional list, all events are enabled.
◆
disable: Disable generation of specified event. If no event is
specified in the optional list, all events are disabled.
Optional list of event’s category is :
• -service: Generate service events
• -task: Generate task events
• -access: Generate SMB Hash access events.
-smbhash -service {enable | disable}
Enables or disables the SMB hash generation service (default is
started). If CIFS service is started, this command is taken into account
immediately. If CIFS is not running, this command is executed at the
next “cifs start”.
-smbhash -cleanup <fs_name> [-all | -unusedfor <days> |
-unusedsince <date>]
Cleans up the SMB Hash Files of the specified file system.
◆
If no option is specified, only obsolete SMB Hash Files are
removed.
server_cifs
495
The server Commands
◆
If -all option is specified, the entire “smbhash” directory is
removed.
◆
If -unusedfor <days> option is specified, obsolete SMB Hash
Files plus SMB Hash Files not accessed since the specified
number of days are removed.
◆
If -unusedsince <date> option is specified, SMB Hash Files not
accessed since the specified date are removed. The format of the
date is <YYMMDDHHMM>.
-setspn {-list [server=<full_comp_name>]
| -add <SPN>compname=<comp_name>,
domain=<full_domain_name>,admin=<admin_name>
| -delete <SPN> compname=<comp_name>,
domain=<full_domain_name>,admin=<admin_name}
Displays all SPNs for the specified FQDN server, both for the Data
Mover and for the KDC Windows Active Directory entry. If no server
is specified, then the SPNs for all joined CIFS Servers for the specified
movername is displayed. The command fails if an error occurs. For
example, unable to connect to the Active Directory, the specified
server is not joined to the domain.
When the -add and -delete sub-options are used, the user is
prompted for the password associated with the admin name. The
SPN must be the full value to use, including the realm.
The -add sub-option attempts to add the specified SPN to both the
Data Mover and Active Directory. The operation succeeds if the SPN
is added to both the Data Mover and Active Directory. In an entry
already exists in one of these places, it is not duplicated. Otherwise,
the operation fails if an error occurs. For example, unable to connect
to the Active Directory, the specified server is not joined to the
domain, or incorrect admin password.
The -delete sub-option attempts to remove the specified SPN from
both the Data Mover and Active Directory. The operation succeeds if
the SPN is removed from both the Data Mover and Active Directory.
If the entry has already been deleted, it is not considered an error.
Otherwise, the operation fails if an error occurs. For example, unable
to connect to the Active Directory, the specified server is not joined to
the domain, or incorrect admin password.
SEE ALSO
496
Using EMC Utilities for the CIFS Environment, Managing a
Multiprotocol Environment on VNX, Using VNX Replicator, Using
Command Line Interface Reference for File 8.1
The server Commands
International Character Sets on VNX for File, server_date, server_export,
server_mount, and server_setup.
OUTPUT NOTE
EXAMPLE #1
The network interface that appears in the output is dependent on the
type of network interface cards that are installed. Dates appearing in
the output are in UTC format.
To display the number and names of open files on server_2, type:
$ server_cifs server_2 -o audit,full
AUDIT Ctx=0xdffcc404, ref=2, Client(fm-main07B60004) Port=36654/139
NS40_1[BRCSLAB] on if=cge0_new
CurrentDC 0xceeab604=W2K3PHYAD
Proto=NT1, Arch=UNKNOWN, RemBufsz=0xfefb, LocBufsz=0xffff, popupMsg=1
0 FNN in FNNlist NbUsr=1 NbCnx=0
Uid=0x3f NTcred(0xcf156a04 RC=1 NTLM Capa=0x401) 'BRCSLAB\gustavo' CHECKER
AUDIT Ctx=0xde05cc04, ref=2, XP Client(BRCSBARREGL1C) Port=1329/445
NS40_1[BRCSLAB] on if=cge0_new
CurrentDC 0xceeab604=W2K3PHYAD
Proto=NT1, Arch=Win2K, RemBufsz=0xffff, LocBufsz=0xffff, popupMsg=1
0 FNN in FNNlist NbUsr=1 NbCnx=2
Uid=0x3f NTcred(0xceeabc04 RC=3 NTLMSSP Capa=0x11001) 'BRCSLAB\gustavo'
CHECKER
Cnxp(0xceeaae04), Name=IPC$, cUid=0x3f Tid=0x3f, Ref=1, Aborted=0
readOnly=0, umask=22, opened files/dirs=0
Cnxp(0xde4e3204), Name=gustavo, cUid=0x3f Tid=0x41, Ref=1, Aborted=0
readOnly=0, umask=22, opened files/dirs=2
Fid=64, FNN=0x1b0648f0(FREE,0x0,0), FOF=0x0 DIR=\
Notify commands received:
Event=0x17, wt=0, curSize=0x0, maxSize=0x20, buffer=0x0
Tid=0x41, Pid=0xb84, Mid=0xec0, Uid=0x3f, size=0x20
Fid=73, FNN=0x1b019ed0(FREE,0x0,0), FOF=0xdf2ae504 (CHECK) FILE=\New Wordpad
Document.doc
EXAMPLE #2
To configure CIFS service on server_2 with a NetBIOS name of
dm110-cge0, in the NT4 domain NASDOCS, with a NetBIOS alias of
dm110-cge0a1, hiding the NetBIOS name in the Network
Neighborhood, with the interface for CIFS service as cge0, the WINS
server as 172.24.102.25, and with the comment string EMC VNX,
type:
$ server_cifs server_2 -add
netbios=dm110-cge0,domain=NASDOCS,alias=dm110-cge0a1,hid
den=y,interface=cge0,wins=172.24.102.25 -comment "EMC
Celerra"
server_2 : done
server_cifs
497
The server Commands
EXAMPLE #3
To enable the home directory on server_2, type:
$ server_cifs server_2 -option homedir
server_2 : done
EXAMPLE #4
To add the WINS servers, 172.24.103.25 and 172.24.102.25, type:
$ server_cifs server_2 -add
wins=172.24.103.25,wins=172.24.102.25
server_2 : done
EXAMPLE #5
To rename the NetBIOS name from dm110-cge0 to dm112-cge0, type:
$ server_cifs server_2 -rename -netbios dm110-cge0
dm112-cge0
server_2 : done
EXAMPLE #6
To display the CIFS configuration for NT4 with Internal Usermapper,
type:
$ server_cifs server_2
server_2 :
256 Cifs threads started
Security mode = NT
Max protocol = NT1
I18N mode = UNICODE
Home Directory Shares ENABLED, map=/.etc/homedir
Usermapper auto broadcast enabled
Usermapper[0] = [127.0.0.1] state:active port:14640 (auto discovered)
Default WINS servers = 172.24.103.25:172.24.102.25
Enabled interfaces: (All interfaces are enabled)
Disabled interfaces: (No interface disabled)
DOMAIN NASDOCS RC=3
SID=S-1-5-15-99589f8d-9aa3a5f-338728a8-ffffffff
>DC=WINSERVER1(172.24.102.66) ref=2 time=0 ms
CIFS Server DM112-CGE0[NASDOCS] RC=2 (Hidden)
Alias(es): DM110-CGE0A1
Comment='EMC Celerra'
if=cge0 l=172.24.102.242 b=172.24.102.255 mac=0:60:16:4:35:4f
wins=172.24.102.25
Password change interval: 0 minutes
498
Command Line Interface Reference for File 8.1
The server Commands
Where:
Value
Cifs threads started
Security mode
Max protocol
I18N mode
Home Directory Shares
map
Usermapper auto
broadcast enabled
Usermapper
state
Default WINS servers
Enabled interfaces
Disabled interfaces
Unused Interface(s)
RC
SID
DC
ref
time
Alias(es)
if
Password change
interval:
EXAMPLE #7
Definition
Number of CIFS threads used when the CIFS service was started.
User authorization mechanism used by the Data Mover.
Maximum dialect supported by the security mode.
L18N mode (unicode or ASCII).
Whether Home Directory shares are enabled.
Home directory used by the Data Mover.
Usermapper is using its broadcast mechanism to discover its servers. This only displays
when the mechanism is active. It is disabled when you manually set the Usermapper
server addresses.
IP address of the servers running the Usermapper service.
Current state of Usermapper.
Addresses of the default WINS servers.
Data Mover’s enabled interfaces.
Data Mover’s disabled interfaces.
Interfaces not currently used by the Data Mover.
Reference count indicating the number of internal objects (such as client contexts) using
the CIFS server.
Security ID of the domain.
Domain controllers used by the Data Mover. Depending on the number of DCs in the
domain, this list may be large.
Number of internal objects using the Domain Controller.
Domain Controller response time.
Alternate NetBIOS names assigned to the CIFS server configuration.
Interfaces used by the CIFS server.
The amount of time between password changes.
To display the CIFS configuration for NT4, type:
$ server_cifs server_2
server_2 :
256 Cifs threads started
Security mode = NT
Max protocol = NT1
I18N mode = UNICODE
Home Directory Shares ENABLED, map=/.etc/homedir
Usermapper auto broadcast suspended
Usermapper[0] = [172.24.102.20] state:available
Default WINS servers = 172.24.103.25:172.24.102.25
Enabled interfaces: (All interfaces are enabled)
server_cifs
499
The server Commands
Disabled interfaces: (No interface disabled)
DOMAIN NASDOCS RC=3
SID=S-1-5-15-99589f8d-9aa3a5f-338728a8-ffffffff
>DC=WINSERVER1(172.24.102.66) ref=2 time=0 ms
CIFS Server DM112-CGE0[NASDOCS] RC=2 (Hidden)
Alias(es): DM110-CGE0A1
Comment='EMC Celerra'
if=cge0 l=172.24.102.242 b=172.24.102.255 mac=0:60:16:4:35:4f
wins=172.24.102.25
Password change interval: 0 minutes
EXAMPLE #8
To add a Windows server using the compname dm112-cge0, in the
Active Directory domain nasdocs.emc.com, with a NetBIOS alias of
dm112-cge0a1, hiding the NetBIOS name in the Network
Neighborhood, with the interface for CIFS service as cge0, the WINS
servers as 172.24.102.25 and 172.24.103.25, in the DNS domain
nasdocs.emc.com, and with the comment string EMC VNX, type:
$ server_cifs server_2 -add
compname=dm112-cge0,domain=nasdocs.emc.com,alias=dm112-c
ge0a1,hidden=y,interface=cge0,wins=172.24.102.25:172.24.
103.25,dns=nasdocs.emc.com -comment "EMC Celerra"
server_2 : done
EXAMPLE #9
To join dm112-cge0 into the Active Directory domain
nasdocs.emc.com, using the Administrator account, and to add this
server to Engineering\Computers organizational unit, type:
$ server_cifs server_2 -Join
compname=dm112-cge0,domain=nasdocs.emc.com,admin=adminis
trator,ou="ou=Computers:ou=Engineering"
server_2 : Enter Password:********
done
EXAMPLE #10
To add the NFS service to the CIFS server to make it possible for NFS
users to access the Windows KDC, type:
$ server_cifs server_2 -Join
compname=dm112-cge0,domain=nasdocs.emc.com,admin=adminis
trator -option addservice=nfs
server_2 : Enter Password:********
done
EXAMPLE #11
To enable the cge1 interface, type:
$ server_cifs server_2 -Enable cge1
server_2 : done
500
Command Line Interface Reference for File 8.1
The server Commands
EXAMPLE #12
To display CIFS information for a Data Mover in a Windows domain
with internal usermapper, type:
$ server_cifs server_2
server_2 :
256 Cifs threads started
Security mode = NT
Max protocol = NT1
I18N mode = UNICODE
Home Directory Shares ENABLED, map=/.etc/homedir
Usermapper auto broadcast enabled
Usermapper[0] = [127.0.0.1] state:active (auto discovered)
Default WINS servers = 172.24.103.25:172.24.102.25
Enabled interfaces: (All interfaces are enabled)
Disabled interfaces: (No interface disabled)
Unused Interface(s):
if=cge1 l=172.24.102.243 b=172.24.102.255 mac=0:60:16:4:35:4e
DOMAIN NASDOCS FQDN=nasdocs.emc.com SITE=Default-First-Site-Name RC=3
SID=S-1-5-15-99589f8d-9aa3a5f-338728a8-ffffffff
>DC=WINSERVER1(172.24.102.66) ref=3 time=1 ms (Closest Site)
CIFS Server DM112-CGE0[NASDOCS] RC=2 (Hidden)
Alias(es): DM112-CGEA1
Full computer name=dm112-cge0.nasdocs.emc.com realm=NASDOCS.EMC.COM
Comment='EMC Celerra'
if=cge0 l=172.24.102.242 b=172.24.102.255 mac=0:60:16:4:35:4f
wins=172.24.102.25:172.24.103.25
FQDN=dm112-cge0.nasdocs.emc.com (Updated to DNS)
Password change interval: 30 minutes
Last password change: Thu Oct 27 15:59:17 2005
Password versions: 2
EXAMPLE #13
To display CIFS information for a Data Mover in a Window domain,
type:
$ server_cifs server_2
server_2 :
256 Cifs threads started
Security mode = NT
Max protocol = NT1
I18N mode = UNICODE
Home Directory Shares ENABLED, map=/.etc/homedir
Usermapper auto broadcast suspended
Usermapper[0] = [172.24.102.20] state:available
Default WINS servers = 172.24.103.25:172.24.102.25
server_cifs
501
The server Commands
Enabled interfaces: (All interfaces are enabled)
Disabled interfaces: (No interface disabled)
Unused Interface(s):
if=cge1 l=172.24.102.243 b=172.24.102.255 mac=0:60:16:4:35:4e
DOMAIN NASDOCS FQDN=nasdocs.emc.com SITE=Default-First-Site-Name RC=3
SID=S-1-5-15-99589f8d-9aa3a5f-338728a8-ffffffff
>DC=WINSERVER1(172.24.102.66) ref=3 time=1 ms (Closest Site)
CIFS Server DM112-CGE0[NASDOCS] RC=2 (Hidden)
Alias(es): DM112-CGEA1
Full computer name=dm112-cge0.nasdocs.emc.com realm=NASDOCS.EMC.COM
Comment='EMC Celerra'
if=cge0 l=172.24.102.242 b=172.24.102.255 mac=0:60:16:4:35:4f
wins=172.24.102.25:172.24.103.25
FQDN=dm112-cge0.nasdocs.emc.com (Updated to DNS)
Password change interval: 30 minutes
Last password change: Thu Oct 27 16:29:21 2005
Password versions: 3, 2
EXAMPLE #14
To display CIFS information for a Data Mover when CIFS service is
not started, type:
$ server_cifs server_2
server_2 :
Cifs NOT started
Security mode = NT
Max protocol = NT1
I18N mode = UNICODE
Home Directory Shares ENABLED, map=/.etc/homedir
Usermapper auto broadcast suspended
Usermapper[0] = [172.24.102.20] state:available
Default WINS servers = 172.24.103.25:172.24.102.25
Enabled interfaces: (All interfaces are enabled)
Disabled interfaces: (No interface disabled)
Unused Interface(s):
if=cge1 l=172.24.102.243 b=172.24.102.255 mac=0:60:16:4:35:4e
CIFS Server DM112-CGE0[NASDOCS] RC=2 (Hidden)
Alias(es): DM112-CGEA1
Full computer name=dm112-cge0.nasdocs.emc.com realm=NASDOCS.EMC.COM
Comment='EMC Celerra'
if=cge0 l=172.24.102.242 b=172.24.102.255 mac=0:60:16:4:35:4f
wins=172.24.102.25:172.24.103.25
FQDN=dm112-cge0.nasdocs.emc.com (Updated to DNS)
Password change interval: 30 minutes
502
Command Line Interface Reference for File 8.1
The server Commands
Last password change: Thu Oct 27 16:29:21 2005
Password versions: 3, 2
EXAMPLE #15
To add a Windows server named, dm112-cge0, in the Active
Directory domain nasdocs.emc.com, with the interface for CIFS
service as cge0, and enable local users support, type:
$ server_cifs server_2 -add
compname=dm112-cge0,domain=nasdocs.emc.com,interface=cge
0,local_users
server_2 : Enter Password:********
Enter Password Again:********
done
EXAMPLE #16
To set a security mode to NT for a Data Mover, type:
$ server_cifs server_2 -add security=NT
server_2 : done
EXAMPLE #17
To disable a CIFS interface, type:
$ server_cifs server_2 -Disable cge1
server_2 : done
EXAMPLE #18
To display CIFS audit information for a Data Mover, type:
$ server_cifs server_2 -option audit
server_2 :
|||| AUDIT Ctx=0xad3d4820, ref=1, W2K3 Client(WINSERVER1) Port=1638/139
||| DM112-CGE0[NASDOCS] on if=cge0
||| CurrentDC 0xad407620=WINSERVER1
||| Proto=NT1, Arch=Win2K, RemBufsz=0xffff, LocBufsz=0xffff
||| 0 FNN in FNNlist NbUsr=1 NbCnx=1
||| Uid=0x3f NTcred(0xad406a20 RC=2 KERBEROS Capa=0x2) 'NASDOCS\administrator'
|| Cnxp(0xad3d5420), Name=IPC$, cUid=0x3f Tid=0x3f, Ref=1, Aborted=0
| readOnly=0, umask=22, opened files/dirs=1
|||| AUDIT Ctx=0xad43c020, ref=1, W2K3 Client(172.24.102.67) Port=1099/445
||| DM112-CGE0[NASDOCS] on if=cge0
||| CurrentDC 0xad407620=WINSERVER1
||| Proto=NT1, Arch=Win2K, RemBufsz=0xffff, LocBufsz=0xffff
||| 0 FNN in FNNlist NbUsr=1 NbCnx=1
||| Uid=0x3f NTcred(0xad362c20 RC=2 KERBEROS Capa=0x2) 'NASDOCS\user1'
|| Cnxp(0xaec21020), Name=IPC$, cUid=0x3f Tid=0x3f, Ref=1, Aborted=0
| readOnly=0, umask=22, opened files/dirs=2
server_cifs
503
The server Commands
Where:
Value
Ctx
ref
Port
CurrentDC
Proto
Arch
RemBufsz
LocBufsz
FNN/FNNlist
NbUsr
NbCnx
Uid/NTcred
Definition
Address in memory of the Stream Context.
Reference counter of components using this context at this time.
The client port and the Data Mover port used in the current TCP connection.
Specifies the address of the Domain Controller that is currently used.
Dialect level that is currently used.
Type of the client OS.
Max buffer size negotiated by the client.
Max buffer size we have negotiated.
Number of blocked files that has not yet been checked by Virus Checker.
Number of sessions connected to the stream context (TCP connection).
Number of connections to shares for this TCP connection.
User Id (this number is not related to the UNIX UID used to create a file), the credential address,
and the type of authentication.
Cnxp/Name
Share connection address’ and the name of the share the user is connecting to.
cUid
User Id who has opened the connection first.
Tid
Tree Id (number which represents the share connection in any protocol request).
Aborted
Status of the connection.
readOnly
If the share connection is read only.
umask
A user file-creation mask.
opened files/dirs Number of files or directories opened on this share connection.
EXAMPLE #19
To unjoin the computer dm112-cge0 from the nasdocs.emc.com
domain, type:
$ server_cifs server_2 -Unjoin
compname=dm112-cge0,domain=nasdocs.emc.com,admin=adminis
trator
server_2 : Enter Password:********
done
EXAMPLE #20
To delete WINS servers, 172.24.102.25, and 172.24.103.25, type:
$ server_cifs server_2 -delete
wins=172.24.102.25,wins=172.24.103.25
server_2 : done
EXAMPLE #21
To delete a NetBIOS name, dm112-cge0, type:
$ server_cifs server_2 -delete netbios=dm112-cge0
server_2 : done
504
Command Line Interface Reference for File 8.1
The server Commands
EXAMPLE #22
To delete the compname, dm112-cge0, type:
$ server_cifs server_2 -delete compname=dm112-cge0
server_2 : done
EXAMPLE #23
To delete the usermapper, 172.24.102.20, type:
$ server_cifs server_2 -delete usrmapper=172.24.102.20
server_2 : done
EXAMPLE #24
To add and join a Windows server in disjoint DNS and Windows
domains, type:
$ server_cifs server_2 -add
compname=dm112-cge0,domain=nasdocs.emc.com,interface=cge0,dns=eng.emc.com
-comment "EMC Celerra"
$ server_cifs server_2 -Join
compname=dm112-cge0.eng.emc.com,domain=nasdocs.emc.com,admin=Administrator
EXAMPLE #25
To add a Windows server using a delegated account from a trusted
domain, type:
$ server_cifs server_2 -Join
compname=dm112-cge0,domain=nasdocs.emc.com,admin=delegat
eduser@it.emc.com
server_2 : Enter Password:********
done
EXAMPLE #26
To add a Windows server in the Active Directory domain using a
pre-created computer account, type:
$ server_cifs server_2 -Join
compname=dm112-cge0,domain=nasdocs.emc.com,admin=adminis
trator -option reuse
server_2 : Enter Password:********
done
EXAMPLE #27
To update the directory /ufs1/users with a new minimum directory
size of 8192, type:
$ server_cifs server_2 -update /ufs1/users
mindirsize=8192
server_2 : done
EXAMPLE #28
To migrate all SIDs in the ACL database for file system, ufs1, from the
<src_domain>, eng.emc.com:nb=dm112-cge1:if=cge1 to the
<dst_domain>, nasdocs.emc.com:nb=dm112-cge0:if=cge0, type:
server_cifs
505
The server Commands
$ server_cifs server_2 -Migrate ufs1 -acl
eng.emc.com:nb=dm112-cge1:if=cge1
nasdocs.emc.com:nb=dm112-cge0:if=cge0
server_2 : done
EXAMPLE #29
To migrate SIDs of members of the local group defined for the
specified NetBIOS name, from the <src_domain>,
eng.emc.com:nb=dm112-cge1:if=cge1 to the <dst_domain>,
nasdocs.emc.com:nb=dm112-cge0:if=cge0, type:
$ server_cifs server_2 -Migrate dm112-cge1 -localgroup
eng.emc.com:nb=dm112-cge1:if=cge1
nasdocs.emc.com:nb=dm112-cge0:if=cge0
server_2 : done
EXAMPLE #30
To replace the SIDs for ufs1, type:
$ server_cifs server_2 -Replace ufs1 -acl
:nb=dm112-cge0:if=cge0
server_2 : done
EXAMPLE #31
To configure a stand-alone CIFS server on server_2 with a NetBIOS
name of dm112-cge0, in the workgroup NASDOCS, with a NetBIOS
alias of dm112-cge0a1, hiding the NetBIOS name in the Network
Neighborhood, with the interface for CIFS service as cge0, the WINS
servers as 172.24.102.25 and 172.24.103.25, and with enabled local
users support, type:
$ server_cifs server_2 -add
standalone=dm112-cge0,workgroup=NASDOCS,alias=dm112-cge0
a1,hidden=y,interface=cge0,wins=172.24.102.25:172.24.103
.25,local_users
server_2 : Enter Password:********
Enter Password Again:********
done
EXAMPLE #32
To delete the stand-alone CIFS server, dm112-cge0, type:
$ server_cifs server_2 -delete standalone=dm112-cge0
server_2 : done
EXAMPLE #33
To display a summary of SMB statistics, type:
$ server_cifs server_2 -stats -summary
server_2 :
State info:
Open connection Open files
2
2
SMB total requests:
totalAllSmb
totalSmb
506
totalTrans2Smb
Command Line Interface Reference for File 8.1
totalTransNTSmb
The server Commands
10038
6593
EXAMPLE #34
3437
8
To display all non-zero CIFS statistics, type:
$ server_cifs server_2 -stats
server_2 :
SMB statistics:
proc
Close
Rename
Trans
Echo
ReadX
WriteX
Trans2Prim
TreeDisco
NegProt
SessSetupX
UserLogoffX
TreeConnectX
TransNT
CreateNTX
CancelNT
ncalls
1305
2
314
21
231
3697
9375
10
29
47
9
13
8
1338
1
%totcalls
7.96
0.01
1.91
0.13
1.41
22.54
57.16
0.06
0.18
0.29
0.05
0.08
0.05
8.16
0.01
maxTime
46.21
0.81
0.77
0.01
0.03
39.96
34.27
0.06
0.42
60.55
0.01
0.39
0.01
47.11
0.03
ms/call
2.16
0.50
0.08
0.00
0.00
0.98
0.46
0.00
0.24
5.81
0.00
0.23
0.00
0.81
0.00
Trans2 SMBs:
proc
FindFirst
QFsInfo
QPathInfo
QFileInfo
SetFileInfo
ncalls
22
3154
1113
2077
3007
%totcalls
0.23
33.65
11.87
22.16
32.08
maxTime
0.22
0.08
6.73
0.04
34.26
ms/call
0.09
0.05
0.15
0.02
1.28
NT SMBs:
proc
NotifyChange
ncalls
8
%totcalls
100.00
maxTime
0.01
ms/call
0.00
Performance info:
Read
Re/s
231
231000.00
State info:
Open connection Open files
2
2
Shadow info:
Reads
0
Writes
0
SMB total requests:
totalAllSmb
totalSmb
25783
16400
Write
3697
Wr/s
1021.27
All
25783
Ops/sec
1575.40
Splits
0
Extinsert
0
Truncates
0
totalTrans2Smb
9375
totalTransNTSmb (unsupported)
8
2
server_cifs
507
The server Commands
Where:
Value
proc
ncalls
%totcalls
maxTime
ms/call
failures
Read
Re/s
Write
Wr/s
Definition
Name of CIFS requests received.
Number of requests received.
Percentage of this type of request compared to all requests.
Maximum amount of time used.
Average time in milliseconds taken to service calls.
Number of times the call has failed.
Total number of read operations.
Number of read operations per second.
Total number of write operations.
Number of write operations per second.
EXAMPLE #35
To reset to zero the values for all SMB statistics, type:
$ server_cifs server_2 -stats -zero
server_2 : done
EXAMPLE #36
To configure CIFS service in a language that uses multibyte
characters (such as Japanese), type:
$ server_cifs server_2 -add
compname=<computer_name_in_local_language_text>
domain=nasdocs.emc.com, -comment
<comment_in_local_language_text>
server_2 : done
EXAMPLE #37
To enable the SMB3 protocol, type:
$ server_cifs server_2 -add security=NT,dialect=SMB3
server_2 :
256 Cifs threads started
Security mode = NT
Max protocol = SMB3.0
I18N mode = UNICODE
EXAMPLE #38
To disable both SMB2 and SMB3, type:
$ server_cifs server_2 -add security=NT,dialect=NT1
server_2 : done
508
Command Line Interface Reference for File 8.1
The server Commands
server_cifssupport
Provides support services for CIFS users.
SYNOPSIS
server_cifssupport {<movername>|ALL}
-accessright
{-name <name> [-domain <domain_name>]
| -sid <SID>|-uname <unix_name>|-uid <user_id>}
{-path <pathname>|-share <sharename>}
[-policy {mixed|native|secure|nt|unix}]
[-build [-admin <admin_name>]]
[-netbios <netbios_servername>|-compname <comp_name>
| -standalone <netbios_ name>]
| -acl {-path <pathname>|-share <sharename>} [-verbose]
| -cred
{-name <name> [-domain <domain_name>]
| -sid <SID>|-uname <unix_name>|-uid <user_id>}
[-build [-ldap][-admin <admin_name>]]
[-netbios <netbios_servername>|-compname <comp_name>
| -standalone <netbios_name>]
| -pingdc
{-netbios <netbios_servername>|-compname <comp_name>}
[-dc <netbios_Dcname>]
[-verbose]}
| -secmap
-list
[-name <name> [-domain <domain_name>
| -domain <domain_name>
| -sid <SID>
| -uid <user_id>
| -gid <group_id>]
| -create
{-name <name> [-domain <domain_name>]}
| -sid <SID>}
| -verify
{-name <name> [-domain <domain_name>]}
| -sid <SID>}
|-update
{-name <name> [-domain <domain_name>]}
| -sid <SID>}
| -delete
{-name <name> [-domain <domain_name>]}
| -sid <SID>}
| -export [-file <filename>]
| -import -file <filename>
| -report
| -migration
server_cifssupport
509
The server Commands
DESCRIPTION
server_cifssuport checks network connectivity between a CIFS server
and domain controller, manages access rights, generates credentials,
and manages secure mapping cache.
The -accessright option:
◆
Displays user access rights to a file, directory, or share in a
Windows permission mask.
◆
Rebuilds and displays a credential for users of a file, directory, or
share, who do not have a session opened in one of the CIFS
servers.
◆
Without impact on the actual access-checking policy of a file
system object, shows how user permissions are reset if you were
to change the policy on that object.
The -acl option displays the access control list (ACL) of files,
directories, or shares in plain text form.
The -cred option generates a credential containing all groups to
which a user belongs, including local groups, without the user being
connected to a CIFS server. This allows you to verify if user’s SIDs are
being correctly mapped to UNIX UIDs and GIDs and to troubleshoot
any user access control issues.
The -pingdc option checks the network connectivity between a CIFS
server and a domain controller then verifies that a CIFS server can
access and use the following domain controller services:
◆
IPC$ share logon
◆
Secure Channel when verifying domain users during NT LAN
Manager (NTLM) authentication
◆
Local Security Authority (LSA) pipe information when mapping
Windows SIDs to UNIX UIDs and GIDs
◆
SAMR (Remote Security Account Manager) pipe when merging a
user’s UNIX and Windows groups together to create a credential
◆
Trusted domain information
◆
Privilege names for internationalization: pingdc
The -secmap option manages the secure mapping (secmap) cache.
Secmap contains all mapping between SIDs and UID/GIDs used by a
Data Mover or Virtual Data Mover (VDM). The Data Mover
permanently caches all mappings it receives from any mapping
mechanism (local files, NIS, iPlanet, Active Directory, and
Usermapper) in the secmap database, making the response to
510
Command Line Interface Reference for File 8.1
The server Commands
subsequent mapping requests faster and less susceptible to network
problems. Reverse mapping provides better quota support.
ACCESS RIGHT
OPTIONS
-accessright {-name <name> [-domain <domain_name>]|
-sid <SID>|-uname <unix_name>|-uid <user_id>}
{-path <pathname>|-share <sharename>}
Displays user access rights to a file, directory, or share in a Windows
permission mask for the specified:
◆
Windows username and the optional domain to which the user
belongs
or
◆
<SID> which is the the user’s Windows security identifier
or
◆
<unix_name>
or
◆
<user_id> which is the user’s UNIX identifier
The -path option specifies the path of the file or directory to check for
user permissions, or the absolute path of the share to check for user
permissions.
[-policy {mixed|native|secure|nt|unix}]
Specifies an access-checking policy for the specified file, directory,
or share. This does not change the current access-checking policy,
instead it helps you anticipate any access problems before
actually resetting the policy on a file system object. server_cifs
provides more information.
[-build [-admin <admin_name>]]
Rebuilds a credential for a user of a file, directory, or share, who
does not have a session opened in one of the CIFS servers. If
-build is not specified, the system searches the known user
credentials in cache. If none are found, an error message is
generated. The -admin option specifies the name of an
administrative user to use for creating the access right list. The
password of the admin_name user is prompted when executing
the command.
[-netbios <netbios_servername>|-compname <comp_
name>|-standalone <netbios_ name>]}
Indicates the CIFS server, as specified by its NetBIOS name or
computer name to use when rebuilding the user credential.
server_cifssupport
511
The server Commands
The -standalone option specifies the stand-alone CIFS server, as
specified by its name, to use when rebuilding a user credential.
Note: If no CIFS server is specified, the system uses the default CIFS
server, which uses all interfaces not assigned to other CIFS servers on the
Data Mover.
ACL OPTIONS
-acl {-path <pathname>|-share <sharename>} [-verbose]
Displays the ACL of a file, directory, or a share in plain text form.
Windows or UNIX access control data are both displayed in their
native forms. The -verbose option displays the ACE access rights
mask in plain text form in addition to their native forms.
CREDENTIAL OPTIONS
-cred {-name <name> [-domain <domain_name>]|-sid
<SID>|-uname <unix_name>|-uid <user_id>}
Generates a credential containing all of the groups to which a user
belongs without being connected to a CIFS server. The credential is
specifies by the user’s:
◆
Windows username and the domain to which the user belongs
or
◆
Windows security identifier
or
◆
UNIX name
or
◆
UNIX identifier
[-build [-ldap][-admin <admin_name>]]
Rebuilds a user credential. If -build is not specified, the system
searches the known user credentials in cache. If none are found,
an error message is generated. The -ldap option retrieves the
user’s universal groups to be included in the credential. If none
are found, no universal groups are incorporated into the
credential. The -admin option indicates the name of an
administrative user for creating the credential. The password of
the <admin_name> is prompted when executing the command.
[-netbios <netbios_servername>|-compname <comp_
name>|-standalone <netbios_ name>]}
Indicates the CIFS server, as specified by its NetBIOS name or
computer name to use when rebuilding the user credential.
512
Command Line Interface Reference for File 8.1
The server Commands
The -standalone option specifies the stand-alone CIFS server to
use when rebuilding a user credential.
Note: If no CIFS server is specified, the system uses the default CIFS
server, which uses all interfaces not assigned to other CIFS servers on the
Data Mover.
PINGDC OPTIONS
-pingdc {-netbios <netbios_servername>|-compname
<comp_ name>}
Checks the network connectivity for the CIFS server as specified by
its NetBIOS name or by its computer name with a domain controller.
Once connectivity is established, it verifies that a CIFS server can
access and use the domain controller services.
Note: An IP address can be used for the <netbios_servername> and the
<comp_name>.
[-dc <netbios_Dcname>]
Indicates the domain controller to ping for network and resource
connectivity with the CIFS server. If not specified, the command
checks the domain controllers currently in use by the CIFS server.
Note: An IP address can be used for the <netbios_Dcname>.
[-verbose]
Adds troubleshooting information to the command output.
SECMAP OPTIONS
-secmap -list
Lists the secmap mapping entries.
-secmap -list -name <name> -domain <domain_name>
Lists the secmap mapping entries with the specified name and
domain name.
-secmap -list -domain <domain_name>
Lists the secmap mapping entries with the specified domain name.
-secmap -list -sid <SID>
Lists the secmap mapping entries with the specified SID.
-secmap -list -uid <user_id>
Lists the secmap mapping entries with the specified UID (reverse
mapping).
server_cifssupport
513
The server Commands
-secmap -list -gid <group_id>
Lists the secmap mapping entries with the specified GID (reverse
mapping).
-secmap -create {-name <name> [-domain <domain_name>]}
Creates the secmap mapping entry with the specified name and
domain name.
-secmap -create -sid <SID>
Creates the secmap mapping entry with the specified SID.
-secmap -verify {-name <name> [-domain <domain_name>]}
Checks the mapping entry stored in secmap with the specified name
and optional domain name with what is currently available in the
mapping sources. If a mapping has changed, it is marked.
-secmap -verify -sid <SID>
Checks the secmap mapping entry with the specified SID.
-secmap -update {-name <name> [-domain <domain_name>]}
Updates the specified mapping entry stored in secmap with the
mappings currently available in the mapping sources. Once this
option is performed, force an update of the Data Mover’s file systems
ACLs so that the new mappings are recognized.
-secmap -update -sid <SID>
Updates the secmap mapping entry with the specified SID.
-secmap -delete -name <name> [-domain <domain_name>]
Deletes the secmap mapping entry with the specified name and
domain name.
-secmap -delete -sid <SID>
Deletes the secmap mapping entry with the specified SID.
-secmap -export [-file <filename>]
Exports the secmap mapping entry to the specified file.
Note: If no filename is specified, the secmap database is displayed on the
screen.
-secmap -import -file <filename>
Imports secmap mapping entries from the specified file.
-secmap -report
Displays current secmap status, including database state, domains
handled by secmap, and resource usage.
514
Command Line Interface Reference for File 8.1
The server Commands
-secmap -migration
Displays secmap database migration information like start date and
end date of the operation, and migration status.
EXAMPLE #1
To display user access rights to a file for user1, type:
$ server_cifssupport server_2 -accessright -name user1
-domain NASDOCS -path /ufs1/test/test.txt
server_2 : done
ACCOUNT GENERAL INFORMATIONS
Name
Domain
Path
Allowed mask
Action
Action
Action
Action
Action
EXAMPLE #2
:
:
:
:
:
:
:
:
:
user1
NASDOCS
/ufs1/test/test.txt
0x200a9
List Folder / Read data
Read Extended Attributes
Traverse Folder / Execute File
Read Attributes
Read Permissions
To rebuild a credential for a user to a file using the SID, type:
$ server_cifssupport server_2 -accessright -sid
S-1-5-15-b8e641e2-33f0942d-8f03a08f-1f4 -path
/ufs1/test/test.txt -build -compname dm102-cge0
server_2 : done
ACCOUNT GENERAL INFORMATIONS
Path
Allowed mask
Action
Action
Action
Action
Action
Action
Action
Action
Action
Action
Action
EXAMPLE #3
:
:
:
:
:
:
:
:
:
:
:
:
:
/ufs1/test/test.txt
0x301ff
List Folder / Read data
Create Files / Write data
Create Folders / Append Data
Read Extended Attributes
Write Extended Attributes
Traverse Folder / Execute File
Delete Subfolders and Files
Read Attributes
Write Attributes
Delete
Read Permissions
To display user access rights to a file for user1 with access-checking
policy UNIX, type:
$ server_cifssupport server_2 -accessright -name user1
-domain NASDOCS -path /ufs1/test/test.txt -policy unix
server_2 : done
server_cifssupport
515
The server Commands
ACCOUNT GENERAL INFORMATIONS
Name
Domain
Path
Allowed mask
Action
Action
Action
Action
:
:
:
:
:
:
:
:
EXAMPLE #4
user1
NASDOCS
/ufs1/test/test.txt
0x20089
List Folder / Read data
Read Extended Attributes
Read Attributes
Read Permissions
To rebuild a credential for user1 to a file using an administrative
account, type:
$ server_cifssupport server_2 -accessright -name user1
-domain NASDOCS -path /ufs1/test/test.txt -build -admin
administrator
server_2 : Enter Password:*******
done
ACCOUNT GENERAL INFORMATIONS
Name
Domain
Path
Allowed mask
Action
Action
Action
Action
Action
:
:
:
:
:
:
:
:
:
EXAMPLE #5
user1
NASDOCS
/ufs1/test/test.txt
0x200a9
List Folder / Read data
Read Extended Attributes
Traverse Folder / Execute File
Read Attributes
Read Permissions
To display the verbose ACL information of a file, type:
$ server_cifssupport server_2 -acl -path
/ufs1/test/test.txt -verbose
server_2 : done
ACL DUMP REPORT
Path
UID
GID
Rights
acl ID
acl size
owner SID
group SID
:
:
:
:
:
:
:
:
/ufs1/test/test.txt
32770
32797
rw-r--r-0x4
174
S-1-5-20-220
S-1-5-15-b8e641e2-33f0942d-8f03a08f-201
DACL
Owner
516
: USER 32770 S-1-5-15-b8e641e2-33f0942d-8f03a08f-1f4
Command Line Interface Reference for File 8.1
The server Commands
Access
Rights
: ALLOWED 0x0 0x1f01ff RWXPDO
: List Folder / Read data
Create Files / Write data
Create Folders / Append Data
Read Extended Attributes
Write Extended Attributes
Traverse Folder / Execute File
Delete Subfolders and Files
Read Attributes
Write Attributes
Delete
Read Permissions
Change Permissions
Take Ownership
Synchronize
Owner
Access
Rights
: USER 32771 S-1-5-15-b8e641e2-33f0942d-8f03a08f-a59
: ALLOWED 0x0 0x1200a9 R-X--: List Folder / Read data
Read Extended Attributes
Traverse Folder / Execute File
Read Attributes
Read Permissions
Synchronize
EXAMPLE #6
To display the access control level of a share, type:
$ server_cifssupport server_2 -acl -share ufs1
server_2 : done
ACL DUMP REPORT
Share
UID
GID
Rights
:
:
:
:
ufs1
0
1
rwxr-xr-x
EXAMPLE #7
To generate a credential for user1, type:
$ server_cifssupport server_2 -cred -name user1 -domain
NASDOCS
server_2 : done
ACCOUNT GENERAL INFORMATIONS
Name
Domain
Primary SID
UID
GID
Authentification
Credential capability
:
:
:
:
:
:
:
user1
NASDOCS
S-1-5-15-b8e641e2-33f0942d-8f03a08f-a59
32771
32768
KERBEROS
0x2
server_cifssupport
517
The server Commands
Privileges
System privileges
Default Options
NT administrator
Backup administrator
Backup
NT credential capability
:
:
:
:
:
:
:
0x8
0x2
0x2
False
False
False
0x2
ACCOUNT GROUPS INFORMATIONS
Type UNIX ID
Name
Domain
NT
32797
S-1-5-15-b8e641e2-33f0942d-8f03a08f-201
NT
32798
S-1-5-15-b8e641e2-33f0942d-8f03a08f-e45
NT
4294967294
NT
4294967294
NT
4294967294
NT
2151678497
UNIX 32797
UNIX 32798
UNIX 4294967294
UNIX 2151678497
EXAMPLE #8
SID
S-1-1-0
S-1-5-2
S-1-5-b
S-1-5-20-221
To rebuild a user credential including the user’s universal groups for
a user using SID, type:
$ server_cifssupport server_2 -cred -sid
S-1-5-15-b8e641e2-33f0942d-8f03a08f-1f4 -build -ldap
-compname dm102-cge0
server_2 : done
ACCOUNT GENERAL INFORMATIONS
Name
Domain
Server
Primary SID
UID
GID
Authentification
Credential capability
Privileges
System privileges
Default Options
NT administrator
Backup administrator
Backup
NT credential capability
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
NASDOCS
dm102-cge0
S-1-5-15-b8e641e2-33f0942d-8f03a08f-1f4
32770
32768
NTLM
0x0
0x7f
0x1
0xe
True
True
False
0x0
ACCOUNT GROUPS INFORMATIONS
518
Command Line Interface Reference for File 8.1
The server Commands
Type UNIX ID
Name
NT 32794
Group Policy Cre
NT
32795
Schema Admins
NT 32796
Enterprise Admin
NT
32797
Domain Users
NT
32793
Domain Admins
NT
4294967294 Everyone
NT
4294967294 NETWORK
NT
4294967294 ANONYMOUS LOGON
NT
2151678496 Administrators
NT
2151678497 Users
NT
1
UNIX GID=0x1 &ap
UNIX 32794
UNIX 32795
UNIX 32796
UNIX 32797
UNIX 32793
EXAMPLE #9
Domain
SID
NASDOCS S-1-5-15-b8e641e2-33f0942d-8f03a08f-208
NASDOCS S-1-5-15-b8e641e2-33f0942d-8f03a08f-206
NASDOCS S-1-5-15-b8e641e2-33f0942d-8f03a08f-207
NASDOCS S-1-5-15-b8e641e2-33f0942d-8f03a08f-201
NASDOCS S-1-5-15-b8e641e2-33f0942d-8f03a08f-200
S-1-1-0
NT AUTHORITY
S-1-5-2
NT AUTHORITY
S-1-5-7
BUILTIN
S-1-5-20-220
BUILTIN
S-1-5-20-221
S-1-5-12-2-1
To check the network connectivity for the CIFS server with netbios
dm102-cge0, type:
$ server_cifssupport server_2 -pingdc -netbios dm102-cge0
server_2 : done
PINGDC GENERAL INFORMATIONS
DC SERVER:
Netbios name
: NASDOCSDC
CIFS SERVER :
Compname
: dm102-cge0
Domain
: nasdocs.emc.com
EXAMPLE #10
To check the network connectivity between the domain controller
and the CIFS server with compname dm102-cge0, type:
$ server_cifssupport server_2 -pingdc -compname
dm102-cge0 -dc NASDOCSDC -verbose
server_2 : done
PINGDC GENERAL INFORMATIONS
DC SERVER:
Netbios name
: NASDOCSDC
CIFS SERVER :
Compname
: dm102-cge0
Domain
: nasdocs.emc.com
EXAMPLE #11
To display the secmap mapping entries, type:
$ server_cifssupport server_2 -secmap -list
server_cifssupport
519
The server Commands
server_2 : done
SECMAP USER MAPPING TABLE
UID
Origin
Date
Name
32772
usermapper Tue Sep 18 19:08:40 2007 NASDOCS\user2
S-1-5-15-b8e641e2-33f0942d-8f03a08f-452
32771
usermapper Tue Sep 18 17:56:53 2007 NASDOCS\user1
S-1-5-15-b8e641e2-33f0942d-8f03a08f-a59
32770
usermapper Sun Sep 16 07:50:39 2007 NASDOCS\Administrator
S-1-5-15-b8e641e2-33f0942d-8f03a08f-1f4
SID
SECMAP GROUP MAPPING TABLE
GID
Origin
Date
Name
SID
32793
usermapper Wed Sep 12 14:16:18 2007 NASDOCS\Domain Admins
S-1-5-15-b8e641e2-33f0942d-8f03a08f-200
32797
usermapper Sun Sep 16 07:50:40 2007 NASDOCS\Domain Users
S-1-5-15-b8e641e2-33f0942d-8f03a08f-201
32799
usermapper Mon Sep 17 19:13:16 2007 NASDOCS\Domain Guests
S-1-5-15-b8e641e2-33f0942d-8f03a08f-202
32800
usermapper Mon Sep 17 19:13:22 2007 NASDOCS\Domain Computers
S-1-5-15-b8e641e2-33f0942d-8f03a08f-203
32795
usermapper Sun Sep 16 07:50:40 2007 NASDOCS\Schema Admins
S-1-5-15-b8e641e2-33f0942d-8f03a08f-206
32796
usermapper Sun Sep 16 07:50:40 2007 NASDOCS\Enterprise Admins
S-1-5-15-b8e641e2-33f0942d-8f03a08f-207
32794
usermapper Sun Sep 16 07:50:40 2007 NASDOCS\Group Policy Creator
Owners S-1-5-15-b8e641e2-33f0942d-8f03a08f-208
32798
usermapper Mon Sep 17 19:13:15 2007 NASDOCS\CERTSVC_DCOM_ACCESS
S-1-5-15-b8e641e2-33f0942d-8f03a08f-e45
32801
usermapper Tue Sep 18 19:08:41 2007 NASDOCS\NASDOCS Group
S-1-5-15-b8e641e2-33f0942d-8f03a08f-45b
EXAMPLE #12
To display the secmap mapping entry for a user user1 in a domain
NASDOCS, type:
$ server_cifssupport server_2 -secmap -list -name user1
-domain NASDOCS
server_2 : done
SECMAP USER MAPPING TABLE
UID
Origin
Date
Name
32771
usermapper Tue Sep 18 17:56:53 2007 NASDOCS\user1
S-1-5-15-b8e641e2-33f0942d-8f03a08f-a59
EXAMPLE #13
SID
To display the secmap mapping entry for a user with UID 32771,
type:
$ server_cifssupport server_2 -secmap -list -uid 32771
server_2 : done
520
Command Line Interface Reference for File 8.1
The server Commands
SECMAP USER MAPPING TABLE
UID
Origin
Date
Name
32771
usermapper Tue Sep 18 17:56:53 2007 NASDOCS\user1
S-1-5-15-b8e641e2-33f0942d-8f03a08f-a59
EXAMPLE #14
SID
To create the secmap mapping entry for user3 in a domain
NASDOCS, type:
$ server_cifssupport server_2 -secmap -create -name user3
-domain NASDOCS
server_2 : done
SECMAP USER MAPPING TABLE
UID
Origin
Date
Name
32773
usermapper Tue Sep 18 19:21:59 2007 NASDOCS\user3
S-1-5-15-b8e641e2-33f0942d-8f03a08f-a3d
EXAMPLE #15
SID
To check the secmap mapping for user1 in a domain NASDOCS,
type:
$ server_cifssupport server_2 -secmap -verify -name user1
-domain NASDOCS
server_2 : done
EXAMPLE #16
To update the secmap mapping entry for a user using SID, type:
$ server_cifssupport server_2 -secmap -update -sid
S-1-5-15-b8e641e2-33f0942d-8f03a08f-a3d
server_2 : done
EXAMPLE #17
To delete the secmap mapping entry for user3, type:
$ server_cifssupport server_2 -secmap -delete -name user3
-domain NASDOCS
server_2 : done
EXAMPLE #18
To display current secmap status, type:
$ server_cifssupport server_2 -secmap -report
server_2 : done
SECMAP GENERAL INFORMATIONS
Name
State
Fs
Used nodes
Used blocks
:
:
:
:
:
server_2
Enabled
/
12
8192
SECMAP MAPPED DOMAIN
server_cifssupport
521
The server Commands
Name
NASDOCS
SID
S-1-5-15-b8e641e2-33f0942d-8f03a08f-ffffffff
EXAMPLE #19
To export the secmap mapping entries to the display, type:
$ server_cifssupport server_2 -secmap -export
server_2 : done
SECMAP MAPPING RECORDS
S-1-5-15-b8e641e2-33f0942d-8f03a08f-200:2:96:8019:8019:NASDOCS\Domain Admins
S-1-5-15-b8e641e2-33f0942d-8f03a08f-201:2:96:801d:801d:NASDOCS\Domain Users
S-1-5-15-b8e641e2-33f0942d-8f03a08f-202:2:96:801f:801f:NASDOCS\Domain Guests
S-1-5-15-b8e641e2-33f0942d-8f03a08f-203:2:96:8020:8020:NASDOCS\Domain Computers
S-1-5-15-b8e641e2-33f0942d-8f03a08f-206:2:96:801b:801b:NASDOCS\Schema Admins
S-1-5-15-b8e641e2-33f0942d-8f03a08f-207:2:96:801c:801c:NASDOCS\Enterprise Admins
S-1-5-15-b8e641e2-33f0942d-8f03a08f-208:2:96:801a:801a:NASDOCS\Group Policy
Creator Owners
S-1-5-15-b8e641e2-33f0942d-8f03a08f-e45:2:96:801e:801e:NASDOCS\CERTSVC_DCOM_ACCE
SS
S-1-5-15-b8e641e2-33f0942d-8f03a08f-452:1:96:8004:8000:NASDOCS\user2
S-1-5-15-b8e641e2-33f0942d-8f03a08f-a59:1:96:8003:8000:NASDOCS\user1
S-1-5-15-b8e641e2-33f0942d-8f03a08f-45b:2:96:8021:8021:NASDOCS\NASDOCS Group
S-1-5-15-b8e641e2-33f0942d-8f03a08f-1f4:1:96:8002:8000:NASDOCS\Administrator
EXAMPLE #20
To export the secmap mapping entries to a file, type:
$ server_cifssupport server_2 -secmap -export -file
exportfile.txt
server_2 : done
EXAMPLE #21
To import the secmap mapping entries from a file, type:
$ server_cifssupport server_2 -secmap -import -file
exportfile.txt
server_2 :
Secmap import in progress : #
done
522
Command Line Interface Reference for File 8.1
The server Commands
server_cpu
Performs an orderly, timed, or immediate halt or reboot of a Data
Mover.
SYNOPSIS
server_cpu {<movername>|ALL}
{-halt|-reboot[cold|warm]} [-monitor] <time>
DESCRIPTION
server_cpu performs an orderly halt or reboot of the specified Data
Mover.
The ALL option executes the command for all Data Movers.
OPTIONS
-halt
Performs an orderly shutdown of a Data Mover for the VNX. To
restart a Data Mover, perform a -reboot. For the NS series, a -halt
causes a system reboot.
-reboot
Performs an orderly shutdown, and restarts a Data Mover. The
default parameter of the -reboot option will be the warm parameter.
In case the warm rebooting fails, the -reboot option will use the cold
parameter to reboot the Data Mover.
[cold]
A cold reboot or a hardware reset shuts down the Data Mover
completely before restarting, including a Power on Self Test
(POST).
[warm]
A warm reboot or a software reset performs a partial shutdown of
the Data Mover, and skips the POST after restarting. A software
reset is faster than the hardware reset.
!
CAUTION
Performing a reboot for ALL Data Movers can be time consuming
relative to the size of the mounted file system configuration.
-monitor
Polls and displays the boot status until completion of the halt or
reboot.
server_cpu
523
The server Commands
<time>
Specifies the time when the Data Mover is to be halted or
rebooted. Time is specified as {now|+<min>|<hour>:<min>}.
The now option is used for an immediate shutdown or reboot.
After a power fail and crash recovery, the system reboots itself at
power-up unless previously halted.
SEE ALSO
EXAMPLE #1
VNX System Operations.
To monitor an immediate reboot of server_2, type:
$ server_cpu server_2 -reboot -monitor now
server_2 : reboot in progress 0.0.0.0.0.0.0.0.0.0.0.3.3.3.3.3.3.4.done
Where:
Value
0
1
2
3
4
5
7
9
Definition
Reset
DOS booted
SIB failed
Loaded
Configured
Contacted
Panicked
Reboot pending
EXAMPLE #2
To immediately halt server_2, type:
$ server_cpu server_2 -halt now
server_2 : done
EXAMPLE #3
To immediately reboot server_2, type:
$ server_cpu server_2 -reboot now
server_2 : done
EXAMPLE #4
To monitor a reboot of server_2, that is set to take place in one
minute, type:
$ server_cpu server_2 -reboot -monitor +1
server_2 : reboot in progress ........3.3.3.3.3.done
524
Command Line Interface Reference for File 8.1
The server Commands
server_date
Displays or sets the date and time for a Data Mover, and synchronizes
time between a Data Mover and an external time source.
SYNOPSIS
server_date {<movername>|ALL}
[+<format>][<yymmddhhmm>[<ss>]]
| timesvc start ntp [-sync_delay][-interval <hh>[:<mm>]][<host>[<host>...]]
| timesvc update ntp
| timesvc stop ntp
| timesvc delete ntp
| timesvc set ntp
| timesvc stats ntp
| timesvc
| timezone [<timezonestr>]
| timezone -name <timezonename>
DESCRIPTION
server_date sets and displays the current date and time for the
specified Data Movers.
The server_date timesvc commands control the synchronization of
the Data Mover with external timing sources and gets and sets the
time zone.
The ALL option executes the command for all Data Movers.
OPTIONS
No arguments
Displays the current date and time for the specified Data Mover.
+<format>
Displays the date information in the format specified by each field
descriptor. Each field descriptor is preceded by percent and is
replaced in the output by its corresponding value. A single percent is
encoded by double percent (%%).
If the argument contains embedded blanks, it must be quoted.
The complete listing of all field descriptors can be viewed using the
Linux strftime (3C) man page.
<yymmddhhmm>[<ss>]
Sets a two-digit number for the year, month, day, hour, minutes, and
seconds in this order where <yy> is the year; the first <mm> is the
month; <dd> is the day; <hh> is the hour (in 24-hour system); and the
second <mm> is the minute, and <ss> is the second.
server_date
525
The server Commands
timesvc start ntp <host> [<host>...]
Starts time synchronization immediately between a Data Mover and
a host, which is the IP address of the time server hosts, and adds an
entry to the database. The host must be running the NTP protocol.
Only four host entries are allowed.
Other options include:
-sync_delay
Indicates that the clock should not be synchronized when the
time server is activated. Instead, when the first poll is taken,
latency adjustments are handled slowly. This option is generally
used if time service is started after the Data Mover has already
started, or if synchronization is starting after other services have
already started.
Note: If -sync_delay is not typed, by default, the clock is set at Data
Mover startup. The clock is synchronized after the first poll.
-interval <hh>[:<mm>]
Sets the delay in hours (or hours and minutes) between polls
(default=1 hour which is entered 01 or 00:60). Interval is
displayed in minutes.
timesvc update ntp
Immediately polls the external source and synchronizes the time on
the Data Mover.
timesvc stop ntp
Stops timing synchronization between the Data Mover and an
external timing host for the NTP protocol, and does not remove the
entry from the database.
Note: A stop of time services takes about 12 seconds. If time service is
restarted within this time, a "busy" status message is returned.
timesvc delete ntp
Stops time synchronization and deletes the NTP protocol from the
database.
timesvc set ntp
Immediately polls the external source and synchronizes the time on
the Data Mover without slewing the clock.
timesvc stats ntp
Displays the statistical information of time synchronization for the
526
Command Line Interface Reference for File 8.1
The server Commands
Network Time Protocol such as time differences between the Data
Mover and the time server. Also provides information about the
current state of NTP service on the Data Mover.
timesvc
Displays the current time service configuration.
timezone
Displays the current time zone on the specified Data Mover.
[<timezonestr>]
Sets the current time zone on the specified Data Mover. The
<timezonestr> is a POSIX style time zone specification with the
following formats:
<std><offset> (no daylight savings time)
<std><offset><dst>[offset],start[/time],end[/time] (adjusts
for daylight savings time)
Note: The Linux man page for tzset provides information about the format.
timezone -name <timezonename>
Sets the time zone on the Data Mover to the specified
<timezonename>. The <timezonename> is in Linux style time zone
specification. A list of valid Linux timezones is located in the
/usr/share/zoneinfo directory. The third column in the table in this
file lists the valid timezones.
Note: The timezone -name option does not reset time on the Data Mover to
the specified <timezonename> time.
SEE ALSO
EXAMPLE #1
server_2 : Thu Jan
Configuring Time Services on VNX, server_dns, and server_nis.
To display the current date and time on a Data Mover, type:
$ server_date server_2
6 16:55:09 EST 2005
EXAMPLE #2
To customize the display of the date and time on a Data Mover, type:
$ server_date server_2 "+%Y-%m-%d %H:%M:%S"
server_2 : 2005-01-06 16:55:58
server_date
527
The server Commands
EXAMPLE #3
To start time synchronization between a Data Mover and an external
source, type:
$ server_date server_2 timesvc start ntp -interval 06:00
172.24.102.20
server_2 : done
EXAMPLE #4
To set the time service without slewing the clock, type:
$ server_date server_2 timesvc set ntp
server_2 : done
EXAMPLE #5
To display statistical information, type:
$ server_date server_2 timesvc stats ntp
server_2 :
Time synchronization statistics since start:
hits= 2, misses= 0, first poll hit= 2, miss= 0
Last offset: 0 secs, 0 usecs
Current State: Running, connected, interval=360
Time sync hosts:
0 1 172.24.102.20
Where:
Value
hits
misses
first poll hit
miss
Last offset
Current State
Time sync hosts
Definition
When a client sends a request to the server requesting the current time, if there is a reply, that is a
hit.
No reply from any of the time servers.
First poll hit which sets the first official time for the Data Mover.
First poll miss.
Time difference between the time server and the Data Mover.
State of the time server.
IP address of the time server.
EXAMPLE #6
To update time synchronization between a Data Mover and an
external source, type:
$ server_date server_2 timesvc update ntp
server_2 : done
EXAMPLE #7
To get the time zone on the specified Data Mover, type:
$ server_date server_2 timezone
server_2 : Local timezone: GMT
EXAMPLE #8
To set the time zone to Central Time for a Data Mover when you do
not have to adjust for daylight savings time, type:
$ server_date server_2 timezone CST6
server_2 : done
528
Command Line Interface Reference for File 8.1
The server Commands
EXAMPLE #9
To set the time zone to Central Time and adjust for daylight savings
time for a Data Mover, type:
$ server_date server_2 timezone CST6CDT5,M4.1.0,M10.5.0
server_2 : done
EXAMPLE #10
To set the time zone to Central Time and adjust the daylight savings
time for a Data Mover using the Linux method, type:
$ server_date server_2 timezone -name America/Chicago
server_2 : done
EXAMPLE #11
To display the time service configuration for a Data Mover, type:
$ server_date server_2 timesvc
server_2 :
Timeservice State
time:
type:
sync delay:
interval:
hosts:
Thu Jan 6 17:04:28 EST 2005
ntp
off
360
172.24.102.20,
Where:
Value
time
type
sync delay
interval
hosts
Definition
Date and time known to the Data Mover.
Time service protocol configured on the Data Mover.
Whether sync delay is on or off.
Time interval between polls.
Specifies the IP address of the time server.
EXAMPLE #12
To stop time services for a Data Mover, type:
$ server_date server_2 timesvc stop ntp
server_2 : done
EXAMPLE #13
To delete the time service configuration for a Data Mover, type:
$ server_date server_2 timesvc delete ntp
server_2 : done
EXAMPLE #14
To set the timezone on a Data Mover to Los Angeles, type:
$ server_date server_2 timezone -n America/Los_Angeles
server_2 : done
server_date
529
The server Commands
server_dbms
Enables backup and restore of databases, displays database
environment statistics.
SYNOPSIS
server_dbms {<movername>|ALL}
{-db
{-list [<db_name>]
| -delete <db_name>
| -check [<db_name>]
| -repair [<db_name>]
| -compact [<db name>]
| -fullbackup -target <pathname>
| -incrbckup -previous <pathname> -target <pathname>
| -restore [<db_name>] -source <pathname>
| -stats [<db_name> [-table <name>]][-reset]}
| -service -stats [transaction|memory|log|lock|mutex][-reset]
}
DESCRIPTION
server_dbms provides recovery of media failure or application
corruption, displays database information, checks application
database consistency, and fixes inconsistencies.
The ALL option executes the command for all Data Movers.
OPTIONS
-db -list [<db_name>]
Gets the list of all application databases and their status. If
<db_name> is specified, displays the list of all tables belonging to
that database.
-db -delete <db_name>
Deletes the target application database.
Note: This command will fail if the target database is not closed.
-db -check [<db_name>]
Checks the consistency of the target database at application level.
-db -repair [<db_name>]
Fixes the application level inconsistencies in the database.
-db -compact [<db_name> [-table <name>]]
Frees up disc space by compacting the target environment or
database.
530
Command Line Interface Reference for File 8.1
The server Commands
-db -fullbackup -target <pathname>
Performs an online full backup of the VDM database environment.
The target parameter specifies the location to copy the database files.
The <pathname> specifies the local path of the database environment
on the Control Station.
-db -incrbackup -previous <pathname> -target <pathname>
Downloads the transactional logs from the VDM and replays them on
a copy of the previous VDM backup specified by previous
<pathname>.
The -target option specifies the location to which the database files
are copied.
-db -restore [<db_name>] -source <pathname>
Restores the environment or database specified by <db_name>.
source <pathname> specifies the location for the backup of the
environment/database to be restored.
Note: The database must be closed before the command is executed.
-db -stats [<db_name> [-table <name>]][-reset]
Displays statistics related to the specified databases and tables. If
-reset is specified, resets the statistics.
-service -stats [transaction|memory|log|lock|
mutex][-reset]
Displays transaction, memory, logging, locking or mutex statistics of
the VDM database environment. If -reset is specified, resets all or
specified statistics.
Note: For this command to be executed, the VDM on which the target
environment resides must be up.
EXAMPLE #1
To get the list of all application databases and their status, type:
$ server_dbms server_3 -db -list
server_3 : done
BASE NAME : Secmap
Version : 1
Comment : CIFS Secure mapping database.
This is a cache of the sid to uid/gid mapping of the VDM.
This database is part of the CIFS application.
It can closed with the command server_setup
Size : 16384
Modification time : Fri May 25 09:58:21 2007
Creation time : Fri May 25 09:58:21 2007
server_dbms
531
The server Commands
TABLE NAME : Mapping
Version : 1
Comment : Sid to uid/gid mapping table with one secondary key on xid ((1,uid) &amp;
(2,gid))
Size : 16384
Modification time : Fri May 25 09:58:21 2007
Creation time : Fri May 25 09:58:21 2007
BASE NAME : V4NameSpace
Version : 1
Comment : NFSv4 namespace database, this represents the pseudofs and referrals.
Size : 8192
Modification time : Tue Jun 5 08:57:12 2007
Creation time : Tue Jun 5 08:57:12 2007
TABLE NAME : pseudofs
Version : 1
Comment : Pseudofs-table, this holds the export tree heirarchy
Size : 8192
Modification time : Mon Jun 11 11:06:23 2007
Creation time : Mon Jun 11 11:06:23 2007
BASE NAME : Usermapper
Version : 1
Comment : Usermapper database. It allows to assign a new uid or gid to a given SID.
Size : 57344
Modification time : Tue Jun 12 09:14:31 2007
Creation time : Tue Jun 12 09:14:31 2007
TABLE NAME : aliases
Version : 1
Comment : This table allows to retrieve a domain name from one of his aliases
Size : 8192
Modification time : Tue Jun 12 09:14:31 2007
Creation time : Tue Jun 12 09:14:31 2007
TABLE NAME : usrmapc
Version : 1
Comment : Store the uid &amp; gid ranges allocations for domains.
Size : 8192
Modification time : Tue Jun 12 09:14:31 2007
Creation time : Tue Jun 12 09:14:31 2007
TABLE NAME : idxname
Version : 1
Comment : Store the reverse mapping uid/gid to sid.
Size : 8192
Modification time : Tue Jun 12 09:14:31 2007
Creation time : Tue Jun 12 09:14:31 2007
TABLE NAME : usrmapusrc
Version : 1
Comment : Store the mapping SID -&gt; (uid, name).
Size : 8192
Modification time : Tue Jun 12 09:14:31 2007
Creation time : Tue Jun 12 09:14:31 2007
TABLE NAME : usrgrpmapnamesid
Version : 1
Comment : Store the mapping user.domain -&gt; SID.
532
Command Line Interface Reference for File 8.1
The server Commands
Size : 8192
Modification time : Tue Jun 12 09:14:31 2007
Creation time : Tue Jun 12 09:14:31 2007
TABLE NAME : usrmapgrpc
Version : 1
Comment : Store the mapping SID -&gt; (gid, name).
Size : 8192
Modification time : Tue Jun 12 09:14:31 2007
Creation time : Tue Jun 12 09:14:31 2007
TABLE NAME : groupmapnamesid
Version : 1
Comment : Store the mapping group.domain -&gt; SID.
Size : 8192
Modification time : Tue Jun 12 09:14:31 2007
Creation time : Tue Jun 12 09:14:31 2007
EXAMPLE #2
To display Secmap statistics, type:
$ server_dbms server_3 -db -stats Secmap
server_3 : done
STATISTICS FOR DATABASE : Secmap
TABLE : Mapping
NAME VALUE COMMENT
magic 340322 Magic number.
version 9 Table version number.
metaflags 0 Metadata flags.
nkeys 14 Number of unique keys.
ndata 14 Number of data items.
pagesize 4096 Page size.
minkey 2 Minkey value.
re_len 0 Fixed-length record length.
re_pad 32 Fixed-length record pad.
levels 1 Tree levels.
int_pg 0 Internal pages.
leaf_pg 1 Leaf pages.
dup_pg 0 Duplicate pages.
over_pg 0 Overflow pages.
empty_pg 0 Empty pages.
free 0 Pages on the free list.
int_pgfree 0 Bytes free in internal pages.
leaf_pgfree 2982 Bytes free in leaf pages.
dup_pgfree 0 Bytes free in duplicate pages.
over_pgfree 0 Bytes free in overflow pages.
EXAMPLE #3
To display statistics of the VDM database environment, type:
$ server_dbms server_3 -service -stats
STATISTICS FOR MODULE : LOG
NAME
magic
version
VALUE
264584
12
COMMENT
Log file magic number.
Log file version number.
server_dbms
533
The server Commands
mode
lg_bsize
lg_size
record
w_bytes
w_mbytes
wc_bytes
wc_mbytes
wcount
wcount_fill
rcount
scount
region_wait
region_nowait
cur_file
cur_offset
disk_file
disk_offset
regsize
maxcommitperflush
mincommitperflush
0
32768
5242880
96
16001
0
0
0
31
0
137
31
0
0
3
16001
3
16001
98304
1
1
Log file mode.
Log buffer size.
Log file size.
Records entered into the log.
Bytes to log.
Megabytes to log.
Bytes to log since checkpoint.
Megabytes to log since checkpoint.
Total writes to the log.
Overflow writes to the log.
Total I/O reads from the log.
Total syncs to the log.
Region lock granted after wait.
Region lock granted without wait.
Current log file number.
Current log file offset.
Known on disk log file number.
Known on disk log file offset.
Region size.
Max number of commits in a flush.
Min number of commits in a flush.
STATISTICS FOR MODULE : LOCK
NAME
last_id
cur_maxid
maxlocks
maxlockers
maxobjects
nmodes
nlocks
maxnlocks
nlockers
maxnlockers
nobjects
maxnobjects
nrequests
nreleases
nupgrade
ndowngrade
lock_wait
lock_nowait
ndeadlocks
locktimeout
nlocktimeouts
txntimeout
ntxntimeouts
region_wait
region_nowait
regsize
534
VALUE
91
2147483647
1000
1000
1000
9
20
21
49
49
20
21
65711
65691
0
20
0
0
0
0
0
0
0
0
0
352256
Command Line Interface Reference for File 8.1
COMMENT
Last allocated locker ID.
Current maximum unused ID.
Maximum number of locks in table.
Maximum num of lockers in table.
Maximum num of objects in table.
Number of lock modes.
Current number of locks.
Maximum number of locks so far.
Current number of lockers.
Maximum number of lockers so far.
Current number of objects.
Maximum number of objects so far.
Number of lock gets.
Number of lock puts.
Number of lock upgrades.
Number of lock downgrades.
Lock conflicts w/ subsequent wait.
Lock conflicts w/o subsequent wait.
Number of lock deadlocks.
Lock timeout.
Number of lock timeouts.
Transaction timeout.
Number of transaction timeouts.
Region lock granted after wait.
Region lock granted without wait.
Region size.
The server Commands
STATISTICS FOR MODULE : TXN
NAME
last_ckp
time_ckp
last_txnid
maxtxns
naborts
nbegins
ncommits
nactive
nsnapshot
nrestores
maxnactive
maxnsnapshot
region_wait
region_nowait
regsize
VALUE
COMMENT
3/15945
lsn of the last checkpoint.
Fri Aug 3 09:38:36 2007 time of last checkpoint.
0x8000001a
last transaction id given out.
20
maximum txns possible.
0
number of aborted transactions.
26
number of begun transactions.
26
number of committed transactions.
0
number of active transactions.
0
number of snapshot transactions.
0
number of restored transactions
after recovery.
2
maximum active transactions.
0
maximum snapshot transactions.
0
Region lock granted after wait.
0
Region lock granted without wait.
16384
Region size.
STATISTICS FOR MODULE : MPOOL
NAME
gbytes
bytes
ncache
regsize
mmapsize
maxopenfd
maxwrite
maxwrite_sleep
map
cache_hit
cache_miss
page_create
page_in
page_out
ro_evict
rw_evict
page_trickle
pages
page_clean
page_dirty
hash_buckets
hash_searches
hash_longest
hash_examined
hash_nowait
hash_wait
hash_max_nowait
hash_max_wait
region_nowait
VALUE
0
10487684
1
10493952
0
0
0
0
0
65672
36
0
36
2
0
0
0
36
36
0
1031
65744
1
65672
0
0
0
0
0
COMMENT
Total cache size: GB.
Total cache size: B.
Number of caches.
Region size.
Maximum file size for mmap.
Maximum number of open fd's.
Maximum buffers to write.
Sleep after writing max buffers.
Pages from mapped files.
Pages found in the cache.
Pages not found in the cache.
Pages created in the cache.
Pages read in.
Pages written out.
Clean pages forced from the cache.
Dirty pages forced from the cache.
Pages written by memp_trickle.
Total number of pages.
Clean pages.
Dirty pages.
Number of hash buckets.
Total hash chain searches.
Longest hash chain searched.
Total hash entries searched.
Hash lock granted with nowait.
Hash lock granted after wait.
Max hash lock granted with nowait
Max hash lock granted after wait.
Region lock granted with nowait.
server_dbms
535
The server Commands
region_wait
mvcc_frozen
mvcc_thawed
mvcc_freed
alloc
alloc_buckets
alloc_max_buckets
alloc_pages
alloc_max_pages
io_wait
0
0
0
0
123
0
0
0
0
0
Region lock granted after wait.
Buffers frozen.
Buffers thawed.
Frozen buffers freed.
Number of page allocations.
Buckets checked during allocation.
Max checked during allocation.
Pages checked during allocation.
Max checked during allocation.
Thread waited on buffer I/O.
STATISTICS FOR MODULE : MUTEX
NAME
mutex_align
mutex_tas_spins
mutex_cnt
mutex_free
mutex_inuse
mutex_inuse_max
region_wait
region_nowait
regsize
536
VALUE
4
1
3254
1078
2176
2176
0
0
278528
Command Line Interface Reference for File 8.1
COMMENT
Mutex alignment.
Mutex test-and-set spins.
Mutex count.
Available mutexes.
Mutexes in use.
Maximum mutexes ever in use.
Region lock granted after wait.
Region lock granted without wait.
Region size.
The server Commands
server_devconfig
Queries, saves, and displays the SCSI over Fibre Channel device
configuration connected to the specified Data Movers.
SYNOPSIS
server_devconfig {<movername>|ALL}
| -create -scsi {-disks|-nondisks|-all}
[-discovery {y|n}][-monitor {y|n}][-Force {y|n}]
| -list -scsi [<chain_number>] {-disks|-nondisks|-all}
| -probe -scsi [<chain_number>] {-disks|-nondisks|-all}
| -rename <old_name> <new_name>
DESCRIPTION
!
server_devconfig queries the available system device and tape
device configuration, and saves the device configuration into the
Data Mover’s database. server_devconfig renames the device name,
and lists SCSI devices.
CAUTION
It is recommended that all Data Movers have the same device
configuration. When adding devices to the device table for a single
Data Mover only, certain actions such as standby failover will not
be successful unless the standby Data Mover has the same disk
device configuration as the primary Data Mover.
The ALL option executes the command for all Data Movers.
OPTIONS
-create -scsi [<chain_number>] {-disks|-nondisks| -all}
Queries SCSI devices and saves them into the device table database
on the Data Mover. The <chain_number> specifies a SCSI chain
number.
The -disks option limits operations to disks. The -nondisks option
limits operations to non-disks such as tapes, juke boxes, and
gatekeeper devices. The -all option permits all SCSI devices (disks
and non-disks).
Note: The -create option modifies VNX for lock LUN names to the
VNX_<vnx-hostname>_<lun-id>_<vnx-dvol-name> format, if the LUNs use
the default Unisphere name.
server_devconfig
537
The server Commands
!
CAUTION
The time taken to complete this command might be lengthy,
dependent on the number and type of attached devices.
[-discovery {y|n}]
Enables or disables the storage discovery operation.
!
CAUTION
Disabling the -discovery option should only be done under the
direction of an EMC Customer Service Engineer.
[-monitor {y|n}]
Displays the progress of the query and discovery operations.
[-Force {y|n}]
Overrides the health check failures and changes the storage
configuration.
!
CAUTION
High availability can be lost when changing the storage
configuration. Changing the storage configuration should only
be done under the direction of an EMC Customer Service
Engineer.
-list -scsi [<chain_number>] {-disks|-nondisks| -all}
Lists the SCSI device table database that has been saved on the Data
Mover. The <chain_number> specifies a SCSI chain number.
Note: Fibre Channel devices appear as SCSI devices. Therefore, chain
numbers might be different for Fibre Channel.
The -disks option limits operations to disks. The -nondisks option
limits operations to non-disks such as tapes, juke boxes, and
gatekeeper devices. The -all option permits all SCSI devices (disks
and non-disks).
-probe -scsi [<chain_number>] {-disks|-nondisks| -all}
Queries and displays the SCSI devices without saving them into the
database. The <chain_number> specifies a SCSI chain number.
Note: Fibre Channel devices appear as SCSI devices, therefore, chain
numbers may be different for Fibre Channel.
538
Command Line Interface Reference for File 8.1
The server Commands
The -disks option limits operations to disks. The -nondisks option
limits operations to non-disks such as tapes, juke boxes, and
gatekeeper devices. The -all option permits all SCSI devices (disks
and non-disks).
-rename <old_name> <new_name>
Renames the specified non-disk from the <old_name> to
<new_name>. The -rename option is available for non-disks only.
SEE ALSO
SYSTEM OUTPUT
EXAMPLE #1
VNX System Operations, nas_disk, and nas_storage.
The number associated with the storage device is dependent on the
attached system. VNX for block displays a prefix of APM before a set
of integers, for example, APM00033900124-0019. For example,
Symmetrix systems appear as 002804000190-003C.
For the VNX system, to list all devices, type:
$ server_devconfig server_2 -list -scsi -all
server_2:
Scsi Disk Table
name
root_disk
root_disk
root_ldisk
root_ldisk
d3
d3
d4
d4
d5
d5
d6
d6
d7
d7
d8
d8
name
gk01
ggk01
gk161
addr
c0t0l0
c16t0l0
c0t0l1
c16t0l1
c0t0l2
c16t0l2
c0t0l3
c16t0l3
c0t0l4
c16t0l4
c0t0l5
c16t0l5
c0t1l0
c16t1l0
c16t1l1
c0t1l1
Scsi Device Table
addr
c0t0l0
c0t1l0
c16t1l1
Director
num type
type
disk
disk
disk
Port
num sts stor_id
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
APM00043807043
stor_dev
0000
0000
0001
0001
0002
0002
0003
0003
0004
0004
0005
0005
0010
0010
0011
0011
info
5 020700000000APM00043807043
5 020710001000APM00043807043
5 020711001100APM00043807043
For the VNX with a Symmetrix system, to list all the devices in the
SCSI table, type:
$ server_devconfig server_2 -list -scsi -all
server_devconfig
539
The server Commands
server_2 :
Scsi Disk Table
name
root_disk
root_disk
root_ldisk
root_ldisk
d3
d3
d4
d4
d5
d5
d6
d6
d7
d7
<... removed ...>
d377
d377
rootd378
rootd378
rootd379
rootd379
rootd380
rootd380
rootd381
rootd381
name
gk01
gk161
addr
c0t0l0
c16t0l0
c0t0l1
c16t0l1
c0t1l0
c16t1l0
c0t1l1
c16t1l1
c0t1l2
c16t1l2
c0t1l3
c16t1l3
c0t1l4
c16t1l4
Director
num type
16C FA
01C FA
16C FA
01C FA
16C FA
01C FA
16C FA
01C FA
16C FA
01C FA
16C FA
01C FA
16C FA
01C FA
Port
num
0
0
0
0
0
0
0
0
0
0
0
0
0
0
sts
On
On
On
On
On
On
On
On
On
On
On
On
On
On
stor_id
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
stor_dev
0000
0000
0001
0001
0006
0006
0007
0007
0008
0008
0009
0009
000A
000A
c1t8l6
c17t8l6
c1t8l7
c17t8l7
c1t8l8
c17t8l8
c1t8l9
c17t8l9
c1t8l10
c17t8l10
16C
01C
16C
01C
16C
01C
16C
01C
16C
01C
0
0
0
0
0
0
0
0
0
0
On
On
On
On
On
On
On
On
On
On
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
000187940268
017C
017C
0180
0180
0181
0181
0182
0182
0183
0183
info
56706817D480
56706817D330
000187940268
000187940268
FA
FA
FA
FA
FA
FA
FA
FA
FA
FA
Scsi Device Table
addr
type
c0t0l15
disk
c16t0l15
disk
Note: This is a partial display due to the length of the output.
540
Command Line Interface Reference for File 8.1
The server Commands
Where:
Value
name
addr
Director num
type
Port num
sts
Definition
A unique name for each device in the chain.
SCSI chain, target, and LUN information.
Director number. This output is applicable for Symmetrix systems only.
Device type, as specified in the SCSI spec. for peripherals. This output is applicable for Symmetrix
systems only.
Port number. This output is applicable for Symmetrix systems only.
Indicates the port status. Possible values are: On, Off, WD (write disabled), and NA. This output is
applicable for Symmetrix systems only.
System ID.
System device ID.
stor_id
stor_dev
EXAMPLE #2
For the VNX, to list all SCSI-attached non-disk devices, type:
$ server_devconfig server_2 -list -scsi -nondisks
server_2 :
Scsi Device Table
addr
c0t0l0
c0t1l0
c16t1l1
name
gk01
ggk01
gk161
type
disk
disk
disk
info
5 020700000000APM00043807043
5 020710001000APM00043807043
5 020711001100APM00043807043
For the VNX with a Symmetrix system, to list all SCSI-attached
non-disk devices, type:
$ server_devconfig server_2 -list -scsi -nondisks
server_2 :
Scsi Device Table
addr
type
c0t0l15
disk
c16t0l15
disk
name
gk01
gk161
info
56706817D480
56706817D330
000187940268
000187940268
For info=56706817D480, the following breakdown applies:
5670
68
Symm code
Last 2 digits in the Symm S/N Symm Device ID# Symm SA #
EXAMPLE #3
17D
48
0
SA Port # (0=a, 1=b)
To rename a device, type:
$ server_devconfig server_2 -rename gk161 gk201
server_2 : done
server_devconfig
541
The server Commands
EXAMPLE #4
For the VNX, to discover SCSI disk devices, without saving them to
the database table, type:
$ server_devconfig server_2 -probe -scsi -disks
server_2 :
SCSI disk devices :
chain= 0, scsi-0
stor_id= APM00043807043 celerra_id= APM000438070430000
tid/lun= 0/0 type= disk sz= 11263 val= 1 info= DGC RAID 5 02070000000000NI
tid/lun= 0/1 type= disk sz= 11263 val= 2 info= DGC RAID 5 02070100010001NI
tid/lun= 0/2 type= disk sz= 2047 val= 3 info= DGC RAID 5 02070200020002NI
tid/lun= 0/3 type= disk sz= 2047 val= 4 info= DGC RAID 5 02070300030003NI
tid/lun= 0/4 type= disk sz= 2047 val= 5 info= DGC RAID 5 02070400040004NI
tid/lun= 0/5 type= disk sz= 2047 val= 6 info= DGC RAID 5 02070500050005NI
tid/lun= 1/0 type= disk sz= 245625 val= 7 info= DGC RAID 5 02071000100010NI
tid/lun= 1/1 type= disk sz= 0 val= -5 info= DGC RAID 5 02071100110011NI
tid/lun= 1/2 type= disk sz= 273709 val= 9 info= DGC RAID 5 02071200120012NI
tid/lun= 1/3 type= disk sz= 0 val= -5 info= DGC RAID 5 02071300130013NI
tid/lun= 1/4 type= disk sz= 273709 val= 10 info= DGC RAID 5 02071400140014NI
tid/lun= 1/5 type= disk sz= 0 val= -5 info= DGC RAID 5 02071500150015NI
tid/lun= 1/6 type= disk sz= 273709 val= 11 info= DGC RAID 5 02071600160016NI
tid/lun= 1/7 type= disk sz= 0 val= -5 info= DGC RAID 5 02071700170017NI
tid/lun= 1/8 type= disk sz= 273709 val= 12 info= DGC RAID 5 02071800180018NI
tid/lun= 1/9 type= disk sz= 0 val= -5 info= DGC RAID 5 02071900190019NI
chain= 1, scsi-1 : no devices on chain
chain= 2, scsi-2 : no devices on chain
chain= 3, scsi-3 : no devices on chain
chain= 4, scsi-4 : no devices on chain
chain= 5, scsi-5 : no devices on chain
chain= 6, scsi-6 : no devices on chain
chain= 7, scsi-7 : no devices on chain
chain= 8, scsi-8 : no devices on chain
chain= 9, scsi-9 : no devices on chain
chain= 10, scsi-10 : no devices on chain
chain= 11, scsi-11 : no devices on chain
chain= 12, scsi-12 : no devices on chain
chain= 13, scsi-13 : no devices on chain
chain= 14, scsi-14 : no devices on chain
chain= 15, scsi-15 : no devices on chain
For the VNX with a Symmetrix system, to discover SCSI disk devices,
without saving them to the database table, type:
$ server_devconfig server_2 -probe -scsi -disks
server_2 :
SCSI disk devices :
chain= 0, scsi-0 : no devices on chain
chain= 1, scsi-1 : no devices on chain
chain= 2, scsi-2
stor_id= 000190102173 celerra_id= 0001901021730041
tid/lun= 0/0 type= disk sz= 11507 val= 1 info= 577273041291SI00041
tid/lun= 0/1 type= disk sz= 11507 val= 2 info= 577273042291SI00042
542
Command Line Interface Reference for File 8.1
The server Commands
tid/lun= 1/0 type= disk sz= 11501 val= 3 info= 57727304F291SI0004F
tid/lun= 1/1 type= disk sz= 11501 val= 4 info= 577273050291SI00050
tid/lun= 1/2 type= disk sz= 11501 val= 5 info= 577273051291SI00051
tid/lun= 1/3 type= disk sz= 11501 val= 6 info= 577273052291SI00052
tid/lun= 1/4 type= disk sz= 11501 val= 7 info= 577273053291SI00053
tid/lun= 1/5 type= disk sz= 11501 val= 8 info= 577273054291SI00054
tid/lun= 1/6 type= disk sz= 11501 val= 9 info= 577273055291SI00055
tid/lun= 1/7 type= disk sz= 11501 val= 10 info= 577273056291SI00056
tid/lun= 1/8 type= disk sz= 11501 val= 11 info= 577273057291SI00057
tid/lun= 1/9 type= disk sz= 11501 val= 12 info= 577273058291SI00058
tid/lun= 1/10 type= disk sz= 11501 val= 13 info= 577273059291SI00059
tid/lun= 1/11 type= disk sz= 11501 val= 14 info= 57727305A291SI0005A
tid/lun= 1/12 type= disk sz= 11501 val= 15 info= 57727305B291SI0005B
tid/lun= 1/13 type= disk sz= 11501 val= 16 info= 57727305C291SI0005C
tid/lun= 1/14 type= disk sz= 11501 val= 17 info= 57727305D291SI0005D
tid/lun= 1/15 type= disk sz= 11501 val= 18 info= 57727305E291SI0005E
tid/lun= 2/0 type= disk sz= 11501 val= 19 info= 57727305F291SI0005F
tid/lun= 2/1 type= disk sz= 11501 val= 20 info= 577273060291SI00060
tid/lun= 2/2 type= disk sz= 11501 val= 21 info= 577273061291SI00061
<... removed ...>
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
tid/lun=
7/6 type= disk sz= 11501 val= 105 info= 577273517291SI00517
7/7 type= disk sz= 11501 val= 106 info= 577273518291SI00518
7/8 type= disk sz= 11501 val= 107 info= 577273519291SI00519
7/9 type= disk sz= 11501 val= 108 info= 57727351A291SI0051A
7/10 type= disk sz= 11501 val= 109 info= 57727351B291SI0051B
7/11 type= disk sz= 11501 val= 110 info= 57727351C291SI0051C
7/12 type= disk sz= 11501 val= 111 info= 57727351D291SI0051D
7/13 type= disk sz= 11501 val= 112 info= 57727351E291SI0051E
7/14 type= disk sz= 11501 val= 113 info= 57727351F291SI0051F
7/15 type= disk sz= 11501 val= 114 info= 577273520291SI00520
chain= 3, scsi-3
chain= 4, scsi-4
chain= 5, scsi-5
chain= 6, scsi-6
<... removed ...>
:
:
:
:
no
no
no
no
devices
devices
devices
devices
on
on
on
on
chain
chain
chain
chain
chain= 18, scsi-18
stor_id= 000190102173 celerra_id= 0001901021730041
tid/lun= 0/0 type= disk sz= 11507 val= 1 info= 577273041201SI00041
tid/lun= 0/1 type= disk sz= 11507 val= 2 info= 577273042201SI00042
tid/lun= 1/0 type= disk sz= 11501 val= 3 info= 57727304F201SI0004F
tid/lun= 1/1 type= disk sz= 11501 val= 4 info= 577273050201SI00050
tid/lun= 1/2 type= disk sz= 11501 val= 5 info= 577273051201SI00051
tid/lun= 1/3 type= disk sz= 11501 val= 6 info= 577273052201SI00052
tid/lun= 1/4 type= disk sz= 11501 val= 7 info= 577273053201SI00053
Note: This is a partial listing due to the length of the output.
server_devconfig
543
The server Commands
EXAMPLE #5
To discover and save all SCSI devices, type:
$ server_devconfig server_2 -create -scsi -all
Discovering storage (may take several minutes)
server_2 : done
EXAMPLE #6
To discover and save all non-disk devices, type:
$ server_devconfig server_2 -create -scsi -nondisks
Discovering storage (may take several minutes)
server_2 : done
EXAMPLE #7
To save all SCSI devices with the discovery operation disabled, and
display information regarding the progress, type:
$ server_devconfig ALL -create -scsi -all -discovery n
-monitor y
server_2 :
server_2:
chain 0 ..........
chain 16 .....
done
server_3 :
server_3:
chain 0 ..........
chain 16 .....
done
server_4 :
server_4:
chain 0 ..........
chain 16 .....
done
server_5 :
server_5:
chain 0 ..........
chain 16 .....
done
544
Command Line Interface Reference for File 8.1
The server Commands
server_df
Reports free and used disk space and inodes for mounted file systems
on the specified Data Movers.
SYNOPSIS
server_df {<movername>|ALL}
[-inode][<pathname>|<fs_name>]
DESCRIPTION
server_df reports the amount of used and available disk space for a
Data Mover or file system, how much of a file system’s total capacity
has been used, and the number of used and free inodes.
The ALL option executes the command for all Data Movers.
OPTIONS
No arguments
Displays the amount of disk space in kilobytes used by file systems.
-inode
Reports used and free inodes.
[<pathname>|<fs_name>]
Gets file system information. If <fs_name> specified, gets
information for file system only.
SEE ALSO
EXAMPLE #1
Managing Volumes and File Systems for VNX Manually, nas_disk, and
nas_fs.
To display the amount of used and available disk space on a Data
Mover, type:
$ server_df server_2
server_2 :
Filesystem
ufs1
ufs4
ufs2
nmfs1
root_fs_common
root_fs_2
kbytes
1075686032
101683184
206515184
308198368
13624
114592
used
477816
584
600
1184
5264
760
avail capacity Mounted on
1075208216
0%
/ufs1
101682600
0%
/nmfs1/ufs4
206514584
0%
/nmfs1/ufs2
308197184
0%
/nmfs1
8360
39%
/.etc_common
113832
1%
/
Where:
Value
Filesystem
kbytes
used
Definition
Name of the file system.
Total amount of space in kilobytes for the file system.
Amount of kilobytes used by the file system.
server_df
545
The server Commands
Value
avail
capacity
Mounted on
Definition
Amount of space in kilobytes available for the file system.
Percentage capacity that is used.
Mount point of the file system.
EXAMPLE #2
To display the amount of disk space and the amount of free and
unused inodes on a Data Mover, type:
$ server_df server_2 -inode
server_2 :
Filesystem
ufs1
ufs4
ufs2
nmfs1
root_fs_common
root_fs_2
inodes
131210494
25190398
25190398
50380796
21822
130942
EXAMPLE #3
used
140
10
11
21
26
66
avail capacity Mounted on
131210354
0%
/ufs1
25190388
0%
/nmfs1/ufs4
25190387
0%
/nmfs1/ufs2
50380775
0%
/nmfs1
21796
0%
/.etc_common
130876
0%
/
To display the amount of disk space and the amount of free and
unused inodes on a file system, type:
$ server_df server_2 -inode ufs1
server_2 :
Filesystem
ufs1
546
inodes
131210494
used
140
Command Line Interface Reference for File 8.1
avail capacity Mounted on
131210354
0%
/ufs1
The server Commands
server_dns
Manages the Domain Name System (DNS) lookup server
configuration for the specified Data Movers.
SYNOPSIS
server_dns {<movername>|ALL}
[[-protocol {tcp|udp}] <domainname> {<ip_addr>,...}]
| [-delete <domainname>]
| [-option {start|stop|flush|dump}]
DESCRIPTION
server_dns provides connectivity to the DNS lookup servers for the
specified Data Movers to convert hostnames and IP addresses. Up to
three DNS lookup servers are supported for each domain on the Data
Mover.
server_dns also provides the ability to clear the cache that has been
saved on the Data Mover as a result of the DNS lookup servers.
The ALL option executes the command for all Data Movers.
OPTIONS
No arguments
Displays the DNS configuration.
-protocol {tcp|udp} <domainname> {<ip_addr>,...}
Sets the protocol for the DNS lookup servers (udp is the default).
<domainname> {<ip_addr>,...}
Creates list of up to three IP addresses to be used as the DNS lookup
servers for the specified <domainname>.
-delete <domainname>
Deletes the DNS lookup servers in the DNS domain name.
-option {start|stop|flush|dump}
The start option activates the link for the DNS lookup servers. The
stop option halts access to the DNS lookup servers. After DNS service
has been halted, the flush option can be used to clear the cache that
has been saved on the Data Mover, and the dump option displays the
DNS cache.
SEE ALSO
EXAMPLE #1
Configuring VNX Naming Services and server_nis.
To connect to a DNS lookup server, type:
$ server_dns server_2 prod.emc.com 172.10.20.10
server_2 : done
server_dns
547
The server Commands
EXAMPLE #2
To display the DNS configuration, type:
$ server_dns server_2
server_2 :
DNS is running.
prod.emc.com
proto:udp server(s):172.10.20.10
EXAMPLE #3
To change the protocol to TCP from UDP, type:
$ server_dns server_2 -protocol tcp prod.emc.com
172.10.20.10
server_2 : done
EXAMPLE #4
To halt access to the DNS lookup servers, type:
$ server_dns server_2 -option stop
server_2 : done
EXAMPLE #5
To flush the cache on a Data Mover, type:
$ server_dns server_2 -option flush
server_2 : done
EXAMPLE #6
To dump the DNS cache, type:
$ server_dns server_2 -option dump
server_2 :
DNS cache size for one record type: 64
DNS cache includes 6 item(s):
dm102-cge0.nasdocs.emc.com
Type:A
TTL=184 s dataCount:1
172.24.102.202 (local subnet)
--winserver1.nasdocs.emc.com
Type:A
TTL=3258 s dataCount:1
172.24.103.60
--_ldap._tcp.Default-First-Site-Name._sites.dc._msdcs.nasdocs.emc.com
Type:SRV
TTL=258 s dataCount:1
priority:0
weight:100
port:389
server:winserver1.nasdocs.emc.com
--_kerberos._tcp.Default-First-Site-Name._sites.dc._msdcs.nasdocs.emc.com
Type:SRV
TTL=258 s dataCount:1
priority:0
weight:100
port:88
server:winserver1.nasdocs.emc.com
--Expired item(s): 2
EXAMPLE #7
To delete the DNS lookup servers, type:
$ server_dns server_2 -delete prod.emc.com
server_2 : done
548
Command Line Interface Reference for File 8.1
The server Commands
server_export
Exports file systems, and manages access on the specified Data
Movers for NFS and CIFS clients.
SYNOPSIS
server_export {<movername>|ALL}
operations on all cifs and/or nfs entries:
| [-Protocol {cifs|nfs}] -list -all
| [-Protocol {cifs|nfs}] -all
| [-Protocol {cifs|nfs}] -unexport [-perm] -all
nfs operations per entry:
| -list <pathname>
| [-Protocol nfs [-name <name>]][-ignore][-option <options>]
[-comment <comment>] <pathname>
| -unexport [-perm] <pathname>
cifs operations per entry:
| -list -name <sharename> [-option <options>]
| -name <sharename> [-ignore][-option <options>][-comment <comment>]
<pathname>
| -unexport -name <sharename> [-option <options>]
-option type = {
CA[:] Encrypted[:] [ABE [:] HASH [:][OCAutoI|OCVDO|OCNONE]]|NONE
}
DESCRIPTION
server_export provides user access by exporting an NFS pathname,
or creating a CIFS share. Allows specification of multiple clients
identified by hostnames or network and subnet addresses separated
by a colon.
server_export removes access by unexporting an NFS pathname,
deleting a CIFS share, and displaying the exported entries and
available shares for the specified Data Mover.
The ALL option executes the command for all Data Movers.
Note: NFSv4 does not support the -name option.
GENERAL OPTIONS
FOR CIFS AND NFS
OPERATIONS
No arguments
Displays all exported NFS entries and CIFS shares.
[-Protocol {cifs|nfs}] -list -all
Lists all exported entries as defined by the protocol. The default is
NFS.
server_export
549
The server Commands
[-Protocol {cifs|nfs}] -all
Exports all entries on a Data Mover as defined by the protocol. The
default is NFS.
[-Protocol {cifs|nfs}] -unexport [-perm] -all
Unexports all entries as defined by the protocol. By default,
unexports are permanent for CIFS, and temporary for NFS, unless
-perm is specified. If -perm is specified, removes all entries from the
export table. When the entry is temporarily unexported, clients are
denied access to the entry until it is re-exported or the system is
rebooted, but the entries are not removed from the export table. The
default is NFS.
FOR NFS OPERATIONS
-list <pathname>
Lists a specific NFS entry. If there are extra spaces in the <pathname>,
the entire pathname must be enclosed by quotes. By using the
server_export command, IPv6 addresses can be specified and the
hosts configured with these addresses can mount and access file
systems over NFS.
Note: If you are configuring an IPv6 address for ro, rw, access, and root, it
must be enclosed in [ ] or square brackets. This is to properly handle the colon
used to separate entries. Link local addresses are not supported.
-Protocol nfs [-name <name>] <pathname>
Exports an NFS <pathname> by default as read-write for everyone. If
specified, assigns an optional file system name for the <name>.
Pathname length is limited to 1024 bytes (represented as 1024 ASCII
characters or a variable number of Unicode multibyte characters),
and must be enclosed by quotes, if spaces are used. Name length is
limited to 255 bytes.
Note: In a nested mount file system hierarchy, users can export the mount
point path of the component file system. Subdirectories of the component file
system cannot be exported. In a multilevel file system hierarchy, users can
export any part of a file system independent of existing exports.
[-ignore] <pathname>
Overwrites previous options and comments in the export table for
the entry.
[-comment <comment>] <pathname>
Adds a comment for the specified NFS export entry. The comment is
displayed when listing the exported entries.
550
Command Line Interface Reference for File 8.1
The server Commands
[-option <options>] <pathname>
Specifies the following comma-separated options:
sec=[sys|krb5|krb5i|krb5p]:<mode> [,<mode>,...]
Specifies a user authentication or security method with an access
mode setting. The sys (default) security option specifies
AUTH_SYS security. The access mode can be one, or a
combination of the following: ro, rw=, ro=, root=, access=, anon=,
webroot, public.
If the sec option is specified, it must always be the first option
specified in the string.
krb5 security specifies Kerberos user and data authentication.
krb5i checks for the integrity of the data by adding a signature to
each NFS packet and krb5p encrypts the data before sending it
over the network.
For krb5, krb5i, and krb5p security, the access mode can be one,
or a combination of the following: ro, rw=, ro=, root=, access=.
ro
Exports the <pathname> for all NFS clients as read-only.
ro=<client>[:<client>]...
Exports the <pathname> for the specified NFS clients as
read-only.
Note: If <client> is an IPv6 address, it must be enclosed in square
brackets or [ ].
ro=<-client>[:<-client>]...
Excludes the specified NFS clients from ro privileges. Clients
must be preceded with dash (-) to specify exclusion.
Note: If <client> is an IPv6 address, it must be enclosed in square
brackets or [ ].
rw=<client>[:<client>]...
Exports the <pathname> as read-mostly for the specified NFS
clients. Read-mostly means exported read-only to most machines,
but read-write to those specified. The default is read-write to all.
Note: If <client> is an IPv6 address, it must be enclosed in square
brackets or [ ].
server_export
551
The server Commands
rw=<-client>[:<-client>]...
Excludes the specified NFS clients from rw privileges. The
description of read-mostly provides information. Clients must be
preceded with - (dash) to specify exclusion.
Note: If <client> is an IPv6 address, it must be enclosed in square
brackets or [ ].
root=<client>[:<client>]...
Provides root privileges for the specified NFS clients. By default,
no host is granted root privilege.
Note: If <client> is an IPv6 address, it must be enclosed in square
brackets or [ ].
root=<-client>[:<-client>]...
Excludes the specified NFS clients from root privileges. Clients
must be preceded with - (dash) to specify exclusion.
Note: If <client> is an IPv6 address, it must be enclosed in square
brackets or [ ].
anon=<uid>
If a request comes from an unknown user, the UID should be used
as the effective user ID. Root users (UID =0) are considered
“unknown” by the NFS server unless they are included in the
root option. The default value for anon=<uid> is the user
“nobody.” If the user “nobody” does not exist, then the value
65534 is used.
!
CAUTION
Using anon=0 is not recommended for security concerns.
access=<client>[:<client>]...
Provides mount access for the specified NFS clients.
Note: If <client> is an IPv6 address, it must be enclosed in square
brackets or [ ].
552
Command Line Interface Reference for File 8.1
The server Commands
access=<-client>[:<-client>]...
Excludes the specified NFS clients from access even if they are
part of a subnet or netgroup that is allowed access. Clients must
be preceded with - (dash) to specify exclusion.
Note: If <client> is an IPv6 address, it must be enclosed in square
brackets or [ ].
nfsv4only
Specifies that the NFS export can be accessed only when a client is
using NFSv4.
vlan=<vlanid>[,<vlanid>,...]
Specifies that all hosts belonging to the VLAN ID will have access
to the exported file system. Hosts on other VLANs will be denied.
The VLANs are seperated by : or colon, just as any other
server_export option values.
Clients can be a hostname, netgroup, subnet, or IP address and
must be colon-separated, without spaces. A subnet is an IP
address/netmask (for example, 168.159.50.0/255.255.255.0). A
hostname is first checked for in the Data Mover’s local hosts
database, then on the NIS (host database) or DNS server (if
enabled). A netgroup is searched in the local netgroup database
and then on the NIS server's netgroup database. If the client name
does not exist in any case, then access is denied.
Note: Netgroups are supported. The hosts and netgroup files can be
created on the Control Station using your preferred method (for example,
with an editor, or by copying from another node), then copied to the Data
Mover.
-unexport [-perm] <pathname>
Temporarily unexports a <pathname> unless -perm is specified. If
-perm is specified, removes the entry from the export table.
FOR CIFS OPERATIONS
-list -name <sharename>
Displays the specified CIFS share.
[-option <options>]
Specifies the following comma-separated list of options:
server_export
553
The server Commands
[netbios=<netbios_name>]
When the share has an associated NetBIOS name that name is
required to locate the entry. Multiple CIFS entries can have same
<sharename> when belonging to a different NetBIOS name.
-name <sharename> [-ignore][-option <options>]
[-comment <comment>] <pathname>
Creates a CIFS share. Share name length is limited to 12 ASCII
characters unless Unicode is enabled, in which case the limit is 80
multibyte characters. Share names cannot include the following
characters: /, \, %, ", NUL (Null character), STX (start of header),
SOT (start of text), and LF (line feed). Share names can contain spaces
and other nonalphanumeric characters, but must be enclosed by
quotes if spaces are used. Share names cannot begin with a (hyphen). Share names are case-sensitive.
Comment length is limited to 256 bytes (represented as 256 ASCII
characters or a variable number of Unicode multibyte characters). A
comment cannot include the following characters: NUL (Null
character), STX (start of header), and SOT (start of text). Comments
can contain spaces and other nonalphanumeric characters, but must
be enclosed by quotes if spaces are used. Pathname length is limited
to 1024 bytes.
The -ignore option overwrites the previous options and comment in
the export table.
[-option <options>]
Specifies the following comma-separated options:
ro
Exports the <pathname> for CIFS clients as read-only.
rw=<client>[:<client>]...
Creates the share for CIFS clients as read-mostly. Read-mostly
means shared read-only to most clients, but read-write to those
specified. By default, the <pathname> is shared read-write to all.
A client may be either a <user_name> or <group_name>.
Note: If <client> is an IPv6 address, it must be enclosed in [ ] or square
brackets.
umask=<mask>
Specifies a user file-creation mask for the umask allowing NFS
permissions to be determined for the share.
554
Command Line Interface Reference for File 8.1
The server Commands
Note: The value of the umask is specified in octal and is XORed with the
permissions of 666 for files and 777 for directories. Common values
include 002, which gives complete access to the group, and read (and
directory search) access to others or 022 (default) which gives read (and
directory search), but not write permission to the group and others.
user=<default_user>
When using share level access (server_checkup provides
information), specifies a <default_user> which must be entered as
a character string. The user must be defined in the Data Mover’s
password file. There is a 20 character limit for the username.
group=<default_group>
When using share level access (server_checkup provides
information), indicates a <default_group> which must be entered
as a character string. There is a 256 character limit for group
names.
ropasswd=<share_passwd>
When using share level access (server_checkup provides
information), creates a read-only password to allow clients access
to the share. Passwords can be viewed in the list of shared entries.
rwpasswd=<share_rw_passwd>
When using share level access (server_checkup provides
information), creates a read-write password to allow clients
access to the share. Passwords are displayed in the list of shared
entries.
Note: Users from any client machine who know the value of the
ropasswd or rwpasswd can access the share for read-only and read-write
operations.
maxusr=<maxusr>
Sets the maximum number of simultaneous users permitted for a
share.
netbios=<netbiosName>[,netbios=<netbiosName>]...
Associates a share on a single domain with one or more NetBIOS
names created with server_checkup. By default, if a NetBIOS
name is not specified for a share, the share is visible to all
NetBIOS names.
server_export
555
The server Commands
-comment
Adds a comment for the specified CIFS share. The comment is
displayed when listing the shared entries.
-unexport -name <sharename>
Permanently removes access to a share by removing the entry from
the export table.
[-option <options>]
Specifies the following comma-separated options:
netbios=<netbios_name>
When the share has an associated NetBIOS name the NetBIOS
name is required to locate the entry. This is required because
multiple CIFS entries can have same <sharename> when
belonging to different NetBIOS name.
-option
type={CA[:]Encrypted[:][ABE[:]HASH[:][OCAutoI|OCVDO|OCNO
NE]]|NONE}
Specifies the following colon-separated list of options:
◆
Continuous Availability (CA): Indicates continuous availability of
data on the specific share.
◆
Encrypted: The server requires encrypted messages to access the
share.
◆
Access Based Enumeration (ABE): Only files and directories to
which the user has read access are visible (Access Based
Enumeration).
◆
HASH: Indicates that the share supports hash generation for
BranchCache retrieval.
◆
Offline Caching Attributes (OC): User MUST allow only manual
caching for the files open from this share by default.
• OCAutoI: The user MAY cache every file that it opens from
this share.
• OCVDO: The user MAY cache every file that it opens from this
share. Also, the user MAY satisfy the file requests from its local
cache.
• OCNone: Indicates no files or programs from the shared
folder are available offline.
556
Command Line Interface Reference for File 8.1
The server Commands
◆
SEE ALSO
EXAMPLE #1
HASH: The share supports hash generation for BranchCache
retrieval.
Configuring NFS on VNX, Managing Volumes and File Systems for VNX
Manually, server_checkup, and server_mount.
To export a specific NFS entry, type:
$ server_export server_2 -Protocol nfs /ufs1
server_2 : done
EXAMPLE #2
To export an NFS entry and overwrite existing settings, type:
$ server_export server_2 -Protocol nfs -ignore -option
access=172.24.102.0/255.255.255.0,root=172.24.102.240
-comment 'NFS Export for ufs1' /ufs1
server_2 : done
EXAMPLE #3
To export an NFS entry dir1, a subdirectory of the exported entry
/ufs1 in a multilevel file system hierarchy, type:
$ server_export server_2 -Protocol nfs /ufs1/dir1
server_2 : done
EXAMPLE #4
To assign a name to an NFS export, type:
$ server_export server_2 -Protocol nfs -name nasdocsfs
/ufs1
server_2 : done
EXAMPLE #5
To export an NFS entry using Kerberos authentication, type:
$ server_export server_2 -Protocol nfs -option
sec=krb5:ro,root=172.24.102.240,access=172.24.102.0/255.
255.255.0 /ufs2
server_2 : done
EXAMPLE #6
To export an NFS entry for NFSv4 only, type:
$ server_export server_2 -Protocol nfs -option nfsv4only
/ufs1
server_2 : done
EXAMPLE #7
To list all NFS entries, type:
$ server_export server_2 -Protocol nfs -list -all
server_2 :
export "/ufs2" sec=krb5 ro root=172.24.102.240 access=172.24.102.0/255.255.255.0
export "/ufs1" name="/nasdocsfs" access=172.24.102.0/255.255.255.0
root=172.24.102.240 nfsv4only comment="NFS Export for ufs1"
export "/" anon=0
access=128.221.252.100:128.221.253.100:128.221.252.101:128.221.253.101
server_export
557
The server Commands
EXAMPLE #8
To list NFS entries for the specified path, type:
$ server_export server_2 -list /ufs1
server_2 :
export "/ufs1" name="/nasdocsfs" access=172.24.102.0/255.255.255.0
root=172.24.102.240 nfsv4only comment="NFS Export for ufs1"
EXAMPLE #9
To temporarily unexport an NFS entry, type:
$ server_export server_2 -Protocol nfs -unexport /ufs2
server_2 : done
EXAMPLE #10
To export all NFS entries, type:
$ server_export server_2 -Protocol nfs -all
server_2 : done
EXAMPLE #11
To export a specific NFS entry in a language that uses multibyte
characters (such as Japanese), type:
$ server_export server_2 -Protocol nfs
/<nfs_entry_in_local_language_text>
server_2 : done
EXAMPLE #12
To permanently unexport an NFS entry, type:
$ server_export server_2 -unexport -perm /ufs1
server_2 : done
EXAMPLE #13
To permanently unexport all NFS entries, type:
$ server_export server_2 -Protocol nfs -unexport -perm
-all
server_2 : done
EXAMPLE #14
To provide access to a CIFS share, type:
$ server_export server_2 -name ufs1 /ufs1
server_2 : done
EXAMPLE #15
To create a CIFS share and overwrite existing settings, type:
$ server_export server_2 -name ufs1 -ignore -option
ro,umask=027,maxusr=200,netbios=dm112-cge0 -comment
'CIFS share' /ufs1
server_2 : done
558
Command Line Interface Reference for File 8.1
The server Commands
EXAMPLE #16
To create a CIFS share in a language that uses multibyte characters
(such as Japanese), type:
$ server_export server_2 -P cifs -name <name_in _local
_language_text> -comment <comment_in
local_language_text>
server_2 : done
EXAMPLE #17
To list all CIFS entries, type:
$ server_export server_2 -Protocol cifs -list
server_2 :
share "ufs1" "/ufs1" ro umask=027 maxusr=200 netbios=DM112-CGE0 comment="CIFS
share"
share "ufs2" "/ufs2" umask=022 maxusr=4294967295
EXAMPLE #18
To display a specific CIFS share, type:
$ server_export server_2 -list -name ufs1 -option
netbios=dm112-cge0
server_2 :
share "ufs1" "/ufs1" ro umask=027 maxusr=200 netbios=DM112-CGE0 comment="CIFS
share"
EXAMPLE #19
To export all CIFS entries, type:
$ server_export server_2 -Protocol cifs -all
server_2 : done
EXAMPLE #20
To list all NFS and CIFS entries, type:
$ server_export server_2
server_2 :
export "/ufs2" sec=krb5 ro root=172.24.102.240 access=172.24.102.0/255.255.255.0
export "/ufs1" nfsv4only
export "/" anon=0
access=128.221.252.100:128.221.253.100:128.221.252.101:128.221.253.101
share "ufs2" "/ufs2" umask=022 maxusr=4294967295
share "ufs1" "/ufs1" ro umask=027 maxusr=200 netbios=DM112-CGE0 comment="CIFS
share"
server_export
559
The server Commands
Where:
Value
export
sec
ro
root
access
share
ro
umask
maxuser
netbios
comment
Definition
A file system entry to be exported.
Security mode for the file system.
File system is to be exported as read-only.
IP address with root access.
Access is permitted for those IP addresses.
Entry to be shared.
File system is to be shared as read-only.
User creation mask.
Maximum number of simultaneous users.
NetBIOS name for the share.
Comment specified for the share.
EXAMPLE #21
To permanently unexport all CIFS and NFS entries, type:
$ server_export server_2 -unexport -perm -all
server_2 : done
EXAMPLE #22
To delete a CIFS share, type:
$ server_export server_2 -unexport -name ufs1 -option
netbios=dm112-cge0
server_2 : done
EXAMPLE #23
To delete all CIFS shares, type:
$ server_export server_2 -Protocol cifs -unexport -all
server_2 : done
EXAMPLE #24
To export a file system for NFS that specifies an IPv4 and IPv6
address, type:
$ server_export server_2 -Protocol nfs -option
access=172.24.108.10:[1080:0:0:0:8:800:200C:417A] /fs1
server_2 : done
EXAMPLE #25
To export a file system for NFS that specifies two IPv6 addresses,
type:
$ server_export server_2 -Protocol nfs -option
rw=[1080:0:0:0:8:80:200C:417A]:[1080:0:0:0:8:800:200C:41
7B] /fs1
server_2 : done
560
Command Line Interface Reference for File 8.1
The server Commands
EXAMPLE #26
To verify that the file system was exported, type:
$ server_export server_2 -list /fs1
server_2 :
export "/fs1" rw=[1080:0:0:0:8:80:200C:417A]:[1080:0:0:0:8:800:200C:417B]
EXAMPLE #27
To export the fs42 file system of the VDM vdm1, type:
$ server_export vdm1 –P nfs /fs42
done
EXAMPLE #28
To create a share foo on the server PALIC with HASH and ABE
enabled, type:
$ server_export server_3 -name foo -option netbios=PALIC,
type=ABE:HASH /fs3/foo
server_3 : done
EXAMPLE #29
To change attributes to this share to ABE only, type:
$ server_export server_3 -name foo -option netbios=PALIC,
type=ABE /fs3/foo
server_3 : done
EXAMPLE #30
To remove all the attributes, type:
server_export server_3 -name foo -ignore -option
netbios=PALIC,type=None /fs3/fro
server_3 : done
EXAMPLE #31
To view the attributes, type:
server_export server_3 share "foo" "/fs3/fro"
type=ABE:HASH umask=022 maxusr=4294967295 netbios=PALIC
server_3 : done
EXAMPLE #32
To create a share foo on the server palic with CA and ABE enabled,
type:
$ server_export server_3 -name foo -option netbios=PALIC,
type=CA:ABE /fs3/foo
server_3 : done
server_export
561
The server Commands
EXAMPLE #33
To change attributes of the share foo to CA only, type:
$ server_export server_3 -name foo -option netbios=PALIC,
type=CA /fs3/foo
server_3 : done
EXAMPLE #34
To view the attributes, type:
$ server_export server_3 share "foo" "/fs3/fro" type=CA
umask=022 maxusr=4294967295 netbios=PALIC
server_3 : done
EXAMPLE #35
To create a share share10 accessible only through encrypted SMB
messages, type:
$ server_export vdm1 -P cifs -name share10 -o
type=Encrypted /fs42/protected_dir1
server_3 : done
562
Command Line Interface Reference for File 8.1
The server Commands
server_file
Copies files between the Control Station and the specified Data
Movers.
SYNOPSIS
server_file {<movername>|ALL}
{-get|-put} <src_file> <dst_file>
DESCRIPTION
server_file copies the source file from the specified Data Mover (or
Control Station) to the destination file on the Control Station (or
specified Data Mover). The <src_file> indicates the source file, and
the name <dst_file> indicates destination file. By default, if a
directory is not specified on the Data Mover, the /.etc directory is
used.
The ALL option executes the command for all Data Movers.
OPTIONS
-get <src_file> <dst_file>
Copies the source file on Data Mover to destination file on Control
Station. Both the <src_file> and <dst_file> may be full pathnames.
-put <src_file> <dst_file>
Copies source file on the Control Station to the destination file on the
Data Mover. Both the <src_file> and <dst_file> must be full
pathnames.
!
CAUTION
This command overwrites existing files of the same name without
notification. Use care when copying files.
EXAMPLE #1
To copy a file from the Control Station to a Data Mover, type:
$ server_file server_2 -put passwd passwd
server_2 : done
EXAMPLE #2
To copy a file from the Data Mover to the Control Station, type:
$ server_file server_2 -get passwd /home/nasadmin/passwd
server_2 : done
server_file
563
The server Commands
server_fileresolve
Starts, deletes, stops, checks, and displays the fileresolve service for
the specified Data Mover. Filereseolve service facilitates
inode-to-filename translation. This translation is required when
administrator monitors the "fs.qtreeFile" and "fs.filesystem" statistics.
SYNOPSIS
server_fileresolve <movername>
-service { -start [-maxlimit <1M>]
| -stop
| -delete
| -status }
| -list
| -add <path_name> [,...]
| -drop <path_name> [,...]
| -lookup { -filesystem <fs_name> -inode <inode>[,...]
| -qtree <qt_name> -inode <inode>[,...] } [...]
DESCRIPTION
OPTIONS
Controls and manages the fileresolve service, which crawls through
file systems specified by the user. To have the fileresolve service
started at boot time, it is recommended that this command be added
to the eof config file for the Data Mover.
-service {-start [-maxlimit <1M>]
Starts the fileresolve service on the specified Data Mover. By default,
the fileresolve service caches upto 1 million files (this takes about
32MB of memory on the Data Mover). By increasing the maximum
limit of the inode-to-filename translation cache, from 1M to 2M, it
would increase memory consumed by the service to 64M.
To change the maxlimit, use the following command:
server_fileresolve <movername> -service -start –maxlimit
<new_value>
This new limit will be preserved across Data Mover reboots.
However, when a new limit is applied, the entire inode-to-filename
cache will be flushed and rebuilt. The File system crawler adds files
to its cache in the order they are traversed. Hence, the first 1 million
files traversed (by default) go in the cache.
-stop
Flushes the inode-to-filename cache and stops the service. Deleting
the service also would free up the memory consumed by the
564
Command Line Interface Reference for File 8.1
The server Commands
fileresolve service and deletes the configuration files created by the
service.
-delete
Deletes the fileresolve service on the specified Data Mover.
-status
Checks the status of the files that are added to the cache on the
specified Data Mover.
-list
Displays the file systems/ directories that are in the configuration
and used for crawling.
-add <path_name> [,...]
Adds specified path to the server_fileresolver configuration. Crawls
the specified path and buildsthe inode-to-filename cache.
To add a specific file that should be included in the inode-to-filename
map, the followingcommand should be used:
server_fileresolve server_X –add <path for file>
-drop <path_name>[,...]
Drops specified path to the server_fileresolver configuration.
Inode-to-filename cache for the specified path is not cleared until the
service is restarted.
-lookup {-filesystem <fs_name> -inode <inode> [,...]
Performs an on demand crawl of the specified file system to translate
the inode to a pathname. If the pathname is not found for the inode,
the inode value is returned. For example, server_stats displays this
inode value instead of a path name in its output.
The user can do a 'deep, non-cached' lookup of the inode to discover
the pathname (if it still exists). However, this could take time (in the
order of minutes). Hence, server_stats only attempts to lookup in the
cache and does not attempt a full File system crawl.
Note: If the file name is successfully resolved, full pathname is
returned. Even if the file name is the same as the inode path is
appended.
-lookup -qtree <qt_name> -inode <inode> [,...] } [...]
server_fileresolve
565
The server Commands
Performs an on demand crawl of the specified quota tree to translate
the inode to a pathname.
EXAMPLE #1
To display the new paths added, type:
$ server_fileresolve server_2 -add /server_2/ufs_0
server_2 :
New paths are added
EXAMPLE #2
To list the specified file paths that are included in the
inode-to-filename map, type:
$ server_fileresolve server_2 -list
server_2 :
PATH
/server_2/ufs_5
/server_2/ufs_4
/server_2/ufs_3
/server_2/ufs_2
/server_2/ufs_1
/server_2/ufs_0
EXAMPLE #3
To check the status of the fileresolve services, type:
$ server_fileresolve server_2 -service -status
server_2 :
FileResolve service is running :Max Limit of the cache:1000000 Entries used:10
Dropped entries:0
EXAMPLE #4
To drop the specified path to the server_fileresolver configuration,
type
$ server_fileresolve server_2 -drop /server_2/ufs_0
server_2 :
Paths are dropped
Warning: Restart service to remove the cached entries of dropped paths.
EXAMPLE #5
To lookup multiple inodes within the same file system, type:
$ server_fileresolve server_2 -lookup -filesystem ufs_0
-inode 61697,61670,61660
server_2 :
Filesystem/QTree
ufs_0
Inode
61660
ufs_0
61670
ufs_0
61697
EXAMPLE #6
566
Path
/server_2/ufs_0/dir00000
/testdir/yYY_0000039425.tmp
/server_2/ufs_0/dir00000/
testdir/kNt_0000028175.tmp
/server_2/ufs_0/dir00000
/testdir/gwR_0000058176.tmp
To lookup multiple inodes within a Quota Tree, type:
Command Line Interface Reference for File 8.1
The server Commands
$ server_fileresolve
-inode 61697
server_2 :
Filesystem/QTree
dir00000
Inode
61697
server_2 -lookup -qtree dir00000
Path
/server_2/ufs_0/dir00000/
testdir/gwR_0000058176.tmp
server_fileresolve
567
The server Commands
server_ftp
Configures the FTP server configuration for the specified Data
Movers.
SYNOPSIS
server_ftp {<movername>|ALL}
-service {-status|-start|-stop|{-stats [-all|-reset]}}
| -info
| -modify
[-controlport <controlport>]
[-dataport <dataport>]
[-defaultdir <path>]
[-homedir {enable|disable}]
[-keepalive <keepalive>]
[-highwatermark <highwatermark>]
[-lowwatermark <lowwatermark>]
[-deniedusers [<path>]]
[-welcome [<path>]]
[-motd [<path>]]
[-timeout <timeout>]
[-maxtimeout <maxtimeout>]
[-readsize <readsize>]
[-writesize <writesize>]
[-maxcnx <maxcnx>]
[-umask <umask>]
[-sslcontrol {no|allow|require|requireforauth}]
[-ssldata {allow|require|deny}]
[-sslpersona {anonymous|default|<persona_name>}]
[-sslprotocol {default|ssl3|tls1|all}]
[-sslcipher {default|<cipherlist>}]
[-sslcontrolport <sslcontrolport>]
[-ssldataport <ssldataport>]
DESCRIPTION
OPTIONS
server_ftp configures the ftp daemon. Optional SSL security support
is available. The modifications are performed when the ftp daemon is
stopped and are reflected after restart of the ftp daemon. There is no
need to reboot the Data Mover for the changes to be reflected.
server_ftp {<movername>|ALL}
Sends a request to the Data Mover to get all the parameters of the ftp
daemon.
ALL option executes the command for all Data Movers.
-service {-status|-start|-stop|{-stats [-all| -reset]}}
568
Command Line Interface Reference for File 8.1
The server Commands
-status
Retrieves the current status of the ftp daemon.
-start
Starts the ftp daemon. The start option persists after the daemon
is rebooted.
-stop
Stops the ftp daemon.
-stats [all|reset]
Displays the statistics of the ftp daemon. The reset option resets
all the ftp server statistics. The all option displays detailed
statistics.
-info
Retrieves all the parameters for the ftp daemon along with its current
status.
-modify
Modifies the ftp daemon configuration. The ftp daemon has to be
stopped to carry out the changes. The modifications are taken into
account when the service is restarted.
-controlport <controlport>
Sets the local tcp port for control connections. By default, the port
is 21. When control port is set to 0, it disables unsecure ftp usage
and validates port for implicit secure connection on SSL port
(default 990).
Note: This default port can be changed using the sslcontrolport option.
-dataport <dataport>
Sets the local tcp port for active data connections. By default, the
port is 20. When <dataport> is set to 0, the port is allocated
dynamically by the server in active mode.
-defaultdir <path>
Sets the default user directory when the user home directory is
not accessible. This option replaces "ftpd.defaultdir.” By default,
"/" is used.
-homedir {enable|disable}
Restricts or allows user access to their home directory tree. When
enabled the user is allowed access to their home directory only. If
server_ftp
569
The server Commands
the user home directory is not accesible, access is denied. During
connection the user is denied access to data outside of their home
directory space. By default, this feature is disabled.
Note: Using FTP on VNX provides more information about how the home
directory of a user is managed.
-umask <umask>
Defines the mask to set the mode bits on file or directory creation.
By default the mask is 027, which means that xrw-r---- mode bits
are assigned.
-keepalive <keepalive>
Sets TCP keepalive value for the ftp daemon. This value is given
in seconds. By default, the value is 60. The value 0 disables the
TCP keepalive option. The maximum value is 15300 (255
minutes).
-highwatermark <highwatermark>
Sets TCP high watermark value (amount of data stored without
knowledge of the client) for the ftp daemon. By default, the value
is 65536. The minimum value is 8192, and the maximum value is
1048576 (1 MB).
!
CAUTION
Do not modify this parameter without a thorough knowledge of
the impact on FTP client behavior.
-lowwatermark <lowwatermark>
Sets TCP low watermark value (amount of the data to be added,
after the highwatermark has been reached and new data can be
accepted from the client) for the ftp daemon. The minimum value
is 8192, maximum value is 1048576 (1 MB), and default value is
32768.
!
CAUTION
Do not modify this parameter without a thorough knowledge of
the impact on FTP client behavior.
570
Command Line Interface Reference for File 8.1
The server Commands
-deniedusers <deniedusers_file>
Denies FTP access to specific users on a Data Mover. Specifies the
path and text file containing the list of usernames to be denied
access. Places each username on a separate line. By default, all
users are allowed.
-welcome <welcome_file>
Specifies path of the file to be displayed on the welcome screen.
For example, this file can display a login banner before the user is
requested for authentication data. By default, no welcome
message is displayed.
-motd <motd_file>
Specifies path of the file to be displayed on the welcome screen.
Users see a welcome screen ("message of the day") after they
successfully log in. By default, no message of the day is
displayed.
-timeout <timeout>
Specifies the default inactivity time-out period (when not set by
the client). The value is given in seconds. After the specified time
if there is no activity, the client is disconnected from the server
and will have to re-open a connection. By default, the <timeout>
value is 900 seconds. The minimum value is 10 seconds, and the
maximum value is 7200.
-maxtimeout <maxtimeout>
Sets the maximum time-out period allowed by the client. The
value is given in seconds and any value larger than maximum
time-out period is not allowed. By default, the <maxtimeout>
value is 7200 seconds. The minimum value is 10 seconds, and the
maximum value is 7200.
-readsize <readsize>
Sets the size for reading files from the disk. The value must be
greater than 8192, and a multiple of 8K. By default, the
<readsize> is 8192 bytes. The minimum value is 8192, and the
maximum value is 1048576 (1 MB).
-writesize <writesize>
Sets the size for writing files from the disk. The value must be
greater than 8192, and a multiple of 8K. By default, the
<writesize> is 49152 (48 KB). The minimum value is 8192, and the
maximum value is 1048576 (1 MB).
server_ftp
571
The server Commands
-maxcnx <maxcnx>
Sets the maximum number of control connections the ftp daemon
will support. By default, the <maxcnx> value is set to 65535
(64K-1). The minimum value is 1, and the maximum value is
65535 (64K-1).
-sslcontrol {no|allow|require|requireforauth}
Uses SSL for the ftp control connection depending on the
attributes specified. By default, SSL is disabled. The no option
disables SSL control. The allow option specifies that SSL is
enabled, but the user can still connect without SSL. The require
option specifies that SSL is required for the connection. The
requireforauth option specifies that SSL is required for
authentication. The control path goes back to unsecure after this
authentication. When the client is behind a firewall, this helps the
firewall to filter the ftp commands requiring new port access.
Note: Before the server can be configured with SSL, the Data Mover must be
set up with a private key and a public certificate. This key and certificate are
identified using a persona. In addition, the necessary Certificate Authority
(CA) certificates used to identify trusted servers must be imported into the
Data Mover. Use the system’s PKI feature to manage the use of certificates
prior to configuring SSL operation.
-ssldata {no|allow|require}
Uses SSL for the data connection depending on the attributes
specified. The no option disables SSL. The allow option specifies
that SSL is enabled, but the user can also transfer data without
SSL. The require option specifies that SSL is required for data
connection. The ssldata value cannot be set to allow or require if
sslcontrol is set to no. By default, SSL is disabled.
Note: These options are set on the server but are dependent on ftp client
capabilities. Some client capabilities may be incompatible with server
settings. Using FTP on VNX provides information on validating compatibility.
-sslpersona {anonymous|default|<persona_name>}
Specifies the persona associated with the Data Mover. Personas
are used to identify the private key and public certificate used by
SSL. The default value specified is default (each Data Mover is
configured with a persona named default). The anonymous value
specifies that SSL can operate without using a certificate. This
implies that the communication between client and server is
encrypted and data integrity is guaranteed.
572
Command Line Interface Reference for File 8.1
The server Commands
Note: Use server_certificate to configure the persona before using server_ftp.
-sslprotocol {default|ssl3|tls1|all}
Specifies the SSL protocol version that the ftp daemon on the
server accepts:
◆
ssl3 — Only SSLv3 connections
◆
tls1 — Only TLSv1 connections
◆
all — Both SSLv3 and TLSv1 connections
◆
default — Uses the value set in the ssl.protocol parameter which,
by default, is 0 (SSLv3 and TLSv1)
-sslcipher {default|<cipherlist>}
Specifies the SSL cipher suite. The value of default is the value set
in the ssl.cipher parameter. This value means that all ciphers are
supported by VNX except the Anonymous Diffie-Hellman,
NULL, and SSLv2 ciphers and that the supported ciphers are
sorted by the size of the encryption key.
-sslcontrolport <sslcontrolport>
Sets the implicit control port for FTP connections over SSL. By
default, the port is 990. To disable implicit FTP connections over
SSL, the <sslcontrolport> must be set to 0.
-ssldataport <ssldataport>
Sets the local tcp port for active data connections using implicit
FTP connections over SSL. By default, the port is 899. If the
ssldataport is set to 0, the Data Mover will use a port allocated by
the system.
SEE ALSO
EXAMPLE #1
“server_certificate”.
To retrieve all the parameters for the ftp daemon and its status, type:
$ server_ftp server_2 -info
FTP started
=========
controlport
dataport
defaultdir /
homedir
umask
21
20
.etc/ftpd/pub
disable
027
tcp keepalive
tcp high watermark
tcp low watermark
1 minute
65536 bytes
32768 bytes
server_ftp
573
The server Commands
readsize
writesize
denied users file path
welcome file path
motd file path
8192 bytes
49152 bytes
/.etc/ftpd/conf/ftpusers
/.etc/ftpd/conf/welcome
/.etc/ftpd/conf/motd
session timeout
max session timeoutQ
900 seconds
7200 seconds
Security Options
=============
sslpersona
sslprotocol
sslcipher
default
default
default
FTP over TLS explicit Options
---------------------------------------sslcontrol
SSL require for authentication
ssldata
allow SSL
FTP over SSL implicit Options
----------------------------------------sslcontrolport
990
ssldataport
989
EXAMPLE #2
574
To display the statistics of the ftp daemon, type:
Login Type
==========
Anonymous
Unix
CIFS
$ server_ftp server_2 -service -stats
Successful
Failed
==========
=======
10
0
3
2
7
1
Data transfers
==============
Write Bin
Read Bin
Write ASCII
Read ASCII
SSL Write Bin
SSL Read Bin
SSL Write ASCII
SSL Read ASCII
Count
=====
10
0
2
0
5
15
0
0
min
====
10.00
---1.00
---5.00
7.00
-------
Command Line Interface Reference for File 8.1
Throughput (MBytes/sec)
average
=======================
19.00
---1.50
--- 17.00
25.00
-------
max
====
20.00
---2.00
---18.00
35.00
-------
The server Commands
Where:
Value
Throughput
(MBytes/sec)
average
Data transfers
Count
min
max
EXAMPLE #3
Commands
========
USER
PASS
QUIT
PORT
EPRT
...
FEAT
Definition
Throughput is calculated using the size of the file (Mbytes) divided by the duration of the transfer
(in seconds).
Average is the average of the throughputs (sum of the throughputs divided by the number of
transfers).
Defines the type of transfer.
Number of operations for a transfer type.
Minimum time in milliseconds required to execute the operation (with regards to Data Mover).
Maximum time in milliseconds required to execute the operation (with regards to Data Mover).
To display the statistics of the ftp daemon with details, type:
$ server_ftp server_2 -service -stats -all
Count
=====
23
23
23
45
10
....
23
SITE Commands
=============
UMASK
IDLE
CHMOD
HELP
BANDWIDTH
KEEPALIVE
PASV
Count
=====
0
10
0
0
0
10
56
OPTS Commands
=============
UTF8
Count
=====
10
Login Type
==========
Anonymous
Unix
CIFS
Successful
==========
10
3
7
Connections
===========
Non secure
---------Control
Count
=====
Failed
=======
0
2
1
10
server_ftp
575
The server Commands
Data
44
Explicit SSL
-----------Control Auth
Control
Data
3
8
20
Implicit SSL
-----------Control
Data
0
0
Data transfers
==============
Write Bin
Read Bin
Write ASCII
Read ASCII
SSL Write Bin
SSL Read Bin
SSL Write ASCII
SSL Read ASCII
Count
=====
10
0
2
0
5
15
0
0
min
========
10.00
---1.00
---5.00
7.00
-------
Throughput (MBytes/sec)
average
====================
19.00
---1.50
---17.00
25.00
-------
Where:
Value
Commands
Count
SITE Commands
POTS Commands
EXAMPLE #4
Definition
FTP protocol command name.
Number of commands received by Data Mover.
Class of command in FTP protocol.
Class of command in FTP protocol.
To retrieve the status of the ftp daemon, type:
$ server_ftp server_3 -service -status
server_3 : done
State : running
EXAMPLE #5
To start the ftp daemon, type:
$ server_ftp server_2 -service -start
server_2 : done
EXAMPLE #6
To stop the ftp daemon, type:
$ server_ftp server_2 -service -stop
server_2 : done
576
Command Line Interface Reference for File 8.1
max
=======
20.00
---2.00
---18.00
35.00
-------
The server Commands
EXAMPLE #7
To set the local tcp port for the control connections, type:
$ server_ftp server_2 -modify -controlport 256
server_2 :done
FTPD CONFIGURATION
==================
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
Low watermark
Timeout
Max timeout
Read size
Write size
Umask
Max connection
:
:
:
:
:
:
:
:
:
:
:
:
:
:
stopped
256
20
/
disable
1
65536
32768
900
7200
8192
49152
27
65535
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
Protocol
Cipher
Control port
Data port
:
:
:
:
:
:
:
disable
disable
default
default
default
990
989
EXAMPLE #8
To set the local tcp port for active data connections, type:
$ server_ftp server_2 -modify -dataport 257
server_2 : done
FTPD CONFIGURATION
==================
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
Low watermark
Timeout
Max timeout
Read size
Write size
Umask
:
:
:
:
:
:
:
:
:
:
:
:
:
stopped
256
257
/
disable
1
65536
32768
900
7200
8192
49152
27
server_ftp
577
The server Commands
Max connection
: 65535
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
Protocol
Cipher
Control port
Data port
:
:
:
:
:
:
:
EXAMPLE #9
disable
disable
default
default
default
990
989
To change the default directory of a user when the home directory is
not accessible, type:
$ server_ftp server_2 -modify -defaultdir /big
server_2 : done
FTPD CONFIGURATION
==================
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
Low watermark
Timeout
Max timeout
Read size
Write size
Umask
Max connection
:
:
:
:
:
:
:
:
:
:
:
:
:
:
stopped
256
257
/big
disable
1
65536
32768
900
7200
8192
49152
27
65535
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
Protocol
Cipher
Control port
Data port
:
:
:
:
:
:
:
disable
disable
default
default
default
990
989
EXAMPLE #10
To allow users access to their home directory tree, type:
$ server_ftp server_2 -modify -homedir enable
server_2 : done
FTPD CONFIGURATION
==================
578
Command Line Interface Reference for File 8.1
The server Commands
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
Low watermark
Timeout
Max timeout
Read size
Write size
Umask
Max connection
:
:
:
:
:
:
:
:
:
:
:
:
:
:
stopped
256
257
/big
enable
1
65536
32768
900
7200
8192
49152
27
65535
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
Protocol
Cipher
Control port
Data port
:
:
:
:
:
:
:
disable
disable
default
default
default
990
989
EXAMPLE #11
To restrict users access to their home directory tree, type:
$ server_ftp server_2 -modify -homedir disable
server_2 : done
FTPD CONFIGURATION
==================
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
Low watermark
Timeout
Max timeout
Read size
Write size
Umask
Max connection
:
:
:
:
:
:
:
:
:
:
:
:
:
:
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
: disable
: disable
: default
stopped
256
257
/big
disable
1
65536
32768
900
7200
8192
49152
27
65535
server_ftp
579
The server Commands
Protocol
Cipher
Control port
Data port
EXAMPLE #12
:
:
:
:
default
default
990
989
To set the default umask for creating a file or a directory by means of
the ftp daemon, type:
$ server_ftp server_2 -modify -umask 077
server_2 : done
FTPD CONFIGURATION
==================
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
Low watermark
Timeout
Max timeout
Read size
Write size
Umask
Max connection
:
:
:
:
:
:
:
:
:
:
:
:
:
:
stopped
256
257
/big
disable
1
65536
32768
900
7200
8192
49152
77
65535
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
Protocol
Cipher
Control port
Data port
:
:
:
:
:
:
:
disable
disable
default
default
default
990
989
EXAMPLE #13
To set the TCP keepalive for the ftp daemon, type:
$ server_ftp server_2 -modify -keepalive 120
server_2 : done
FTPD CONFIGURATION
==================
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
580
:
:
:
:
:
:
:
stopped
256
257
/big
disable
120
65536
Command Line Interface Reference for File 8.1
The server Commands
Low watermark
Timeout
Max timeout
Read size
Write size
Umask
Max connection
:
:
:
:
:
:
:
32768
900
7200
8192
49152
77
65535
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
Protocol
Cipher
Control port
Data port
:
:
:
:
:
:
:
disable
disable
default
default
default
990
989
EXAMPLE #14
To set the TCP high water mark for the ftp daemon, type:
$ server_ftp server_2 -modify -highwatermark 90112
server_2 : done
FTPD CONFIGURATION
==================
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
Low watermark
Timeout
Max timeout
Read size
Write size
Umask
Max connection
:
:
:
:
:
:
:
:
:
:
:
:
:
:
stopped
256
257
/big
disable
120
90112
32768
900
7200
8192
49152
77
65535
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
Protocol
Cipher
Control port
Data port
:
:
:
:
:
:
:
disable
disable
default
default
default
990
989
server_ftp
581
The server Commands
EXAMPLE #15
To set the TCP low water mark for the ftp daemon, type:
$ server_ftp server_2 -modify -lowwatermark 32768
server_2 : done
FTPD CONFIGURATION
==================
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
Low watermark
Timeout
Max timeout
Read size
Write size
Umask
Max connection
:
:
:
:
:
:
:
:
:
:
:
:
:
:
stopped
256
257
/big
disable
120
90112
32768
900
7200
8192
49152
77
65535
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
Protocol
Cipher
Control port
Data port
:
:
:
:
:
:
:
disable
disable
default
default
default
990
989
EXAMPLE #16
To restrict FTP server access to specific users, type:
$ server_ftp server_2 -modify -deniedusers
/.etc/mydeniedlist
server_2 : done
FTPD CONFIGURATION
==================
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
Low watermark
Denied users conf file
Timeout
Max timeout
Read size
582
:
:
:
:
:
:
:
:
:
:
:
:
stopped
256
257
/big
disable
120
90112
32768
/.etc/mydeniedlist
900
7200
8192
Command Line Interface Reference for File 8.1
The server Commands
Write size
Umask
Max connection
: 49152
: 77
: 65535
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
Protocol
Cipher
Control port
Data port
:
:
:
:
:
:
:
EXAMPLE #17
disable
disable
default
default
default
990
989
To set the path of the file displayed before the user logs in, type:
$ server_ftp server_2 -modify -welcome
/.etc/mywelcomefile
server_2 : done
FTPD CONFIGURATION
==================
State
Control Port
Data Port
Default dir
Home dir
Keepalive
High watermark
Low watermark
Welcome file
Timeout
Max timeout
Read size
Write size
Umask
Max connection
:
:
:
:
:
:
:
:
:
:
:
:
:
:
:
stopped
256
257
/big
disable
120
90112
32768
/.etc/mywelcomefile
900
7200
8192
49152
77
65535
SSL CONFIGURATION
=================
Control channel mode
Data channel mode
Persona
Protocol
Cipher
Control port
Data port
:
:
:
:
:
:
:
disable
disable
default
default
default
990
989
server_ftp
583
The server Commands
server_http
Configures the HTTP configuration file for independent services,
such as VNX FileMover, for the specified Data Movers.
SYNOPSIS
server_http {<movername>|ALL}
-info [<feature>]
| -service <feature> {-start|-stop}
| -service [<feature>] -stats [-reset]
| -modify <feature>
[-threads <threads>]
[-users {valid|<user>[,<user>,<user>...]}]
[-hosts <ip>[,<ip>,<ip>...]]
[-port <port_number>]
[-timeout <max_idle_time>]
[-maxrequests <maxrequests>]
[-authentication {none|basic|digest}]
[-realm <realm_name>]
[-ssl {required|off}]
[-sslpersona {anonymous|default|<persona_name>}]
[-sslprotocol {default|ssl3|tls1|all}]
[-sslcipher {default|<cipherlist>}]
| -append <feature>
[-users {valid|<user>[,<user>,<user>...]}]
[-hosts <ip>[,<ip>,<ip>...]]
| -remove <feature>
[-users {valid|<user>[,<user>,<user>...]}]
[-hosts <ip>[,<ip>,<ip>...]]
DESCRIPTION
server_http manages user and host access to HTTP servers for
independent services such as FileMover.
ALL option executes the command for all Data Movers.
OPTIONS
-info [<feature>]
Displays information about the specified feature or all features
including server status, port, threads, requests allowed, timeout,
access control, and SSL configuration.
-service <feature> {-start|-stop}
Stops or starts the HTTP server for the specified feature.
-service [<feature>] -stats [-reset}
Lists the usage statistics of the HTTP server for the specified feature
or all features. If -reset is specified, statistics are reset to zero.
584
Command Line Interface Reference for File 8.1
The server Commands
-modify <feature>
Displays the current HTTP protocol connection for the specified
feature. When issued with options, -modify sets the HTTP protocol
connection for the specified option. Any options previously set will
be overwritten.
[-threads <threads>]
Sets the number of threads (default=20) for incoming service
requests. The minimum value is 4, the maximum 99. The HTTP
threads are started on the Data Mover at boot time.
[-users [valid|<user>[,<user>,<user>…]]]
Allows the users who correctly authenticate as defined in the
Data Mover passwd file (server_user provides more information)
to execute commands for the specified <feature>.
If valid is entered, all users in the passwd file are allowed to use
digest authentication. A comma-separated list of users can also
be given. If no users are given, digest authentication is turned off.
[-hosts <ip>[,<ip>,<ip>…]]
Specifies hosts by their IP addresses that are allowed to execute
commands for the specified <feature>.
[-port <port_number>]
Specifies the port on which the HTTP server listens for incoming
service requests. By default, the HTTP server instance for
FileMover listens on port 5080.
[-timeout <max_idle_time>]
Specifies the maximum time the HTTP server waits for a request
before disconnecting from the client. The default value is 60
seconds.
[-maxrequests <max_requests>]
Specifies the maximum number of requests allowed. The default
value is 300 requests.
[-authentication {none|basic|digest}]
Specifies the authentication method. none disables user
authentication, allowing for anonymous access (that is, no
authentication). basic authentication uses a clear text password.
digest authentication uses a scripted password. The default value
is digest authentication.
server_http
585
The server Commands
[-realm <realm_name>]
Specifies the realm name. This information is required when
authentication is enabled (that is, the -authentication option is set
to basic or digest). The default realm name for FileMover is
DHSM_authorization.
[-ssl {required|off}]
Specifies whether the HTTP server runs in secure mode, that is,
only accepts data receive on encrypted SSL sessions. The default
value is off.
Note: Before the HTTP server can be configured with SSL, the Data
Mover must be set up with a private key and public certificate. This key
and certificate are identified using a persona. In addition, the necessary
Certificate Authority (CA) certificates to identify trusted servers must be
imported into the Data Mover. Use the system’s PKI feature to manage
the use of certificates prior to configuring SSL operation.
[-sslpersona {default|anonymous|<persona_name>}]
Specifies the persona associated with the Data Mover. Personas
are used to identify the private key and public certificate used by
SSL. The default value is default (each Data Mover is currently
configured with a single persona named default). anonymous
specifies that SSL can operate without using a certificate.
[-sslprotocol {default|ssl3|tls1|all}]
Specifies the SSL protocol version the HTTPS server accepts:
•
•
•
•
ssl3 — Only SSLv3 connections
tls1 — Only TLSv1 connections
all — Both SSLv3 and TLSv1 connections
default — Uses the value set in the ssl.protocol parameter
which, by default, is 0 (SSLv3 and TLSv1)
[-sslcipher {default|<cipherlist>}]
Specifies the SSL cipher suite. The value of default is the value set
in the ssl.cipher parameter which, by default, is
ALL:!ADH:!SSLv2:@STRENGTH. This value means that all
ciphers are supported by VNX except the Anonymous
Diffie-Hellman, NULL, and SSLv2 ciphers and that the supported
ciphers are sorted by the size of the encryption key.
-append <feature> [-users {valid|<user>[,<user>,
<user>…]}][-hosts <ip>[,<ip>,<ip>…]}]
Adds the specified users or hosts to the list of those who can execute
commands for the specified <feature> without having to re-enter the
586
Command Line Interface Reference for File 8.1
The server Commands
existing list. The users and hosts descriptions provide information. If
users or hosts are not specified, displays the current HTTP
configuration.
-remove <feature> [-users {valid|<user>[,<user>,
<user>…]}][-hosts <ip>[,<ip>,<ip>…]]
Removes the specified users and hosts from the list of those who can
execute commands for the specified <feature> without impacting
others in the list. The users and hosts descriptions provide
information. If users or hosts are not specified, displays the current
HTTP configuration.
SEE ALSO
Using VNX FileMover, Security Configuration Guide for File, fs_dhsm,
server_certificate, and nas_ca_certificate.
EXAMPLE #1
To display information about the HTTP protocol connection for the
FileMover service, type:
$ server_http server_2 -info dhsm
server_2 : done
DHSM FACILITY CONFIGURATION
Service name
: EMC File Mover service
Comment
: Service facility for getting DHSM attributes
Active
: False
Port
: 5080
Threads
: 16
Max requests
: 300
Timeout
: 60 seconds
ACCESS CONTROL
Allowed IPs
: any
Authentication
: digest ,Realm : DHSM_Authorization
Allowed user
: nobody
SSL CONFIGURATION
Mode
: OFF
Persona
: default
Protocol
: default
Cipher
: default
Where:
Value
Service name
active
port
threads
max requests
timeout
allowed IPs
Definition
Name of the FileMover service.
Whether VNX FileMover is enabled or disabled on the file system.
TCP port of the FileMover service.
Number of threads reserved for the FileMover service.
Maximum number of HTTP requests the service allows to keep the connection alive.
The time in seconds until which the service is kept alive after a period of no activity.
List of client IP addresses that are allowed to connect to the service.
server_http
587
The server Commands
authentication
allowed user
mode
persona
protocol
cipher
The HTTP authentication method used by the service.
Users allowed to connect to the service.
The SSL mode.
Name of the persona associated with the Certificate for establishing a secure connection.
The level of SSL protocol used for the service.
The cipher suite the service is negotiating, for establishing a secure connection with the client.
EXAMPLE #2
To display statistical information about the HTTP protocol
connection for the FileMover service, type:
$ server_http server_2 -service dhsm -stats
server_2 : done
Statistics report for HTTPD facility DHSM :
Thread activity
Maximum in use count
: 0
Connection
IP filtering rejection count : 0
Request
Authentication failure count : 0
SSL
Handshake failure count
: 0
EXAMPLE #3
To configure an HTTP protocol connection for FileMover using SSL,
type:
$ server_http server_2 -modify dhsm -ssl required
server_2 : done
EXAMPLE #4
To modify the threads option of the HTTP protocol connection for
FileMover, type:
$ server_http server_2 -modify dhsm -threads 40
server_2 : done
DHSM FACILITY CONFIGURATION
Service name
: EMC File Mover service
Comment
: Service facility for getting DHSM attributes
Active
: False
Port
: 5080
Threads
: 40
Max requests
: 300
Timeout
: 60 seconds
ACCESS CONTROL
Allowed IPs
: any
Authentication
: digest ,Realm : DHSM_Authorization
Allowed user
: nobody
SSL CONFIGURATION
Mode
: OFF
Persona
: default
Protocol
: default
588
Command Line Interface Reference for File 8.1
The server Commands
Cipher
EXAMPLE #5
: default
To allow specific users to manage the HTTP protocol connection for
FileMover, type:
$ server_http server_2 -modify dhsm -users valid -hosts
10.240.12.146
server_2 : done
EXAMPLE #6
To add specific users who can manage the existing HTTP protocol
connection for FileMover, type:
$ server_http server_2 -append dhsm -users
user1,user2,user3
server_2 : done
EXAMPLE #7
To add a specific user who can manage the existing HTTP protocol
connection for FileMover, type:
$ server_http server_2 -append dhsm -users user4 -hosts
172.24.102.20,172.24.102.21
server_2 : done
EXAMPLE #8
To remove the specified users and hosts so they can no longer
manage the HTTP connection for FileMover, type:
$ server_http server_2 -remove dhsm -users user1,user2
-hosts 10.240.12.146
server_2 : done
server_http
589
The server Commands
server_ifconfig
Manages the network interface configuration for the specified Data
Movers.
SYNOPSIS
server_ifconfig {<movername>|ALL}
-all [-ip4|-ip6]
| -delete <if_name>
| -create -Device <device_name> -name <if_name>
-protocol { IP <ipv4_addr> <ipmask> <ipbroadcast>
| IP6 <ipv6_addr> [/PrefixLength]}
| <if_name> {up|down
| mtu=<MTUbytes> [vlan=<vlanID>]
| vlan=<vlanID> [mtu=<MTUbytes>]}
DESCRIPTION
server_ifconfig creates a network interface, assigns an IP address to a
network interface, enables and disables an interface, sets the MTU
size and the VLAN ID, and displays network interface parameters for
the specified Data Mover.
server_ifconfig is used to define the network address of each
interface existing on a machine, to delete and recreate an interface’s
address and operating parameters.
The ALL option executes the command for all Data Movers.
OPTIONS
-all [ip4|ip6]
Displays parameters for all configured interfaces. The -ip4 option
displays all ipv4 only interfaces, and the -ip6 option displays all ipv6
only interfaces.
-delete <if_name>
Deletes a network interface configuration. However, the
autogenerated link-local interfaces cannot be deleted.
-create -Device <device_name> -name <if_name>
-protocol IP <ipv4_addr> <ipmask> <ipbroadcast> |IP6
<ipv6_addr> [/PrefixLength]}
Creates a network interface configuration on the specified device
with the specified name and assigns a protocol to the interface. The
<if_name> must not contain a colon (:).
590
Command Line Interface Reference for File 8.1
The server Commands
Available protocols are:
IP <ipv4_addr> <ipmask> <ipbroadcast>
|IP6 <ipv6_addr> [/PrefixLength]}
IPv4 assigns the IP protocol with the specified IP address, mask, and
broadcast address. The IP address is the address of a particular
interface. Multiple interfaces are allowed for each device, each
identified by a different IP address. The IP mask includes the
network part of the local address and the subnet, which is taken from
the host field of the address. For example, 255.255.255.0 would be a
mask for a Class C network. The IP broadcast is a special destination
address that specifies a broadcast message to a network. For example,
x.x.x.255 is the broadcast address for a Class C network.
IP6 assigns the IPv6 address and prefix length. When prefix length is
not specified, the default value of 64 is used.
When creating the first IPv6 interface with a global unicast address
on a broadcast domain, the system automatically creates an
associated IPv6 link-local interface. Similarly, when deleting the last
remaining IPv6 interface on a broadcast domain, the system
automatically deletes the associated IPv6 link-local interface.
For CIFS users, when an interface is created, deleted, or marked up or
down, use the server_setup command to stop and then restart the
CIFS service in order to update the CIFS interface list.
<if_name> up
Allows the interface to receive and transmit data, but does not enable
the physical port. Interfaces are marked up automatically when
initially setting up the IP address.
<if_name> down
Stops data from being transmitted through that interface. If possible,
the interface is reset to disable reception as well. This does not
automatically disable routes using the interface.
<if_name> mtu=<MTUbytes>
Resets the maximum transmission unit (MTU) size in bytes for the
specified interface. By default, the MTU is automatically set
depending on the type of network interface card installed.
Regardless of whether you have Ethernet or Gigabit Ethernet, the
initial default MTU size is 1500 bytes. To take advantage of the
capacity of Gigabit Ethernet, the MTU size can be increased up to
9000 bytes if your switch supports jumbo frames. Jumbo frames
server_ifconfig
591
The server Commands
should be used only when the entire infrastructure, including client
NICs, supports them.
For UDP, it is important that both the client and server use the same
MTU size. TCP negotiates the MTU size when the connection is
initialized. The switch’s MTU must be greater than or equal to the
host’s MTU.
Note: The MTU size specified here is for the interface. The MTU size
specified in server_netstat applies to the device and is automatically set.
<if_name> vlan=<vlanID>
Sets the ID for the virtual LAN (VLAN). Valid inputs are 0 (default) to
4094. When a VLAN ID other than 0 is set, the interface only accepts
packets tagged with that specified ID. Outbound packets are also
tagged with the specified ID.
Note: IEEE 802.1Q VLAN tagging is supported. VLAN tagging is not
supported on ana interfaces.
SEE ALSO
FRONT-END OUTPUT
EXAMPLE #1
Configuring and Managing Networking on VNX and Configuring and
Managing Network High Availability on VNX, server_netstat,
server_setup, and server_sysconfig.
The network device name is dependent on the front end of the system
(for example, NS series Data Mover, 514 Data Movers, 510 Data
Movers, and so on) and the network device type. NS series and 514
Data Movers network device name display a prefix of cge, for
example, cge0. 510 or earlier Data Movers display a prefix of ana or
ace, for example, ana0, ace0. Internal network devices on a Data
Mover are displayed as el30, el31.
To display parameters of all interfaces on a Data Mover, type:
$ server_ifconfig server_2 -all
server_2 :
loop protocol=IP device=loop
inet=127.0.0.1 netmask=255.0.0.0 broadcast=127.255.255.255
UP, loopback, mtu=32768, vlan=0, macaddr=0:0:0:0:0:0 netname=localhost
cge0 protocol=IP device=cge0
inet=172.24.102.238 netmask=255.255.255.0 broadcast=172.24.102.255
UP, ethernet, mtu=1500, vlan=0, macaddr=0:60:16:4:29:87
el31 protocol=IP device=cge6
inet=128.221.253.2 netmask=255.255.255.0 broadcast=128.221.253.255
UP, ethernet, mtu=1500, vlan=0, macaddr=0:60:16:4:11:a6 netname=localhost
el30 protocol=IP device=fxp0
592
Command Line Interface Reference for File 8.1
The server Commands
inet=128.221.252.2 netmask=255.255.255.0 broadcast=128.221.252.255
UP, ethernet, mtu=1500, vlan=0, macaddr=8:0:1b:43:7e:b8 netname=localhost
EXAMPLE #2
To create an IP interface for Gigabit Ethernet, type:
$ server_ifconfig server_2 -create -Device cge1 -name
cge1 -protocol IP 172.24.102.239 255.255.255.0
172.24.102.255
server_2 : done
EXAMPLE #3
To create an interface for network device cge0 with an IPv6 address
with a nondefault prefix length on server_2, type:
$ server_ifconfig server_2 -create -Device cge0 -name
cge0_int1 -protocol IP6
3ffe:0000:3c4d:0015:0435:0200:0300:ED20/48
server_2 : done
EXAMPLE #4
To create an interface for network device cge0 with an IPv6 address
on server_2, type:
$ server_ifconfig server_2 -create -Device cge0 -name
cge0_int1 -protocol IP6
3ffe:0000:3c4d:0015:0435:0200:0300:ED20
server_2 : done
EXAMPLE #5
To verify that the settings for the cge0_int1 interface for server_2 are
correct, type:
$ server_ifconfig server_2 cge0_int1
server_2 :
cge0_int1 protocol=IP6 device=cge0
inet=3ffe:0:3c4d:15:435:200:300:ed20 prefix=48
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:c:5:5
Note: The bold item in the output highlights the nondefault 48-bit prefix.
EXAMPLE #6
To verify that the interface settings for server_2 are correct, type:
$ server_ifconfig server_2 -all
server_2 :
el30 protocol=IP device=mge0
inet=128.221.252.2 netmask=255.255.255.0 broadcast=128.221.252.255
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:d:30:b1 netname=localhost
el31 protocol=IP device=mge1
inet=128.221.253.2 netmask=255.255.255.0 broadcast=128.221.253.255
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:d:30:b2 netname=localhost
loop6 protocol=IP6 device=loop
inet=::1 prefix=128
UP, Loopback, mtu=32768, vlan=0, macaddr=0:0:0:0:0:0 netname=localhost
server_ifconfig
593
The server Commands
loop protocol=IP device=loop
inet=127.0.0.1 netmask=255.0.0.0 broadcast=127.255.255.255
UP, Loopback, mtu=32768, vlan=0, macaddr=0:0:0:0:0:0 netname=localhost
cge0_int1 protocol=IP6 device=cge0
inet=3ffe:0:3c4d:15:435:200:300:ed20 prefix=64
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:c:2:5
cge0_0000_ll protocol=IP6 device=cge0
inet=fe80::260:16ff:fe0c:205 prefix=64
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:c:2:5
Note: The first bold item in the output highlights the default 64-bit prefix.
The second and third bold items highlight the link-local name and address
that are automatically generated when you configure a global address for
cge0. The automatically created link-local interface name is made by
concatinating the device name with the four digit VLAN ID between 0 and
4094. Note that the interface you configured with the IPv6 address
3ffe:0:3c4d:15:435:200:300:ed20 and the address with the link-local address
fe80::260:16ff:fe0c:205 share the same MAC address. The link-local address is
derived from the MAC address.
EXAMPLE #7
To verify that the interface settings for server_2 are correct, type:
$ server_ifconfig server_2 -all
server_2 :
cge0_int2 protocol=IP device=cge0
inet=172.24.108.10 netmask=255.255.255.0 broadcast=172.24.108.255
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:c:2:5
cge0_int1 protocol=IP6 device=cge0
inet=3ffe:0:3c4d:15:435:200:300:ed20 prefix=64
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:c:2:5
cge0_0000_ll protocol=IP6 device=cge0
inet=fe80::260:16ff:fe0c:205 prefix=64
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:c:2:5
el30 protocol=IP device=mge0
inet=128.221.252.2 netmask=255.255.255.0 broadcast=128.221.252.255
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:d:30:b1 netname=localhost
el31 protocol=IP device=mge1
inet=128.221.253.2 netmask=255.255.255.0 broadcast=128.221.253.255
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:d:30:b2 netname=localhost
loop6 protocol=IP6 device=loop
inet=::1 prefix=128
UP, Loopback, mtu=32768, vlan=0, macaddr=0:0:0:0:0:0 netname=localhost
loop protocol=IP device=loop
inet=127.0.0.1 netmask=255.0.0.0 broadcast=127.255.255.255
UP, Loopback, mtu=32768, vlan=0, macaddr=0:0:0:0:0:0 netname=localhost
Note: The bold items in the output highlight the IPv4 interface, cge0_int2,
and the IPv6 interface, cge0_int1.
594
Command Line Interface Reference for File 8.1
The server Commands
EXAMPLE #8
To disable an interface, type:
$ server_ifconfig server_2 cge0_int2 down
server_2 : done
EXAMPLE #9
To enable an interface, type:
$ server_ifconfig server_2 cge0_int2 up
server_2 : done
EXAMPLE #10
To reset the MTU for Gigabit Ethernet, type:
$ server_ifconfig server_2 cge0_int2 mtu=9000
server_2 : done
EXAMPLE #11
To set the ID for the Virtual LAN, type:
$ server_ifconfig server_2 cge0_int1 vlan=40
server_2 : done
EXAMPLE #12
To verify that the VLAN ID in the interface settings for server_2 are
correct, type:
$ server_ifconfig server_2 -all
server_2 :
cge0_int1 protocol=IP6 device=cge0
inet=3ffe:0:3c4d:15:435:200:300:ed20 prefix=64
UP, Ethernet, mtu=1500, vlan=40, macaddr=0:60:16:c:2:5
cge0_0040_ll protocol=IP6 device=cge0
inet=fe80::260:16ff:fe0c:205 prefix=64
UP, Ethernet, mtu=1500, vlan=40, macaddr=0:60:16:c:2:5
cge0_int2 protocol=IP device=cge0
inet=172.24.108.10 netmask=255.255.255.0 broadcast=172.24.108.255
UP, Ethernet, mtu=1500, vlan=20, macaddr=0:60:16:c:2:5
el30 protocol=IP device=mge0
inet=128.221.252.2 netmask=255.255.255.0 broadcast=128.221.252.255
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:d:30:b1 netname=localhost
el31 protocol=IP device=mge1
inet=128.221.253.2 netmask=255.255.255.0 broadcast=128.221.253.255
UP, Ethernet, mtu=1500, vlan=0, macaddr=0:60:16:d:30:b2 netname=localhost
loop6 protocol=IP6 device=loop
inet=::1 prefix=128
UP, Loopback, mtu=32768, vlan=0, macaddr=0:0:0:0:0:0 netname=localhost
loop protocol=IP device=loop
inet=127.0.0.1 netmask=255.0.0.0 broadcast=127.255.255.255
UP, Loopback, mtu=32768, vlan=0, macaddr=0:0:0:0:0:0 netname=localhost
Note: The bold items in the output highlight the VLAN tag.
Note that the link-local address uses the VLAN tag as part of its name.
EXAMPLE #13
To delete an IP interface, type:
server_ifconfig
595
The server Commands
$ server_ifconfig server_2 -delete cge1_int2
server_2 : done
Note: The autogenerated link local interfaces cannot be deleted.
596
Command Line Interface Reference for File 8.1
The server Commands
server_ip
Manages the IPv6 neighbor cache and route table for the VNX.
SYNOPSIS
server_ip {ALL|<mover>}
-neighbor {
-list [<v6addr> [-interface <ifname>]]
| -create <v6addr> -lladdress <macaddr> [-interface <ifname>]
| -delete {-all|<v6addr> [-interface <ifname>]}
}
|-route {
-list
| -create {
-destination <destination> -interface <ifname>
| -default -gateway <v6gw> [-interface <ifname>]
}
| -delete {
-destination <destination>
| -default -gateway <v6gw> [-interface <ifname>]
| -all
}
}
DESCRIPTION
OPTIONS
server_ip creates, deletes, and lists the neighbor cache and route
tables.
server_ip {<movername>|ALL}
Sends a request to the Data Mover to get IPv6 parameters related to
the IPv6 routing table and neighbor cache.
ALL option executes the command for all Data Movers.
-neighbor {-list|-create|-delete}
Lists, creates, or deletes the neighbor cache entries from the neighbor
cache table.
-list
Displays the neighbor cache entries.
-create
Creates a neighbor cache table entry with the specified details.
-delete
Deletes the specified neighbor cache table entries or all entries.
-route {-list|-create|-delete}
Lists, creates, or deletes entries in the IPv6 route table.
server_ip
597
The server Commands
-list
Displays the IPv6 route table.
-create
Creates a route table entry with the specified details.
-delete
Deletes the specified route table entries.
EXAMPLE #1
To view a list of neighbor cache entries on Data Mover server_2, type:
$
server_2:
Address
fe80::204:23ff:fead:4fd4
fe80::216:9cff:fe15:c00
fe80::216:9cff:fe15:c00
fe80::216:9cff:fe15:c00
fe80::216:9cff:fe15:c00
3ffe::1
server_ip server_2 -neighbor -list
Link layer address
0:4:23:ad:4f:d4
0:16:9c:15:c:0
0:16:9c:15:c:0
0:16:9c:15:c:0
0:16:9c:15:c:0
0:16:9c:15:c:10
Interface
cge1_0000_ll
cge1_0000_ll
cge4_0000_ll
cge3_2998_ll
cge2_2442_ll
cge3_0000_ll
Type
host
router
router
router
router
router
State
STALE
STALE
STALE
STALE
STALE
REACHABLE
Where:
Value
Definition
Address
The neighbor IPv6 address.
Link layer address The link layer address of the neighbor.
Interface
Interface name of the interface connecting to the neighbor.
Type
Type of neighbor. The neighbor can be either host or router.
State
The state of the neighbor such as REACHABLE, INCOMPLETE, STALE, DELAY, or PROBE.
EXAMPLE #2
To view a list of neighbor cache entries for a specific IP address on
Data Mover server_2, type:
$ server_ip server_2 -neighbor -list
fe80::216:9cff:fe15:c00
server_2:
Address
fe80::216:9cff:fe15:c00
fe80::216:9cff:fe15:c00
fe80::216:9cff:fe15:c00
fe80::216:9cff:fe15:c00
598
Link layer address
0:16:9c:15:c:0
0:16:9c:15:c:0
0:16:9c:15:c:0
0:16:9c:15:c:0
Interface
cge1_0000_ll
cge4_0000_ll
cge3_2998_ll
cge2_2442_ll
Command Line Interface Reference for File 8.1
Type
router
router
router
router
State
STALE
STALE
STALE
STALE
The server Commands
EXAMPLE #3
To view a list of neighbor cache entries for a specific IP address and
interface type on Data Mover server_2, type:
$ server_ip server_2 -neighbor -list
fe80::216:9cff:fe15:c00 -interface cge1_0000_ll
server_2:
Address
fe80::216:9cff:fe15:c00
EXAMPLE #4
Link layer address Interface
0:16:9c:15:c:0
cge1_0000_ll
Type
router
State
STALE
To add an entry to the neighbor cache for a global unicast IPv6
address on Data Mover server_2, type:
$ server_ip server_2 -neighbor -create 2002:8c8:0:2310::2
-lladdress 0:16:9c:15:c:15
OK
EXAMPLE #5
To add an entry to the neighbor cache for a link-local IPv6 address on
Data Mover server_2, type:
$ server_ip server_2 -neighbor -create fe80::2 -lladdress
0:16:9c:15:c:12 -interface cge1v6
OK
EXAMPLE #6
To delete an entry from the neighbor cache for a global unicast IPv6
address on Data Mover server_2, type:
$ server_ip server_2 -neighbor -delete
2002:8c8:0:2310:0:2:ac18:f401
OK
EXAMPLE #7
To delete an entry from the neighbor cache for a link-local IPv6
address on all Data Movers, type:
$ server_ip ALL -neighbor -delete fe80::1 -interface
cge1v6
OK
EXAMPLE #8
To delete entries from the neighbor cache on Data Mover server_2,
type:
$ server_ip server_2 -neighbor -delete -all
OK
EXAMPLE #9
To view a list of route table entries on Data Mover server_2, type:
$ server_ip server_2 -route -list
server_ip
599
The server Commands
server_2:
Destination
2002:8c8:0:2310::/64
2002:8c8:0:2311::/64
2002:8c8:0:2312::/64
2002:8c8:0:2313::/64
default
default
default
selected default
Gateway
fe80::260:16ff:fe05:1bdd
fe80::260:16ff:fe05:1bdc
2002:8c8:0:2314::1
fe80::260:16ff:fe05:1bdd
Interface
cge1v6
cge1v6
cge1v6
cge1v6
cge1_0000_ll
cge1_0000_ll
cge4v6
cge1_0000_ll
Expires (secs)
0
0
0
0
1785
1785
0
1785
Where:
Value
Destination
Gateway
Interface
Expires
Definition
The prefix of the destination or the default route entry. There can be multiple default routes, but
only one is active and shown as “selected default”. The default sorting of the destination column
displays the default routes at the bottom of the list and the selected default at the end of the list.
The default gateway for default route entries. This value is blank for prefix destination entries.
Interface name of the interface used for the route.
The time until the route entry is valid. Zero denotes route is permanent and does not have an
expiry.
EXAMPLE #10
To add a default route table entry on the Data Mover server_2 to the
destination network with the specified prefix, type:
$ server_ip server_2 -route -create -destination
2002:8c8:0:2314::/64 -interface cge4v6
OK
EXAMPLE #11
To add a default route table entry on the Data Mover server_2
through the specified gateway, type:
$ server_ip server_2 -route -create -default -gateway
2002:8c8:0:2314::1
OK
EXAMPLE #12
To add a default route table entry on the Data Mover server_2
through the specified gateway using the link-local interface, type:
$ server_ip server_2 -route -create -default -gateway
fe80::1 -interface cge1v6
OK
600
Command Line Interface Reference for File 8.1
The server Commands
EXAMPLE #13
To delete an entry from the route table with an IPv6 prefix route
destination for all Data Movers, type:
$ server_ip ALL -route -delete -destination
2002:8c8:0:2314::/64
OK
EXAMPLE #14
To delete an entry from the route table for a global unicast IPv6
address on Data Mover server_2, type:
$ server_ip server_2 -route -delete -default -gateway
2002:8c8:0:2314::1
OK
EXAMPLE #15
To delete an entry from the route table for a link-local IPv6 address on
Data Mover server_2, type:
$ server_ip server_2 -route -delete -default -gateway
fe80::1 -interface cge1v6
OK
EXAMPLE #16
To delete all entries from the IPv6 route table on Data Mover
server_2, type:
$ server_ip server_2 -route -delete -all
OK
server_ip
601
The server Commands
server_kerberos
Manages the Kerberos configuration within the specified Data
Movers.
SYNOPSIS
server_kerberos {<movername>|ALL}
-add realm=<realm_name>,kdc=<fqdn_kdc_name>[:<port>]
[,kdc=<fqdn_kdc_name>[:<port>]...]
[,kpasswd=<fqdn_kpasswd_server_name>]
[,kadmin=<kadmin_server>]
[,domain=<domain_name>][,defaultrealm]
| -add -Disjoint realm=<realm_name>,domain=<domain_name>
| -delete realm=<realm_name>
| -delete -Disjoint realm=<realm_name>,domain=<domain_name>
| -keytab
| -ccache [-flush]
| -list
| -kadmin [<kadmin_options>]
DESCRIPTION
server_kerberos adds, deletes, lists the realms within the Kerberos
configuration of a Data Mover, and manages the Data Movers service
principals and keys.
server_kerberos displays the key table content, and specifies a
kadmin server.
OPTIONS
-add realm=<realm_name>,kdc=<fqdn_kdc_name>
Adds the specified realm to the Kerberos configuration on the
specified Data Mover. The <realm_name> is the fully qualified
domain name of the Kerberos realm to be added to the key
distribution center (KDC) configuration. The <fqdn_kdc_name> is
the fully qualified domain name of the KDC for the specified realm.
Note: The -add option is relevant only if you are using a UNIX/Linux
Kerberos KDC.
[:<port>]
Specifies a port that the KDC listens on.
[,kdc=<fqdn_kdc_name[:<port>]...]
Specifies additional KDCs with ports that KDCs listen on.
602
Command Line Interface Reference for File 8.1
The server Commands
[,kpasswd=<fqdn_kpasswd_server_name>]
Specifies a password server for the KDC. The
<fqdn_kpasswd_server_name> must be a fully qualified domain
name for the server.
[,kadmin=<kadmin_server>]
Specifies the kadmin server.
[,domain=<domain_name>]
The <domain_name> is the full name of the DNS domain for the
realm.
[,defaultrealm]
Indicates that the default realm is to be used.
-delete realm=<realm_name>
Deletes the specified realm from the Kerberos configuration for the
specified Data Mover.
Note: The -delete option is relevant only if you are using a UNIX/Linux
Kerberos KDC.
-keytab
Displays the principal names for the keys stored in the keytab file.
-ccache
Displays the entries in the Data Movers Kerberos credential cache.
Note: The -ccache option can also be used to provide EMC Customer Support
with information for troubleshooting user access problems.
[-flush]
Flushes the Kerberos credential cache removing all entries.
Credential cache entries are automatically flushed when they
expire or during a Data Mover reboot.
Once the cache is flushed, Kerberos obtains new credentials when
needed. The repopulation of credentials may take place
immediately, over several hours, or be put off indefinitely if no
Kerberos activity occurs.
-list
Displays a listing of all configured realms on a specified Data Mover
or on all Data Movers.
-kadmin [<kadmin_options>]
Invokes the kadmin tool with the following specified options:
server_kerberos
603
The server Commands
[-r <realm>]
Specifies a realm as the default database realm.
[-p <principal>]
Specifies the principal for authentication. Otherwise, kadmin will
append "/admin" to the primary principal name of the default
cache, the value of the USER environment variable, or the
username as obtained with getpwuid, in order of preference.
[-q <query>]
Runs kadmin in non-interactive mode. This passes the query
directly to kadmin, which performs the query, then exits.
[-w <password>]
Uses a specified password instead of prompting for a password.
[-s <admin_server> [:<port>]]
Specifies the ka