Version 2 Release 4
GC23-3900-05
Program Number: 5765-529
Note |
---|
Before using this information and the product it supports, be sure to read the general information under "Notices". |
Fifth Edition (February 1998)
This is a major revision of GC23-3900-04.
This edition applies to Version 2 Release 4 of the IBM Parallel System Support Programs for AIX (PSSP) Licensed Program, program number 5765-529, and to all subsequent releases and modifications until otherwise indicated in new editions. Significant changes or additions to the text and illustrations are indicated by a vertical line (|) to the left of the change.
Order publications through your IBM representative or the IBM branch office serving your locality. Publications are not stocked at the address below.
IBM welcomes your comments. A form for readers' comments may be provided at the back of this publication, or you may address your comments to the following address:
If you would like a reply, be sure to include your name, address, telephone number, or FAX number.
Make sure to include the following in your comment or note:
When you send information to IBM, you grant IBM a nonexclusive right to use or distribute the information in any way it believes appropriate without incurring any obligation to you.
© Copyright International Business Machines Corporation 1995, 1998. All rights reserved.
Note to U.S. Government Users -- Documentation related to restricted rights -- Use, duplication or disclosure is subject to restrictions set forth in GSA ADP Schedule contract with IBM Corp.
References in this publication to IBM products, programs, or services do not imply that IBM intends to make these available in all countries in which IBM operates. Any reference to an IBM product, program, or service is not intended to state or imply that only IBM's product, program, or service may be used. Any functionally equivalent product, program, or service that does not infringe any of IBM's intellectual property rights may be used instead of the IBM product, program, or service. Evaluation and verification of operation in conjunction with other products, except those expressly designated by IBM, are the user's responsibility.
IBM may have patents or pending patent applications covering subject matter in this document. The furnishing of this document does not give you any license to these patents. You can send license inquiries, in writing, to:
Licensees of this program who wish to have information about it for the purpose of enabling: (i) the exchange of information between independently created programs and other programs (including this one) and (ii) the mutual use of the information which has been exchanged, should contact:
Such information may be available, subject to appropriate terms and conditions, including in some cases, payment of a fee.
The following terms are trademarks of the International Business machines Corporation in the United States and/or countries:
Microsoft, Windows, and the Windows 95 logo are trademarks or registered trademarks of Microsoft Corporation.
UNIX is a registered trademark in the United States and other countries licensed exclusively through X/Open Company Limited.
Other company, product, and service names, which may be denoted by a double asterisk (**), may be trademarks or service marks of others.
This product includes software that is publicly available:
This book discusses the use of these products only as they apply specifically to the SP system. The distribution for these products includes the source code and associated documentation. (Kerberos does not ship source code.) /usr/lpp/ssp/public contains the compressed tar files of the publicly available software. (IBM has made minor modifications to the versions of Tcl and Tk used in the SP system to improve their security characteristics. Therefore, the IBM-supplied versions do not match exactly the versions you may build from the compressed tar files.) All copyright notices in the documentation must be respected. You can find version and distribution information for each of these products that are part of your selected install options in the /usr/lpp/ssp/README/ssp.public.README file.
The IBM Parallel System Support Programs for AIX: Command and Technical Reference provides detailed syntax and parameter information for all commands you can use to install, customize, and maintain the IBM RS/6000 SP system.
Other books that help you administer and use the SP system include:
This book applies to PSSP Version 2 Release 4. To find out what version of PSSP is running on your control workstation, enter the following:
SDRGetObjects SP code_version
In response, the system displays something similar to:
code_version PSSP-2.4
If the response indicates PSSP-2.4, this book applies to the version of PSSP that is running on your system.
To find out what version of PSSP is running on the nodes of your system, enter the following from your control workstation:
splst_versions -G -t
In response, the system displays something similar to:
1 PSSP-2.4 2 PSSP-2.4 7 PSSP-2.3 8 PSSP-2.3
If the response indicates PSSP-2.4, this book applies to the version of PSSP that is running on your system.
If you are running mixed levels of PSSP, be sure to maintain and refer to the appropriate documentation for whatever versions of PSSP you are running.
This book is intended for anyone not familiar with the syntax and use of the RS/6000 SP commands.
This book consists of three parts. Part 1 of this book is the Command Reference. It contains RS/6000 SP commands which are organized alphabetically. Part 2 of this book is the Technical Reference. It contains RS/6000 SP files, subroutines, and other technical information. Part 3 of this book is the Appendix. It lists Perspectives colors and fonts.
The back of the book includes a glossary and an index.
Vertical bars (|) to the left of the text in this book indicate changes or additions.
The commands in this book are in the following format:
This book uses the following typographic
conventions:
Typographic | Usage |
---|---|
Bold |
|
Italic |
|
Constant width | Examples and information that the system displays appear in constant width typeface. |
[ ] | Brackets enclose optional items in format and syntax descriptions. |
{ } | Braces enclose a list from which you must choose an item in format and syntax descriptions. |
| | A vertical bar separates items in a list of choices. (In other words, it means "or.") |
< > | Angle brackets (less-than and greater-than) enclose the name of a key on the keyboard. For example, <Enter> refers to the key on your terminal or workstation that is labeled with the word Enter. |
... | An ellipsis indicates that you can repeat the preceding item one or more times. |
<Ctrl-x> | The notation <Ctrl-x> indicates a control character sequence. For example, <Ctrl-c> means that you hold down the control key while pressing <c>. |
In order to use the PSSP man pages or access the PSSP online (HTML) publications, the ssp.docs file set must first be installed. To view the PSSP online publications, you also need access to an HTML document browser such as Netscape. An index to the HTML files that are provided with the ssp.docs file set is installed in the /usr/lpp/ssp/html directory.
You can view this book or download a PostScript version of it from the IBM RS/6000 web site at http://www.rs6000.ibm.com. At the time this manual was published, the full path was http://www.rs6000.ibm.com/resource/aix_resource/sp_books. However, the structure of the RS/6000 web site can change over time.
Here are some related publications.
As an alternative to ordering the individual books, you can use SBOF-8587 to order the entire SP software library.
As an alternative to ordering the individual books, you can use SBOF-8588 to order the entire Parallel Environment for AIX library.
Here are some other IBM publications that you may find helpful.
Order according to RS/6000 SP Switch Router model:
You can order the RS/6000 SP Switch Router as the IBM 9077.
Here are some non-IBM publications that you may find helpful.
The following manual pages for public code are available in this product:
Manual pages and other documentation for Tcl, TclX, Tk, and expect can be found in the compressed tar files located in /usr/lpp/ssp/public.
This part of the book contains the RS/6000 SP commands.
To access the RS/6000 SP online manual pages, set the MANPATH environment variable as follows:
export MANPATH=$MANPATH:/usr/lpp/ssp/man
setenv MANPATH $MANPATH\:/usr/lpp/ssp/man
When you partition your system, you create one or more system partitions which, for most tasks, function as separate and distinct logical RS/6000 SP systems. Most commands function within the boundary of the system partition in which they are executed. A number of commands, however, continue to treat the RS/6000 SP as a single entity and do not respect system partition boundaries. That is, in their normal function they may affect a node or other entity outside of the current system partition. In addition, some commands which normally function only within the current system partition have been given a new parameter which, when used, allows the scope of that command to exceed the boundaries of the current system partition.
On the control workstation, the administrator is in an environment for one system partition at a time. The SP_NAME environment variable identifies the system partition to subsystems. (If this environment variable is not set, the system partition is defined by the primary: stanza in the /etc/SDR_dest_info file.) Most tasks performed on the control workstation that get information from the System Data Repository (SDR) will get the information for that particular system partition.
In managing multiple system partitions, it is helpful to open a window for each system partition. You can set and export the SP_NAME environment variable in each window and set up the window title bar or shell prompt with the system partition name. The following script is an example:
sysparenv: # !/bin/ksh for i in 'splst_syspars' do syspar='host $i | cut -f 1 -d"."' echo "Opening the $syspar partition environment" sleep 2 export SP_NAME=$syspar aixterm -T "Work Environment for CWS 'hostname -s' - View: $syspar" -ls -sb & done exit .profile addition: # Added for syspar environment setup if [ "'env | grep SP_NAME | cut -d= -f1'" = SP_NAME ] then PS1="['hostname -s'<p>"$SP_NAME] ['$PWD]> ' else PS1="['hostname -s']["'$PWD]< ' fi export ENV
As a user, you can check what system partition you're in with the command:
spget_syspar -n
The following table summarizes those commands which can exceed the boundary
of the current system partition. Unless otherwise stated, commands not listed
in this table have as their scope the current system partition.
Command | Effect |
---|---|
arp | Can reference any node (by its host name) in any system partition. |
Automounter commands | Host names need not be in the current system partition. |
crunacct | Merges accounting data from all nodes regardless of system partition boundaries. |
cshutdown -G | The -G flag allows specification of target nodes outside of the current system partition. |
cstartup -G | The -G flag allows specification of target nodes outside of the current system partition. |
dsh dsh -w{hostname | -} | Hosts added to the working collective by host name need not be in the current system partition. |
dsh -aG | The -G flag modifies the -a flag (all nodes in the current system partition) by expanding the scope to all nodes in the entire physical SP system. |
Eclock | There is a single switch clock for the SP regardless of the number of system partitions. |
Efence -G | The -G flag allows specification of nodes outside of the current system partition. |
emonctrl -c | The system partition-sensitive control script for the emon subsystem supports the -c option, which crosses system partitions. |
Eunfence -G | The -G flag allows specification of nodes outside of the current system partition. |
haemctrl -c haemctrl -u | The system partition-sensitive control script for the haem subsystem supports the -c and -u options, which cross system partitions. |
hagsctrl -c hagsctrl -u | The system partition-sensitive control script for the hags subsystem supports the -c and -u options, which cross system partitions. |
hatsctrl -c hatsctrl -u | The system partition-sensitive control script for the hats subsystem supports the -c and -u options, which cross system partitions. |
hbctrl -c | The system partition-sensitive control script for the hb subsystem supports the -c option, which crosses system partitions. |
hmcmds -G | The -G flag allows the hmcmds commands to be sent to any hardware on the SP system. |
hmmon -G | The -G flag allows for the specification of hardware outside of the current system partition. |
hostlist hostlist -f filename hostlist -w hostname | Host names need not be in the current system partition. |
hostlist -aG | -nG | -sG | The -G flag modifies the -a, -n, or -s flag by expanding the scope to the entire physical SP system. |
hrctrl -c | The system partition-sensitive control script for the hr subsystem supports the -c option, which crosses system partitions. |
hsdatalst -G | The -G flag causes the display of HSD information to be for all system partitions. |
lppdiff -aG | The -G flag modifies the -a flag (all nodes in the current system partition) by expanding the scope to all nodes in the entire physical SP system. |
nodecond -G | The -G flag allows specification of a node outside of the current system partition. |
psyslrpt -w hostnames | The host names supplied with the -w flag can be in any system partition (the -a flag will select all nodes in the current system partition). |
psyslclr -w hostnames | The host names supplied with the -w flag can be in any system partition (the -a flag will select all nodes in the current system partition). |
penotify -w hostnames | The host names supplied with the -w flag can be in any system partition (the -a flag will select all nodes in the current system partition). |
pmanctrl -c | The system partition-sensitive control script for the pman subsystem supports the -c option, which crosses system partitions. |
Parallel commands:
| Parallel commands can take the following options and will behave
accordingly:
|
SDRArchive, SDRRestore | Archives/restores the SDR representing the entire SP. |
SDRGetObjects -G | The -G flag allows for retrieval of partitioned class objects from partitions other than the current system partition. Without the -G, objects which are in a partitioned class are retrieved from the current system partition only. |
SDRMoveObjects | Moves objects from one system partition to another. |
Other SDR commands | SDR commands that create, change or delete values work within the system partition. Note though that System classes (Frame, for example) are shared among all system partitions. Changes to system classes will affect other system partitions. |
Security commands:
| The function of these security commands is unchanged under system partitioning. That is, if they previously affected the entire SP, they continue to do so even if the system has been partitioned. If they previously had the ability to affect a remote node (rsh, rcp, for example), that function is unchanged in a system partitioned environment. |
spapply_config | Applies a system partition configuration to the entire SP. |
spbootins | If a boot server outside of the current system partition is specified, that node is prepared appropriately. |
sp_configdctrl -c | The system partition-sensitive control script for the sp_configd subsystem supports the -c option, which crosses system partitions. |
spframe | Configures data for one or more frames across the entire SP. |
splm | The target nodes defined in the input table can include nodes from any system partition. |
splst_versions -G | The -G flag allows for retrieval of PSSP version information from nodes outside the current system partition. |
splstdata -G | The -G flag allows display of information on nodes and adapters outside of the current system partition. |
splstadapters -G | The -G flag lists information about target adapters outside of the current system partition. |
splstnodes -G | The -G flag lists information about target nodes outside of the current system partition. |
spmon -G | The -G flag allows specification of nodes outside of the current system partition. The -G flag is required when performing operations on any frame or switch. |
sprestore_config | Restores the entire SP SDR from a previously made archive. |
spsitenv | Site environment variables are specified for the SP system as a whole. The specification of acct_master= can be any node in the SP regardless of system partition. The specification of install_image= may cause boot server nodes outside of the current system partition to refresh the default installation image they will serve to their nodes. |
spverify_config | Verifies the configuration of all system partitions in the SPsystem. |
supper | File collections are implemented and managed without respect to system partition boundaries. |
sysctl | The Sysctl client can send requests to any node in the SP. |
syspar_ctrl -c -G | The -c and -G flags allow for the crossing of system partitions in providing a single interface to the control scripts for the system partition-sensitive subsystems. |
s1term -G | The -G flag allows specification of a node outside of the current system partition. |
vsdatalst -G | The -G flag causes the display of IBM Virtual Shared Disk information to be for all system partitions. |
vsdsklst -G | The -G flag specifies the display of information for disks outside the current system partition. |
Purpose
add_principal - Creates principals in the authentication database.
Syntax
add_principal [-r realm_name] [-v] file_name
Flags
Operands
Description
This command provides an interface to the authentication database to add an entry for a user or service instance, supplying the password used to generate the encrypted private key. The add_principal command is suitable for mass addition of users or multiple instances of servers (for example, SP nodes).
This command operates noninteractively if you have a valid ticket-granting-ticket (TGT) for your admin instance in the applicable realm. A TGT can be obtained using the kinit command. If you do not have a TGT for the admin instance for the realm in which you are adding principals, or if the add_principal command cannot obtain a service ticket for changing passwords using the admin TGT, the user is prompted for the password for the user's admin instance.
Administrators use the add_principal command to register new users and services instances to the authentication database. An administrator must have a principal ID with an instance of admin. Also, user_name.admin must appear in the admin_acl.add Access Control List (ACL).
The add_principal program communicates over the network with the kadmind program, which runs on the machine housing the primary authentication database. The kadmind program creates new entries in the database using data provided by this command.
When using the add_principal command, the principal's expiration date and maximum ticket lifetime are set to the default values. To override the defaults, the root user must use the kdb_edit command to modify those attributes.
Input to the program is read from the file specified by the file_name argument. It contains one line of information for each principal to be added, in the following format:
name[.instance][@realm] password
Note: | The @realm cannot be different from the local realm or the realm argument if the -r option is specified. |
For user entries with a NULL instance, this format matches that of the log file created by the spmkuser command. Any form of white space can surround the two fields. Blank lines are ignored. Any line containing a # as the first nonwhite space character, is treated as a comment.
Since the input file contains principal identifiers and their passwords, ensure that access to the file is controlled. You should remove the input file containing the unencrypted passwords after using it, or delete the passwords from it.
The add_principal command does not add principals to an AFS authentication database. If authentication services are provided through AFS, use the AFS kas command to add principals to the database. Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for an overview.
Files
Exit Values
Related Information
Commands: kadmin, kinit, kpasswd, ksrvutil
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Purpose
allnimres - Allocates Network Installation Management (NIM) resources from a NIM master to a NIM client.
Syntax
allnimres -h | -l node_list
Flags
Operands
None.
Description
Use this command to allocate all necessary NIM resources to a client based on the client's bootp_response in the System Data Repository (SDR). This includes executing the bos_inst command for allocation of the boot resource and nimscript resource. At the end of this command, nodes are ready to netboot to run installation, diagnostics, or maintenance. If the node's bootp_response is "disk", all NIM resources are deallocated from the node.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/allnimres
Related Information
Commands: setup_server, unallnimres
Examples
To allocate boot/installation resources to boot/install client nodes 1, 3, and 5 from their respective boot/install servers, enter:
allnimres -l 1,3,5
Purpose
/usr/lpp/ssp/css/arp - Displays and modifies address resolution.
Syntax
Parameters
type host_name adapter_address [route] [temp] [pub]
type host_name adapter_address [route] [temp] [pub]
where:
Description
The arp command has been modified to add support for the switch. This command is valid only on an SP system.
The arp command displays and modifies the Internet-to-adapter address translation tables used by ARP. The arp command displays the current ARP entry for the host specified by the host_name variable. The host can be specified by name or number, using Internet dotted decimal notation.
Related Information
SP Command: ifconfig
AIX Commands: crash, netstat
AIX Daemon: inetd
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for additional information on the SP Switch and the High Performance Switch.
Refer to "TCP/IP Protocols" in AIX Version 4.1 System Management Guide: Communications and Networks.
Examples
arp -s switch host2 1
arp -d host1
Purpose
cfghsd - Configures a data striping device (HSD) for an IBM Virtual Shared Disk.
Syntax
cfghsd -a hsd_name ...
Flags
Operands
Description
This command configures the already defined HSDs and makes them available. The command extracts information from the System Data Repository (SDR).
Files
Security
You must have root privilege to run this command.
Restrictions
If you have the IBM Recoverable Virtual Shared Disk product installed and operational, do not use this command. The results may be unpredictable.
See IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: defhsd, hsdatalst, lshsd, ucfghsd
Examples
To make the HSD hsd1 available, enter:
cfghsd hsd1
Purpose
cfghsdvsd - Configures a data striping device (HSD) and the IBM Virtual Shared Disks that comprise it on one node.
Syntax
cfghsdvsd -a | {hsd_name...}
Flags
Operands
Description
Use this command to configure already-defined HSDs and their underlying IBM Virtual Shared Disks and make them available. Note all of the IBM Virtual Shared Disks go to the active state.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit hsd_mgmtand select the Configure an HSD and its underlying IBM Virtual Shared Disks option.
Files
Security
You must have sysctl and sysctl.vsd access and authorization from your system administrator to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfghsd, cfgvsd, ucfghsdvsd
Examples
To configure the data striping device hsd1 and the IBM Virtual Shared Disks that comprise it, enter:
cfghsdvsd hsd1
Purpose
cfgvsd - Configures an IBM Virtual Shared Disk.
Syntax
cfgvsd -a | vsd_name ...
Flags
Operands
Description
Use this command to configure the already defined IBM Virtual Shared Disks and bring them to the stopped state. It does not make the IBM Virtual Shared Disk available. The command extracts information from the System Data Repository (SDR).
You can use the System Management Interface Tool (SMIT) to run the cfgvsd command. To use SMIT, enter:
smit vsd_mgmtand select the Configure an IBM Virtual Shared Disk option.
Files
Security
You must have root privilege to run this command.
Restrictions
If you have the IBM Recoverable Virtual Shared Disk product installed and operational, do not use this command. The results may be unpredictable.
See IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: ctlvsd, lsvsd, preparevsd, resumevsd, startvsd, stopvsd, suspendvsd, ucfgvsd
Examples
To bring the IBM Virtual Shared Disk vsd1vg1n1 from the defined state to the stopped state, enter:
cfgvsd vsd1vg1n1
Purpose
chgcss - Applies configuration changes to a Scalable POWERparallel Switch (SP Switch) Communications Adapter (Type 6-9) or a High Performance Switch Communications Adapter Type 2.
Implementation Note |
---|
Configuration changes are later applied to the device when it is configured at system reboot. |
Syntax
chgcss -l name -a attr=value [-a attr=value]
Flags
where:
Operands
None.
Description
Use this command to apply configuration changes to an SP Switch Communications Adapter (Type 6-9) or a High Performance Switch Communications Adapter Type 2.
Security
You must have root privilege to run this command.
Prerequisite Information
For additional information on values for the rpoolsize and spoolsize attributes, refer to the "Tuning the SP System" chapter in IBM Parallel System Support Programs for AIX: Administration Guide.
Related Information
AIX Command: lsattr
Examples
chgcss -l css0 -a rpoolsize=0x100000
chgcss -l css0 -a rpoolsize=1048576 -a spoolsize=1048576
Purpose
chkp - Changes Kerberos principals.
Syntax
chkp -h
chkp [-e expiration] [-l lifetime] name[.instance] ...
Flags
lifetime operand - Approximate duration 141 1 day 151 2 days 170 1 week 180 2 weeks 191 1 month
At least one flag must be specified.
Operands
Description
Use this command to change principals in the local Kerberos database. It allows the current expiration date and maximum ticket lifetime to be redefined. It cannot be used to change the principal's password. To do that, the administrator must use the kpasswd, kadmin, or kdb_edit commands. The chkp command should normally be run only on the primary server. If there are secondary authentication servers, the push-kprop command is invoked to propagate the change to the other servers. The command can be used to update a secondary server's database, but the changes may be negated by a subsequent update from the primary.
Files
Exit Values
Security
The chkp command can be run by the root user logged in on a Kerberos server host. It can be invoked indirectly as a Sysctl procedure by a Kerberos database administrator who has a valid ticket and is listed in the admin_acl.mod file.
Location
/usr/kerberos/etc/chkp
Related Information
Commands: kadmin, kdb_edit, lskp, mkkp, rmkp, sysctl
Examples
chkp -l 171 default
chkp -l 181 -e 2003-06-30 franklin jtjones root.admin susan
Purpose
cksumvsd - Views and manipulates the IBM Virtual Shared Disk checksum parameters.
Syntax
cksumvsd [-s] [-R] [-i | -I]
Flags
If no flags are specified, the current setting of all IBM Virtual Shared Disk checksum parameters and counters are displayed.
Operands
None.
Description
The IBM Virtual Shared Disk IP device driver can calculate and send checksums on remote packets it sends. It also can calculate and verify checksums on remote packets it receives. The cksumvsd command is used to tell the device driver whether to perform checksum processing. The default is no checksumming.
Issuing cksumvsd -i turns on checksumming on the node on which it is run. cksumvsd -i must be issued on all IBM Virtual Shared Disk nodes in the system partition, or the IBM Virtual Shared Disk software will stop working properly on the system partition. If node A has cksumvsd -i (checksumming turned on) and node B has cksumvsd -I (checksumming turned off, the default), then A will reject all messages from B (both requests and replies), since A's checksum verification will fail on all B's messages. The safe way to run cksumvsd -i is to make sure that all IBM Virtual Shared Disks on all nodes are in the STOPPED or SUSPENDED states, issue cksumvsd -i on all nodes, then resume the needed IBM Virtual Shared Disks on all nodes.
In checksumming mode, the IBM Virtual Shared Disk IP device driver keeps a counter of the number of packets received with good checksums, and the number received with bad checksums. cksumvsd and statvsd both display these values (statvsd calls cksumvsd -s).
cksumvsd dynamically responds to the configuration of the IBM Virtual Shared Disk IP device driver loaded in the kernel. Its output and function may change if the IBM Virtual Shared Disk IP device driver configuration changes.
Files
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Command: cfgvsd
Examples
cksumvsd
You should receive output similar to the following:
VSD cksum: current values: do_ip_checksum: 0 ipcksum_cntr: 350 good, 0 bad, 0 % bad.
The IBM Virtual Shared Disk checksumming is currently turned off on the node. Prior to this, checksumming was turned on and 350 IBM Virtual Shared Disk remote messages were received, all with good checksumming.
cksumvsd -i
You should receive output similar to the following:
VSD cksum: current values: do_ip_checksum: 0 ipcksum_cntr: 350 good, 0 bad, 0 % bad. VSD cksum: new values: do_ip_checksum: 1 ipcksum_cntr: 350 good, 0 bad, 0 % bad.
The command displays old and new values. As before, the node has received 350 IBM Virtual Shared Disk remote messages with good checksums.
cksumvsd -s
You should receive output similar to the following:
ipcksum_cntr: 350 good, 0 bad, 0 % bad.
Purpose
cmonacct - Performs monthly or periodic SP accounting.
Syntax
cmonacct [number]
Flags
None.
Operands
Description
The cmonacct command performs monthly or periodic SP system accounting. The intervals are set in the crontab file. You can set the cron daemon to run the cmonacct command once each month or at some other specified time period. By default, if accounting is enabled for at least one node, cmonacct executes on the first day of every month.
The cmonacct command creates summary files under the /var/adm/cacct/fiscal directory and restarts summary files under the /var/adm/cacct/sum directory, the cumulative summary to which daily reports are appended.
Purpose
cprdaily - Creates an ASCII report of the previous day's accounting data.
Syntax
cprdaily [-c] [[-l] [yyyymmdd]]
Flags
Operands
Description
This command is called by the crunacct command to format an ASCII report of the previous day's accounting data for all nodes. The report resides in the /var/adm/cacct/sum/rprtyyyymmdd file, where yyyymmdd specifies the year, month, and day of the report.
Purpose
cptuning - Copies a file to /tftpboot/tuning.cust.
Syntax
cptuning -h | file_name
Flags
Operands
Description
Use this command to copy the specified file to the /tftpboot/tuning.cust file. IBM ships the following four predefined tuning parameter files in /usr/lpp/ssp/install/config:
This command is intended for use in copying one of these files to /tftpboot/tuning.cust on the control workstation for propagation to the nodes in the SP. It can also be used on an individual node to copy one of these files to /tftpboot/ tuning.cust.
Standard Output
When the command completes successfully, a message to that effect is written to standard output.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Output Files
Upon successful completion, the /tftpboot/tuning.cust file is updated.
Consequences of Errors
If the command does not run successfully, it terminates with an error message and a nonzero return code.
Security
Use of this command by other than the root user is not restricted. However, this command will fail if the user does not have read permission to the specified file and write permission to the /tftpboot directory.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/cptuning
Related Information
SP Files: tuning.commercial, tuning.default, tuning.development, tuning.scientific
IBM Parallel System Support Programs for AIX: Installation and Migration Guide
Examples
cptuning /tmp/my-tuning-file
cptuning tuning.commercial
Purpose
create_krb_files - Creates the necessary krb_srvtab and tftp access files on the Network Installation Management (NIM) master.
Syntax
create_krb_files [-h]
Flags
Operands
None.
Description
Use this command on a boot/install server (including the control workstation). On the server, it creates the Kerberos krb_srvtab file for each boot/install client of that server and also updates the /etc/tftpaccess.ctl file on the server.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/create_krb_files
Related Information
Commands: setup_server
Examples
To create or update the krb_srvtab and tftp access files on a boot/install server, enter the following command on that server:
create_krb_files
Purpose
createhsd - Creates one data striping device (HSD) that encompasses two or more IBM Virtual Shared Disks.
Syntax
Flags
Note: | Some examples shown in this list do not contain enough flags to be executable. They are shown in an incomplete form to illustrate specific flags. |
The sequence in which nodes are listed determines the names given to the IBM Virtual Shared Disks; for example:
createhsd -n 1,6,4 -d DATA(with the hsd_prefix DATA) creates IBM Virtual Shared Disks DATA1n1 on node 1, DATA2n6 on node 6, and DATA3n4 on node 4, which make up a single HSD named DATA. To create volume groups that span specified disks on nodes 1, 2, and 3 of a system with backup on nodes 4, 5, and 6 of the same system, and that make up a single HSD, enter:
createhsd -n 1/4:hdisk1,hdisk2,hdisk3/,2/5:hdisk5,hdisk6, \ hdisk7/,3/6:hdisk2,hdisk4,hdisk6/ -d DATA -s 12 -g OLD -t 4096This command is shown on two lines here, but you must enter it without any spaces between the items in node_list.
The command creates:
If a volume group is already created and the combined physical hdisk lists contain disks that are not needed for the logical volume, those hdisks are added to the volume group. If the volume group has not already been created, createhsd creates a volume group that spans hdisk_list1+hdisk_list2.
Backup nodes cannot use the same physical disk as the primary does to serve IBM Virtual Shared Disks.
createhsd -n 6 -g VSDVGcreates a new volume group with the local AIX volume group name VSDVG and the IBM Virtual Shared Disk global volume group name VSDVGn6. The node number is added to the local volume group name to create a unique global volume group name within a system partition to avoid name conflicts with the name used for volume groups on other nodes. If a backup node exists, the global volume group name will be created by concatenating the backup node number as well as the primary node number to the local volume group name. For example:
createhsd -n 6/3/ -g VSDVGcreates VSDVGn6b3, where the primary node is node 6 and the backup node for this global volume group is node 3. The local AIX volume group name will still be VSDVG. You can specify a local volume group that already exists. You do not need to use the -T flag if you specify a volume group name that already exists.
createhsd -n 1,6 -c 2 -d DATAcreates IBM Virtual Shared Disks DATA1n1 on node 1, DATA2n6 on node 6, DATA3n1 on node 1, and DATA4n6 on node 6 and uses them to make up the HSD DATA.
createhsd -n 1,6 -c 2 -g DATAcreates DATA1n1 and DATA2n1 on node 1, and DATA3n6 and DATA4n6 on node 6.
The command:
createhsd -n 1,2 -h DATAcreates two IBM Virtual Shared Disks, DATA1n1 and DATA2n2. These IBM Virtual Shared Disks make up one HSD named DATA.
createhsd -n 1 -v DATAcreates one IBM Virtual Shared Disk on node 1 named DATA1n1 with an underlying logical volume lvDATA1n1. If the command
createhsd -n 1 -v DATA -l newis used, the IBM Virtual Shared Disk on node 1 is still named DATA1n1, but the underlying logical volume is named lvnew1n1.
It is usually more helpful not to specify -l, so that your lists of IBM Virtual Shared Disk names and logical volume names are easy to associate with each other and you avoid naming conflicts.
The Logical Volume Manager limits the number of physical partitions to 1016 per disk. If a disk is greater than 4 gigabytes in size, the physical partition size must be greater than 4MB to keep the number of partitions under the limit.
is not done as part of the createvsd processing that underlies the createhsd command. This speeds the operation of the command and avoids unnecessary processing in the case where several IBM Virtual Shared Disks are being created on the same primary/secondary nodes. In that case, however, you need to explicitly issue the volume group commands listed previously.
Operands
None.
Description
This command utilizes the sysctl facility.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit createhsd_dialogor
smit vsd_dataand Select the Create an HSD option with the vsd_data fastpath.
Files
Standard Output
For the following command:
createhsd -n 1/:hdisk2,hdisk3/ -g twinVG -s 1600 -t 8 -S -l \ twinLV -d twinHSD -c 4The messages returned to standard output are:
OK:0:vsdvg -g twinVGn1 twinVG 1 OK:0:defvsd twinLV1n1 twinVGn1 twinHSD1n1 nocache OK:0:defvsd twinLV2n1 twinVGn1 twinHSD2n1 nocache OK:0:defvsd twinLV3n1 twinVGn1 twinHSD3n1 nocache OK:0:defvsd twinLV4n1 twinVGn1 twinHSD4n1 nocache OK:createvsd { -n 1/:hdisk2,hdisk3/ -s 401 -T 4 -g twinVG -c 4 -v twinHSD -l twinLV -o cache -K } OK:0:defhsd twinHSD not_protect_lvcb 8192 twinHSD1n1 twinHSD2n1 twinHSD3n1 twinHSD4n1
Exit Values
Security
You must have sysctl and sysctl.vsd access and authorization from your system administrator to run this command.
Restrictions
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: createvsd, defhsd, vsdvg
Examples
To create six 4MB IBM Virtual Shared Disks and their underlying logical volumes with a prefix of TEMP, as well as an HSD comprising those IBM Virtual Shared Disks (24MB overall) with a stripe size of 32KB, enter the following (assuming that no previous IBM Virtual Shared Disks are defined with the TEMP prefix):
createhsd -n 3,4,7/8/ -c 2 -s 1024 -g vsdvg -d TEMP -t 32
This creates the following IBM Virtual Shared Disks:
and the HSD:
Note: | TEMP does not write to the first 32KB of each of its IBM Virtual Shared Disks. |
Purpose
createvsd - Creates a set of IBM Virtual Shared Disks, with their associated logical volumes, and puts information about them into the System Data Repository (SDR).
Syntax
Note: | Some examples shown in this list do not contain enough flags to be executable. They are shown in an incomplete form to illustrate specific flags. |
Flags
createvsd -n 1,6,4 -v PRE(with the vsd_prefix PRE) creates IBM Virtual Shared Disks PRE1n1 on node 1, PRE2n6 on node 6, and PRE3n4 on node 4.
To create a volume group that spans hdisk2, hdisk3, and hdisk4 on node 1, with a backup on node 3, enter:
createvsd -n 1/3:hdisk2,hdisk3,hdisk4/ -v DATAThis command creates:
To create volume groups just like that one on nodes 1, 2, and 3 of a system with backup on nodes 4, 5, and 6 of the same system, enter:
createvsd -n 1/4:hdisk1,hdisk2,hdisk3/,2/5:hdisk5,hdisk6, \ hdisk7/,3/6:hdisk2,hdisk4,hdisk6/ -v DATA
This command is shown on two lines here, but you must enter it without any spaces between the items in node_list.
The command creates:
To create an IBM Virtual Shared Disk where the logical volume spans only two of the physical disks in the volume group, enter:
createvsd -n 1/3:hdisk1,hdisk2+hdisk3/ -v DATAThis command creates the IBM Virtual Shared Disk DATA1n1 with logical volume lvDATA1n1 spanning hdisk1 and hdisk2 in the volume group DATA, which includes hdisk1, hdisk2, and hdisk3. It exports the volume group DATA to node 3.
If a volume group is already created and the combined physical hdisk lists contain disks that are not needed for the logical volume, those hdisks are added to the volume group. If the volume group has not already been created, createvsd creates a volume group that spans hdisk_list1+hdisk_list2.
Backup nodes cannot use the same physical disk as the primary does to serve IBM Virtual Shared Disks.
createvsd -n 6 -g VSDVGcreates a volume group with the local volume group name VSDVG and the global volume group name VSDVG1n6 on node 6. The node number is added to the prefix to avoid name conflicts when a backup node takes over a volume group. If a backup node exists, the global volume group name will be concatenated with the backup node number as well as the primary. For example:
createvsd -n 6/3/ -g VSDVGcreates a volume group with the local volume group name VSDVG and the global volume group name VSDVGn6b3. The primary node is node 6 and the backup node for this volume group is node 3.
createvsd -n 1,6 -c 2 -v DATAcreates IBM Virtual Shared Disks DATA1n1 on node 1, DATA2n6 on node 6, DATA3n1 on node 1, and DATA4n6 on node 6.
createvsd -n 1,6 -c 2 -h DATAcreates DATA1n1 and DATA2n1 on node 1, and DATA3n6 and DATA4n6 on node 6.
If -v is not specified, the prefix vsd is used.
Note: | The last character of the vsd_name_prefix cannot be a digit. Otherwise, the 11th IBM Virtual Shared Disk with the prefix PRE would have the same name as the first IBM Virtual Shared Disk with the prefix PRE1. Nor can the vsd_name_prefix contain the character '.' because '.' can be any character in regular expressions. |
createvsd -n 1 -v DATAcreates one IBM Virtual Shared Disk on node 1 named DATA1n1 with an underlying logical volume lvDATA1n1. If the command
createvsd -n 1 -v DATA -l newis used, the IBM Virtual Shared Disk on node 1 is still named DATA1n1, but the underlying logical volume is named lvnew1n1.
It is usually more helpful not to specify -l, so that your lists of IBM Virtual Shared Disk names and logical volume names are easy to associate with each other and you avoid naming conflicts.
The Logical Volume Manager limits the number of physical partitions to 1016 per disk. If a disk is greater than 4 gigabytes in size, the physical partition size must be greater than 4MB to keep the number of partitions under the limit.
is not done as part of the createvsd processing. This speeds the operation of the command and avoids unnecessary processing in the case where several IBM Virtual Shared Disks are being created on the same primary/secondary nodes. In this case, however, you should either not specify -x on the last createvsd in the sequence or issue the volume group commands listed above explicitly.
Operands
None.
Description
Use this command to create a volume group with the specified name (if one does not already exist) and creates a logical volume of size s within that volume group.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit vsd_dataand select the Create an IBM Virtual Shared Disk option.
Files
Standard Output
For the following command:
createvsd -n 1/:hdisk1/ -g testvg -s 16 -T 8 -l lvtest -v test -c 4The messages returned to standard output are:
OK:0:vsdvg -g testvgn1 testvg 1 OK:0:defvsd lvtest1n1 testvgn1 test1n1 nocache OK:0:defvsd lvtest2n1 testvgn1 test2n1 nocache OK:0:defvsd lvtest3n1 testvgn1 test3n1 nocache OK:0:defvsd lvtest4n1 testvgn1 test4n1 nocache
For the following command:
createvsd -n 1/:hdisk1/ -g testvg -s 16 -T 8 -l lvtest -v test -c 4The messages returned to standard output are:
OK:0:defvsd lvtest5n1 testvgn1 test5n1 nocache OK:0:defvsd lvtest6n1 testvgn1 test6n1 nocache OK:0:defvsd lvtest7n1 testvgn1 test7n1 nocache OK:0:defvsd lvtest8n1 testvgn1 test8n1 nocache
Exit Values
Security
You must have sysctl and sysctl.vsd access and authorization from your system administrator to run this command.
Restrictions
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: defvsd, vsdvg
Examples
To create two 4MB IBM Virtual Shared Disks on each of three primary nodes, one of which has a backup, enter:
createvsd -n 3,4,7/8/ -c 2 -s 4 -g vsdvg -v TEMPThis command creates the following IBM Virtual Shared Disks:
To create three IBM Virtual Shared Disks, where the logical volume created on node 3 spans fewer disks than the volume group does, enter:
createvsd -n 3,4/:hdisk1,hdisk2+hdisk3/,7/8/ -s 4 -g datavg -v USERThis command creates:
Purpose
crunacct - Runs on the acct_master node to produce daily summary accounting reports and to accumulate accounting data for the fiscal period using merged accounting data from each node.
Syntax
Flags
are copied to the acct_master node to the following files:
for all YYYYMMDD prior or equal to the YYYYMMDD being processed.
It also creates an SP system version of the loginlog file, in which each line consists of a date, a user login name and a list of node names. The date is the date of the last accounting cycle during which the user, indicated by the associated login name, had at least one connect session in the SP system. The associated list of node names indicates the nodes on which the user had a login session during that accounting cycle.
Operands
Description
In order for SP accounting to succeed each day, the nrunacct command must complete successfully on each node for which accounting is enabled and then the crunacct command must complete successfully on the acct_master node. However, this may not always be true. In particular, the following scenarios must be taken into account:
From the point of view of the crunacct command, the first scenario results in no accounting data being available from a node. The second scenario results in more than one day's accounting data being available from a node. If it is the case that no accounting data is available from a node, the policy of crunacct is that the error condition is reported and processing continues with data from the other nodes. If data cannot be obtained from at least X percent of nodes, then processing is terminated. "X" is referred to as the spacct_actnode_thresh attribute and can be set via a SMIT panel.
If node data for accounting cycle N is not available when crunacct executes and then becomes available to crunacct during accounting cycle N+1, the node data for both the N and N+1 accounting cycles is merged by crunacct. In general, crunacct merges all data from a node that has not yet been reported into the current accounting cycle, except as in the following case.
If it is the case that crunacct has not run for more than one accounting cycle, such that there are several day's data on each node, then the policy of crunacct is that it processes each accounting cycle's data to produce the normal output for each accounting cycle. For example, if crunacct has not executed for accounting cycles N and N+1, and it is now accounting cycle N+2, then crunacct first executes for accounting cycle N, then executes for accounting cycle N+1 and finally executes for accounting cycle N+2.
However, if the several accounting cycles span from the previous fiscal period to the current fiscal period, then only the accounting cycles that are part of the previous fiscal period are processed. The accounting cycles that are part of the current fiscal period are processed during the next night's execution of crunacct. Appropriate messages are provided in the /var/adm/cacct/active file so that the administrator can execute cmonacct prior to the next night's execution of crunacct.
Restart Procedure
To restart the crunacct command after a failure, first check the /var/adm/cacct/activeYYYYMMDD file for diagnostic messages, and take appropriate actions. For example, if the log indicates that data was unavailable from a majority of nodes, and their corresponding nrunacct state file indicate a state other than complete, check their /var/adm/acct/nite/activeYYYYMMDD files for diagnostic messages and then fix any damaged data files, such as pacct or wtmp.
Remove the lock files and lastdate file (all in the /var/adm/cacct directory), before restarting the crunacct command. You must specify the -r and YYYYMMDD parameters if you are restarting the crunacct command. It specifies the date for which the crunacct command is to rerun accounting. The crunacct procedure determines the entry point for processing by reading the /var/adm/cacct/statefileYYYYMMDD file. To override this default action, specify the desired state on the crunacct command line.
Files
Security
Access Control: This command should grant execute (x) access only to members of the adm group.
Prerequisite Information
For more information about the Accounting System, the preparation of daily and monthly reports, and the accounting files, see IBM Parallel System Support Programs for AIX: Administration Guide.
Related Information
Commands: acctcms, acctcom, acctcon1, acctcon2, acctmerg, acctprc1, acctprc2, accton, crontab, fwtmp, nrunacct
Daemon: cron
The System Accounting information found in AIX Version 4.1 System Management Guide
Examples
nohup /usr/lpp/ssp/bin/crunacct -r 19940601 2>> \ /var/adm/cacct/nite/accterr &This example restarts crunacct for the day of June 1 (0601), 1994. The crunacct command reads the file /var/adm/cacct/statefile19940601 to find out the state with which to begin. The crunacct command runs in the background (&), ignoring all INTERRUPT and QUIT signals (nohup). Standard error output (2) is added to the end (>>) of the /var/adm/cacct/nite/accterr file.
nohup /usr/lpp/ssp/bin/crunacct 19940601 CMS 2>> \ /var/adm/cacct/nite/accterr &This example restarts the crunacct command for the day of June 1 (0601), 1994, starting with the CMS state. The crunacct command runs in the background (&), ignoring all INTERRUPT and QUIT signals (the nohup command). Standard error output (2) is added to the end (>>) of the /var/adm/cacct/nite/accterr file.
Purpose
cshutdown - Specifies the SP system Shutdown command.
Syntax
Flags
If you specify -E, you cannot specify -X.
Notes:
If there are special subsystems, the same waiting procedure applies to subsystem sequencing in the subsystem phase.
Note: | If some critical nodes, but not the entire system, are forced to halt or reboot, the system may not function correctly. |
Operands
Description
Use this command to halt or reboot the entire system or any number of nodes in the system. The SP cshutdown command is analogous to the workstation shutdown command. Refer to the shutdown man page for a description of the shutdown command. The cshutdown command always powers off the nodes except while in Maintenance mode.
Note: |
If you bring a node down to maintenance mode, you must ensure file system
integrity before rebooting the node.
In this case, the cshutdown command, which runs from the control workstation, cannot rsh to the node to perform the node shutdown phase processing. This includes the synchronization of the file systems. Therefore, you should issue the sync command three times in succession from the node console before running the cshutdown command. This is especially important if any files were created while the node was in maintenance mode. To determine which nodes may be affected, issue the spmon -d command and look for a combination of power on and host_responds no. |
The cshutdown command has these advantages over using the shutdown command to shut down each node of an SP:
Using one cshutdown command on the control workstation, you can shut down all or selected nodes.
You can use the /etc/cshutSeq file to control the order in which nodes are shut down, or you can let the system determine the order based on System Data Repository information about /usr servers and clients.
The /etc/subsysSeq file lists these special subsystems and describes any sequencing relationships between them.
Shutdown processing has these phases:
Files
The following files reside on the control workstation:
Restrictions
The cshutdown command can only be issued on the control workstation by root or members of the shutdown group. The root user must issue the kinit command, specifying a principal name for which there is an entry in the hardmon ACLs file with control authorization for the frames to shut down. The hardmon and System Data Repository (SDR) must be running.
Results
The cshutdown command may be gated by the failure of some subsystems or nodes to complete shutdown. In this case, look in the file created: /var/adm/SPlogs/cs/cshut.MMDDhhmmss.pid
If a file with the same name already exists (from a previous year), the cshutdown command overwrites the existing file.
Related Information
Commands: cstartup, init, seqfile, shutdown
Examples
Group1 > Group2 > Group3 Group1: A Group2: B Group3: C
This defines 3 groups, Group1 through Group3, each containing a single node. The nodes names are A, B, and C. The sequence line Group1 > Group2 > Group3 means that Group3 (node C) is shut down first. When Group3 is down, Group2 (node B) is shut down. When Group2 is down, then Group1 (node A) is shut down.
Table 1 shows that the result of a cshutdown command depends on the flags
specified on the command line, the initial state of each node, and the
sequencing rules in /etc/cshutSeq. The shorthand notation
Aup indicates that node A is up and running;
Adnindicates that node A is down.
Table 1. Examples of the cshutdown Command
The subscript the subscript dn means the node is not running. | |||
Initial State | Command Issued | Final State | Explanation |
---|---|---|---|
Aup Bup Cup | cshutdown A B C | Adn Bdn Cdn | The command succeeds; the nodes are all down. |
Aup Bup Cdn | cshutdown B | Aup Bdn Cdn | The command succeeds because C is already not running. |
Aup Bup Cdn | cshutdown A | Unchanged | The command fails because B is still running. |
Aup Bup Cdn | cshutdown -X A | Adn Bup Cdn | The command succeeds because -X considers the sequencing of only the target nodes. |
cshutdown -GXY ALL
cshutdown -G -N 1 9 16-20
The command may fail if any node in the list depends on any node that is not on the list and that node is not shutdown.
cshutdown ALL
The command may fail if any node in the current system partition depends on nodes outside of the current system partition.
cshutdown -N 1 5 6
The command may fail if any node in the list is not in the current system partition or depends on nodes outside of the current system partition.
cshutdown -X -N 1 5 6
cshutdown -F -N 5
cshutdown -kF ALL
cshutdown -rN -C'-W 300' 12-16
cshutdown -rg sleepy_nodes
Purpose
CSS_test - Verifies that the installation and configuration of the Communications Subsystem of the SP system completed successfully.
Syntax
CSS_test
Flags
None.
Operands
None.
Description
Use this command to verify that the Communications Subsystem component ssp.css of the SP system was correctly installed. CSS_test runs on the system partition set in SP_NAME.
A return code of 0 indicates that the test completed without a failure, but unexpected results may be noted on standard output and in the companion log file /var/adm/SPlogs/CSS_test.log. A return code of 1 indicates that a failure occurred.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit SP_verify
Files
Related Information
Commands: jm_install_verify, jm_verify, SDR_test, SYSMAN_test, spmon_ctest, spmon_itest
Examples
To verify the Communication Subsystem following installation, enter:
CSS_test
Purpose
cstartup - Specifies the SP system Startup command.
Syntax
Flags
Note: | Your system may still be usable if one or more nodes fails to complete startup, because the sequencing rules are preserved. |
Note: | If some nodes but not the entire system are forced to start up this way, they may not function properly because of possible resource problems. |
Operands
Description
Caution! |
---|
The cstartup command attempts to power on nodes that are powered off. This has safety implications if someone is working on the nodes. Proper precautions should be taken when using this command. |
The cstartup command starts up the entire system or any number of nodes in the system. If a node is not powered on, startup means powering on the node. If the node is already powered on and not running, startup means resetting the node.
The /etc/cstartSeq file specifies the sequence in which the nodes are started up. See IBM Parallel System Support Programs for AIX: Administration Guide for the format of the /etc/cstartSeq file.
You can use the -SXZ flags to violate the cstartup sequence intentionally. See Table 2 for examples of the effect of these flags.
Files
The following files reside on the control workstation:
Restrictions
The cstartup command can only be issued on the control workstation by root or members of the shutdown group. The root user must issue the kinit command, specifying a principal name for which there is an entry in the hardmon ACLs file with control authorization for the frames to start up. The hardmon and System Data Repository (SDR) must be running.
Results
The /var/adm/SPlogs/cs/cstart.MMDDhhmmss.pid file contains the results of cstartup.
If the command fails, examine this file to see which steps were completed. If a file with the same name already exists (from a previous year), the cstartup command overwrites the existing file.
Related Information
Commands: cshutdown, init, seqfile
Examples
Group1 > Group2 > Group3 > Group4 > Group5 Group1: A Group2: B Group3: C Group4: D Group5: E
This defines five groups, Group1 through Group5, each containing a single node. The nodes names are A, B, C, D, and E. The sequence line Group1 > Group2 > Group3 > Group4 > Group5 means that Group1 (node A) is started first. When Group1 is up, Group2 (node B) is started. When Group2 is up, then Group3 (node C) is started, and so on.
Table 2 shows that the result of a cstartup command depends on the flags
specified on the command line, the initial state of each node, and the
sequencing rules in /etc/cstartSeq. The shorthand notation
Aup indicates that A is powered up and running;
Adnindicates that A is not running.
Table 2. Examples of the cstartup Command
The subscript dn means the node is down. | |||
Initial State | Command Issued | Final State | Explanation |
---|---|---|---|
Adn Bdn Cdn Ddn Edn | cstartup A B C D E | Aup Bup Cup Dup Eup | The command succeeds; the nodes are all up. |
Aup Bup Cdn Ddn Edn | cstartup A B C D E | Aup Bup Cup Dup Eup | The command succeeds, C, D, and E are started up. |
Aup Bup Cdn Dup Edn | cstartup A B C D E | Unchanged | The command fails because D was already up before C. |
Aup Bup Cdn Dup Edn | cstartup -S A B C D E | Aup Bup Cup Dup Eup | The command succeeds because -S ignores sequencing violations. |
Aup Bup Cdn Dup Edn | cstartup -Z A B C D E | Aup Bup Cup Dup Eup | The command succeeds because -Z resets running nodes. |
Aup Bup Cdn Dup Edn | cstartup C E | Unchanged | The command fails because node D was already up before node C. |
Aup Bup Cdn Dup Edn | cstartup -S C E | Aup Bup Cup Dup Eup | The command succeeds because -S ignores sequencing violations. |
Aup Bup Cdn Dup Edn | cstartup -X C E | Aup Bup Cup Dup Eup | The command succeeds because -X considers the sequencing of only the target nodes. |
Aup Bup Cdn Dup Edn | cstartup -Z C E | unchanged | The command fails because resetting C or E does not correct the sequence violation. |
Aup Bup Cdn Ddn Edn | cstartup C E | unchanged | The command fails because D is gating E. Node C is not started either. |
Aup Bup Cdn Ddn Edn | cstartup -S C E | unchanged | The command fails because D is gating E. Node C is not started either. |
Aup Bup Cdn Ddn Edn | cstartup -X C E | Aup Bup Cup Ddn Eup | The command succeeds and starts up only the explicit targets, C and E. |
Aup Bup Cdn Ddn Edn | cstartup -Z C E | unchanged | The command fails because D is gating E. Node C is not started either. |
cstartup -GXZ ALL
cstartup -G -N 1 9 16-20
The command may fail if any node in the list depends on any node that is not on the list and that node is not started up.
cstartup ALL
The command may fail if any node in the current system partition depends on nodes outside of the current system partition.
cstartup -N 1 5 6
The command may fail if any node in the list is not in the current system partition or depends on nodes outside of the current system partition.
cstartup -X -N 1 5 6
cstartup -k ALL
cstartup -E ALL
cstartup -Gg sleepy_nodes
Purpose
ctlhsd - Sets the data striping device (HSD) for the IBM Virtual Shared Disks operational parameters or reset statistics.
Syntax
ctlhsd [-p parallel_level | -v hsd_name ... | -C | -V]
Flags
Operands
None.
Description
Use this command to set the parallelism level and to reset the statistics of the data striping device (HSD) for the IBM Virtual Shared Disk. When specified with no arguments, it displays the the current parallelism level, the number of reworked requests, and the number of requests that were not at a page boundary. When ctlhsd is used to reset the statistics of the device driver, or a particular device, or all the configured data striping devices on the system, it will not suspend all the underlying IBM Virtual Shared Disks. In other words, the user should make sure that there are no I/O activities on the IBM Virtual Shared Disks.
Use lshsd -s to display the statistics on the number of read and write requests at the underlying IBM Virtual Shared Disks in an HSD or all HSDs. Use the -v or -V flag to reset these counters.
Files
Security
You must have root privilege to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfghsd, lshsd, lsvsd, resumevsd, suspendvsd, ucfghsd
Examples
To display the current parallelism level and counter, enter:
ctlhsdThe system displays a message similar to the following:
The current parallelism level is 9. The number of READ requests not at page boundary is 0. The number of WRITE requests not at page boundary is 0.
Purpose
ctlvsd - Sets IBM Virtual Shared Disk operational parameters.
Syntax
Flags
Note: | Before using this flag, refer to the "Restrictions" section that follows. |
Note: | Before using this flag, refer to the "Restrictions" section that follows. |
This value is the buf_cnt parameter on the uphysio call the IBM Virtual Shared Disk IP device driver makes in the kernel. Use statvsd to display the current value on the node on which the command is run.
Operands
None.
Description
The ctlvsd command changes some parameters of the IBM Virtual Shared Disk. When called with no arguments it displays the current and maximum cache buffer count, the request block count, the pbuf count, the minimum buddy buffer size, the maximum buddy buffer size as well as the overall size of the buddy buffer.
Use statvsd to display outgoing and expected sequence numbers and out cast status of other nodes as viewed by the node on which the command is run. It is best to suspendvsd -a on all nodes whose sequence numbers are being reset prior to actually resetting the sequence numbers. Be sure to use resumevsd on all IBM Virtual Shared Disks that were suspended after resetting the sequence numbers.
Initially, all sequence numbers are set to 0 when the first IBM Virtual Shared Disk is configured and the IBM Virtual Shared Disk device driver is loaded. Thereafter, sequence numbers are incremented as requests are sent to (outgoing) and received from (expected) other nodes, and reset via ctlvsd -R | -r commands.
Reloading the IBM Virtual Shared Disk device driver by suspendvsd -a, stopvsd -a, or ucfgvsd -a followed by cfgvsd also resets all sequence numbers to 0.
Initially, all nodes in the IBM Virtual Shared Disk network are cast in. The ctlvsd -k command casts a node out. The local node ignores requests from cast out nodes. The ctlvsd -r command casts nodes back in.
Files
Security
You must have root privilege to run this command.
Restrictions
If you have the IBM Recoverable Virtual Shared Disk product installed and operational, do not use the -k and -K options. The results may be unpredictable.
See IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfgvsd, lsvsd, preparevsd, resumevsd, startvsd, statvsd, stopvsd, suspendvsd, ucfgvsd
Refer to IBM Parallel System Support Programs for AIX: Managing Shared Disks for information on tuning IBM Virtual Shared Disk performance and sequence numbers.
Examples
To display the current parameters, enter:
ctlvsdThe system displays a message similar to the following:
The current cache buffer count is 64. The maximum cache buffer count is 256. The request block count is 256. The pbuf's count is 48. The minimum buddy buffer size is 4096. The maximum buddy buffer size is 65536. The total buddy buffer size is 4 max buffers, 262144 bytes.
To display the mbuf headers and current routing table, enter:
ctlvsd -tThe system displays the following information:
Mbuf Cache Stats: Header Cached 1 Hit 1023 Miss 1 Route cache information: destination interface ref status direct/gateway min managed mbuf 1 css0 2 Up Direct 256
Purpose
defhsd - Defines a data striping device (HSD).
Syntax
defhsd protect_LVCB | not_protect_LVCB hsd_name stripe_size vsd_name...
Flags
None.
Operands
Description
The defhsd command is used to specify the hsd_name, stripe size and underlying IBM Virtual Shared Disks for the new data striping device (HSD).
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit vsd_dataand select the Define a Hashed Shared Disk option.
Files
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: hsdatalst, undefhsd
Refer to IBM Parallel System Support Programs for AIX: Managing Shared Disks for information on tuning IBM Virtual Shared Disk performance and sequence numbers.
Examples
The following example adds SDR information indicating a stripe size of 32768, composed of vsd.vsdn101, vsd.vsdn201, and the name hsd1 is defined.
defhsd hsd1 32768 vsd.vsdn101 vsd.vsdn201
Purpose
defvsd - Defines an IBM Virtual Shared Disk.
Syntax
defvsd logical_volume_name global_group_name vsd_name [nocache | cache]
Flags
None.
Operands
Note: | If you choose a vsd_name that is already the name of another device, the cfgvsd command will fail for that IBM Virtual Shared Disk. This failure ensures that the special device files created for the name do not overlay and destroy files of the same name representing some other device type (such as a logical volume). |
The cache option should only be used if the using application gains performance by avoiding a 4KB read immediately after a 4KB write. Refer to IBM Parallel System Support Programs for AIX: Managing Shared Disks for additional information on IBM Virtual Shared Disk tuning.
Description
This command is run to specify logical volumes residing on globally accessible volume groups to be used as IBM Virtual Shared Disks.
You can use the System Management Interface Tool (SMIT) to run the defvsd command. To use SMIT, enter:
smit vsd_dataand select the Define a Virtual Shared Disk option.
Security
You must have root privilege to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: vsdatalst, vsdvg, undefvsd
Refer to IBM Parallel System Support Programs for AIX: Managing Shared Disks for information regarding IBM Virtual Shared Disk performance enhancements.
Examples
defvsd lv1vg1n1 vg1n1 vsd1vg1n1
defvsd lv2vg1n1 vg1n1 vsd1vg2n1 cache
Purpose
delnimclient - Deletes a Network Installation Management (NIM) client definition from a NIM master.
Syntax
delnimclient -h | -l node_list | -s server_node_list
Flags
Operands
None.
Description
Use this command to undefine a node as a NIM client. This is accomplished by determining the node's boot/install server and unconfiguring that client node as a NIM client on that server. When complete, the entry for the specified client is deleted from the NIM configuration database on the server. This command does not change the boot/install attributes for the nodes in the System Data Repository.
Note: | This command results in no processing on the client node. |
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/delnimclient
Related Information
Commands: mknimclient, setup_server
Examples
To delete the NIM client definition for nodes 1, 3, and 5 from the NIM database on their respective boot/install servers, enter:
delnimclient -l 1,3,5
Purpose
delnimmast - Unconfigures a node as a Network Installation Management (NIM) master.
Syntax
delnimmast -h | -l node_list
Flags
Operands
None.
Description
Use this command to undefine a node as a NIM master. This command does not change the boot/install attributes for the nodes in the System Data Repository.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/delnimmast
Related Information
Commands: mknimmast, setup_server
Examples
To unconfigure nodes 1, 3, and 5 as NIM masters and delete the NIM file sets, enter:
delnimmast -l 1,3,5
Purpose
dsh - Issues commands to a group of hosts in parallel.
Syntax
Flags
If the -a, -w, or -N flags are not specified, the WCOLL environment variable contains the name of a file containing host names for the working collective.
Operands
Description
The dsh executes commands against all or any subset of the hosts in a network. It reads lines from the command line or standard input and executes each as a command on a set of network-connected hosts. These commands are in rsh syntax. Alternatively, a single command in rsh syntax can be specified on the dsh command line.
As each command is read, it is interpreted by passing it to each host in a group called the working collective via the SP rsh command.
The working collective is obtained from the first existence of one of the following:
If neither of these exist, an error has occurred and no commands are issued.
The working collective file should have one host name per line. Blank lines and comment lines beginning with # are ignored.
The path used when resolving the dsh command on the target nodes is the path set by the user with the DSHPATH environment variable. If DSHPATH is not set, the path used is the rsh default path, /usr/ucb:/bin:/usr/bin:. The DSHPATH environment variable only works when the user's remote login shell is the Bourne or Korn shell. An example would be to set DSHPATH to the path set on the source machine (for example, DSHPATH=$PATH).
The maximum number of concurrent rsh's can be specified with the fanout (f) flag or via the FANOUT environment variable. If desired, sequential execution can be obtained by specifying a fanout value of 1. Results are displayed as remote commands complete. All rsh's in a fanout must complete before the next set of rsh's is started. If fanout is not specified via FANOUT or the f flag, rsh's to 64 hosts are issued concurrently. Each rsh that dsh runs requires a reserved TCP/IP port and only 512 such ports are available per host. With large fanouts, it is possible to exhaust all the ports on a host, causing commands that use these ports, such as the SP rlogin and the SP rsh commands, to fail.
Exit values for the rsh commands are displayed in messages from dsh if nonzero. (A nonzero return code from rsh indicates that the rsh has failed; it has nothing to do with the exit code of the remotely executed command.) If an rsh fails, that host is removed from the current working collective (not the current working collective file), unless the c flag was set.
The dsh exit value is 0 if no errors occurred in the dsh command and all rsh's finished with exit codes of 0. The dsh exit value is more than 0 if internal errors occur or the rsh's fail. The exit value is increased by 1 for each rsh failure.
No particular error recovery for command failure on remote hosts is provided. The application or user can examine the command results in dsh's standard error and standard output, and take appropriate action.
The dsh command waits until results are in for each command for all hosts and displays those results before reading more input commands.
The dsh command does not work with interactive commands, including those that read from standard input.
The dsh command output consists of the output (standard error and standard output) of the remotely executed commands. The dsh standard output is the standard output of the remote command. The dsh standard error is the standard error of the remote command. Each line is prefixed with the host name of the host from which that output came. The host name is followed by ":" and a line of the command output.
For example, let's say that a command was issued to a working collective of host1, host2, and host3. When the command was issued on each of the hosts, the following lines were written by the remote commands:
For host1 stdout: h1out1 h1out2 For host2 stdout: h2out1 h2out2 For host3 stdout: h3out1 For host3 stderr: h3err1 h3err2 dsh stdout will be host1: h1out1 host1: h1out2 host2: h2out1 host2: h2out2 host3: h3out1 dsh stderr will be host3: h3err1 host3: h3err2
A filter to display the output lines by the host is provided separately. See the dshbak command.
If a host is detected as down (for example, an rsh returns a nonzero return code), subsequent commands are not sent to it on this invocation of dsh, unless the c (continue) option is specified on the command line.
An exclamation point at the beginning of a command line causes the command to be passed directly to the local host in the current environment. The command is not sent to the working collective.
Signals 2 (INT), 3 (QUIT), and 15 (TERM) are propagated to the remote commands.
Signals 19 (CONT), 17 (STOP), and 18 (TSTP) are defaulted. This means that the dsh command responds normally to these signals, but they do not have an effect on the remotely running commands. Other signals are caught by dsh and have their default effects on the dsh command. In the case of these other signals, all current children, and via propagation their remotely running commands, are terminated (SIGTERM).
Security considerations are the same as for the SP rsh command.
Files
Related Information
Command: dshbak
SP Commands: rsh, sysctl
Examples
WCOLL=./wchosts dsh ps
dsh -q
dsh -w otherhost1,otherhost2,otherhost3
dsh -w host1,host2,host3 -a cat /etc/passwd | dshbak
dsh -w otherhost cat remotefile '>>' otherremotefile
dsh -if 1 -a < commands_file > results 2>&1
dsh ps -ef | grep root
dsh 'ps -ef | grep root'
or
dsh ps -ef "|" grep root
dsh -w host1 cat /etc/passwd | cut -d: -f2- | cut -c2- > myetcpasswd
dsh -N my_nodes ps
Purpose
dshbak - Presents formatted output from the dsh and sysctl commands.
Syntax
dshbak [-c]
Flags
Operands
None.
Description
The dshbak command takes lines in the following format:
host_name: line of output from remote command
The dshbak command formats them as follows and writes them to standard output. Assume that the output from host_name3 and host_name4 is identical and the c option was specified:
HOSTS ----------------------------------------------------------------- host_name1 ----------------------------------------------------------------------- . . lines from dsh or sysctl with host_names stripped off . . HOSTS ----------------------------------------------------------------- host_name2 ----------------------------------------------------------------------- . . lines from dsh or sysctl with host_names stripped off . . HOSTS ----------------------------------------------------------------- host_name3 host_name4 ----------------------------------------------------------------------- . . lines from dsh or sysctl with host_names stripped off . .
When output is displayed from more than one host in collapsed form, the host names are displayed alphabetically.
When output is not collapsed, output is displayed sorted alphabetically by host name.
The dshbak command writes "." for each 1000 lines of output filtered.
Files
Related Information
Commands: dsh, sysctl
Examples
dsh -w host1,host2,host3 cat /etc/passwd | dshbak
dsh -w host1,host2,host3 pwd | dshbak -c
Purpose
Eannotator - Annotates the connection labels in the topology file.
Syntax
Eannotator -F input_file -f output_file -O [yes | no]
Flags
Operands
None.
Description
This command supports all of the following:
This command must be executed whenever a new topology file is selected.
The topology file, which describes the wiring configuration for the High Performance Switch, contains node-to-switch or switch-to-switch cable information. A node-to-switch connection looks like following:
s 25 2 tb0 17 0 E2-S00-BH-J16 to E2-N2
The predefined node-to-switch connections start with an "s" which indicates a switch connection. The next two digits, in this case "25" indicate the switch (2) and switch chip (5) being connected. The next digit, in this case "2", indicates the switch chip port in the connection. The next field, in this case "tb0", specifies the type of adapter present in the SP node. The following field, in this case "17", is the switch node number for the SP node, and the last digit, in this case "0", indicates the adapter port within the connection.
For switch-to-switch connections, the first four fields (switch indicator, switch, switch chip, and switch chip port) are repeated to identify the other end of the connection.
The connection label "E2-S00-BH-J16 to E2-N2" provides physical connection information for a customer's use to identify the bad connection.
Depending on the type of switch installed (High Performance Switch or SP Switch), together with the customer's physical switch frame configuration defined in the SDR, the Eannotator command retrieves switch node and dependent node objects from the SDR and applies proper connection information to the topology file.
If the input topology file contains existing connection information, the Eannotator command replaces the existing connection label with the new connection labels. If the input topology file does not contain connection labels, the Eannotator command appends the proper connection label to each line on the topology file.
The precoded connection labels on the topology file start with an "L" which indicate logical frames. The Eannotator command replaces the "L" character with an "E" which indicates physical frames. The "S" character indicates which slot the switch occupies in the frame, the "BH" characters indicate a Bulk Head connection, the "J" character indicates which jack provides the connection from the switch board, the "N" character indicates the node being connected to the switch, and the "SC" characters indicate the Switch Chip connection.
Files
Security
You must have root privilege to run this command.
Related Information
Commands: Eclock, Eduration, Efence, Eprimary, Equiesce, Estart, Etopology, Eunfence, Eupartition
Refer to IBM RS/6000 SP: Planning, Volume 2, Control Workstation and Software Environment for details about system partition topology files.
Examples
Before: s 15 3 tb0 0 0 L01-S00-BH-J18 to L01-N1 After: s 15 3 tb0 0 0 E01-S17-BH-J18 to E01-N1
Note: | Logical frame L01 is defined as physical frame 1 in the SDR Switch object. |
Before: s 10016 0 s 51 3 L09-S1-BH-J20 to L05-S00-BH-J19 After: s 10016 0 s 51 3 E10-S1-BH-J20 to E05-S17-BH-J19
Note: | Logical frame L09 is defined as physical frame 10 in the SDR Switch object. |
Before: s 15 3 tb0 0 0 L03-S00-BH-J18 to L03-N3 After: s 15 3 tb3 0 0 E03-S17-BH-J18 to E03-N3 # Dependent Node
Note: | Logical frame L03 is defined as physical frame 3 in the SDR Switch object and the node was determined to be a dependent node. |
Eannotator -F expected.top.8nsb.4isb.0 -f expected.top -O no
Eannotator -F expected.top.1nsb.0isb.0 -f expected.top -O yes
Purpose
Eclock - Controls the clock source for each switch board within an SP cluster.
Syntax
Flags
where:
High Performance Switch
SP Switch
If a flag is not specified, the clock input values stored in the SDR are displayed.
Operands
None.
Description
Use this command to set the multiplexors that control the clocking at each switch board within the configuration. One switch board within the configuration is designated as the "Master" switch that provides the clocking signal for all other switch boards within the configuration. The Eclock command reads clock topology information from either the file specified on the command line or the clock topology data within the SDR. If a clock topology file was specified, the Eclock command places the clock topology information into the SDR, so that it can be accessed again during a subsequent Eclock invocation. After processing the clock topology file, Eclock causes the new clock topology to take effect for the switches specified. A clock topology file contains the following information for each switch board within the cluster:
High Performance Switch
SP Switch
High Performance Switch Warning |
---|
Do not change the switch clock multiplexor settings (with Eclock, spmon command/GUI, hmcmds) while the nodes are powered on. Otherwise, AIX must be rebooted and Estart must be run following the clock adjustment. |
SP Switch Warning |
---|
Do not change the switch clock multiplexor settings (with Eclock, spmon command/GUI, hmcmds) while the nodes are powered on. Otherwise, Estart must be run following the clock adjustment. |
To execute the Eclock command, the user must be authorized to access the Hardware Monitor subsystem and, for those frames specified to the command, the user must be granted VFOP (Virtual Front Operator Panel) permission. Commands sent to frames for which the user does not have VFOP permission are ignored. Since the Hardware Monitor subsystem uses SP authentication services, the user must execute the kinit command prior to executing this command. Alternatively, site-specific procedures can be used to obtain the tokens that are otherwise obtained by kinit.
Files
Security
You must have root privilege to run this command.
Related Information
Commands: Eannotator, Eduration, Efence, Eprimary, Equiesce, Estart, Etopology, Eunfence, Eunpartition
Examples
Eclock -f /etc/SP/Eclock.top.8nsb.4isb.0
Eclock
Eclock -s 1 -m 0
Eclock -c /tmp/Eclock.top
Eclock -a /etc/SP/Eclock.top.4nsb.2isb.0
Eclock -d
Purpose
Usage Note |
---|
Do not use this command if you have the SP Switch installed on your system. |
Eduration - Sets the interval that nodes can be added or removed from the High Performance Switch. This interval is called the Run Phase Duration.
Syntax
Eduration [[days day[s]] [hours hour[s]] [minutes minute[s]] ] | [-h]
Flags
Any combination of the three preceding time designations can be used to specify the new Run Phase Duration. Since the duration determines how quickly the system can respond to Efence and Eunfence requests, it should be set to provide the desired response. If none of the time specifiers are present, Eduration will display the current value of the Run Phase Duration.
Operands
None.
Description
The Run Phase Duration controls how frequently Efence and Eunfence requests are handled. This command provides an interface to set that value.
Note: | The Run Phase Duration changes will not take effect until the end of the current Run Phase. If you are changing the Run Phase Duration from a large value to something that is significantly smaller and you do not want to wait for the current the Run Phase to complete, you will have to Estart the switch. |
Security
You must have root privilege to run this command.
Related Information
Commands: Eannotator, Eclock, Efence, Eprimary, Equiesce, Estart, Etopology, Eunfence, Eunpartition
Examples
Eduration 1 minute
Eduration 1 hour 30 minutes
Eduration
Purpose
Efence - Removes an SP node from the current active switch network.
Syntax
Efence [-h] | [-G] [-autojoin] [node_specifier] ...
Flags
If you have an SP Switch installed on your system, such nodes are also rejoined when an Estart command is issued.
Operands
Note: | You cannot fence the primary node on the High Performance Switch and you cannot fence either the primary or primary backup nodes on the SP Switch. |
Description
Use this command to fence a node from the current switch network.
If you have an SP Switch installed on your system, you must do either of the following to bring the node back up onto the switch network:
If you have a High Performance Switch installed on your system, you can issue the Estart command to rejoin all nodes on the switch network.
Note: | If a host name or IP address is used as the node_specifier for a dependent node, it must be a host name or IP address assigned to the adapter that connects the dependent node to the SP Switch. Neither the administrative host name nor the Simple Network Management Protocol (SNMP) agent's host name for a dependent node is guaranteed to be the same as the host name of its switch network interface. |
Security
You must have root privilege to run this command.
Related Information
Commands: Eannotator, Eclock, Eduration, Eprimary, Equiesce, Estart, Etopology, Eunfence, Eunpartition
Examples
Efence
Efence -autojoin
Efence -G
Efence 129.33.34.1 129.33.34.6
Efence r11n01
Efence -autojoin 54 65 32 78
Efence 2,14
Efence 5 6 7 8fences nodes 5 and 6, but not nodes 7 and 8. As a result, the command returns a nonzero return code.
Efence -G 5 6 7 8
Purpose
emconditionctrl - Loads the System Data Repository (SDR) with predefined Event Management conditions.
Syntax
emconditionctrl [-a] [-s] [-k] [-d] [-c] [-t] [-o] [-r] [-h]
Flags
Operands
None.
Description
The emconditionctrl script loads the SDR with some useful conditions that can be used for registering for Event Management events. Currently the SP Perspectives application can make use of conditions.
The emconditionctrl script is not normally executed on the command line. It is normally called by the syspar_ctrl command after the control workstation has been installed or when the system is partitioned. It implements all of the flags that syspar_ctrl can pass to its subsystems, although only the -a flag causes any change to the system. The -a flag causes predefined conditions to be loaded only if run on the control workstation. It has no effect if run elsewhere.
Exit Values
Security
You must be running with an effective user ID of root.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/emconditionctrl
Related Information
Commands: syspar_ctrl
Purpose
emonctrl - A control script that manages the Emonitor subsystem.
Syntax
emonctrl { -a | -s | -k | -d | -c | -t | -o | -r | -h }
Flags
Operands
None.
Description
The Emonitor subsystem monitors designated nodes in an attempt to maximize their availability on the switch network.
The emonctrl control script controls the operation of the Emonitor subsystem. The subsystem is under the control of the System Resource Controller (SRC) and belongs to a subsystem group called emon.
An instance of the Emonitor subsystem can execute on the control workstation for each system partition. Because Emonitor provides its services within the scope of a system partition, it is said to be system partition-sensitive. This control script operates in a manner similar to the control scripts of other system partition-sensitive subsystems. It should be issued from the control workstation and is not functional on the nodes.
From an operational point of view, the Emonitor subsystem group is organized as follows:
The emon group is associated with the Emonitor daemon.
On the control workstation, there are multiple instances of Emonitor, one for each system partition. Accordingly, the subsystem names on the control workstation have the system partition name appended to them. For example, for system partitions named sp_prod and sp_test, the subsystems on the control workstation are named Emonitor.sp_prod and Emonitor.sp_test.
The Emonitor daemon provides switch node monitoring.
The emonctrl script is not normally executed from the command line. It is normally called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system.
The emonctrl script provides a variety of controls for operating the Emonitor daemon:
Before performing any of these functions, the script obtains the current system partition name and IP address (using the spget_syspar command) and the node number (using the node_number) command. If the node number is zero, the control script is running on the control workstation. Since the Emonitor daemon runs only on the control workstation, the script performs no function when run on a node.
Except for the clean function, all functions are performed within the scope of the current system partition.
Adding the Subsystem
When the -a flag is specified, the control script uses the mkssys command to add the Emonitor daemon to the SRC. The control script operates as follows:
Starting the Subsystem
This option is unused since the Emonitor daemon must be started via Estart -m.
Stopping the Subsystem
When the -k flag is specified, the control script uses the stopsrc command to stop the Emonitor daemon in the current system partition.
Deleting the Subsystem
When the -d flag is specified, the control script uses the rmssys command to remove the Emonitor subsystem from the SRC. The control script operates as follows:
Cleaning Up the Subsystems
When the -c flag is specified, the control script stops and removes the Emonitor subsystems for all system partitions from the SRC. The control script operates as follows:
Turning Tracing On
Not currently used.
Turning Tracing Off
Not currently used.
Refreshing the Subsystem
Not currently used.
Logging
While it is running, the Emonitor daemon provides information about its operation and errors by writing entries in a log file. The Emonitor daemon uses log files called /var/adm/SPlogs/css/Emonitor.log and /var/adm/SPlogs/css/Emonitor.Estart.log.
Files
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/emonctrl
Related Information
Commands: Emonitor, Estart, lssrc, startsrc, stopsrc, syspar_ctrl
Examples
emonctrl -a
emonctrl -k
emonctrl -d
emonctrl -c
lssrc -g emon
lssrc -s subsystem_name
lssrc -a
Purpose
Emonitor - Monitors nodes listed in the /etc/SP/Emonitor.cfg file in an to attempt to maximize this availability on the switch.
Syntax
Emonitor
Flags
None.
Operands
None.
Description
Emonitor is a daemon controlled by the System Resource Controller (SRC). It can be used to monitor nodes in a system partition in regard to the their status on the switch. A system-wide configuration file (/etc/SP/Emonitor.cfg) lists all nodes on the system to be monitored. The objective is to bring these nodes back up on the switch network when necessary.
Emonitor is invoked with Estart -m. Once invoked, it is controlled by SRC so it will restart if it is halted abnormally. If the you decide to end monitoring, you must run /usr/lpp/ssp/bin/emonctrl -k to stop the daemon in your system partition.
There is an Emonitor daemon for each system partition. The daemon watches for any node coming up (for example, host_responds goes from 0 to 1). When the daemon detects a node coming up, it performs a review of the nodes in the configuration file to check if any node is off the switch network. If any nodes in the specified system partition are off the switch network, it determines a way to bring them back onto the the switch (for example, via Eunfence or Estart), and takes the appropriate action. In order to avoid the Estart command from being run several times (which can occur if multiple nodes are coming up in sequence), Emonitor waits 3 minutes after a node comes up to be sure no other nodes are in the process of coming up. Each time a new node comes up prior to the 3 minute timeout, Emonitor resets the timer to a maximum wait of 12 minutes.
Emonitor cannot always bring nodes back on the switch. For example, if any of the following occur:
On a High Performance Switch, if a node is faulted off the switch and you are forced to do an Estart, you will lose history of any nodes that you had isolated off the switch. All nodes on a High Performance Switch come back on the switch on an Estart.
Problems can occur if the node that is faulted off the switch is experiencing a recurring error that causes it to come up and then fail repeatedly. The monitor continually attempts to bring this node into the switch network and could jeopardize the stability of the remaining switch network.
Note: | Nodes that will be undergoing hardware or software maintenance should be removed from the Emonitor.cfg file during this maintenance to prevent Emonitor from attempting to to bring them onto the switch network. |
Files
Security
You must have root privilege to run this command.
Related Information
Commands: Eannotator, Eclock, Eduration, Efence, emonctrl, Eprimary, Equiesce, Estart, Etopology, Eunfence, Eupartition
Purpose
enadmin - Changes the desired state of a specified extension node.
Syntax
Flags
Operands
Description
Use this command to change the administrative state of an extension node. Setting the administrative state of an extension node to reconfigure causes configuration data for the extension node to be resent to the extension node's administrative environment. Setting the administrative state of an extension node to reset places the extension node in an initial state in which it is no longer active on the switch network.
This command is invoked internally when choosing the reconfigure option of the endefadapter and endefnode commands or the reset (-r) option of the enrmnode command.
You can use the System Management Interface Tool (SMIT) to run this command by selecting the Extension Node Management panel. To use SMIT, enter:
smit manage_extnode
Standard Output
All informational messages are written to standard output. These messages identify the extension node being changed and indicate when the specified state change has been accepted for processing by the extension node agent (at which point the command is complete). All error messages are also written to standard output.
Exit Values
Security
You must have root privilege to run this command or be a member of the system group.
Restrictions
This command can only be issued on the control workstation.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP) ssp.spmgr file set.
The spmgrd SNMP manager daemon on the SP control workstation allows transfer of extension node configuration data from the SP system to an SNMP agent providing administrative support for the extension node. Version 1 of the SNMP protocol is used for communication between the SNMP manager and the SNMP agent. Limited control of an extension node is also possible. An SNMP set-request message containing an object instantiation representing the requested administrative state for the extension node is sent from the SNMP manager to the SNMP agent providing administrative support for the extension node. After the administrative state of an extension node is received by the SNMP agent, the enadmin command is completed. Requests for configuration information and information about the state of an extension node are sent to the SNMP manager asynchronously in SNMP trap messages.
Prerequisite Information
IBM RS/6000 SP: Planning, Volume 2, Control Workstation and Software Environment
Location
/usr/lpp/ssp/bin/enadmin
Related Information
Commands: endefadapter, endefnode, enrmadapter, enrmnode, spmgrd
Examples
enadmin -a reconfigure 9
enadmin -a reset 9
Purpose
endefadapter - Adds new or changes existing configuration data for an extension node adapter in the System Data Repository (SDR) and optionally performs the reconfiguration request.
Syntax
endefadapter [-a address] [-h] [-m netmask] [-r] node_number
Flags
Operands
Description
Use this command to define extension node adapter information in the SDR. The -a and -m flags and the node_number operand are required.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit enter_extadapter
Environment Variables
The SP_NAME environment variable is used (if set) to direct this command to a system partition. If the SP_NAME environment variable is not set, the default system partition will be used.
Standard Output
This command writes informational messages to standard output.
Standard Error
This command writes all error messages to standard error.
Exit Values
Security
You must have root privilege to run this command or be a member of the system group.
Restrictions
This command can only be issued on the control workstation.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP) ssp.basic file set.
Prerequisite Information
IBM RS/6000 SP: Planning, Volume 2, Control Workstation and Software Environment
Location
/usr/lpp/ssp/bin/endefadapter
Related Information
Commands: enadmin, endefnode, enrmadapter, enrmnode
Examples
endefadapter -a 129.40.158.137 -m 255.255.255.0 10
endefadapter -a 129.40.158.137 -m 255.255.255.0 -r 10
Purpose
endefnode - Adds new or changes existing configuration data for an extension node in the System Data Repository (SDR) and optionally performs the reconfiguration request.
Syntax
Flags
Operands
Description
Use this command to define extension node information in the SDR. When adding a new extension node, the -a, -i, and -s flags and the node_number operand are required. When changing an existing extension node definition, only the node number is required along with the flag corresponding to the field being changed.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit enter_extnode
Environment Variables
The SP_NAME environment variable is used (if set) to direct this command to a system partition. If the SP_NAME environment variable is not set, the default system partition will be used.
Standard Output
This command writes informational messages to standard output.
Standard Error
This command writes all error messages to standard error.
Exit Values
Security
You must have root privilege to run this command or be a member of the system group.
Restrictions
This command can only be issued on the control workstation.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP) ssp.basic file set.
Prerequisite Information
IBM RS/6000 SP: Planning, Volume 2, Control Workstation and Software Environment
Location
/usr/lpp/ssp/bin/endefnode
Related Information
Commands: enadmin, endefadapter, enrmnode, enrmadapter
Refer to the SP Switch Router Adapter Guide for information about attaching an IP router extension node to the SP Switch.
Examples
endefnode -i 13 -a router1 -s router1 -c spenmgmt 2
endefnode -i 02 -a grf.pok.ibm.com -s grf.pok.ibm.com -c spenmgmt -r 7
Purpose
enrmadapter - Removes configuration data for an extension node adapter from the System Data Repository (SDR).
Syntax
enrmadapter [-h] node_number
Flags
Operands
Description
Use this command to remove extension node adapter information from the SDR. The node_number operand is required.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit delete_extadapter
Environment Variables
The environment variable SP_NAME is used (if set) to direct this command to a system partition. If the SP_NAME environment variable is not set, the default system partition will be used.
Standard Output
This command writes informational messages to standard output.
Standard Error
This command writes all error messages to standard error.
Exit Values
Security
You must have root privilege to run this command or be a member of the system group.
Restrictions
This command can only be issued on the control workstation.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP) ssp.basic file set.
Prerequisite Information
IBM RS/6000 SP: Planning, Volume 2, Control Workstation and Software Environment
Location
/usr/lpp/ssp/bin/enrmadapter
Related Information
Commands: enadmin, endefadapter, endefnode, enrmnode
Examples
To remove an extension node adapter with a node number of 12 from the SDR, enter:
enrmadapter 12
Purpose
enrmnode - Removes configuration data for an extension node in the System Data Repository (SDR).
Syntax
enrmnode [-h] [-r] node_number
Flags
Operands
Description
Use this command to remove extension node information from the SDR. When removing information, the node_number operand is required.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit delete_extnode
Environment Variables
The environment variable SP_NAME is used (if set) to direct this command to a system partition. If the SP_NAME environment variable is not set, the default system partition will be used.
Standard Output
This command writes informational messages to standard output.
Standard Error
This command writes all error messages to standard error.
Exit Values
Security
You must have root privilege to run this command or be a member of the system group.
Restrictions
This command can only be issued on the control workstation.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP) ssp.basic file set.
Prerequisite Information
IBM RS/6000 SP: Planning, Volume 2, Control Workstation and Software Environment
Location
/usr/lpp/ssp/bin/enrmnode
Related Information
Commands: enadmin, endefadapter, endefnode, enrmadapter
Examples
To remove an extension node with a node number of 2 from the SDR and reset that extension node, enter:
enrmnode -r 2
Purpose
Eprimary - Assigns or queries the switch primary node and switch primary backup node for a system partition.
***High Performance Switch***
Syntax
Eprimary [-h] [-init] [node_identifier]
Flags
Operands
Note: | If no flags or operands are specified, the current switch primary node is displayed. |
Description
Use this command to assign, change, or query the switch primary node. When the -init option is specified, it can be used to create a switch partition object for a system partition. The primary node should not be changed unless the current primary node is becoming unavailable (for example, if the current primary node is to be serviced). The Estart command must be issued before a change of the primary node (using Eprimary) takes effect. The old primary node must be rebooted or powered off before issuing Estart to remove its inclination to behave as the primary node.
Security
You must have root privilege to run this command.
Related Information
Commands: Eannotator, Eclock, Eduration, Efence, Equiesce, Estart, Etopology, Eunfence
Examples
Eprimary
Eprimary 129.33.34.1
Eprimary 1
Eprimary r11n01
Eprimary -init 1,2
***SP Switch***
Syntax
Eprimary [-h] [-init] [node_identifier] [-backup bnode_identifier]
Flags
Operands
Note: |
If no flags or operands are specified, each of the following is
displayed:
|
Description
Use this command to assign, change, or query the switch primary node or the switch primary backup node. The primary node should not be changed unless the current primary node is becoming unavailable (for example, if the current primary node is to be serviced). The Estart command must be issued before a change of the primary node or the primary backup node (using Eprimary) takes effect.
In an SP Switch network, the primary node takeover facility automatically handles situations (such as a node loss) for each of the primary and primary backup nodes. The primary node replaces a failing primary backup node and the primary backup node automatically takes over for the primary node if the primary node becomes unavailable. Note that the node chosen cannot be a dependent node. The primary backup node should be selected using the following guidelines:
The Eprimary command selects a default oncoming primary or oncoming backup primary node if one is not specified. Users receive a warning in the following situations on the oncoming primary or oncoming backup primary nodes:
Security
You must have root privilege to run this command.
Related Information
Commands: Eannotator, Eclock, Eduration, Efence, Equiesce, Estart, Etopology, Eunfence, Eunpartition
Examples
Eprimary
Eprimary 129.33.34.1
Eprimary 129.33.34.1 -backup 129.33.34.56
Eprimary r11n01 -backup r17n02
Eprimary -init 1,2 -backup 1,6
Purpose
Usage Note |
---|
Use this command only if you have an SP Switch installed on your system. |
Equiesce - Quiesces the switch by causing the primary and primary backup nodes to shut down switch recovery and primary node takeover.
Syntax
Equiesce [-h]
Flags
Operands
None.
Description
Use this command to disable switch error recovery and primary node takeover. It is used to shut down normal switch error actions when global activities affecting nodes are performed. For example, when all nodes are shutdown or rebooted, they are fenced from the switch by the primary node.
If the primary node is not the first node to shut down during a global shutdown or reboot of the entire system, it may fence all the other nodes including the primary backup node. Primary node takeover can also occur if the primary node is shut down and the backup node remains up. Issuing the Equiesce command before the shutdown prevents these situations from occurring.
The Equiesce command causes the primary and primary backup nodes to shut down their recovery actions. Data still flows over the switch, but no faults are serviced and primary node takeover is disabled. Only the Eannotator, Eclock, Eprimary, Estart, and Etopology commands are functional after the Equiesce command is issued.
Estart must be issued when the global activity is complete to reestablish switch recovery and primary node takeover.
Security
You must have root privilege to run this command.
Location
/usr/lpp/ssp/bin/Equiesce
Related Information
Commands: Eannotator, Eclock, Efence, Eprimary, Estart, Etopology, Eunfence, Eunpartition
Examples
To quiesce the switch before shutting down the system, enter:
Equiesce
Purpose
Syntax
Estart [-h] [-m]
Flags
Operands
None.
Description
Use this command to start or restart the current system partition based on its switch topology file. (Refer to the Etopology command for topology file details.) If the -m flag is specified, it will also start the Emonitor daemon to monitor nodes on the switch. Refer to the Emonitor daemon for additional information. If the Estart command is issued when the switch is already running, it causes a switch fault, and messages in flight are lost. Applications using reliable protocols on the switch, such as TCP/IP and the MPI User Space library, recover from switch faults. Applications using unreliable protocols on the switch do not recover from switch faults. For this reason, IBM suggests that you should be aware of what applications or protocols you are running before you issue the Estart command. Since the Estart command uses the SP rsh command, proper authentication and authorization to issue this command is necessary.
SP Switch Notes:
If you have an SP Switch installed on your system, an oncoming primary node as selected via Eprimary is established as primary during Estart. If necessary, the topology file is distributed to partition nodes during Estart. The topology file to be used is distributed to each of the standard nodes in the system partition via the SP Ethernet:
Otherwise, the topology file is already resident on the nodes and does not need to be distributed.
Files
Security
You must have root privilege to run this command.
Related Information
Commands: Eannotator, Eclock, Eduration, Efence, Eprimary, Equiesce, Etopology, Eunfence, Eunpartition
Refer to IBM RS/6000 SP: Planning, Volume 2, Control Workstation and Software Environment for details about system partition topology files.
Examples
Estart
Estart -m
Purpose
Etopology - Stores or reads a switch topology file into or out of the System Data Repository (SDR).
Syntax
Etopology [-h] [-read] switch_topology_file
Flags
Operands
Description
Use this command to store or retrieve the switch_topology_file into or out of the SDR. The switch topology file is used by switch initialization when starting the switch for the current system partition. It is stored in the SDR and can be overridden by having a switch topology file in the /etc/SP directory named expected.top on the switch primary node.
If you have an SP Switch installed on your system, the current topology file is copied to each node of the subject system partition during an Estart and to each targeted node for an Eunfence.
Files
Security
You must have root privilege to run this command.
Related Information
Commands: Eannotator, Eclock, Eduration, Efence, Eprimary, Equiesce, Estart, Eunfence, Eupartition
Refer to the IBM RS/6000 SP: Planning, Volume 2, Control Workstation and Software Environment for information on system partition configurations and topology files.
Examples
Etopology /etc/SP/expected.top.6nsb.4isb.0
Etopology /etc/SP/expected.top.1nsb.0isb.0
Etopology -read /tmp/temporary.top
Purpose
Eunfence - Adds an SP node to the current active switch network that was previously removed from the network.
Syntax
Eunfence [-h | [-G] node_specifier [node_specifier2] ...
Flags
Operands
Description
Use this command to allow a node to rejoin the current switch network that was previously removed with the Efence command.
You can also use this command to allow a node to rejoin the switch network if that node was previously removed from the SP Switch network due to a switch or adapter error.
SP Switch Note:
Eunfence first distributes the current topology file to the nodes before they can be unfenced.
High Performance Switch Note:
The Eunfence command cannot unfence a fenced node if a switch fault occurred or if Estart ran after the node was fenced. You must do another Estart to unfence the node.
Note: | If a host name or IP address is used as the node_specifier for a dependent node, it must be a host name or IP address assigned to the adapter that connects the dependent node to the SP Switch. Neither the administrative host name nor the Simple Network Management Protocol (SNMP) agent's host name for a dependent node is guaranteed to be the same as the host name of its switch network interface. |
Files
Security
You must have root privilege to run this command.
Related Information
Commands: Eannotator, Eclock, Eduration, Efence, Eprimary, Equiesce, Estart, Etopology, Eunpartition
Examples
Eunfence 129.33.34.1
Eunfence r11n01 r11n04
Eunfence 34 43 20 76 40
Eunfence 2,14
Eunfence 5 6 7 8unfences nodes 5 and 6, but not nodes 7 and 8. As a result, the command returns a nonzero return code.
Eunfence -G 5 6 7 8
Purpose
Usage Note |
---|
Use this command only if you have an SP Switch installed on your system. |
Eunpartition - Prepares a system partition for merging with a neighboring system partition.
Syntax
Eunpartition [-h]
Flags
If a flag is not specified, Eunpartition examines the SP_NAME shell variable and selects a system partition based on its current setting.
Operands
None.
Description
Use this command to prepare a partitioned configuration for a new system partition definition within an SP cluster.
This command must be executed for each system partition prior to the spapply_config command to redefine system partitions. Since this command uses the SP rsh command, proper authentication and authorization to issue this command is required.
If you specify Eunpartition in error, it will quiesce the primary and primary backup nodes. If this occurs, you must use Estart to restart the switch.
Security
You must have root privilege to run this command.
Related Information
Commands: Eannotator, Eclock, Eduration, Efence, Eprimary, Equiesce, Estart, Etopology, Eunfence
Examples
To prepare the current system partition for repartitioning as specified by SP_NAME, enter:
Eunpartition
Purpose
export_clients - Creates or updates the Network File System (NFS) export list for a boot/install server.
Syntax
export_clients [-h]
Flags
Operands
None.
Description
Use this command to create or update the NFS export list on a boot/install server node.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/export_clients
Related Information
Commands: setup_server
Examples
To create or update the NFS export list on a boot/install server node, enter:
export_clients
Purpose
ext_srvtab - Extracts service key files from the authentication database.
Syntax
ext_srvtab [-n] [-r realm] [instance ...]
Flags
Operands
Description
The ext_srvtab command extracts service key files from the authentication database. The master key is used to extract service key values from the database. For each instance specified on the command line, the ext_srvtab command creates a new service key file in the current working directory with a file name of instance-new-srvtab which contains all the entries in the database with an instance field of instance. This new file contains all the keys registered for instances of services defined to run on that host. A user must have read access to the authentication database to execute this command. This command can only be issued on the system on which the authentication database resides.
Files
Related Information
Commands: kadmin, ksrvutil
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Examples
If a system has three network interfaces named as follows:
ws3e.abc.org ws3t.abc.org ws3f.finet.abc.orgto re-create the server key file on this workstation (that is an SP authentication server), user root could do the following:
# create a new key file in the /tmp directory for each instance # Combine the instance files into a single file for the hostname. # Delete temporary files and protect key file cd /tmp /usr/kerberos/etc/ext_srvtab -n ws3e ws3t ws3f /bin/cat ws3e-new-srvtab ws3t-new-srvtab ws3f-new-srvtab \ >/etc/krb-srvtab /bin/rm ws3e-new-srvtab ws3t-new-srvtab ws3f-new-srvtab /bin/chmod 400 /etc/krb-srvtab
Purpose
fencevsd - Prevents an application running on a node or group of nodes from accessing an IBM Virtual Shared Disk or group of IBM Virtual Shared Disks.
Syntax
fencevsd -v vsd_name_list -n node_list
Flags
Operands
None.
Description
Under some circumstances, the system may believe a node has failed and begin recovery procedures, when the node is actually operational, but cut off from communication with other nodes running the same application. In this case, the "failed" node must not be allowed to serve requests for the IBM Virtual Shared Disks it normally serves until recovery is complete and the other nodes running the application recognize the failed node as operational. The fencevsd command prevents the failed node from filling requests for its IBM Virtual Shared Disks.
This command can be run from any node.
Note: | This command will fail if you do not specify a current server (primary or backup) to an IBM Virtual Shared Disk with the -v flag. |
Files
Security
You must have root privilege to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: lsfencevsd, lsvsd, unfencevsd, updatevsdtab, vsdchgserver
Refer to IBM Parallel System Support Programs for AIX: Managing Shared Disks for information on how to use this command in writing applications.
Examples
To fence the IBM Virtual Shared Disks vsd1 and vsd2 from node 5, enter:
fencevsd -v vsd1,vsd2 -n 5
Purpose
get_vpd - Consolidates the Vital Product Data (VPD) files for the nodes and writes the information to a file and optionally to a diskette.
Syntax
get_vpd [-h] [-d] -m model_number -s serial_number
Flags
Description
Use this command to consolidate the Vital Product Data (VPD) for the nodes in the RS/6000 SP into a file and to optionally write the file to diskette. The diskette created by this command is sent to IBM manufacturing when an upgrade to the RS/6000 SP hardware is desired. This diskette is used by manufacturing and marketing to configure an upgrade of the RS/6000 SP.
The get_vpd command is issued by IBM field personnel to capture VPD information after an upgrade of the system. All installation and configuration of the RS/6000 SP must be complete prior to issuing the get_vpd command.
Files
Standard Output
This command creates the /var/adm/SPlogs/SPconfig/serial_number.vpd file and optionally writes the file to a diskette.
Standard Error
This command writes all error messages to standard error.
Exit Values
Security
You must have root privilege to run this command.
Restrictions
This command can only be issued on the control workstation.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP) ssp.basic file set.
Prerequisite Information
IBM RS/6000 SP: Planning, Volume 2, Control Workstation and Software Environment
Location
/usr/lpp/ssp/install/bin/get_vpd
Examples
get_vpd -m 204 -s 020077650
get_vpd -m 306 -s 510077730 -d
Purpose
ha_vsd - Starts the rvsd subsystem of IBM Recoverable Virtual Shared Disk (RVSD). This includes configuring IBM Virtual Shared Disks and data striping devices (HSDs) as well as starting the rvsd and hc daemons.
Syntax
ha_vsd [reset]
Flags
None.
Operands
Description
Use this command to start the IBM Recoverable Virtual Shared Disk licensed program after you install it, or, with the reset option, to stop and restart the program.
Exit Values
Security
You must have root privilege to issue the ha_vsd subcommand.
Implementation Specifics
This command is part of the IBM Recoverable Virtual Shared Disk Licensed Program Product (LPP).
Prerequisite Information
See "Using the IBM Recoverable Virtual Shared Disk Software" in IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Location
/usr/lpp/csd/bin/ha_vsd
Related Information
Commands: ha.vsd, hc.vsd
Examples
To stop the rvsd subsystem and restart it, enter:
ha_vsd reset
The system returns the messages:
Starting rvsd subsystem. rvsd subsystem started PID=xxx.
Purpose
ha.vsd - Queries and controls the rvsd subsystem of IBM Recoverable Virtual Shared Disk (RVSD).
Syntax
Flags
None.
Operands
The rvsd subsystem must be restarted for this operand to take effect.
The RVSD subsystem must be restarted for this operand to take effect.
Once debugging is turned on and the RVSD subsystem has been restarted, ha.vsd trace should be issued to turn on tracing.
Use this operand under the direction of your IBM service representative.
Note: the default when the node is booted is to have stdout and stderr routed to the console. If debugging is turned off stdout and stderr will be routed to /dev/null and all further trace messages will be lost. You can determine if debug has been turned on by issuing ha.vsd qsrc. If debug has been turned on the return value will be:
action = "2"
This operand is only meaningful after the debug operand has been used to send stdout and stderr to the console and the rvsd subsystem has been restarted.
Description
Use this command to display information about the rvsd subsystem, to change the number of nodes needed for quorum, and to change the status of the subsystem.
You can start the rvsd subsystem with the VSD Perspective. Type spvsd and select actions for IBM VSD nodes.
Exit Values
Security
You must have root privilege to issue the debug, quorum, refresh, reset, start, stop, trace, mksrc, and rmsrc subcommands.
Implementation Specifics
This command is part of the IBM Recoverable Virtual Shared Disk Licensed Program Product (LPP).
Prerequisite Information
See "Using the IBM Recoverable Virtual Shared Disk Software" in IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Location
/usr/lpp/csd/bin/ha.vsd
Related Information
Commands: ha_vsd, hc.vsd
Examples
ha.vsd reset
The system returns the messages:
Waiting for the rvsd subsystem to exit. rvsd subsystem exited successfully. Starting rvsd subsystem. rvsd subsystem started PID=xxx.
ha.vsd quorum 5
The system returns the message:
Quorum has been changed from 8 to 5.
Purpose
hacws_verify - Verifies the configuration of both the primary and backup High Availability Control Workstation (HACWS) control workstations.
Syntax
hacws_verify
Flags
None.
Operands
None.
Description
Use this command to verify that the primary and backup control workstations are properly configured to provide HACWS services to the SP system. The hacws_verify command inspects both the primary and backup control workstations to verify the following:
Both the primary and backup control workstations must be running and capable of executing remote commands via the /usr/lpp/ssp/rcmd/bin/rsh command.
The system administrator should run the hacws_verify command after HACWS is initially configured. After that, the hacws_verify command can be run at any time.
Exit Values
Prerequisite Information
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for additional information on the HACWS option.
Location
/usr/sbin/hacws/hacws_verify
Related Information
SP Commands: install_hacws, rsh, spcw_addevents
Purpose
haemcfg - Compiles the Event Management objects in the System Data Repository (SDR) and places the compiled information into a binary Event Management Configuration Database (EMCDB) file
Syntax
haemcfg [-c] [-n]
Flags
Operands
None.
Description
The haemcfg utility command builds the Event Management Configuration Database (EMCDB) file for a system partition. If no flags are specified, the haemcfg command:
To place the new EMCDB into production, you must shut down and restart all of this system partition's Event Manager daemons: the daemon on the control workstation and the daemon on each of the system partition's nodes. When the Event Management daemon restarts, it copies the EMCDB from the staging directory to the production directory. The name of the production EMCDB is /etc/ha/cfg/em.syspar_name.cdb.
If you want to test a new EMCDB, IBM recommends that you create a separate system partition for that purpose.
You must create a distinct EMCDB file for each system partition on the IBM RS/6000 SP. To build an EMCDB file, you must be executing on the control workstation and you must set the SP_NAME environment variable to the appropriate system partition name before you issue the command.
Before you build or replace an EMCDB, it is advisable to issue the haemcfg command with the debugging flags.
The -c flag lets you check the validity of the Event Management data that resides in the SDR. This data was previously loaded through the haemloadcfg command. If any of the data is invalid, the command writes an error message that describes the error.
When the -c flag is processed, the command validates the data in the SDR, but does not create a new EMCDB file and does not update the EMCDB version string in the SDR.
The -n flag lets you build a test EMCDB file in the current directory. If anything goes wrong with the creation of the new file, the command writes an error message that describes the error.
When the -n flag is processed, the command uses the data in the SDR to create a test EMCDB file in the current directory, but it does not update the EMCDB version string in the SDR. If any of the data in the SDR is invalid, the command stops at the first error encountered.
If you specify both flags on the command line, the haemcfg command performs the actions of the -c flag.
After you have checked the data and the build process, issue the haemcfg command without any flags. This builds the new EMCDB file, places it in the /spdata/sys1/ha/cfg directory, and updates the EMCDB version string in the SDR.
Files
Standard Output
When the command executes successfully, it writes the following informational messages:
Reading Event Management data for partition syspar_name CDB=new_EMCDB_file_name Version=EMCDB_version_string
Standard Error
This command writes error messages (as necessary) to standard error.
Errors can result from causes that include:
For a listing of the errors that the haemcfg command can produce, see IBM Parallel System Support Programs for AIX: Diagnosis and Messages Guide.
Exit Values
Security
To place an EMCDB file for a system partition into the /spdata/sys1/ha/cfg directory, you must be running with an effective user ID of root on the control workstation. Before running this command, you must set the SP_NAME environment variable to the appropriate system partition name.
Restrictions
To place an EMCDB file for a system partition into the /spdata/sys1/ha/cfg directory, you must be running with an effective user ID of root on the control workstation. Before running this command, you must set the SP_NAME environment variable to the appropriate system partition name.
If you run the haemcfg command without any flags, the command stops at the first error it encounters. With the -c flag on, the command continues, letting you obtain as much debugging information as possible in one pass. To reduce your debugging time, therefore, run the command with the debugging flags first.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
For a general overview of configuring Event Management, see "The Event Management Subsystem" chapter of IBM Parallel System Support Programs for AIX: Administration Guide.
For a description of the SDR classes and attributes that are related to the EMCDB, see IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference.
Location
/usr/lpp/ssp/bin/haemcfg
Related Information
Commands: haemloadcfg
Examples
haemcfg -c
If there are any errors in the data, the command writes appropriate error messages.
To fix the errors, replace the data in the SDR. For more information, see the man page for the haemloadcfg command.
haemcfg -n
If there are any problems in creating the file, the command writes appropriate error messages.
haemcfg
In response, the command creates a new EMCDB file, places it in the staging directory as /spdata/sys1/ha/cfg/em.syspar_name.cdb, where syspar_name is the name of the current system partition, and updates the EMCDB version string in the SDR.
Purpose
haemctrl - A control script that starts the Event Management subsystem.
Syntax
haemctrl {-a | -s | -k | -d | -c | -u | -t | -o | -r | -h}
Flags
Operands
None.
Description
Event Management is a distributed subsystem of PSSP that provides a set of high availability services for the IBM RS/6000 SP. By matching information about the state of system resources with information about resource conditions that are of interest to client programs, it creates events. Client programs can use events to detect and recover from system failures, thus enhancing the availability of the SP system.
The haemctrl control script controls the operation of the Event Management subsystem. The subsystem is under the control of the System Resource Controller (SRC) and belongs to a subsystem group called haem. Associated with each subsystem is a daemon.
An instance of the Event Management subsystem executes on the control workstation and on every node of a system partition. Because Event Management provides its services within the scope of a system partition, its subsystem is said to be system partition-sensitive. This control script operates in a manner similar to the control scripts of other system partition-sensitive subsystems. It can be issued from either the control workstation or any of the system partition's nodes.
From an operational point of view, the Event Management subsystem group is organized as follows:
The haem subsystem is associated with the haemd daemon.
The subsystem name on the nodes is haem. There is one of each subsystem per node and it is associated with the system partition to which the node belongs.
On the control workstation, there are multiple instances of each subsystem, one for each system partition. Accordingly, the subsystem names on the control workstation have the system partition name appended to them. For example, for system partitions named sp_prod and sp_test, the subsystems on the control workstation are named haem.sp_prod and haem.sp_test.
The haemd daemon provides the Event Management services.
The haemctrl script is not normally executed from the command line. It is normally called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system.
The haemctrl script provides a variety of controls for operating the Event Management subsystem:
Before performing any of these functions, the script obtains the current system partition name and IP address (using the spget_syspar command) and the node number (using the node_number) command. If the node number is zero, the control script is running on the control workstation.
Except for the clean and unconfigure functions, all functions are performed within the scope of the current system partition.
Adding the Subsystem
When the -a flag is specified, the control script uses the mkssys command to add the Event Management subsystem to the SRC. The control script operates as follows:
The service name that is entered in the /etc/services file is haem.syspar_name.
For more information about configuring Event Management data, see the IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference.
Then it gets the port number for the subsystem from the SP_ports class of the System Data Repository (SDR) and ensures that the port number is set in the /etc/services file. This port number is used for remote connections to Event Management daemons that are running on the control workstation. If there is no port number in the SDR, the script obtains one and sets it in the /etc/services file. The range of valid port numbers is 10000 to 10100, inclusive.
The service name is haemd.
Starting the Subsystem
When the -s flag is specified, the control script uses the startsrc command to start the Event Management subsystem, haem.
Stopping the Subsystem
When the -k flag is specified, the control script uses the stopsrc command to stop the Event Management subsystem, haem.
Deleting the Subsystem
When the -d flag is specified, the control script uses the rmssys command to remove the Event Management subsystem from the SRC. The control script operates as follows:
Cleaning Up the Subsystems
When the -c flag is specified, the control script stops and removes the Event Management subsystems for all system partitions from the SRC. The control script operates as follows:
Unconfiguring the Subsystems
When the -u flag is specified, the control script performs the function of the -c flag in all system partitions and then removes all port numbers from the SDR allocated by the Event Management subsystems.
Note: | The -u flag is effective only on the control workstation. |
Prior to executing the haemctrl command with the -u flag on the control workstation, the haemctrl command with the -c flag must be executed from all of the nodes. If this subsystem is not successfully cleaned from all of the nodes, different port numbers may be used by this subsystem, leading to undefined behavior.
Turning Tracing On
When the -t flag is specified, the control script turns tracing on for the haemd daemon, using the haemtrcon command.
Turning Tracing Off
When the -o flag is specified, the control script turns tracing off for the haemd daemon, using the haemtrcoff command.
Refreshing the Subsystem
The -r flag has no effect for this subsystem.
Logging
While it is running, the Event Management daemon normally provides information about its operation and errors by writing entries to the AIX error log. If it cannot, errors are written to a log file called /var/ha/log/em.default.syspar_name.
Files
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must be running with an effective user ID of root.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
"The Event Management Subsystem" chapter of IBM Parallel System Support Programs for AIX: Administration Guide
IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/haemctrl
Related Information
Commands: haemcfg, haemd, haemloadcfg, haemtrcoff, haemtrcon, lssrc, startsrc, stopsrc, syspar_ctrl
Examples
haemctrl -a
haemctrl -s
haemctrl -k
haemctrl -d
haemctrl -c
haemctrl -u
haemctrl -t
haemctrl -o
lssrc -g haem
lssrc -s haem
To display the status of an individual Event Management subsystem on the control workstation, enter:
lssrc -s haem. syspar_name
where syspar_name is the system partition name.
lssrc -l -s haem
To display detailed status about an individual Event Management subsystem on the control workstation, enter:
lssrc -l -s haem.syspar_name
where syspar_name is the system partition name.
In response, the system returns information that includes the running status of the subsystem, the settings of trace flags, the version number of the Event Management Configuration Database, the time the subsystem was started, the connection status to Group Services and peer Event Management subsystem, and the connection status to Event Management clients, if any.
lssrc -a
Purpose
haemd - The Event Manager daemon, which observes resource variable instances that are updated by Resource Monitors and generates and reports events to client programs
Syntax
haemd [syspar_IPaddr]
Flags
None.
Operands
Description
The haemd daemon is the Event Manager daemon. The daemon observes resource variable instances that are updated by Resource Monitors and generates and reports events to client programs.
One instance of the haemd daemon executes on the control workstation for each system partition. An instance of the haemd daemon also executes on every node of a system partition. The haemd daemon is under System Resource Controller (SRC) control.
Because the daemon is under SRC control, it cannot be started directly from the command line. It is normally started by the haemctrl command, which is in turn called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system. If you must start or stop the daemon directly, use the haemctrl command.
For more information about the Event Manager daemon, see the haemctrl man page.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
"The Event Management Subsystem" chapter of IBM Parallel System Support Programs for AIX: Administration Guide
IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/haemd
Related Information
Commands: haemctrl
Examples
See the haemctrl command.
Purpose
haemloadcfg - Loads Event Management configuration data into the System Data Repository (SDR)
Syntax
haemloadcfg [-d] [-r] loadlist_file
Flags
Operands
Description
The haemloadcfg utility command loads Event Management configuration data into the SDR. Note that before you invoke haemloadcfg, you must ensure that the SP_NAME environment variable is set to the appropriate system partition name.
The configuration data is contained in a load list file, whose format is described by the man page for the haemloadlist file. For details on the SDR classes and attributes that you can use to specify Event Management configuration data, see IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference.
To load the default Event Management configuration data for PSSP, specify the load list file as /usr/lpp/ssp/install/config/haemloadlist.
To add Event Management configuration data for other Resource Monitors, create a file in load list format and specify its name on the command.
Without any flags, the haemloadcfg command does not replace existing objects in the SDR. The data in the load list file is matched with the existing objects in the SDR based on key attributes, as follows:
Note that the way in which the haemloadcfg command handles existing SDR objects is different from the way in which the SDRCreateObjects command handles them. The SDRCreateObjects command creates a new object as long as the attributes, taken as a group, are unique.
To change a nonkey attribute of an Event Management object that already exists in the SDR, change the attribute in the load list file. Then run the haemloadcfg command using the -r flag and the name of the load list file. All objects in the SDR are replaced by matching objects in the load list file using the key attributes to match. Any unmatched objects in the load list file are added to the SDR.
To delete Event Management objects from the SDR, create a load list file with the objects to be deleted. Only the key attributes need to be specified. Then run the haemloadcfg command using the -d flag and the name of the load list file. All objects in the SDR that match objects in the load list file are deleted. No unmatched objects, if any in the load list file, are added to the SDR.
Under any circumstances, duplicate objects in the load list file, based on matches in key attributes, are ignored. However, such duplicate objects are written to standard output.
Files
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have the appropriate authority to write to the SDR. You should be running on the control workstation. Before running this command, you must set the SP_NAME environment variable to the appropriate system partition name.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
For a general overview of configuring Event Management, see "The Event Management Subsystem" chapter of IBM Parallel System Support Programs for AIX: Administration Guide.
For details on the System Data Repository classes and attributes for Event Management configuration Database, see IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference.
Location
/usr/lpp/ssp/bin/haemloadcfg
Related Information
Commands: haemcfg, SDRCreateObjects, SDRDeleteObjects
Files: haemloadlist
Also, for a description of the SDR classes for Event Management configuration data, see IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference.
Examples
haemloadcfg /usr/lpp/ssp/install/config/haemloadlist
haemloadcfg /usr/local/config/newrmloadlist
If nonkey attributes in this load list file are later changed, update the SDR by entering:
haemloadcfg -r /usr/local/config/newrmloadlist
If this new Resource Monitor is no longer needed, its configuration data is removed from the SDR by entering:
haemloadcfg -d /usr/local/config/newrmloadlist
Purpose
haemtrcoff - Turns tracing off for the Event Manager daemon.
Syntax
haemtrcoff -s subsys_name -a trace_list
Flags
Operands
The following trace arguments may be specified:
Description
The haemtrcoff command is used to turn tracing off for specified activities of the Event Manager daemon. Trace output is placed in an Event Management trace log for the system partition.
Use this command only under the direction of the IBM Support Center. It provides information for debugging purposes and may degrade the performance of the Event Management subsystem or anything else that is running in the system partition. Do not use this command during normal operation.
Files
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
"The Event Management Subsystem" chapter of IBM Parallel System Support Programs for AIX: Administration Guide
Location
/usr/lpp/ssp/bin/haemtrcoff
Related Information
Commands: haemctrl, haemd, haemtrcon
Examples
In the following examples, the SP system has two system partitions named sp_prod and sp_test. The instances of the Event Management subsystem on the control workstation of the SP are named haem.sp_prod and haem.sp_test, respectively. The instance of the Event Management subsystem that runs on any node of either system partition is named haem.
haemtrcoff -s haem.sp_prod -a all
haemtrcoff -s haem -a all
haemtrcoff -s haem.sp_test -a init,config
Purpose
haemtrcon - Turns tracing on for the Event Manager daemon.
Syntax
haemtrcon -s subsys_name -a trace_list
Flags
Operands
The following trace arguments may be specified:
Description
The haemtrcon command is used to turn tracing on for specified activities of the Event Manager daemon. Trace output is placed in an Event Management trace log for the system partition. When used, the regs and dinsts arguments perform a one-time trace. The specified information is placed in the trace log, but no further tracing is done.
Use this command only under the direction of the IBM Support Center. It provides information for debugging purposes and may degrade the performance of the Event Management subsystem or anything else that is running in the system partition. Do not use this command to turn tracing on during normal operation.
Files
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
"The Event Management Subsystem" chapter of IBM Parallel System Support Programs for AIX: Administration Guide
Location
/usr/lpp/ssp/bin/haemtrcon
Related Information
Commands: haemctrl, haemd, haemtrcoff
Examples
In the following examples, the SP system has two system partitions named sp_prod and sp_test. The instances of the Event Management subsystem on the control workstation of the SP are named haem.sp_prod and haem.sp_test, respectively. The instance of the Event Management subsystem that runs on any node of either system partition is named haem.
haemtrcon -s haem.sp_prod -a all
haemtrcon -s haem -a all
haemtrcon -s haem.sp_test -a init,config
Purpose
haemunlkrm - Unlocks and starts a Resource Monitor.
Syntax
haemunlkrm -s subsys_name -a resmon_name
Flags
Description
If the Event Manager daemon cannot successfully start a Resource Monitor after three attempts within a two hour interval, the Resource Monitor is "locked" and no further attempts are made to start it. Once the cause of the failure is determined and the problem corrected, the haemunlkrm command can be used to unlock the Resource Monitor and attempt to start it.
The status of the Event Manager daemon, as displayed by the lssrc command, indicates if a Resource Monitor is locked.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
"The Event Management Subsystem" chapter of IBM Parallel System Support Programs for AIX: Administration Guide
Location
/usr/lpp/ssp/bin/haemunlkrm
Examples
If the output of the lssrc command indicates that the hardware Resource Monitor IBM.PSSP.hmrmd is locked, then after correcting the condition that prevented the Resource Monitor from being started, enter:
haemunlkrm -s haem -a IBM.PSSP.hmrmd
Note: | This example applies to unlocking a Resource Monitor on a node. |
Purpose
hagsctrl - A control script that starts the Group Services subsystems.
Syntax
hagsctrl {-a | -s | -k | -d | -c | -u | -t | -o | -r | -h}
Flags
Operands
None.
Description
Group Services provides distributed coordination and synchronization services for other distributed subsystems running on a set of nodes on the IBM RS/6000 SP. The hagsctrl control script controls the operation of the subsystems that are required for Group Services. These subsystems are under the control of the System Resource Controller (SRC) and belong to a subsystem group called hags. Associated with each subsystem is a daemon.
An instance of the Group Services subsystem executes on the control workstation and on every node of a system partition. Because Group Services provides its services within the scope of a system partition, its subsystems are said to be system partition-sensitive. This control script operates in a manner similar to the control scripts of other system partition-sensitive subsystems. It can be issued from either the control workstation or any of the system partition's nodes.
From an operational point of view, the Group Services subsystem group is organized as follows:
The hags subsystem is associated with the hagsd daemon. The hagsglsm subsystem is associated with the hagsglsmd daemon.
The subsystem names on the nodes are hags and hagsglsm. There is one of each subsystem per node and it is associated with the system partition to which the node belongs.
On the control workstation, there are multiple instances of each subsystem, one for each system partition. Accordingly, the subsystem names on the control workstation have the system partition name appended to them. For example, for system partitions named sp_prod and sp_test, the subsystems on the control workstation are named hags.sp_prod, hags.sp_test, hagsglsm.sp_prod, and hagsglsm.sp_test.
The hagsd daemon provides the majority of the Group Services functions.
The hagsglsmd daemon provides global synchronization services for the switch adapter membership group.
The hagsctrl script is not normally executed from the command line. It is normally called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system.
The hagsctrl script provides a variety of controls for operating the Group Services subsystems:
Before performing any of these functions, the script obtains the current system partition name (using the spget_syspar command) and the node number (using the node_number) command. If the node number is zero, the control script is running on the control workstation.
Except for the clean and unconfigure functions, all functions are performed within the scope of the current system partition.
Adding the Subsystem
When the -a flag is specified, the control script uses the mkssys command to add the Group Services subsystems to the SRC. The control script operates as follows:
The service name that is entered in the /etc/services file is hags.syspar_name.
Starting the Subsystem
When the -s flag is specified, the control script uses the startsrc command to start the Group Services subsystems, hags and hagsglsm.
Stopping the Subsystem
When the -k flag is specified, the control script uses the stopsrc command to stop the Group Services subsystems, hags and hagsglsm.
Deleting the Subsystem
When the -d flag is specified, the control script uses the rmssys command to remove the Group Services subsystems from the SRC. The control script operates as follows:
Cleaning Up the Subsystems
When the -c flag is specified, the control script stops and removes the Group Services subsystems for all system partitions from the SRC. The control script operates as follows:
Unconfiguring the Subsystems
When the -u flag is specified, the control script performs the function of the -c flag in all system partitions and then removes all port numbers from the SDR allocated by the Group Services subsystems.
Note: | The -u flag is effective only on the control workstation. |
Prior to executing the hagsctrl command with the -u flag on the control workstation, the hagsctrl command with the -c flag must be executed from all of the nodes. If this subsystem is not successfully cleaned from all of the nodes, different port numbers may be used by this subsystem, leading to undefined behavior.
Turning Tracing On
When the -t flag is specified, the control script turns tracing on for the hagsd daemon, using the traceson command. Tracing is not available for the hagsglsmd daemon.
Turning Tracing Off
When the -o flag is specified, the control script turns tracing off (returns it to its default level) for the hagsd daemon, using the tracesoff command. Tracing is not available for the hagsglsmd daemon.
Refreshing the Subsystem
The -r flag has no effect for this subsystem.
Logging
While they are running, the Group Services daemons provide information about their operation and errors by writing entries in a log file in the /var/ha/log directory.
Each daemon limits the log size to a pre-established number of lines (by default, 5,000 lines). When the limit is reached, the daemon appends the string .bak to the name of the current log file and begins a new log. If a .bak version already exists, it is removed before the current log is renamed.
Files
The file names include the following variables:
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must be running with an effective user ID of root.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
"The Group Services Subsystem" chapter of IBM Parallel System Support Programs for AIX: Administration Guide
IBM Parallel System Support Programs for AIX: Group Services Programming Guide and Reference
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/hagsctrl
Related Information
Commands: hagsd, hagsglsmd, lssrc, startsrc, stopsrc, syspar_ctrl
Examples
hagsctrl -a
hagsctrl -s
hagsctrl -k
hagsctrl -d
hagsctrl -c
hagsctrl -u
hagsctrl -t
hagsctrl -o
lssrc -g hags
lssrc -s subsystem_name
lssrc -l -s subsystem_name
In response, the system returns information that includes the running status of the subsystem, the number and identity of connected GS clients, information about the Group Services domain, and the number of providers and subscribers in established groups.
lssrc -a
Purpose
hagsd - A Group Services daemon that provides a general purpose facility for coordinating and monitoring changes to the state of an application that is running on a set of nodes.
Syntax
hagsd daemon_name
Flags
None.
Operands
Description
The hagsd daemon is part of the Group Services subsystem, which provides a general purpose facility for coordinating and monitoring changes to the state of an application that is running on a set of nodes. This daemon provides most of the services of the subsystem.
One instance of the hagsd daemon executes on the control workstation for each system partition. An instance of the hagsd daemon also executes on every node of a system partition. The hagsd daemon is under System Resource Controller (SRC) control.
Because the daemon is under SRC control, it is better not to start it directly from the command line. It is normally called by the hagsctrl command, which is in turn called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system. If you must start or stop the daemon directly, use the startsrc or stopsrc command.
For more information about the Group Services daemons, see the hagsctrl man page.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
"The Group Services Subsystem" chapter of IBM Parallel System Support Programs for AIX: Administration Guide
IBM Parallel System Support Programs for AIX: Group Services Programming Guide and Reference
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/hagsd
Related Information
Commands: hagsctrl, hagsglsmd
Examples
See the hagsctrl command.
Purpose
hagsglsmd - A Group Services daemon that provides global synchronization services for the switch adapter membership group.
Syntax
hagsglsmd daemon_name
Flags
None.
Operands
Description
The hagsglsmd daemon is part of the Group Services subsystem, which provides a general purpose facility for coordinating and monitoring changes to the state of an application that is running on a set of nodes. This daemon provides global synchronization services for the High Performance Switch adapter membership group.
One instance of the hagsglsmd daemon executes on the control workstation for each system partition. An instance of the hagsglsmd daemon also executes on every node of a system partition. The hagsglsmd daemon is under System Resource Controller (SRC) control.
Because the daemon is under SRC control, it is better not to start it directly from the command line. It is normally called by the hagsctrl command, which is in turn called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system. If you must start or stop the daemon directly, use the startsrc or stopsrc command.
For more information about the Group Services daemons, see the hagsctrl man page.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
"The Group Services Subsystem" chapter of IBM Parallel System Support Programs for AIX: Group Services Programming Guide and Reference
IBM Parallel System Support Programs for AIX: Group Services Programming Guide and Reference
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/hagsglsmd
Related Information
Commands: hagsctrl, hagsd
Examples
See the hagsctrl command.
Purpose
hardmon - Monitors and controls the state of the SP hardware.
Syntax
hardmon [-B] [-r poll_rate] [-d debug_flag] ...
Flags
Operands
None.
Description
hardmon is the Hardware Monitor daemon. The daemon monitors and controls the state of the SP hardware contained in one or more SP frames. This command is not normally executed from the command line. Access to the Hardware Monitor is provided by the hmmon, hmcmds, spmon, s1term, and nodecond commands. Control of the Hardware Monitor daemon is provided by the hmadm command. These commands are the Hardware Monitor "client" commands.
The Hardware Monitor daemon executes on the Monitor and Control Node (MACN). The MACN is that IBM RS/6000 workstation to which the RS-232 lines are connected to the frames. The MACN is one and the same as the control workstation. The daemon is managed by the System Resource Controller (SRC). When the MACN is booted, an entry in /etc/inittab invokes the startsrc command to start the daemon. The daemon is configured in the SRC to be restarted automatically if it terminates for any reason other than the stopsrc command. The SRC subsystem name for the Hardware Monitor daemon is hardmon.
hardmon obtains configuration information from the System Data Repository (SDR). The SP_ports object class specifies the port number that the daemon is to use to accept TCP/IP connections from the client commands. The port number is obtained from the object whose daemon attribute value matches hardmon and whose host_name attribute value matches the host name of the workstation on which the daemon is executing. There must be one hardmon object in SP_ports for the MACN. The Frame object class contains an object for each frame in the SP system.
The attributes of interest to the daemon are frame_number, tty, and MACN. When started, the daemon fetches all those objects in the Frame class whose MACN attribute value matches the host name of the workstation on which the daemon is executing. For each frame discovered in this manner, the daemon saves the frame number and opens the corresponding tty device. When all frames have been configured, the daemon begins to poll the frames for state information. Current state and changed state can then be obtained using the hmmon and spmon commands. The hmcmds and spmon commands can be used to control the hardware within the frames.
The daemon also reads the file /spdata/sys1/spmon/hmthresholds for values used to check boundary conditions for certain state variables. This file should only be changed on request from IBM support. Finally, the /spdata/sys1/spmon/hmacls file is read for Access Control List (ACL) information. Refer to the hmadm command and the /spdata/sys1/spmon/hmacls file for more information on ACLs.
All errors detected by the Hardware Monitor daemon are written to the AIX error log.
The flags in the SRC subsystem object for the hardmon subsystem should not normally be changed. For example, if the poll rate is more than 5 seconds, the nodecond command can fail with unpredictable results. Upon request from IBM support for more information to aid in problem determination, debug flags can be set using the hmadm command.
If the High Availability Control Workstation (HACWS) Frame Supervisor (type 20) or the SEPBU HACWS Frame Supervisor (type 22) is installed in the SP frames, the -B flag is used to run the Hardware Monitor daemon in diagnostic mode. This diagnostic mode is used to validate that the frame ID written into the Supervisor matches the frame ID configured in the SDR for that frame. Normally, the frame ID is automatically written into the Supervisor during system installation. The frame ID is written into the frame to detect cabling problems in an HACWS configuration. In a non-HACWS SP configuration, the -B flag is useful whenever the RS232 cables between the frames and MACN are changed (but only if one or more frames contain a type 20 or type 22 supervisor). The hardmon command can be executed directly from the command line with the -B flag, but only after the currently running daemon is stopped using the stopsrc command. Diagnostic messages are written to the AIX error log. The daemon exits when all frames are validated.
Frame ID validation is also performed every time the daemon is started by the System Resource Controller. Any frame that has a frame ID mismatch can be monitored, but any control commands to the frame are ignored until the condition is corrected. A frame with a mismatch is noted in the System Monitor Graphical User Interface as well as in the AIX error log. The hmcmds command can be used to set the currently configured frame ID into a type 20 or type 22 supervisor after it is verified that the frame is correctly connected to the MACN.
Additional Configuration Information: The Hardware Monitor subsystem also obtains information from the system partition and the Syspar_map object classes in the SDR. While this information is not used by the hardmon daemon itself, it is used by the hardmon client commands listed under Related Information. Each of these commands executes in the environment of one system partition. If the SP system is not partitioned, these commands execute in the environment of the entire system. In any case, the Syspar_map object class is used to determine which nodes are contained in the current environment. The attributes of interest are syspar_name and node_number.
Starting and Stopping the hardmon Daemon
The hardmon daemon is under System Resource Controller (SRC) control. It uses the signal method of communication in SRC. The hardmon daemon is a single subsystem and not associated with any SRC group. The subsystem name is hardmon. In order to start the hardmon daemon, use the startsrc -s hardmon command. This starts the daemon with the default arguments and SRC options. The hardmon daemon is setup to be respawnable and be the only instance of the hardmon daemon running on a particular node or control workstation. Do not start the hardmon daemon from the command line without using the startsrc command to start it.
To stop the hardmon daemon, use the stopsrc -s hardmon command. This stops the daemon and does not allow it to respawn.
To display the status of the hardmon daemon, use the lssrc -s hardmon command.
If the default startup arguments need to be changed, use the chssys command to change the startup arguments or the SRC options. Refer to AIX Version 4 Commands Reference and AIX Version 4 General Programming Concepts: Writing and Debugging Programs for more information about daemons under SRC control and how to modify daemon arguments when under SRC.
To view the current SRC options and daemon arguments, use the odmget -q 'subsysname=hardmon' SRCsubsys command.
Files
Related Information
Commands: hmadm, hmcmds, hmmon, nodecond, spmon, s1term
File: /spdata/sys1/spmon/hmacls
Examples
startsrc -s hardmon
stopsrc -s hardmon
lssrc -s hardmon
lssrc -a
odmget -q 'subsysname=hardmon' SRCsubsys
Purpose
hats - Starts or restarts Topology Services on a node or on the control workstation.
Syntax
hats
Flags
None.
Operands
None.
Description
Use this command to start the operation of Topology Services for a system partition (the hatsd daemon) on the control workstation or on a node within a system partition.
The hats script is not normally executed from the command line. It is normally called by the hatsctrl command, which is in turn called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system.
The Topology Services subsystem provides internal services to PSSP components.
Note that the hats script issues the no -o nonlocsrcroute=1 command, which enables IP source routing. Do not change this setting, because the Topology Services subsystem requires this setting to work properly. If you change the setting, the Topology Services subsystem and a number of other subsystems that depend on it will no longer operate properly.
The hatsd daemon is initially started on the control workstation with the System Resource Controller (SRC), regardless of the level of the system partition. It is respawned automatically if the hatsd daemon fails. The SP_NAME environment variable causes selection of the correct topology configuration.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
The "Starting Up and Shutting Down the SP System" chapter and "The System Data Repository" appendix in IBM Parallel System Support Programs for AIX: Administration Guide
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/hats
Related Information
Commands: hatsctrl, lssrc, startsrc, stopsrc, syspar_ctrl
Examples
See the hatsctrl command.
Purpose
hatsctrl - A control script that starts the Topology Services subsystem.
Syntax
hatsctrl {-a | -s | -k | -d | -c | -u | -t | -o | -r | -h}
Flags
Operands
None.
Description
Topology Services is a distributed subsystem of PSSP that provides information to other PSSP subsystems about the state of the nodes and adapters on the IBM RS/6000 SP.
The hatsctrl control script controls the operation of the Topology Services subsystem. The subsystem is under the control of the System Resource Controller (SRC) and belongs to a subsystem group called hats. Associated with each subsystem is a daemon and a script that configures and starts the daemon.
An instance of the Topology Services subsystem executes on the control workstation and on every node of a system partition. Because Topology Services provides its services within the scope of a system partition, its subsystem is said to be system partition-sensitive. This control script operates in a manner similar to the control scripts of other system partition-sensitive subsystems. It can be issued from either the control workstation or any of the system partition's nodes.
From an operational point of view, the Topology Services subsystem group is organized as follows:
The hats subsystem is associated with the hatsd daemon and the hats script. The hats script configures and starts the hatsd daemon.
The subsystem name on the nodes is hats. There is one of each subsystem per node and it is associated with the system partition to which the node belongs.
On the control workstation, there are multiple instances of each subsystem, one for each system partition. Accordingly, the subsystem names on the control workstation have the system partition name appended to them. For example, for system partitions named sp_prod and sp_test, the subsystems on the control workstation are named hats.sp_prod and hats.sp_test.
The hatsd daemon provides the Topology Services. The hats script configures and starts the hatsd daemon.
The hatsctrl script is not normally executed from the command line. It is normally called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system.
The hatsctrl script provides a variety of controls for operating the Topology Services subsystem:
Before performing any of these functions, the script obtains the current system partition name and IP address (using the spget_syspar command) and the node number (using the node_number) command. If the node number is zero, the control script is running on the control workstation.
Except for the clean and unconfigure functions, all functions are performed within the scope of the current system partition.
Adding the Subsystem
When the -a flag is specified, the control script uses the mkssys command to add the Topology Services subsystem to the SRC. The control script operates as follows:
The service name that is entered in the /etc/services file is hats.syspar_name.
Starting the Subsystem
When the -s flag is specified, the control script uses the startsrc command to start the Topology Services subsystem, hats.
Stopping the Subsystem
When the -k flag is specified, the control script uses the stopsrc command to stop the Topology Services subsystem, hats.
Deleting the Subsystem
When the -d flag is specified, the control script uses the rmssys command to remove the Topology Services subsystem from the SRC. The control script operates as follows:
Cleaning Up the Subsystems
When the -c flag is specified, the control script stops and removes the Topology Services subsystems for all system partitions from the SRC. The control script operates as follows:
Unconfiguring the Subsystems
When the -u flag is specified, the control script performs the function of the -c flag in all system partitions and then removes all port numbers from the SDR allocated by the Topology Services subsystems.
Note: | The -u flag is effective only on the control workstation. |
Prior to executing the hatsctrl command with the -u flag on the control workstation, the hatsctrl command with the -c flag must be executed from all of the nodes. If this subsystem is not successfully cleaned from all of the nodes, different port numbers may be used by this subsystem, leading to undefined behavior.
Turning Tracing On
When the -t flag is specified, the control script turns tracing on for the hatsd daemon, using the traceson command.
Turning Tracing Off
When the -o flag is specified, the control script turns tracing off (returns it to its default level) for the hatsd daemon, using the tracesoff command.
Refreshing the Subsystem
When the -r flag is specified, the control script refreshes the subsystem, using the hats refresh command and the refresh command.
It rebuilds the information about the node and adapter configuration in the SDR and signals the daemon to read the rebuilt information.
Logging
While it is running, the Topology Services daemon provides information about its operation and errors by writing entries in a log file. The hatsd daemon in the system partition named syspar_name uses a log file called /var/ha/log/hats.syspar_name.
Files
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must be running with an effective user ID of root.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/hatsctrl
Related Information
Commands: hats, lssrc, startsrc, stopsrc, syspar_ctrl
Examples
hatsctrl -a
hatsctrl -s
hatsctrl -k
hatsctrl -d
hatsctrl -c
hatsctrl -u
hatsctrl -t
hatsctrl -o
lssrc -g hats
lssrc -s subsystem_name
lssrc -l -s subsystem_name
In response, the system returns information that includes the running status of the subsystem, the number of defined and active nodes, the required number of active nodes for a quorum, the status of the group of nodes, and the IP addresses of the source node, the group leader, and the control workstation.
lssrc -a
Purpose
hb - Starts or restarts heartbeat services on a node or on the control workstation.
Syntax
hb [-spname syspar_name] [-splevel pssp_level]
{ [start | resume] | [stop | quiesce] | reset |
[query | qall | qsrc] | refresh | mksrc optional_flags | rmsrc restore |
[debug | debug off] | [trace on | trace off] }
Flags
Operands
Description
Use this command to control the operation of heartbeat services for a system partition (the hbd daemon) on the control workstation or on a node within a system partition.
The hb script is not normally executed from the command line. It is normally called by the hbctrl command, which is in turn called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system.
The heartbeat server provides input to the host_responds function within a system partition for the System Monitor through the System Monitor hr daemons. It also provides input to the IBM Recoverable Virtual Shared Disk daemons, if that product is installed on the nodes. This involves the following daemons:
Note: | The hrd daemon is controlled by the hr script. The hbd daemon is controlled with this script. |
The hbd daemon is initially started on the control workstation with the System Resource Controller (SRC), regardless of the level of the system partition. It is respawned automatically if the hbd daemon fails. The SP_NAME environment variable causes selection of the correct heartbeat daemon.
The hbd daemons communicate with their counterparts on other nodes over the SP Ethernet. The udp heartbeat entry in /etc/services on all nodes must specify the same port number.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
The "Starting Up and Shutting Down the SP System" chapter and "The System Data Repository" appendix in IBM Parallel System Support Programs for AIX: Administration Guide
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/hb
Related Information
Commands: hbctrl, lssrc, startsrc, stopsrc, syspar_ctrl
Examples
See the hbctrl command.
Purpose
hbctrl - A control script that starts the Heartbeat subsystem.
Syntax
hbctrl { -a | -s | -k | -d | -c | -t | -o | -r | -h }
Flags
Operands
None.
Description
The Heartbeat subsystem communicates with several PSSP subsystems as part of providing information about the state of the nodes and adapters on the IBM RS/6000 SP.
The hbctrl control script controls the operation of the Heartbeat subsystem. The subsystem is under the control of the System Resource Controller (SRC) and belongs to a subsystem group called hb. Associated with each subsystem is a daemon and a script that configures and starts the daemon.
An instance of the Heartbeat subsystem executes on the control workstation and on every node of a system partition. Because Heartbeat provides its services within the scope of a system partition, its subsystem is said to be system partition-sensitive. This control script operates in a manner similar to the control scripts of other system partition-sensitive subsystems. It can be issued from either the control workstation or any of the system partition's nodes.
From an operational point of view, the Heartbeat subsystem group is organized as follows:
The hb subsystem is associated with the hbd daemon and the hb script. The hb script configures and starts the hbd daemon.
The subsystem name on the nodes is hb. There is one of each subsystem per node and it is associated with the system partition to which the node belongs.
On the control workstation, there are multiple instances of each subsystem, one for each system partition. Accordingly, the subsystem names on the control workstation have the system partition name appended to them. For example, for system partitions named sp_prod and sp_test, the subsystems on the control workstation are named hb.sp_prod and hb.sp_test.
The hbd daemon provides the Heartbeat services. The hb script configures and starts the hbd daemon.
The hbctrl script is not normally executed from the command line. It is normally called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system.
The hbctrl script provides a variety of controls for operating the Heartbeat subsystem:
Before performing any of these functions, the script obtains the current system partition name and IP address (using the spget_syspar command) and the node number (using the node_number) command. If the node number is zero, the control script is running on the control workstation.
Except for the clean function, all functions are performed within the scope of the current system partition.
Adding the Subsystem
When the -a flag is specified, the control script uses the mkssys command to add the Heartbeat subsystem to the SRC. The control script operates as follows:
The service name that is entered in the /etc/services file is heartbeat.
Starting the Subsystem
When the -s flag is specified, the control script uses the hb command to start the Heartbeat subsystem, hb.
Stopping the Subsystem
When the -k flag is specified, the control script uses the hb command to stop the Heartbeat subsystem, hb.
Deleting the Subsystem
When the -d flag is specified, the control script uses the rmssys command to remove the Heartbeat subsystem from the SRC. The control script operates as follows:
Cleaning Up the Subsystems
When the -c flag is specified, the control script stops and removes the Heartbeat subsystems for all system partitions from the SRC. The control script operates as follows:
Turning Tracing On
When the -t flag is specified, the control script turns tracing on for the hbd daemon, using the traceson command.
Turning Tracing Off
When the -o flag is specified, the control script turns tracing off (returns it to its default level) for the hbd daemon, using the tracesoff command.
Refreshing the Subsystem
When the -r flag is specified, the control script refreshes the subsystem, using the hb refresh command.
Logging
While it is running, the Heartbeat daemon provides information about its operation and errors by writing entries in a log file. The hbd daemon in the system partition named syspar_name uses a log file called /var/ha/log/hb.syspar_name.
Files
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must be running with an effective user ID of root.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/hbctrl
Related Information
Commands: hb, lssrc, startsrc, stopsrc, syspar_ctrl
Examples
hbctrl -a
hbctrl -s
hbctrl -k
hbctrl -d
hbctrl -c
hbctrl -t
hbctrl -o
lssrc -g hb
lssrc -s subsystem_name
lssrc -l -s subsystem_name
In response, the system returns information that includes the running status of the subsystem, the number of defined and active nodes, the required number of active nodes for a quorum, the status of the group of nodes, the frequency and sensitivity values in use for the subsystem, and the IP addresses of the source node, the group leader, and the control workstation.
lssrc -a
Purpose
hc.vsd - Queries and controls the hc subsystem of IBM Recoverable Virtual Shared Disk.
Syntax
Flags
None.
Operands
The hc subsystem must be restarted for this operand to take effect.
Once debugging is turned on and the hc subsystem has been restarted, hc.vsd trace should be issued to turn on tracing.
Use this operand under the direction of your IBM service representative.
Note: the default when the node is booted is to have stdout and stderr routed to the console. If debugging is turned off stdout and stderr will be routed to /dev/null and all further trace messages will be lost. You can determine if debug has been turned on by issuing hc.vsd qsrc. If debug has been turned on the return value will be:
action = "2"
This operand is only meaningful after the debug operand has been used to send stdout and stderr to the console and the hc subsystem has been restarted.
Description
Use this command to display information about the hc subsystem and to change the status of the subsystem.
You can restart the hc subsystem with the VSD Perspective. Type spvsd and select actions for IBM VSD nodes.
Exit Values
Note: | The query and qsrc subcommands have no exit values. |
Security
You must have root privilege to issue the debug, mksrc, reset, start, and stop commands.
Implementation Specifics
This command is part of the IBM Recoverable Virtual Shared Disk Licensed Program Product (LPP).
Prerequisite Information
See "Using the IBM Recoverable Virtual Shared Disk Software" in IBM Parallel System Support Programs for AIX: Managing Shared Disks
Location
/usr/lpp/csd/bin/hc.vsd
Related Information
Commands: ha_vsd, ha.vsd
Examples
To stop the hc subsystem and restart it, enter:
hc.vsd reset
The system returns the messages:
Waiting for the hc subsystem to exit. hc subsystem exited successfully. Starting hc subsystem. hc subsystem started PID=xxx.
Purpose
hmadm - Administers the Hardware Monitor daemon.
Syntax
hmadm [ {-d debug_flag} ... ] operation
Flags
Operands
The operation must be one of the following:
This operation must by invoked by the administrator after the administrator modifies the ACL configuration file.
Description
The hmadm command is used to administer the Hardware Monitor daemon. The Hardware Monitor daemon executes on the control workstation and is used to monitor and control the SP hardware. Five administrative actions are supported, as specified by the operation operand.
Normally when the daemon exits, it is automatically restarted by the system. If frame configuration information is changed, the quit operation can be used to update the system.
The daemon writes debug information and certain error information to its log file. The log file is located in /var/adm/SPlogs/spmon and its name is of the form hmlogfile.nnn, where nnn is the Julian date of the day the log file was opened by the daemon. The clog operation causes the daemon to close its current log file and create a new one using the name hmlogfilennn, where nnn is the current Julian date. If this name already exists, a name of the form hmlogfile.nnn_m is used, where m is a number picked to create a unique file name.
There are 15 debug flags supported by the daemon:
This command uses the SP Hardware Monitor. Therefore, the user must be authorized to access the Hardware Monitor subsystem and must have administrative permission. Since the Hardware Monitor subsystem uses SP authentication services, the user must execute the kinit command prior to executing this command. Alternatively, site-specific procedures can be used to obtain the tokens that are otherwise obtained by kinit.
Files
Related Information
File: /spdata/sys1/spmon/hmacls
Purpose
hmcmds - Controls the state of the SP hardware.
Syntax
Flags
Operands
Description
Use this command to control the state of the SP hardware. Control is provided via the Virtual Front Operator Panel (VFOP). VFOP is a set of commands that can be sent to the hardware components contained in one or more SP frames. Each frame consists of 18 slots, numbered 0 through 17, where slot 0 represents the frame itself, slot 17 can contain a switch and slots 1 through 16 can contain thin or wide processing nodes. Wide nodes occupy two slots and are addressed by the odd slot number. In a switch only frame, slots 1 through 16 can contain switches; the switches occupy two slots and are addressed by the even slot number.
Normally, commands are only sent to the hardware components in the current system partition. A system partition only contains processing nodes. The switches and the frames themselves are not contained in any system partition. To send VFOP commands to hardware components not in the current system partition or to any frame or switch, use the -G flag.
The following list describes the VFOP command set. Commands that require the -G flag are marked by an asterisk (*). Commands marked by a double asterisk (**) are primarily used by the Eclock command and are not intended for general use since an in-depth knowledge of switch clock topology is required to execute these commands in the proper sequence.
Before issuing these commands, refer to the "Using a Switch" chapter in the IBM Parallel System Support Programs for AIX: Administration Guide for detailed descriptions.
High Performance Switch
SP Switch
Any Frame, Node, or Switch that Supports Microcode Download
Note: | You must issue this command before issuing the microcode command. |
Note: | You must issue the basecode command before issuing this command. |
Refer to the s1term command for information on making serial connections.
Any Node
Any Frame
Any Frame, Node, or Switch
Any Node or Switch
One of these commands must be specified using the command operand. The command is sent to the hardware specified by the slot_spec operands. However, the command is not sent to any hardware that is not in the current system partition unless the -G flag is specified. If the -G flag is not specified and the slot_spec operands specify no hardware in the current system partition, an error message is displayed.
The slot_spec operands are interpreted as slot ID specifications. A slot ID specification names one or more slots in one or more SP frames and it has either of two forms:
fidlist:sidlist or nodlist
where:
The first form specifies frame numbers and slot numbers. The second form specifies node numbers. A fval is a frame number or a range of frame numbers of the form a-b. A sval is a slot number from the set 0 through 17 or a range of slot numbers of the form a-b. A nval is a node number or a range of node numbers of the form a-b.
The relationship of node numbers to frame and slot numbers is shown in the following formula:
node_number = ((frame_number - 1) × 16) + slot_number
Note: | Node numbers can only be used to specify slots 1 through 16 of any frame. |
The following are some examples of slot ID specifications.
To specify slot 1 in frames 1 through 10, enter:
1-10:1
To specify frames 2, 4, 5, 6, and 7, enter:
2,4-7:0
To specify slots 9 through 16 in frame 5, enter:
5:9-16
If frame 5 contained wide nodes, the even slot numbers are ignored.
To specify specifies slots 1, 12, 13, 14, 15, and 16 in each of frames 3 and 4, enter:
3,4:1,12-16
To specify slot 17 in frame 4, enter:
4:17
To specify the nodes in slots 1 through 16 of frame 2, enter:
17-32
To specify the nodes in slot 1 of frame 1, slot 1 of frame 2 and slot 1 of frame 3, enter:
1,17,33
To specify the node in slot 6 of frame 1, enter:
6
Optionally, slot ID specifications can be provided in a file rather than as command operands. The file must contain one specification per line. The command requires that slot ID specifications be provided. If the command is to be sent to all SP hardware, the keyword all must be provided in lieu of the slot_spec operands. However, the all keyword can only be specified if the -G flag is specified and if the VFOP command is on or off, since on or off are the only commands common to all hardware components.
Commands sent to hardware for which they are not appropriate, or sent to hardware which does not exist, are silently ignored by the Hardware Monitor subsystem.
By default, and except for the reset, flash, and run_post commands, the hmcmds command does not terminate until the state of the hardware to which the command was sent matches the command or until 15 seconds have elapsed. If 15 seconds have elapsed, the hmcmds command terminates with a message stating the number of nodes whose state was expected to match the VFOP command sent and the number of nodes which actually are in that state. The state of hardware for which the VFOP command is inappropriate, or where the hardware does not exist, is ignored.
To execute the hmcmds command, the user must be authorized to access the Hardware Monitor subsystem and, for those frames specified to the command, the user must be granted VFOP permission. Commands sent to frames for which the user does not have VFOP permission are ignored. Since the Hardware Monitor subsystem uses SP authentication services, the user must execute the kinit command prior to executing this command. Alternatively, site-specific procedures can be used to obtain the tokens that are otherwise obtained by kinit.
Files
Related Information
Command: hmmon, spsvrmgr
Examples
hmcmds -G off all
hmcmds secure 1-5:1-16
hmcmds -G extclk3 1-8:17
hmcmds normal 6 2,3:2
Purpose
hmmon - Monitors the state of the SP hardware.
Syntax
Flags
Operands
Description
Use this command to monitor the state of the SP hardware contained in one or more SP frames. Each frame consists of 18 slots, numbered 0 through 17, where slot 0 represents the frame itself, slot 17 can contain a switch and slots 1 through 16 can contain thin or wide processing nodes. Wide nodes occupy two slots and are addressed by the odd slot number. In a switch only frame, slots 1 through 16 can contain switches; the switches occupy two slots and are addressed by the even slot number.
With no flags and operands, the command prints to standard output descriptive text of all hardware state changes in the current system partition as they occur, from the time the command is invoked. The command does not terminate, unless the -Q flag or the -V flag is specified, and must be interrupted by the user. To monitor all of the hardware in the SP system, the -G flag must be specified. Note that the switches and the frames themselves are not contained in any system partition.
When one or more slot_spec operands are present, each operand is interpreted as a slot ID specification. A slot ID specification names one or more slots in one or more SP frames and it has either of two forms:
fidlist:[sidlist] or nodlist
where:
The first form specifies frame numbers and slot numbers. The second form specifies node numbers. A fval is a frame number or a range of frame numbers of the form a-b. A sval is a slot number from the set 0 through 17 or a range of slot numbers of the form a-b. An nval is a node number or a range of node numbers of the form a-b. If a sidlist is not specified, all hardware in the frames specified by the fidlist is monitored.
The relationship of node numbers to frame and slot numbers is given by the following formula:
node_number = ((frame_number - 1) × 16) + slot_number
Note: | The node numbers can only be used to specify slots 1 through 16 of any frame. |
The following are some examples of slot ID specifications.
To specify all hardware in frames 1 through 10, enter:
1-10:
To specify frames 2, 4, 5, 6, and 7, enter:
2,4-7:0
To specify slots 9 through 16 in frame 5, enter:
5:9-16
If frame 5 contained wide nodes, the even slot numbers are ignored.
To specify slots 1, 12, 13, 14, 15, and 16 in each of frames 3 and 4, enter:
3,4:1,12-16
To specify slot 17 in frame 4, enter:
4:17
To specify the nodes in slots 1 through 16 of frame 2, enter:
17-32
To specify the nodes in slot 1 of frame 1, slot 1 of frame 2 and slot 1 of frame 3, enter:
1,17,33
To specify the node in slot 6 of frame 1, enter:
6
Optionally, slot ID specifications may be provided in a file rather than as command operands. The file must contain one specification per line. When slot ID specifications are provided to the command, only the hardware named by the specifications is monitored. Furthermore, of the hardware named by these specifications, only that which is located in the current system partition is monitored. To monitor hardware not contained in the current system partition, the -G flag must be specified. If the -G flag is not specified and the slot ID specifications name no hardware in the current system partition, an error message is displayed.
The default output displays hardware state information on a slot-by-slot basis. The state information for each slot is captioned by its frame ID and slot ID and consists of two columns. Each column contains state variable information, one variable per line. Each variable is displayed as descriptive text and a value. Boolean values are displayed as TRUE or FALSE. Integer values are displayed in hexadecimal.
The command provides two other output formats, raw and symbolic. Both write the information for one state variable per line. The raw format consists of four fields separated by white space as follows:
The symbolic format consists of six fields separated by white space as follows:
The alternative output formats are suitable for input to post-processing programs, such as awk or scripts.
Output in any format can be limited to display only information from the specified hardware that corresponds to a list of state variables supplied to the command with the -v flag.
To execute the hmmon command, the user must be authorized to access the Hardware Monitor subsystem and, for those frames specified to the command, the user must be granted "Monitor" permission. State information is not returned for frames for which the user does not have "Monitor" permission. Since the Hardware Monitor subsystem uses SP authentication services, the user must execute the kinit command prior to executing this command. Alternatively, site specific procedures may be used to obtain the tokens that are otherwise obtained by kinit.
The user can monitor nonexistent nodes in an existing frame in order to detect when a node is added while the system is up and running. No information is returned for nonexistent nodes when the -q or -Q flag is specified.
Files
Related Information
Command: hmcmds
Examples
The following is an example of default output from hmmon -G -Q 1:0,1. The command returns similar output, depending on your system configuration.
frame 001, slot 00: node 01 I2C not responding FALSE node 02 I2C not responding TRUE node 03 I2C not responding FALSE node 04 I2C not responding TRUE switch I2C not responding FALSE node 01 serial link open TRUE node 02 serial link open FALSE node 03 serial link open TRUE frame LED 1 (green) 0x0001 frame LED 2 (green) 0x0001 frame LED 3 (yellow) 0x0000 frame LED 4 (yellow) 0x0000 AC-DC section A power off FALSE AC-DC section B power off FALSE AC-DC section C power off FALSE AC-DC section D power off FALSE supervisor timer ticks 0x88f2 +48 voltage 0x0078 temperature 0x0036 supervisor serial number 0x1234 supervisor type 0x0011 supervisor code version 0x5ff5 frame 001, slot 01: serial 1 DTR asserted TRUE -12 volt low warning TRUE -12 volt low shutdown FALSE -12 volt high warning TRUE +4 volt low shutdown FALSE +4 volt high warning TRUE fan 1 shutdown FALSE fan 2 warning TRUE DC-DC power on > 10 secs TRUE +5 DC-DC output good TRUE 7 segment display flashing FALSE node/switch LED 1 (green) 0x0001 reset button depressed FALSE serial link open TRUE diagnosis return code 0x00dd 7 segment LED A 0x00ff +5 I/O voltage 0x007f +12 voltage 0x0096
The following is an example of raw output from hmmon -G -Q -r 1:0. The command returns similar output, depending on your system configuration.
1 0 0x880f 32 1 0 0x881c 0 1 0 0x881d 4 1 0 0x8834 54 1 0 0x8839 4660 1 0 0x883a 17 1 0 0x88a8 1 1 1 0x9097 16 1 1 0x9098 0 1 1 0x9047 1 1 1 0x909d 128 1 1 0x9023 221 1 1 0x90a1 255 1 1 0x90a2 127 1 1 0x903b 24565
The following is an example of symbolic output from hmmon -G -Q -s 1:0. The command returns similar output, depending on your system configuration.
1 0 nodefail1 FALSE 0x8802 node 01 I2C not responding 1 0 nodeLinkOpen1 TRUE 0x8813 node 01 serial link open 1 0 frACLED 1 0x8824 frame LED 1 (green) 1 0 frNodeComm 0 0x8827 frame LED 4 (yellow) 1 0 frPowerOff_B FALSE 0x882d AC-DC section B power off 1 0 timeTicks 34881 0x8830 supervisor timer ticks 1 0 voltP48 46.800 0x8831 +48 voltage 1 0 type 17 0x883a supervisor type 1 0 codeVersion 24565 0x883b supervisor code version 1 0 controllerResponds TRUE 0x88a8 Frame responding to polls 1 0 rs232DCD TRUE 0x88a9 RS232 link DCD active 1 0 rs232CTS TRUE 0x88aa RS232 link CTS active 1 1 fanfail2 FALSE 0x9050 fan 2 shutdown 1 1 nodePowerOn10Sec TRUE 0x904b DC-DC power on > 10 secs 1 1 P5DCok TRUE 0x9097 +5 DC-DC output good 1 1 powerLED 1 0x9047 node/switch LED 1 (green) 1 1 envLED 0 0x9048 node/switch LED 2 (yellow) 1 1 keyModeSwitch 0 0x909b key switch 1 1 serialLinkOpen TRUE 0x909d serial link open 1 1 LED7SegA 255 0x909f 7 segment LED A 1 1 voltP5i 4.978 0x90a2 +5 I/O voltage
The raw and symbolic formats output by the hmmon command contain the variable ID of each state variable. Refer to Appendix D in IBM Parallel System Support Programs for AIX: Administration Guide.
Purpose
hmreinit - Stops and starts the Hardware Monitor daemon and modifies the System Data Repository (SDR) as necessary.
Syntax
hmreinit
Flags
None.
Operands
None.
Description
Use this command to reinitialize the Hardware Monitor daemon when changes to the SP system occur. Specifically, hmreinit determines if there are any changes in the switch configuration (such as, adding, deleting, or upgrading switches). The hmreinit command then calls SDR_config -u to update the switch information in the SDR and generates switch node numbers based on this change. If a switch configuration change is detected, hmreinit will test to see if a single system partition exists. If only one system partition exists, hmreinit will delete the Syspar_map entries from the SDR and then calls create_par_map to generate the correct objects. If more than one system partition exists, hmreinit will issue a message to that effect and exits.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command and have a valid ticket.
Implementation Specifics
This command is part of the Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/install/bin/hmreinit
Related Information
Commands: spframe, SDR_config
For additional information, refer to the "Reconfiguring the IBM RS/6000 SP System" chapter in IBM Parallel System Support Programs for AIX: Installation and Migration Guide.
Purpose
hostlist - Lists SP host names to standard output based on criteria.
Syntax
Flags
Operands
None.
Description
The hostlist command writes SP host names to standard output. The arguments to the command indicate the host names to be written. More than one flag can be specified, in which case, the hosts indicated by all the flags are written.
If no arguments are specified, hostlist writes the contents of a file specified by the WCOLL environment variable. If the WCOLL environment variable does not exist, the MP_HOSTFILE environment variable is used as the name of a POE host file to use for input. Finally, ./host.list is tried. If none of these steps are successful, an error has occurred. The input file is in dsh-working-collective-file or POE-host-list-file format. Node pool specifications in POE host files are not supported.
Files
Related Information
Commands: dsh, sysctl
Examples
hostlist -av -e badhost > ./working
hostlist -s 1-4:1 | dsh -w - program
hostlist -n 1-16,33-35 -w otherone | dsh -w - program
export WCOLL=./wcoll hostlist | sysctl -c - sysctl_app args
Purpose
hr - Controls the host_responds monitor daemon, hrd, on the control workstation.
Syntax
hr [-spname syspar_name]
{ [start | resume] | [stop | quiesce] | reset |
[query | qall | qsrc] | refresh | mksrc optional_flags | rmsrc | clean | restore |
[debug | debug off ] | [trace on | trace off ] }
Flags
Operands
Description
Use this command to control the operation of hrd, the host_responds daemon on the control workstation within a system partition. The heartbeat server provides input to the host_responds function within a system partition for the System Monitor through the hrd daemons.
The hr script is not normally executed from the command line. It is normally called by the hrctrl command, which is in turn called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system.
The hrd daemon is initially started on the control workstation with the System Resource Controller (SRC). It is respawned automatically if the hrd daemon fails. The SP_NAME environment variable causes selection of the correct daemon.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
The "Starting Up and Shutting Down the SP System" chapter and "The System Data Repository" appendix in IBM Parallel System Support Programs for AIX: Administration Guide
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/hr
Related Information
Commands: hb, hrctrl, lssrc, startsrc, stopsrc, syspar_ctrl
Examples
See the hrctrl command.
Purpose
hrctrl - A script that controls the Host_Responds subsystem.
Syntax
hrctrl { -a | -s | -k | -d | -c | -t | -o | -r | -h }
Flags
Operands
None.
Description
The Host_Responds subsystem provides to other PSSP subsystems information about the state of the nodes on the IBM RS/6000 SP.
The hrctrl control script controls the operation of the Host_Responds subsystem. The subsystem is under the control of the System Resource Controller (SRC) and belongs to a subsystem group called hr. Associated with each subsystem is a daemon and a script that configures and starts the daemon.
An instance of the Host_Responds subsystem executes on the control workstation for every system partition. Because Host_Responds provides its services within the scope of a system partition, its subsystem is said to be system partition-sensitive. This control script operates in a manner similar to the control scripts of other system partition-sensitive subsystems. The script should be issued on the control workstation. If it is issued on a node, it has no effect.
From an operational point of view, the Host_Responds subsystem group is organized as follows:
The hr subsystem is associated with the hrd daemon and the hr script. The hr script configures and starts the hrd daemon.
On the control workstation, there are multiple instances of each subsystem, one for each system partition. Accordingly, the subsystem names on the control workstation have the system partition name appended to them. For example, for system partitions named sp_prod and sp_test, the subsystems on the control workstation are named hr.sp_prod and hr.sp_test.
The subsystem does not run on the nodes.
The hrd daemon provides the Host_Responds services. The hr script configures and starts the hrd daemon.
The hrctrl script is not normally executed from the command line. It is normally called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system.
The hrctrl script provides a variety of controls for operating the Host_Responds subsystem:
Before performing any of these functions, the script obtains the node number (using the node_number) command. If the node number is not zero, the control script is running on a node and it exits immediately. Otherwise, it is executing on the control workstation and it calls the hr script with an operand that specifies the action to be performed.
Adding the Subsystem
When the -a flag is specified, the control script uses the hr command with the mksrc operand to add the Host_Responds subsystem to the SRC.
Starting the Subsystem
When the -s flag is specified, the control script uses the hr command with the start operand to start the Host_Responds subsystem, hr.
Stopping the Subsystem
When the -k flag is specified, the control script uses the hr command with the stop operand to stop the Host_Responds subsystem, hr.
Deleting the Subsystem
When the -d flag is specified, the control script uses the hr command with the rmsrc operand to remove the Host_Responds subsystem from the SRC.
Cleaning up the Subsystems
When the -c flag is specified, the control script uses the hr command with the clean operand to stop and remove the Host_Responds subsystems for all system partitions from the SRC.
Turning Tracing On
When the -t flag is specified, the control script turns tracing on for the hrd daemon, using the hr command with the trace on operand.
Turning Tracing Off
When the -o flag is specified, the control script turns tracing off (returns it to its default level) for the hrd daemon, using the hr command with the trace off operand.
Refreshing the Subsystem
When the -r flag is specified, the control script refreshes the subsystem, using the hr refresh command.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must be running with an effective user ID of root.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/hrctrl
Related Information
Commands: hr, lssrc, startsrc, stopsrc, syspar_ctrl
Examples
hrctrl -a
hrctrl -s
hrctrl -k
hrctrl -d
hrctrl -c
hrctrl -t
hrctrl -o
lssrc -g hr
lssrc -s subsystem_name
lssrc -l -s subsystem_name
In response, the system returns information that includes the running status of the subsystem and the status of the nodes within the system partition.
lssrc -a
Purpose
hsdatalst - Displays data striping device (HSD) data for the IBM Virtual Shared Disks from the System Data Repository (SDR).
Syntax
hsdatalst [-G]
Flags
Operands
None.
Description
This command is used to display defined HSD information in the system.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit list_vsdand select the List Defined Hashed Shared Disk option.
Files
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: defhsd, undefhsd
Examples
To display SDR HSD data, enter:
hsdatalst
Purpose
hsdvts - Verifies that a data striping device (HSD) for the IBM Virtual Shared Disks has been correctly configured and works.
Syntax
hsdvts hsd_name
Flags
None.
Operands
Description
Attention |
---|
Data on hsd_name will be overwritten and, therefore, destroyed. Use this command after you have defined your HSD, IBM Virtual Shared Disks, and logical volumes, but before you have loaded your application data onto any of them. |
This command writes /unix to hsd_name, reads it from hsd_name to a temporary file, and compares the temporary file to the original to make sure the I/O was successful. If the files compare exactly, the test was successful.
hsdvts writes to the raw hsd_name device /dev/rhsd_name. Since raw devices can only be written in multiples of 512-sized blocks, hsdvts determines the number of full 512-byte blocks in /unix file, and writes that number to hsd_name via dd command. It makes a copy of /unix that contains this number of 512-byte blocks for comparison to the copy read from hsd_name. The dd command is used for all copy operations.
Files
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfghsd, cfgvsd, dd, defhsd, startvsd
Purpose
/usr/lpp/ssp/css/ifconfig - Configures or displays network interface parameters for a network using TCP/IP.
Syntax
Flags
None.
Operands
Include a numeral after the abbreviation to identify the specific interface (for example, tr0).
The mask variable includes both the network part of the local address and the subnet part, which is taken from the host field of the address. The mask can be specified as a single hexadecimal number beginning with 0x, in standard Internet dotted decimal notation, or beginning with a name or alias that is listed in the /etc/networks file.
The mask contains 1's (ones) for the bit positions in the 32-bit address that are reserved for the network and subnet parts, and 0's (zeros) for the bit positions that specify the host. The mask should contain at least the standard network portion, and the subnet segment should be contiguous with the network segment.
Note: | If the ARP is enable, offset is not used. |
Description
The ifconfig command has been modified to add support for the switch. This command is valid only on an SP system.
The ifconfig command can be used from the command line either to assign an address to a network interface, or to configure or display the current network interface configuration information. The ifconfig command must be used at system start up to define the network address of each interface present on a machine. It can also be used at a later time to redefine an interface's address or other operating parameters. The network interface configuration is held on the running system and must be reset at each system restart.
An interface can receive transmissions in differing protocols, each of which may require separate naming schemes. It is necessary to specify the address_family parameter, which can change the interpretation of the remaining parameters. The address families currently supported are inet and ns.
For the DARPA Internet family, inet, the address is either a host name present in the host name database, that is, the /etc/hosts file, or a DARPA Internet address expressed in the Internet standard dotted decimal notation.
For the Xerox Network Systems (XNS) family, ns, addresses are net:a.b.c.d.e.f., where net is the assigned network number (in decimal), and each of the six bytes of the host number, a through f, are specified in hexadecimal. The host number can be omitted on 10-Mbps Ethernet interfaces, which use the hardware physical address, and on interfaces other than the first interface.
While any user can query the status of a network interface, only a user who has administrative authority can modify the configuration of those interfaces.
Related Information
AIX Command: netstat
AIX Files: /etc/host, /etc/networks
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for additional information on the SP Switch and the High Performance Switch.
Refer to AIX Version 4 System Management Guide: Communications and Networks for additional information on TCP/IP protocols.
Refer to AIX Version 4 General Programming Concepts: Writing and Debugging Programs for an overview on Xerox Network Systems (XNS).
Examples
The following are examples using the ifconfig command on a TCP/IP network and an XNS network, respectively:
Inet Examples
ifconfig sl1
In this example, the interface to be queried is sl1. The result of the command looks similar to the following:
sl1: flags=51<UP,POINTOPOINT,RUNNING> inet 192.9.201.3 --> 192.9.354.7 netmask ffffff00
ifconfig lo0 inet 127.0.0.1 up
ifconfig tr0 inet down
In this example, the interface to be marked is token0.
Note: | Only a user with root user authority can modify the configuration of a network interface. |
ifconfig css0 inet 127.0.0.1 netmask 255.255.255.0 alias
XNS Examples
ifconfig en0 ns 110:02.60.8c.2c.a4.98 up
In this example, ns is the XNS address family, 110 is the network number and 02.60.8c.2c.a4.98 is the host number, which is the Ethernet address unique to each individual interface. Specify the host number when there are multiple Ethernet hardware interfaces, as the default may not correspond to the proper interface. The Ethernet address can be obtained by the commands:
ifconfig en0 netstat -v
The XNS address can be represented by several means, as can be seen in the following examples:
123#9.89.3c.90.45.56 5-124#123-456-900-455-749 0x45:0x9893c9045569:90 0456:9893c9045569H
The first example is in decimal format, and the second example, using minus signs, is separated into groups of three digits each. The 0x and H examples are in hexadecimal format. Finally, the 0 in front of the last example indicates that the number is in octal format.
ifconfig et0 ns 120:02.60.8c.2c.a4.98 up
The en0 and et0 interfaces are considered as separate interfaces even though the same Ethernet adapter is used. Two separate networks can be defined and used at the same time as long as they have separate network numbers. Multiple Ethernet adapters are supported.
Note: | The host number should correspond to the Ethernet address on the hardware adapter. A system can have multiple host numbers. |
ifconfig en0 inet 11.0.0.1 up ifconfig en0 ns 110:02.60.8c.2c.a4.98 up ifconfig en0 ns 130:02.60.8c.34.56.78 ipdst 11.0.0.10
The first command brings up the Internet with the inet address 11.0.0.1. The second command configures the en0 interface to be network 110 and host number 02.60.8c.2c.a4.98 in the ns address family. This defines the host number for use when the XNS packet is encapsulated within the Internet packet. The last command defines network 130, host number 02.60.8c.34.56.78, and destination Internet address 11.0.0.10. This last entry creates a new network interface, nsip. Use the netstat -i command for information about this interface.
Purpose
install_cw - Completes the installation of system support programs in the control workstation.
Syntax
install_cw
Flags
None.
Operands
None.
Description
Use this command at installation to perform the following tasks:
Purpose
install_hacws - Creates and configures a High Availability Control Workstation (HACWS) configuration from a regular control workstation configuration.
Syntax
install_hacws -p host_name -b host_name [-s]
Flags
Operands
None.
Description
Use this command to perform configuration and installation tasks on HACWS. This command is used instead of install_cw once the configuration has been made an HACWS configuration. This command is valid only when issued on the control workstation. When the command is executed and the calling process is not on a control workstation, an error occurs.
Note: | The install_hacws command permanently alters a control workstation to an HACWS. The only way to go back to a single control workstation is to have a mksysb image of the primary control workstation before the install_hacws command is executed. |
Both the primary and backup control workstations must be running and capable of executing remote commands via the /usr/lpp/ssp/rcmd/bin/rsh command.
Exit Values
Standard output consists of messages indicating the progress of the command as it configures the control workstations.
Prerequisite Information
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for information on the HACWS option.
Location
/usr/sbin/hacws/install_hacws
Related Information
SP Commands: install_cw, rsh, setup_logd
Examples
install_hacws -p primary_cw -b backup_cw -s
On the primary control workstation, enter:
install_hacws -p primary_cw -b backup_cw
After the preceding command completes on the primary control workstation, enter the following on the backup control workstation:
install_hacws -p primary_cw -b backup_cw
Purpose
jm_config - Reconfigures the Resource Manager.
Syntax
jm_config
Flags
None.
Operands
None.
Description
Use this command to reconfigure the Resource Manager (RM) servers.
This command must be executed by root on the control workstation. It reads the Resource Manager configuration data from the /etc/jmd_config.syspar_name file, where syspar_name represents the current system partition environment. This current working environment can be determined by issuing spget_syspar -n. The jm_config command then contacts the correct primary Resource Manager and sends a message telling the server to update its configuration data from the System Data Repository (SDR). The new configuration takes effect with the next client request.
Note: | 604 High Nodes cannot be configured as part of a parallel pool. Therefore, the Resource Manager will not allocate these nodes for parallel jobs. |
The Resource Manager can also be reconfigured via the System Management Interface Tool (SMIT). To use SMIT, enter:
smit RM_optionsand select the Reconfigure the Resource Manager option. Refer to IBM Parallel System Support Programs for AIX: Administration Guide for additional information on configuring the Resource Manager and system partitioning.
Files
Related Information
Commands: jm_start, jm_status, jm_stop, locate_jm, spget_syspar
Examples
To reconfigure the Resource Manager in the current system partition, enter:
jm_configThe following response shows the configuration data file being used by the jm_config command:
jm_config: Using /etc/jmd_config.k42sp1 for Resource Manager configuration data. jm_config: jm_config successful
Purpose
jm_install_verify - Verifies that the installation of the Resource Manager component of the SP system completed successfully.
Syntax
jm_install_verify [-h] [-q] [-l logfile]
Flags
Operands
None.
Description
Use this command to perform various tests to determine whether the Resource Manager component of the SP system is installed correctly. This command checks that:
This command must be executed by root and operates within the scope of the current system partition environment. When executed on the control workstation, installation will be verified for the control workstation and all of the nodes within the current system partition. When executed on a node, installation is verified on that node only.
If you do not specify the -l flag, more detailed information is recorded in /var/adm/SPlogs/jm_install_verify.log.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit SP_verifyand select the Resource Manager Installation option.
Files
Exit Values
If you do not specify the -q flag, a message is displayed on standard output indicating the success or failure of the tests. If errors are detected, a message is displayed stating how many errors were found and more detailed information is recorded in the log file.
Related Information
Commands: CSS_test, jm_verify, SDR_test, SYSMAN_test, spmon_ctest, spmon_itest
Examples
To verify installation of the Resource Manager on a single node, saving error messages in the jminst.err file, from that node, enter:
jm_install_verify -l jminst.errUpon successful completion, the following message is displayed:
Verifying installation of Resource Manager on node 9. Resource Manager installation verification SUCCESSFUL on node 9. Check ./jminst.err file for details.
Purpose
jm_start - Starts the primary and backup Resource Manager servers in the current system partition.
Syntax
jm_start [-h | -i]
Flags
Operands
None.
Description
Use this command to start the primary and backup Resource Manager servers.
This command must be executed by root on the control workstation. It reads the Resource Manager configuration data from the /etc/jmd_config.syspar_name file, where syspar_name represents the current system partition environment. This current working environment can be determined by issuing spget_syspar -n. This command reads the configuration data into the System Data Repository (SDR) and starts the primary Resource Manager server on one of the nodes indicated. This server then starts up a backup on the next available node in the list of server candidates.
Note: | 604 High Nodes cannot be configured as part of a parallel pool. Therefore, the Resource Manager will not allocate these nodes for parallel jobs. |
If jm_start successfully starts the primary Resource Manager server, it returns a message indicating its location. If a backup is not successfully started, an additional message is issued and the primary Resource Manager server continues to run, attempting periodically to start the backup. If you do not have a backup running, and the primary Resource Manager server dies for any reason, there is no recovery and the Resource Manager server must be restarted manually.
Use the -i flag only if the jm_start command cannot determine whether the Resource Manager server is already running. The -i flag does not check if the Resource Manager server is already running, therefore, it is best not to use this option unless you are sure the Resource Manager server is not running on any node.
The Resource Manager can also be reconfigured via the System Management Interface Tool (SMIT). To use SMIT, enter:
smit RM_optionsand select the Start the Resource Manager option. Refer to IBM Parallel System Support Programs for AIX: Administration Guide for additional information on the Resource Manager and system partitioning.
Files
Related Information
Commands: jm_config, jm_status, jm_stop, locate_jm, spget_syspar
Examples
To start the Resource Manager in the current system partition, enter:
jm_startThe following response shows the configuration data file being used by the jm_start command:
jm_start: Using /etc/jmd_config.k42sp1 for Resource Manager configuration data. jm_start: jmd started on k42n09:
Purpose
jm_status - Gets information about defined pools or about jobs running on nodes allocated by the Resource Manager.
Syntax
Flags
Operands
None.
Description
Use the jm_status command to get information about defined pools or about jobs running on nodes allocated by the Resource Manager. Any user can execute jm_status on any workstation.
Files
Related Information
Commands: locate_jm, spget_syspar
Examples
jm_status -n k42sp1 -jThe response should be similar to this:
Job 1: time_allocated=Mon_Apr__4_16:09:28_1994 description=interpret_alloc_imp_test requestor=root requestor_pid=11787 requestor_node=r07n03.hpssl.kgn.ibm.com Adapter type=HPS_US Node: r07n01.hpssl.kgn.ibm.com Usage: cpu=SHARED adapter=DEDICATED virtual task ids: 0 1 Node: r07n09.hpssl.kgn.ibm.com Usage: cpu=SHARED adapter=DEDICATED virtual task ids: 2 3 4 Job 2: time_allocated=Mon_Apr__4_16:13:08_1994 description=interpret_alloc_imp_test requestor=root requestor_pid=12355 requestor_node=r07n03.hpssl.kgn.ibm.com Adapter type=ETHERNET Node: r07n01.hpssl.kgn.ibm.com Usage: cpu=SHARED adapter=DEDICATED virtual task ids: 0 1 Node: r07n09.hpssl.kgn.ibm.com Usage: cpu=SHARED adapter=DEDICATED virtual task ids: 2 3 4 Job 3: time_allocated=Mon_Apr__4_16:13:38_1994 description=interpret_alloc_imp_test requestor=root requestor_pid=14404 requestor_node=r07n03.hpssl.kgn.ibm.com Adapter type=HPS_US Node: r07n15.hpssl.kgn.ibm.com Usage: cpu=SHARED adapter=SHARED virtual task ids: 0 1
jm_status -PThe response should be similar to this:
Pool 1: Ethernet_test_pool_1 Subpool: INTERACTIVE Node: r07n01.hpssl.kgn.ibm.com Subpool: BATCH Node: r07n03.hpssl.kgn.ibm.com Node: r07n13.hpssl.kgn.ibm.com Node: r07n11.hpssl.kgn.ibm.com Subpool: GENERAL Node: r07n15.hpssl.kgn.ibm.com Pool 2: Ethernet_test_pool_2 Subpool: INTERACTIVE Node: r07n09.hpssl.kgn.ibm.com Subpool: BATCH Node: r07n05.hpssl.kgn.ibm.com Subpool: GENERAL Node: r07n07.hpssl.kgn.ibm.com
jm_status -p1 -vshould get a response similar to this:
Pool 1: Ethernet_test_pool_1 Subpool: INTERACTIVE Node: r07n01.hpssl.kgn.ibm.com Job 1: time allocated=Mon_Apr__4_16:09:28_1994 description=interpret_alloc_imp_test requestor=root requestor pid=11787 requestor_node=r07n03.hpssl.kgn.ibm.com Adapter type=HPS_US Usage: cpu=SHARED adapter=DEDICATED virtual task ids: 0 1 Job 2: time allocated=Mon_Apr__4_16:13:08_1994 description=interpret_alloc_imp_test requestor=root requestor pid=12355 requestor_node=r07n03.hpssl.kgn.ibm.com Adapter type=ETHERNET Usage: cpu=SHARED adapter=DEDICATED virtual task ids: 0 1 Subpool: BATCH Node: r07n03.hpssl.kgn.ibm.com Node: r07n13.hpssl.kgn.ibm.com Node: r07n11.hpssl.kgn.ibm.com Subpool: GENERAL Node: r07n15.hpssl.kgn.ibm.com Job 3: time allocated=Mon_Apr__4_16:13:38_1994 description=interpret_alloc_imp_test requestor=root requestor pid=14404 requestor_node=r07n03.hpssl.kgn.ibm.com Adapter type=HPS US Usage: cpu=SHARED adapter=SHARED virtual task ids: 0 1
Purpose
jm_stop - Stops the primary and backup Resource Manager servers in the current system partition.
Syntax
jm_stop [-q | -h]
Flags
Operands
None.
Description
Use this command to stop the primary and backup Resource Manager servers.
This command must be executed by root and operates within the scope of the current system partition environment. It can be issued from any node within a system partition.
The system can be quiesced in a less disruptive fashion by specifying the -q option, which tells the Resource Manager server to first wait for all existing jobs to terminate before exiting. The jm_stop -q command returns immediately. Any subsequent Resource Manager job requests, as well as jm_status and jm_config commands, fail. When all jobs have finished, the primary and backup Resource Manager server exits. Issue locate_jm to determine if the Resource Manager servers have exited. Anytime after jm_stop -q has been issued, you can issue jm_stop (without -q) to force an immediate stop.
The Resource Manager can also be stopped via the System Management Interface Tool (SMIT). To use SMIT, enter:
smit RM_stopand select either the Stop the Resource Manager immediately option (jm_stop), or the Wait for existing clients to exit, then stop the Resource Manager option (jm_stop -q).
Files
Related Information
Commands: jm_config, jm_start, jm_status, locate_jm
Examples
Entering
jm_stopshould get this response:
jm_stop: jm_stop successful.
Purpose
jm_verify - Verifies that the configuration of the Resource Manager component of the SP system completed successfully in the current system partition.
Syntax
jm_verify [-h] [-q] [-l logfile]
Flags
Operands
None.
Description
Use this command to perform various tests to determine whether the Resource Manager component of the SP system is correctly configured.
This command can be executed by any user and operates within the scope of the current system partition environment.
This command utilizes the SP dsh and rsh commands to access remote nodes for configuration information. The same security considerations apply.
If you do not specify the -l flag, more detailed information is recorded in /var/adm/SPlogs/jm_verify.log.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit SP_verifyand select the Resource Manager Installation option.
Files
Exit Values
If you do not specify the -q flag, a message is displayed on standard output indicating the success or failure of the tests. If errors are detected, a message is displayed stating how many errors were found and more detailed information is recorded in the log file.
Related Information
Commands: CSS_test, dsh, jm_install_verify, rsh, SDR_test, SYSMAN_test, spmon_ctest, spmon_itest
Examples
To verify configuration of the Resource Manager, saving error messages in jmconf.err in the current working directory, enter:
jm_verify -l jmconf.errUpon successful completion, the following message is displayed:
Verifying Resource Manager on node 0. Resource Manager verification SUCCESSFUL on node 0. Check ./jmconf.err file for details.
Purpose
jmcmi_accesscontrol - Updates the access_control attribute of the JM_domain_info System Data Repository (SDR) class.
Syntax
jmcmi_accesscontrol [-h] access_control {yes | no} update_config {yes | no}
Flags
Operands
Description
This command is a Perl script used to set the attribute value of access_control within the JM_domain_info SDR class. The access_control attribute determines whether the RM should interact with the system management command, spacs_cntrl, to allow users access to parallel nodes. If yes is specified for the update_config operand, the RM server is notified of the change to access_control and the new value is used after the next client connection. If no is specified, a second action must be taken before the new value is used. If the RM server is not running, the action must be to start the RM server. If the RM server is running, the action would be to reconfigure the RM server.
Executing this script is equivalent to setting the keyword ACCESS_CONTROL within the RM /etc/jmd_config.syspar_name configuration file. The data is updated within the SDR and the /etc/jmd_config.syspar_name file is updated with the new value for ACCESS_CONTROL.
This script must be executed by root from the control workstation and operates within the scope of the current system partition environment. Both operands are required.
The main use of this script is via the System Management Interface Tool (SMIT). To use SMIT, enter:
smit access_control
Files
Related Information
Commands: jm_config, jm_start, jmcmi_createjmdconfig, spacs_cntrl
Examples
To turn access control on and inform the RM server of the changes, enter:
jmcmi_accesscontrol yes yes
Purpose
jmcmi_addpool - Defines a new Resource Manager (RM) pool.
Syntax
Flags
Operands
Description
This command is a Perl script used to define a new RM pool. For more information on RM pools and their definitions, refer to IBM Parallel System Support Programs for AIX: Administration Guide.
After initial RM configuration, the RM Server List must be defined before the definition of any pools.
node_list must be one of the following:
The keyword all when specified for a subpool works in combination with a node_list defined for another subpool, to create a subpool with all the nodes on the system excluding the ones specified by the node_list. See the Examples section for this command.
After this script is executed, the new pool data is updated within the System Data Repository (SDR) and the /etc/jmd_config.syspar_name file is updated to reflect the new pool information.
If the update_config operand is yes, the RM server is notified that a configuration change was made, and it updates all of its configuration data. This data is used at the next client connection. If no is specified, a second action must be taken before the new configuration data is used. If the RM server is not running, the action must be to start the RM server. If the RM server is running, the action would be to reconfigure the RM server.
This script must be executed by root from the control workstation and operates within the scope of the current system partition environment. All operands are required.
The main use of this script is via the System Management Interface Tool (SMIT). To use SMIT, enter:
smit add_pool
Files
Related Information
Commands: jm_config, jm_start, jmcmi_createjmdconfig, jmcmi_servernodes
Examples
jmcmi_addpool 1 pool_for_LL yes batch "r07n01 r07n02 r07n03 r07n04 r07n05 r07n06 r07n07 r07n08" interactive "none" general "r07n09 r07n10 r07n11 r07n12 r07n13 r07n14 r07n15 r07n16" yes
This results in the following definition within the /etc/jmd_config.syspar_name file:
POOL_ID = 1 ATTRIBUTES = pool_for_LL MEMBERS_BATCH = r07n01; r07n02; \ r07n03; r07n04; r07n05; \ r07n06; r07n07; r07n08; MEMBERS_INTERACTIVE = MEMBERS_GENERAL = r07n09; \ r07n10; r07n11;r07n12;r07n13; \ r07n14;r07n15;r07n16;
If the node r07n10 was previously defined within Pool 2, jmcmi_addpool deletes r07n10 from that pool definition. The RM is notified of the new pool definitions and the configuration changes take effect on the next client connection.
jmcmi_addpool 2 General_Pool yes batch "none" interactive "none" general "all" yes
This results in all of the nodes being defined within Pool 2. Because the delete_nodes operand was specified as yes, any other pools that were defined on the system no longer exist. All of the nodes were deleted from the pools and moved to the one being created. The RM is notified of the new pool definitions and the configuration changes take effect at the next client connection.
jmcmi_addpool 2 DeptXYZ yes batch "r07n01 r07n16" interactive "r07n02 r07n15" general "all" yes
This results in the following definition within the /etc/jmd_config.syspar_name file:
POOL_ID = 2 ATTRIBUTES = DeptXYZ MEMBERS_BATCH = r07n01; r07n16; MEMBERS_INTERACTIVE = r07n02; r07n15; MEMBERS_GENERAL = r07n03; \ r07n04; r07n05; \ r07n06; r07n07; r07n08; \ r07n09; r07n10; r07n11; \ r07n12;r07n13; r07n14;
Purpose
jmcmi_changepool - Changes the definition of a Resource Manager (RM) pool.
Syntax
Flags
Operands
Description
This command is a Perl script used to redefine a RM pool. For more information on Resource Manager pools and their definitions, refer to IBM Parallel System Support Programs for AIX: Administration Guide.
The pool_id operand cannot be changed by this script. In order to change a pool ID, the original pool must be deleted and a new one created.
"node_list" must be one of the following:
The keyword all when specified for a subpool works in combination with a "node_list" defined for another subpool, to create a subpool with all the nodes on the system excluding the ones specified by the "node_list." See the Examples section.
After this script is executed, the changed pool data is updated within the SDR and the /etc/jmd_config.syspar_name file is updated to reflect the information.
If the update_config operand is yes, the RM server is notified that a configuration change was made, and updates all of its configuration data. This data is used at the next client connection. If no is specified, a second action must be taken before the new configuration data is used. If the RM server is not running, the action must be to start the RM server. If the RM server is running, the action would be to reconfigure the RM server.
This script must be executed by root from the control workstation and operates within the scope of the current system partition environment. All operands are required.
The main use of this script is via the System Management Interface Tool (SMIT). To use SMIT, enter:
smit configure_pool
Files
Related Information
Commands: jm_config, jm_start jmcmi_addpool, jmcmi_createjmdconfig, jmcmi_deletepool
Examples
jmcmi_changepool 1 pool_for_LL yes batch "r07n01 r07n02 r07n03 r07n04 r07n05 r07n06 r07n07 r07n08" interactive "none" general "r07n09 r07n10 r07n11 r07n12 r07n13 r07n14 r07n15 r07n16" yes
This results in the following definition within the /etc/jmd_config.syspar_name file:
POOL_ID = 1 ATTRIBUTES = pool_for_LL MEMBERS_BATCH = r07n01; r07n02; \ r07n03; r07n04; r07n05; \ r07n06; r07n07; r07n08; MEMBERS_INTERACTIVE = MEMBERS_GENERAL = r07n09; \ r07n10; r07n11;r07n12;r07n13; \ r07n14;r07n15;r07n16;
If the node r07n10 was previously defined within Pool 2, jmcmi_changepool deletes r07n10 from that pool definition. The RM is notified of the new pool definitions and the configuration changes take effect on the next client connection.
jmcmi_changepool 2 General_Pool yes batch "none" interactive "none" general "all" yes
This results in all of the nodes being defined within Pool 2. Because the delete_nodes operand was specified as yes, any other pools that were defined on the system no longer exist. All of the nodes were deleted from the pools and moved to the one being redefined. The RM is notified of the new pool definitions and the configuration changes take effect at the next client connection.
jmcmi_changepool 2 DeptXYZ yes batch "r07n01 r07n16" interactive "r07n02 r07n15" general "all" yes
This results in the following definition within the /etc/jmd_config.syspar_name file:
POOL_ID = 2 ATTRIBUTES = DeptXYZ MEMBERS_BATCH = r07n01; r07n16; MEMBERS_INTERACTIVE = r07n02; r07n15; MEMBERS_GENERAL = r07n03; \ r07n04; r07n05; \ r07n06; r07n07; r07n08; \ r07n09; r07n10; r07n11; \ r07n12;r07n13; r07n14;
Purpose
jmcmi_createjmdconfig - Updates the Resource Manager (RM) configuration file, /etc/jmd_config.syspar_name, with the current configuration data.
Syntax
jmcmi_createjmdconfig
Flags
None.
Operands
None.
Description
This Perl script is executed by other RM configuration scripts to update the /etc/jmd_config.syspar_name file with any changes made during their execution.
The existing /etc/jmd_config.syspar_name is preserved as a unique file name /etc/jmd_config.syspar_name.{timestamp}, so that no previous configuration data is lost. A new /etc/jmd_config.syspar_name file is created by getting the current configuration data from the System Data Repository (SDR).
This script must be executed by root from the control workstation and operates within the scope of the current system partition environment.
Files
Related Information
Commands: jmcmi_accesscontrol, jmcmi_addpool, jmcmi_changepool, jmcmi_deletepool, jmcmi_servernodes
Purpose
jmcmi_deletepool - Deletes a pool definition from the Resource Manager (RM).
Syntax
jmcmi_deletepool [-h] pool_id update_config {yes | no}
Flags
Operands
Description
This command is a Perl script used to delete the configuration data of a RM pool.
If the update_config operand is yes, the RM server is notified of the change to access_control and the new value is used at the next client connection. If no is specified, a second action must be taken before the new value is used. If the RM server is not running, the action must be to start the RM server. If the RM server is running, the action would be to reconfigure the RM server.
After this script is executed, the pool data is deleted from the System Data Repository (SDR) and the /etc/jmd_config.syspar_name file is updated to reflect the current configuration.
This script must be executed by root from the control workstation and operates within the scope of the current system partition environment. Both operands are required.
The main use of this script is via the System Management Interface Tool (SMIT). To use SMIT, enter:
smit delete_pool
Files
Related Information
Command: jmcmi_createjmdconfig
Examples
To delete Pool 1 and inform the RM server of these changes, enter:
jmcmi_deletepool 1 yes
Purpose
jmcmi_servernodes - Defines the Resource Manager (RM) server list.
Syntax
Flags
Operands
Description
This command is a Perl script used to define the SP nodes available for the RM to run its servers on. The order of the nodes listed is important. The RM always tries to start the server on the 1st_server_node if that node is available and then continues trying nodes in the list in the order they are specified until an available node is found. For more information on the RM server list and servers, refer to IBM Parallel System Support Programs for AIX: Administration Guide.
If the update_config operand is yes, the RM server is notified that a configuration change was made, and updates all of its configuration data. This data is used at the next client connection. If no is specified, a second action must be taken before the new configuration data is used. If the RM server is not running, the action must be to start the RM server. If the RM server is running, the action would be to reconfigure the RM server.
The 1st_server_node operand must specify a single node name. The 2nd_server_node operand supports two options: a single node name or the keyword none. If none is specified, no other node names can be defined following it. The additional_server_nodes operand supports the same two options except more than one node name can be specified. If no further nodes need to be defined, the keyword none must be specified.
This script must be executed by root from the control workstation and operates within the scope of the current system partition environment. All operands are required.
After this script is executed, the data is updated within the System Data Repository (SDR) and the /etc/jmd_config.syspar_name file is updated with the new values for JM_LIST.
The main use of this script is via the System Management Interface Tool (SMIT). To use SMIT, enter:
smit server_list
Files
Related Information
Command: jmcmi_createjmdconfig
Examples
jmcmi_servernodes yes r07n01 r07n10 r07n128 r07n150
jmcmi_servernodes no r07n01 r07n10 none
In this case, the RM needs to be informed later in order for this to take affect.
Purpose
kadmin - Provides network access to authentication database administration functions.
Syntax
kadmin [-u admin_name] [-r default_realm] [-m]
Flags
Operands
None.
Description
This command provides an interactive interface to the primary authentication database. Administrators use kadmin to add new users and services to the database, and to change information about existing database entries. For example, an administrator can use kadmin to change a user's password. An administrator is a user with an admin instance whose name appears in at least one of the authentication administration Access Control Lists (ACLs).
The kadmin program communicates over the network with the kadmind program, which runs on the machine housing the primary authentication database. The kadmind program creates new entries and makes modifications to the database.
When you enter the kadmin command, the program displays a message that welcomes you and explains how to ask for help. Then kadmin waits for you to enter commands. After you enter a command, you are prompted to enter your admin password. If the -m option is used, you are prompted for your admin password only for the first command entered. You do not need to issue the kinit command prior to running this command because the necessary tickets are obtained automatically.
When using the kadmin command, the principal's expiration date and maximum ticket lifetime are set to the default values. To override the defaults, the root user must run the kdb_edit command to modify those attributes.
Use the add_new_key (or ank for short) command to add a new principal to the authentication database. The command requires the principal identifier as an argument. The identifier given can be fully qualified using the standard name.instance@realm convention. You are asked to enter your admin password and are then prompted twice to enter the principal's new password. If a realm is not specified, the local realm is used unless another was given on the command line with the r flag. If no instance is specified, a null instance is used. If a realm other than the default realm is specified, you need to supply your admin password for the specified realm.
Use change_password to change a principal's password. The command requires the principal identifier as an argument. You are asked to enter your admin password and are then prompted twice to enter the principal's new password. The identifier given can be fully qualified using the standard name.instance@realm convention.
Use the change_admin_password to change your admin instance password. This command requires no arguments. It prompts you for your old admin password, then prompts you twice to enter the new admin password. If this is your first command, the default realm is used. Otherwise, the realm used in the last command is used.
Use destroy_tickets to destroy any admin tickets obtained by the kadmin command.
Use list_requests to get a list of possible commands.
Use help to display various kadmin help messages. If entered without an argument, help displays a general help message. You can get detailed information on specific kadmin commands by entering help command_name.
To quit the program, type quit.
Files
Related Information
Commands: add_principal, kadmind, kpasswd, ksrvutil
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Examples
The following contains an example of adding a user. To add a user, enter:
kadmin Welcome to the Kerberos Administration Program, version 2 Type "help" if you need it. admin: help Welcome to the Kerberos administration program.Type "?" to get a list of requests that are available. You can get help on each of the commands by typing "help command_name". Some functions of this program requires an "admin" password from you. This is a password private to you, that is used to authenticate requests from this program. You can change this password with the "change_admin_password" (or short form "cap") command. Good Luck! admin: ? Available admin requests: change_password, cpw Change a user's password change_admin_password, cap Change your admin password add_new_key, ank Add new user to kerberos database get_entry, get Get entry from kerberos database destroy_tickets, dest Destroy admin tickets help Request help with this program list_requests, lr, ? List available requests. quit, exit, q Exit program. admin: ank mroz Admin password: Password for mroz: Verifying, please re-enter Password for mroz: mroz added to database. admin: q Cleaning up and exiting.
Note: | Passwords are not echoed back to the user. |
Purpose
kadmind - Contains the daemon for authentication database administration.
Syntax
kadmind [-h] [-n] [-r realm] [-d db_name] [-f file_name] [-a acldir]
Flags
Note: | Use of the -r, -d, and -a flags with values other than the system defaults is not supported on the SP system. |
Operands
None.
Description
The kadmind daemon is the authentication database server for the password-changing and administration tools. It uses the master key for authorization.
The kadmind daemon listens for requests on the kerberos_master/tcp port. If this port is not defined in the /etc/services file, it uses port 751.
When performing requests on behalf of clients, kadmind checks access control lists (ACLs) to determine the authorization of the client to perform the requested action. Currently three distinct access types are supported:
Note: | A principal's private key is never returned by the get functions. |
Principals are always granted authorization to change their own password.
Files
Related Information
Commands: add_principal, kadmin, kpasswd
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Purpose
kdb_destroy - Destroys the authentication database.
Syntax
kdb_destroy
Flags
None.
Operands
None.
Description
The kdb_destroy command removes the authentication database.
You first must reply y or Y to a prompt to confirm the request, or kdb_destroy exits without removing the database files.
This command can only be issued on the system on which the authentication database resides.
Note: | This command does not remove database backup files created by the kdb_util command nor the /.k file created by the kstash command. |
Files
Security
You must have root privilege to run this command.
Related Information
Command: kdb_init
Purpose
kdb_edit - Edits the authentication database.
Syntax
kdb_edit [-n]
Flags
Operands
None.
Description
The kdb_edit command is used to create or change principals in the authentication database. It uses the master key for authorization.
After the master key is verified, kdb_edit begins a prompt loop. The user is prompted for the principal name and instance to be modified. If the entry is not found, the user can create it. After an entry is found or created, the user can set the password, expiration date, maximum ticket lifetime, and attributes. Default expiration dates, maximum ticket lifetimes, and attributes are presented in brackets. If the user presses return, the default is selected. There is no default password. The password RANDOM is interpreted specially, and if entered, the program selects a random key for the principal.
You should use random key generation only if you use the kdb_edit command to replace a deleted service principal (for example, rcmd.host_name).
If you enter a ticket lifetime value, it must be a number between 0 and 255. The actual maximum lifetime value that you choose will be between five minutes and 30 days. Refer to the IBM Parallel System Support Programs for AIX: Administration Guide for a complete list of the possible ticket lifetime values you can enter and the corresponding durations in days, hours, and minutes. The following list shows a representative sample with approximate durations:
Response to kdb_edit Approximate Duration 141 1 day 151 2 days 170 1 week 180 2 weeks 191 1 month
After the entry has been created or changed, "Edit O.K." is printed.
Files
Security
You must have root privilege to run this command.
Related Information
Commands: kadmin, kdb_init
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Examples
To add a service from host mroz, enter:
kdb_edit -n
Opening database... Previous or default values are in [brackets], enter return to leave the same, or new value. Principal name: rcmd Instance: mroz <Not found>, Create [y] ? Y Principal: rcmd, Instance: mroz, kdc_key_ver: 1 New Password: Verifying, please re-enter New Password: Principal's new key version = 1 Expiration date (enter yyyy-mm-dd) [1999-12-31] ? Max ticket lifetime [255] ? Attributes [0] ? Edit O.K. Program re-prompts for another principal "principal name:" Principal name: The program exits when no principal name is entered.
Note: | Passwords are not echoed back to the user. |
Purpose
kdb_init - Initializes the authentication database.
Syntax
kdb_init [realm]
Flags
None.
Operands
Description
Use this command to initialize the authentication database, creating the necessary initial system principals.
After determining the realm to be created, the command prompts for a master key password. The user should choose a nontrivial, not easily-guessable password. The user must remember this password because it is used for other commands. The master key password is used to encrypt every encryption key stored in the database.
Files
Security
You must have root privilege to run this command.
Related Information
Commands: kdb_destroy, kdb_edit, kdb_util
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Purpose
kdb_util - Contains the utility program for managing the authentication database.
Syntax
kdb_util operation file_name
Flags
None.
Operands
Description
The kdb_util command allows the user to perform various utility operations on the authentication database.
Files
Security
You must have root privilege to run this command.
Related Information
Commands: kdb_init, kprop, kpropd
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Purpose
kdestroy - Destroys authentication tickets.
Syntax
kdestroy [-f] [-q]
Flags
Operands
None.
Description
The kdestroy command destroys the user's authentication tickets. The command writes zeros to the user's current ticket cache file and then removes the file from the file system. If the file does not exist or if an error occurs, a message is displayed. The current ticket file is determined by the KRBTKFILE environment variable. If the KRBTKFILE environment variable is undefined, the current ticket file is /tmp/tktuid, where uid specifies your user identification number. If kdestroy cannot destroy the ticket file, the command warns you by making your terminal beep. You can place the kdestroy command in your .logout file (C shell only) so that your tickets are destroyed automatically when you log out.
Files
Related Information
Commands: kinit, klist
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Purpose
kerberos - Contains the authentication ticket-granting service daemon.
Syntax
Flags
Operands
Note: | Specification of a database path name other than the default, /var/kerberos/database/principal, is not supported on the SP system. |
Description
kerberos is the daemon program that provides the Authentication Service and the Ticket Granting Service to client programs that want to obtain tickets for authenticated services.
The kerberos daemon listens for requests on the kerberos4/upd port. If this port is not defined in the /etc/services file, it uses port 750.
When you start the server (normally from init), you can specify a maximum age for the database files. This can be used to ensure that you do not start a secondary server with out-of-date information. This could occur in a situation where a secondary server system was down when a database update was scheduled.
Files
Related Information
Commands: kdb_init, kprop, kpropd
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Purpose
kinit - Contains the SP login authentication program.
Syntax
kinit [-i] [-l] [-r] [-v] [name]
Flags
Operands
Description
The kinit command is used to authenticate the user's identify to the SP authentication service. All previous tickets are discarded.
When you use the kinit command without options, it prompts for your principal name and password, and tries to authenticate your identity within the local realm. If the specified principal name and password are correct, kinit retrieves your initial ticket and puts it in the ticket file specified by your KRBTKFILE environment variable. If the KRBTKFILE variable is undefined, your ticket is stored in the /tmp/tktuid file, where uid specifies your user identification number.
Note: | These tickets are shared by all processes running under the user's IDs. The KRBTKFILE environment variable can be set to change the location of the ticket cache file. |
If you specify the -l flag, the command prompts you to enter a ticket lifetime, in minutes. The actual value you enter will differ somewhat from the actual lifetime, because lifetimes are set to one of a discrete set of values ranging from five minutes to 30 days. kinit rounds the value you enter up to the next higher limit, and applies the maximum that is defined for your Kerberos principal. If you enter a value higher than your allowed limit, kinit does not indicate an error, but simply assigns your maximum lifetime in the ticket it creates. Refer to the IBM Parallel System Support Programs for AIX: Administration Guide for the complete list of maximum lifetime values that the administrator can set. The following list shows a representative sample of lifetimes you can request:
Response to kinit prompt Approximate duration 1500 1 day 3000 2 days 10000 1 week 20000 2 weeks 43000 1 month
Depending on your security policy, you may want to use the kdestroy command to destroy any active tickets before you end your login session. You can place the kdestroy command in your .logout file (C shell only) so that your tickets are destroyed automatically when you logout.
The KRBTKFILE environment variable is used to specify the ticket cache file used by kinit to store authentication tickets.
Files
Related Information
Commands: kdestroy, klist
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Purpose
klist - Lists currently held authentication tickets.
Syntax
klist [-s | -t] [-file name] [-srvtab]
Parameters
Operands
None.
Description
The klist command prints the principal name and the name of the file containing the user's tickets. It also lists the principal name, issue time, and expiration time for each service ticket held by the user. Principal names are listed in the form name.instance@realm. The period (.) is omitted if the instance is null and the at sign (@) is omitted if the realm is null.
Files
Related Information
Commands: kdestroy, kerberos, kinit
Examples
# klist Ticket file: /tmp/tkt0 Principal: root.admin@XYZ.ABC.COM Issued Expires Principal Nov 12 16:26:11 Dec 12 16:26:11 krbtgt.XYZ.ABC.COM@XYZ.ABC.COM Nov 12 16:26:46 Dec 12 16:26:46 hardmon.cwksta@XYZ.ABC.COM Nov 12 16:45:15 Dec 12 16:45:15 rcmd.cwksta@XYZ.ABC.COM #
The second line shows the Kerberos principal acting as client, to whom the tickets belong. This is the user principal you supplied to the kinit command, or the rcmd.instance service principal used by rcmdtgt. The list of tickets always begins with the ticket-granting-ticket. The others are service tickets; in this case for the System Monitor service on the control workstation (hardmon) and the SP Remote Command service also on the control workstation (rcmd).
# klist -srvtab Server key file: /etc/krb-srvtab Service Instance Realm Key Version ------------------------------------------------------ rcmd node3fi XYZ.ABC.COM 1 rcmd node3tr XYZ.ABC.COM 1 rcmd node3sw XYZ.ABC.COM 1 rcmd node3en XYZ.ABC.COM 1 #
You can determine the versions of service keys in the authentication database by locating the entry for the target service principal in a dump of the SP authentication database. If you have secondary authentication servers, or if you use the procedure for backing up your database that IBM suggests using in IBM Parallel System Support Programs for AIX: Administration Guide, the database dump can be found in file /var/kerberos/database/slavesave on the primary server host.
Purpose
kpasswd - Changes the principal's password.
Syntax
Flags
Operands
None.
Description
The kpasswd command changes a principal's password.
It prompts for the principal's current password. If the old password is correct, the user is prompted twice for a new password. A message is printed indicating the success or failure of the password changing operation.
Related Information
Commands: kadmin, kinit, passwd
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Purpose
kprop - Specifies the network utility to propagate the authentication database to secondary servers.
Syntax
kprop [-force] [-realm realm_name] data_file hosts_file
Flags
Operands
Description
The kprop command reads a list of secondary host names and connects to each one in turn using the kprop service provided by the kpropd program. The data_file (the authentication database) is transferred if it has been modified since it was last sent successfully.
Files
Related Information
Commands: kdb_util, kerberos, kpropd
Purpose
kpropd - Contains the daemon to receive updates for a secondary authentication database.
Syntax
kpropd [-r realm] [-s srvtab] [-l log_file] [-d database_name] file_name
Flags
Note: | Use of the -r, -s, and -d flags with values other than the system defaults is not supported on the SP system. |
Operands
Description
kpropd runs as a daemon on secondary authentication database server hosts, listening for a TCP connection on the krb_prop service.
The kpropd daemon listens for requests on the krb_prop/tcp port. If this port is not defined in the /etc/services file, it uses port 754. It validates the connection, which must be from an administrative host as defined in the krb.conf file for the local realm. The service name used for mutual authentication is rcmd.
Files
Related Information
Commands: kdb_util, kerberos, kprop
Purpose
kshd - Provides the server function for remote command execution.
Syntax
kshd kshd [-s] program_name
The kshd daemon is normally started by the inetd daemon.
Flags
Operands
Description
The kshd daemon is the server for the SP rcp and rsh commands. It provides remote execution of shell commands. These commands are based on requests from privileged sockets on trusted hosts. The shell commands must have user authentication. The kshd daemon listens at the socket defined in the /etc/services file and in the InetServ object class with the command entry.
The kshd daemon is started by default. The inetd daemon no longer reads the /etc/inetd.conf file, although this file still exists. Instead, the inetd daemon gets its information from the InetServ object class (stored in the AIX Object Data Manager). This object class is a combination of the information in the /etc/inetd.conf file and the /etc/services file. InetServ is created at install time from information in these two files.
If you have already set up the kshd daemon using the /etc/inetd.conf file, or if you are accustomed to using this file and want to continue doing so, you can. However, the InetServ object class and the /etc/services and /etc/inetd.conf files must be kept synchronized. If you modify the /etc/inetd.conf or the /etc/services file, you need to run the inetimp command to apply those changes to the InetServ object class. Then run the refresh -s inetd command to immediately update the inetd daemon.
When the kshd daemon receives a service request, it initiates the following protocol:
The kshd daemon first attempts to authenticate the requester using the key of the rcmd service instance whose name is the local host name. If that fails, kshd attempts to authenticate using each of the service keys for the other instances of the service. A separate service instance exists for each network interface through which the server may be reached.
This daemon does not support encryption by users.
Files
Prerequisite Information
Related Information
SP Commands: kinit, rcp, rsh
AIX Commands: inetimp, rsh
AIX Daemon: inetd
Purpose
ksrvtgt - Obtains a ticket-granting-ticket using a service key.
Syntax
ksrvtgt name instance [[realm] srvtab]
Flags
None.
Operands
Note: | Specification of a srvtab file other than the system default is not supported on the SP system. |
Description
The ksrvtgt command retrieves a ticket-granting-ticket with a lifetime of five minutes, decrypts the response using the service key found in the service key file, and stores the ticket in the standard ticket cache.
This command is intended primarily for use in shell scripts and other batch-type facilities.
The KRBTKFILE environment variable is used to specify the ticket cache file used by ksrvtgt to store authentication tickets.
Files
Security
You must have root privilege to run this command.
Related Information
Commands: kdestroy, kinit
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Purpose
ksrvutil - Manipulates a server key file.
Syntax
ksrvutil [-afs | -krb] [-k ] [-i] [-f file_name] operation
Flags
If neither -afs nor -krb are specified, the value of the System Data Repository (SDR) authent_server attribute is used. If the value of the SDR authent_server attribute cannot be obtained, the default is -krb.
Note: | Specification of a srvtab file other than the system default is not supported on the SP system. |
Operands
Description
The ksrvutil command allows an administrator to list or change keys currently in the key file or to add new keys to the keyfile.
The ksrvutil command always backs up the key file before making any changes. If ksrvutil fails during a change or add operation, you can recover a usable key file by appending the workfile containing the new and changed keys, file_name.work to the backup copy or the original, file_name.old, and replacing the key file file_name with the result, for example:
cat /etc/krb-srvtab.old /etc/krb-srvtab.work >/etc/krb-srvtab
The recovered key file can be used, but it may contain some out-of-date keys.
Files
Security
You must have root privilege to run this command.
Related Information
Commands: kadmin, ksrvtgt, rcmdtgt
Purpose
kstash - Saves the system's authentication master key.
Syntax
kstash
Flags
None.
Operands
None.
Description
The kstash command saves the system's authentication database master key in the master key cache file. The user is prompted to enter the master key (the same one as specified to kdb_init) to verify the authenticity of the key and authorize caching it.
Files
Security
You must have root privilege to run this command.
Related Information
Purpose
locate_jm - Identifies the nodes on which the primary and backup Resource Manager servers are running within the current system partition.
Syntax
locate_jm
Flags
None.
Operands
None.
Description
Use this command to identify the nodes on which the primary and backup Resource Manager servers are running. This command must be executed by root and operates within the scope of the current system partition environment.
Files
Related Information
Commands: jm_config, jm_install_verify, jm_start, jm_status, jm_stop
Examples
To locate the Resource Manager servers within the current system partition, enter:
locate_jmYou should receive a response similar to the following:
locate_jm: The primary JM server is running on r07n01. locate_jm: The backup JM server is running on r07n03.
Purpose
lppdiff - Queries installed Licensed Program Products (LPPs) on a group of hosts.
Syntax
Flags
Operands
Description
Use this command to query the status of installed LPPs on a group of hosts. The output from each host is collected and identical results are compressed to show the names and a count of the hosts that had identical results.
The dsh command is used to execute the queries on the remote hosts. The lslpp command is used to get the status of the installed LPPs on the remote hosts. The lslpp command is called on each host with the -l, -a, -c, and -q flags.
Output from the lppdiff command consists of one entry for each unique LPP listing information about that LPP. Each LPP's entry is followed by a list of all hosts that have that LPP installed. An LPP is considered unique if any one of the components in its description differ from that of another. For example, consider two hosts that both have ssp.basic installed. On host 1, it is in the APPLY state and on host 2, it is in the COMMITTED state. These LPPs are considered unique and, therefore, each will get its own set of output from lppdiff.
The flags for lppdiff are used to direct the dsh command to certain hosts and to control its behavior. See the dsh command for details on these flags and how to use them.
The fileset operand to lppdiff can be one of two things. It can either be all which queries and displays information about all LPPs installed on the specified hosts, or it can be the name of a file set to query on the specified hosts. The "*" character can be used to specify multiple file sets. For example, lppdiff -Ga ssp.* queries any file sets starting with "ssp." on all hosts in the system.
Examples
[k22s] > lppdiff -a ssp.basic
You should receive output similar to the following:
---------------------------------------------------------------------------- Name Path Level PTF State Type Num ---------------------------------------------------------------------------- LPP: ssp.basic /etc/objrepos 2.2.0.0 COMMITTED I 5 From: k22n09.ppd.pok.ibm.com k22n10.ppd.pok.ibm.com k22n11.ppd.pok. ibm.com k22n12.ppd.pok.ibm.com k22n13.ppd.pok.ibm.com ---------------------------------------------------------------------------- LPP: ssp.basic /usr/lib/objrepos 2.2.0.0 COMMITTED I 5 From: k22n09.ppd.pok.ibm.com k22n10.ppd.pok.ibm.com k22n11.ppd.pok. ibm.com k22n12.ppd.pok.ibm.com k22n13.ppd.pok.ibm.com
[k22s] > lppdiff -w k22n01 X11.base* ----------------------------------------------------------------------------- Name Path Level PTF State Type Num ----------------------------------------------------------------------------- LPP: X11.base.rte /etc/objrepos 4.1.4.0 COMMITTED I 1 From: k22n01 ----------------------------------------------------------------------------- LPP: X11.base.smt /etc/objrepos 4.1.4.0 COMMITTED I 1 From: k22n01 ----------------------------------------------------------------------------- LPP: X11.base.comm /usr/lib/objrepos 4.1.0.0 COMMITTED I 1 on From: k22n01 ----------------------------------------------------------------------------- LPP: X11.base.lib /usr/lib/objrepos 4.1.4.0 COMMITTED I 1 From: k22n01 ----------------------------------------------------------------------------- LPP: X11.base.rte /usr/lib/objrepos 4.1.4.0 COMMITTED I 1 From: k22n01 ----------------------------------------------------------------------------- LPP: X11.base.smt /usr/lib/objrepos 4.1.4.0 COMMITTED I 1 From: k22n01
[k22s] > lppdiff -Ga ssp.clients ssp.bogus ----------------------------------------------------------------------------- Name Path Level PTF State Type Num ----------------------------------------------------------------------------- LPP: ssp.clients /etc/objrepos 2.1.0.0 COMMITTED I 4 From: k22n03.ppd.pok.ibm.com k22n04.ppd.pok.ibm.com k22n07.ppd.pok. ibm.com k22n08.ppd.pok.ibm.com ----------------------------------------------------------------------------- LPP: ssp.clients /etc/objrepos 2.1.0.6 APPLIED F 4 From: k22n03.ppd.pok.ibm.com k22n04.ppd.pok.ibm.com k22n07.ppd.pok. ibm.com k22n08.ppd.pok.ibm.com ----------------------------------------------------------------------------- LPP: ssp.clients /etc/objrepos 2.2.0.0 COMMITTED I 8 From: k22n01.ppd.pok.ibm.com k22n05.ppd.pok.ibm.com k22n06.ppd.pok. ibm.com k22n09.ppd.pok.ibm.com k22n10.ppd.pok.ibm.com k22n11.ppd.pok.ibm.com k22n12.ppd.pok.ibm.com k22n13.ppd. pok.ibm.com ----------------------------------------------------------------------------- LPP: ssp.clients /usr/lib/objrepos 2.1.0.0 COMMITTED I 4 From: k22n03.ppd.pok.ibm.com k22n04.ppd.pok.ibm.com k22n07.ppd.pok. ibm.com k22n08.ppd.pok.ibm.com ----------------------------------------------------------------------------- LPP: ssp.clients /usr/lib/objrepos 2.1.0.6 APPLIED F 4 From: k22n03.ppd.pok.ibm.com k22n04.ppd.pok.ibm.com k22n07.ppd.pok. ibm.com k22n08.ppd.pok.ibm.com ----------------------------------------------------------------------------- LPP: ssp.clients /usr/lib/objrepos 2.2.0.0 COMMITTED I 8 From: k22n01.ppd.pok.ibm.com k22n05.ppd.pok.ibm.com k22n06.ppd.pok. ibm.com k22n09.ppd.pok.ibm.com k22n10.ppd.pok.ibm.com k22n11.ppd.pok.ibm.com k22n12.ppd.pok.ibm.com k22n13.ppd. pok.ibm.com ========================= Errors ============================================ Error: /bin/lslpp: Fileset ssp.bogus not installed. From: k22n01.ppd.pok.ibm.com k22n05.ppd.pok.ibm.com k22n06.ppd.pok. ibm.com k22n03.ppd.pok.ibm.com k22n04.ppd.pok.ibm.com k22n07.ppd.pok.ibm.com k22n08.ppd.pok.ibm.com k22n09.ppd. pok.ibm.com k22n10.ppd.pok.ibm.com k22n11.ppd.pok.ibm.com k22n12.ppd.pok.ibm.com k22n13.ppd.pok.ibm.com -----------------------------------------------------------------------------
[k22s] > lppdiff -Gac ssp.clients ssp.bogus From:Name:Path:Level:PTF:State:Type:Num k22n03.ppd.pok.ibm.com,k22n04.ppd.pok.ibm.com,k22n07.ppd.pok.ibm.com, k22n08.ppd.pok.ibm.com:ssp.clients:/etc/objrepos:2.1.0.0::COMMITTED:I:4 k22n03.ppd.pok.ibm.com,k22n04.ppd.pok.ibm.com,k22n07.ppd.pok.ibm.com, k22n08.ppd.pok.ibm.com:ssp.clients:/etc/objrepos:2.1.0.6::APPLIED:F:4 k22n01.ppd.pok.ibm.com k22n05.ppd.pok.ibm.com k22n06.ppd.pok.ibm.com, k22n09.ppd.pok.ibm.com k22n10.ppd.pok.ibm.com k22n11.ppd.pok.ibm.com, k22n12.ppd.pok.ibm.com k22n13.ppd.pok.ibm.com:ssp.clients:/etc/objrepos: 2.2.0.0::COMMITTED:I:8 k22n03.ppd.pok.ibm.com k22n04.ppd.pok.ibm.com k22n07.ppd.pok.ibm.com, k22n08.ppd.pok.ibm.com:ssp.clients:/usr/lib/objrepos:2.1.0.0:: COMMITTED:I:4 k22n03.ppd.pok.ibm.com k22n04.ppd.pok.ibm.com k22n07.ppd.pok.ibm.com, k22n08.ppd.pok.ibm.com:ssp.clients:/usr/lib/objrepos:2.1.0.6::APPLIED: F:4 k22n01.ppd.pok.ibm.com k22n05.ppd.pok.ibm.com k22n06.ppd.pok.ibm.com, k22n09.ppd.pok.ibm.com k22n10.ppd.pok.ibm.com k22n11.ppd.pok.ibm.com, k22n12.ppd.pok.ibm.com k22n13.ppd.pok.ibm.com:ssp.clients: /usr/lib/objrepos:2.2.0.0::COMMITTED:I:8 From:Error k22n01.ppd.pok.ibm.com k22n05.ppd.pok.ibm.com k22n06.ppd.pok.ibm.com, k22n03.ppd.pok.ibm.com k22n04.ppd.pok.ibm.com k22n07.ppd.pok.ibm.com, k22n08.ppd.pok.ibm.com k22n09.ppd.pok.ibm.com k22n10.ppd.pok.ibm.com, k22n11.ppd.pok.ibm.com k22n12.ppd.pok.ibm.com k22n13.ppd.pok.ibm.com: /bin/lslpp: Fileset ssp.bogus not installed.
Purpose
lsfencevsd - Lists IBM Virtual Shared Disks that are fenced from access by nodes.
Syntax
lsfencevsd
Flags
None.
Operands
None.
Description
Use this command to display a map that shows which IBM Virtual Shared Disks are fenced from which nodes in the system or system partition.
Files
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: fencevsd, unfencevsd
Examples
To display the map of fenced IBM Virtual Shared Disks in the system, enter:
lsfencevsd
The system displays a map similar to the following:
minor Fenced Nodes (13): 13 14 (14): 1 2
Purpose
lshacws - Gets the state of the control workstation.
Syntax
lshacws
Flags
None.
Operands
None.
Description
Use this command to print the current state of the control workstation. It prints to standard output a number string that indicates the state of the primary or backup control workstation and whether the control workstation is a high availability configuration.
This command is valid only when issued on the control workstation.
When the command is executed and the calling process is not on a control workstation, an error occurs.
Note: | The High Availability Cluster Multiprocessing (HACMP) event scripts and installation scripts supplied with the High Availability Control Workstation (HACWS) option of the IBM Parallel System Support Programs for AIX (PSSP), set the control workstation state. The state is changed during fail over or reintegration in the HACWS supplied pre- and post-event scripts for HACMP. The administrator should not normally have to set the control workstation state. |
Exit Values
The following are the valid printed values and their defined control workstation state:
Prerequisite Information
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for information on the HACWS option.
Location
/usr/bin/lshacws
Related Information
Command: sethacws
Subroutines: hacws_set, hacws_stat
Examples
lshacws Results: 32
lshacws Results: 16
lshacws Results: 2
lshacws Results: 1
lshacws Results: 0
lshacws Results: An error occurs and the exit value = 3
Purpose
lshsd - Displays configured data striping device (HSD) for the IBM Virtual Shared Disks and their characteristics.
Syntax
lshsd [-l | -s] [hsd_name ...]
Flags
Operands
Description
This command displays the configured HSDs. If a list of HSDs follow the flag then information about them is displayed. lshsd without any arguments or flag lists the names of all the HSDs currently configured.
Files
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfghsd, ucfghsd
Examples
lshsd
The system displays a message similar to the following:
hsd1 hsd2 hsd3 ·
lshsd -l hsd1 hsd2
The system displays a message similar to the following:
HSD_name=hsd1 Stripe_size=32768 Hsd_minorno=1 numVsds=2 option=protectlvcb size_in_MB=40 vsd.rlv01 vsd.rlv02 HSD_name=hsd2 Stripe_size=32768 Hsd_minorno=1 numVsds=3 option=protectlvcb size_in_MB=40 vsd.rlv03 vsd.rlv04 vsd.rlv05
lshsd -s hsd1The system displays a message similar to the following:
9 hsd parallelism 0 READ requests not at page boundary 0 WRITE requests not at page boundary HSD_name=hsd1 Stripe_size=4096 HSD_minorno=1 numVSDs=2 option=protect_lvcb size_in_MB=40 number_read number_write vsd_name 16 16 vsdn01v1 16 16 vsdn02v1
Purpose
lskp - Lists Kerberos principals.
Syntax
lskp [-h | -p | -s | -c | {name[.instance]|name.|.instance} ...]
Flags
Operands
Note: | The name must be followed by a period and the instance must be preceded by a period. |
Description
Use this command to list principals in the local Kerberos database, displaying for each the principal name and instance, the maximum ticket lifetime, and the expiration date. You can list the entire authentication database, an individual entry, all entries with a specified principal name, or all entries with a specified instance. Or you can list entries in different categories: all client (user) principals, all service principals, or all principals predefined by Kerberos.
Files
Standard Output
For each principal, the lskp command displays the principal identifier as name.instance (on a separate line if its length exceeds twenty characters), and the principal's attributes. The maximum ticket lifetime is the maximum period that a Ticket-Granting-Ticket issued to this principal will be valid. Any ticket lifetime up to this value can be requested using an option on the kinit command. The key version is an integer set to one when the principal is created and incremented each time the password is changed. The principal's expiration date is displayed in local time, based on the setting of the TZ environment variable.
Exit Values
Security
The lskp command can be run by the root user logged in on a Kerberos server host. It can be invoked indirectly as a Sysctl procedure by a Kerberos database administrator who has a valid ticket and is listed in the admin_acl.get file.
Location
/usr/kerberos/etc/lskp
Related Information
Commands: chkp, kadmin, kdb_edit, mkkp, rmkp, sysctl
Examples
lskp -p
You should receive output similar to the following:
krbtgt.ABC.DEF.GHI.COM tkt-life: 30d key-vers: 1 expires: 2037-12-31 23:59 default tkt-life: 30d key-vers: 1 expires: 2037-12-31 23:59 changepw-kerberos tkt-life: 30d key-vers: 1 expires: 2037-12-31 23:59 K.M tkt-life: 30d key-vers: 1 expires: 2037-12-31 23:59
lskp joe.admin lisa
You should receive output similar to the following:
joe.admin tkt-life: 15d+08:46 key-vers: 1 expires: 2005-03-15 23:59 lisa tkt-life: 08:00 key-vers: 1 expires: 1997-06-09 23:59
Purpose
lsvsd - Displays configured IBM Virtual Shared Disks and their characteristics.
Syntax
lsvsd [-l | -s] [-i] [vsd_name...]
Flags
The state field can have one of the following values:
This flag is not compatible with the -s flag.
The local logical operations are requests which were made by a process executing at the local node, whereas the remote logical operations were made by a process executing on a remote node. Client operations are those local logical requests that cannot be satisfied locally, and have to be sent to a remote node. Physical operations are those server operations which must be passed to the underlying disk device. Cache read hits are those server reads which do not require a device read, because the read operation was satisfied from the IBM Virtual Shared Disk cache.
This flag is not compatible with the -l flag.
Operands
Description
The lsvsd command displays information about IBM Virtual Shared Disks currently configured on the node on which the command is run. If a list of IBM Virtual Shared Disks follows the flags, information about those IBM Virtual Shared Disks is displayed. lsvsd with no arguments or flags lists the names of all the IBM Virtual Shared Disks currently configured on the node.
The lsvsd command displays information about both the configuration and the usage of an IBM Virtual Shared Disk.
You can use the System Management Interface Tool (SMIT) to run the lsvsd command. To use SMIT, enter:
smit vsd_mgmtand select the Show All Managed IBM Virtual Shared Disk Characteristics option.
Files
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfgvsd, ctlvsd, preparevsd, resumevsd, startvsd, stopvsd, suspendvsd, ucfgvsd
Examples
lsvsdThe system displays a message similar to the following:
vsd00 vsd01 ·
lsvsd -lThe system displays a message similar to the following:
minor state server lv_major lv_minor vsd_name option size (MB) 83 STP -1 0 0 vsdn08v3 cache 20 84 STP -1 0 0 vsdn08v4 nocache 16
lsvsd -sThe system displays a message similar to the following:
lc-rd lc-wt rm-rd rm-wt c-rd c-wt p-rd p-wt h-rd br bw vsd_name 84 84 2858 169 0 0 348 253 2605 164 184 vsd.vsd1 0 0 0 0 0 0 0 0 0 0 0 vsd.rl01 0 0 0 0 0 0 0 0 0 0 0 vsd.rl02
The following table spells out the names of the headers used in the
displays for the -l and -s options:
Header | Meaning |
---|---|
minor | IBM Virtual Shared Disk minor number |
state | State of this IBM Virtual Shared Disk:active, stopped, suspended |
server | Primary node for this IBM Virtual Shared Disk |
lv major | Logical volume major number |
lv minor | Logical volume minor number |
vsd_name | Name of this IBM Virtual Shared Disk |
option | Option:cache or nocache |
lc-rd | Local logical reads |
lc-wt | Local logical writes |
rm-rd | Remote logical reads |
rm-wt | Remote logical writes |
c-rd | Client logical reads |
c-wt | Client logical writes |
p-rd | Physical reads |
p-wt | Physical writes |
h-rd | Reads from cache |
br | Blocks read |
bw | Blocks written |
Purpose
mkamdent - Creates user home directory entries in the /u automounter map files.
Syntax
mkamdent [-s server_path] user_names
Flags
Operands
Description
Use this command to create user home directory entries in the /u automounter map files. Typically, user home directory entries are generated by the SP User Management Services when a new user is added to the system. However, if SP User Management Services are turned off and SP Automounter Support is still turned on, this command can be used to add user entries to the automounter /u map. This command can also be used to add automounter support for preexisting users that were not added using SP User Management Services and for /u subdirectories that are not associated with SP users.
Files
Examples
To create automounter entries in the /u map file for multiple users, enter:
mkamdent -s hostx:/home/hostx john ken pat paul ron
This assumes the following directories already exist on hostx:
Related Information
The "Managing the Automounter" and "Managing User Accounts" chapters in IBM Parallel System Support Programs for AIX: Administration Guide.
Commands: spsitenv
Purpose
mkautomap - Generates an equivalent Automount map file from an Amd map file.
Syntax
mkautomap [-n] [-o Automount_map] [-f filesystem] [Amd_map]
Flags
Operands
Description
The mkautomap command is a migration command used to generate an Automount map file from the Amd map file Amd_map created by a previous SP release. Only Amd map file entries created by a previous SP release will be recognized. If the Amd map file was modified by the customer, results may be unpredictable. If an Amd map entry cannot be properly interpreted, a message will be written to standard error, and that entry will be ignored. Processing will continue with the next map entry. All recognized entries will be interpreted and equivalent Automount map entries will be written to a temporary file Automount_map.tmp. If no errors were encountered during processing, the temporary file will be renamed to Automount_map.
If all Amd map entries were successfully generated into Automount map entries and written to Automount_map, the /etc/auto.master Automount master file will be updated unless the -n flag is specified. A master map file entry associating the filesystem with the Automount_map will be added. Also, any default mount options specified in Amd_map will be added to the master map file entry for filesystem. This master map file entry will be appended to /etc/auto.master and if the file does not exist, it will be created.
Files
Restrictions
Use this command only with amd.u map files created by PSSP User Management Services. Using other Amd map files or modified amd.u map files as input to this command, will produce unpredictable results.
Related Information
The "Migrating to the Latest Level of PSSP" chapter in IBM Parallel System Support Programs for AIX: Installation and Migration Guide
The "Managing the Automounter" chapter in IBM Parallel System Support Programs for AIX: Administration Guide
Examples
To create the SP Automount /u map file from the Amd map file generated by a previous SP release, enter:
mkautomap
Purpose
mkconfig - Creates the config_info file for each of the boot/install server's clients on the server.
Syntax
mkconfig
Flags
None.
Operands
None.
Description
Use this command to make the config_info files for all the clients of a boot/install server if the client is not set to boot from disk. The mkconfig command is intended to run only on the server node. This command creates a config_info file named /tftpboot/host_name.config_info file.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/mkconfig
Related Information
Commands: setup_server
Examples
To make the config.info files for all boot/install clients of a server, enter on the server:
mkconfig
Purpose
mkinstall - Creates the install_info file for each of the server's clients on the server.
Syntax
mkinstall
Flags
None.
Operands
None.
Description
Use this command on the server node to make the install_info files for all clients of a boot/install server. The mkinstall command creates a /tftpboot/host_name.install_info file.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/mkinstall
Related Information
Commands: setup_server
Examples
To make the install.info files for all boot/install clients of a server, enter on the server:
mkinstall
Purpose
mkkp - Makes Kerberos principals.
Syntax
mkkp -h
mkkp [-e expiration] [-l lifetime] name[.instance] ...
Flags
lifetime operand - Approximate duration 141 1 day 151 2 days 170 1 week 180 2 weeks 191 1 month
Operands
Description
Use this command to create principals in the Kerberos database on the local host. It allows the default values for the expiration date and maximum ticket lifetime to be overridden. Principals created in this way have no passwords. Before a user can kinit as the new principal, an administrator must set your initial password using the kpasswd, kadmin, or kdb_edit command directly. This command should normally be used only on the primary server. If there are secondary authentication servers, the push-kprop command is invoked to propagate the change to the other servers. The command can be used to update a secondary server's database, but the changes may be negated by a subsequent update from the primary.
Files
Exit Values
Security
The mkkp command can be run by the root user logged in on a Kerberos server host. It can be invoked indirectly as a Sysctl procedure by a Kerberos database administrator who has a valid ticket and is listed in the admin_acl.add file.
Location
/usr/kerberos/etc/mkkp
Related Information
Commands: chkp, kadmin, kdb_edit, kpasswd, lskp, rmkp, sysctl
Examples
The following example adds two principals to the database. Both principals are set to expire 30 June 2005. The default value for the maximum ticket lifetime is used.
mkkp -e 2005-06-30 kelly kelly.admin
Purpose
mknimclient - Makes a node a Network Installation Management (NIM) client of its boot/install server.
Syntax
mknimclient -h | -l node_list
Flags
Operands
None.
Description
Use this command to define a node as a NIM client. This is accomplished by determining the node's boot/install server from the System Data Repository (SDR) and configuring that client node as a NIM client on that server. When complete, the NIM configuration database on the server contains an entry for the specified client.
Notes:
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/mknimclient
Related Information
Commands: delnimclient, setup_server
Examples
To define nodes 1, 3, and 5 as NIM clients of their respective boot/install servers, enter:
mknimclient -l 1,3,5
Purpose
mknimint - Creates the necessary Network Installation Management (NIM) interfaces on a NIM master.
Syntax
mknimint -h | -l node_list
Flags
Operands
None.
Description
Use this command to define to NIM new Ethernet network adapters and interfaces on the control workstation and boot/install servers. On the control workstation, any networks not previously defined are defined and NIM interfaces added. On a boot/install server, all the Ethernet networks and interfaces are defined; it then defines all token ring and Ethernet networks that are known on the control workstation (with the netstat -ni command) and defines interfaces for them as well. This is so that resources like the lppsource can be served from the control workstation to a client node by the boot/install server if the client and control workstation are on the same subnetwork.
To serve a resource to a client that is not on the same subnetwork as the control workstation, routing is required. Routing is done in mknimclient.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/mknimint
Related Information
Commands: setup_server
Examples
To make NIM interface definitions for nodes 1, 3, and 5, enter:
mknimint -l 1,3,5
Purpose
mknimmast - Configures a node as a Network Installation Management (NIM) master.
Syntax
mknimmast -h -l node_list
Flags
Operands
None.
Description
Use this command to define a boot/install server node as a NIM master for the subsequent installation of client nodes. It verifies that the listed nodes are defined as boot/install servers in the System Data Repository (SDR). It then installs the NIM master AIX file sets and configures the nodes as NIM masters.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/mknimmast
Related Information
Commands: delnimmast, setup_server
Examples
To define nodes 1, 3, and 5 as NIM masters, enter:
mknimmast -l 1,3,5
Purpose
mknimres - Creates the necessary Network Installation Management (NIM) resources on a NIM master.
Syntax
mknimres -h | -l node_list
Flags
Operands
None.
Description
Use this command to make all the NIM resources for installation, diagnostics, migration, and customization. No resources are allocated to client nodes. The set of resources needed is determined from the list of client nodes found in the System Data Repository (SDR) for the node_list. Any required AIX install and mksysb images are defined as NIM resources. For boot/install server nodes, NIM Shared Product Object Tree (SPOT) directories are created and mksysb images are copied, as required. Because of the large data volumes required for SPOTs and install images, all checking is done before copying data.
Creation of the NIM lppsource resource on a boot/install server will result in setup_server creating a lock in the lppsource directory on the control workstation.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/mknimres
Related Information
Commands: setup_server
Examples
To make NIM resources for boot/install servers 1, 3, and 5, enter:
mknimres -l 1,3,5
Purpose
ngaddto - Adds nodes and node groups to the definition list of the destination node group.
Syntax
Flags
Operands
Description
Use this command to add nodes and node groups to the definition list of the destination node group. If the -G flag is specified, the destination node group must be global. If the -G flag is not specified, the destination node group must belong to the current system partition. If the destination node group does not exist, you will receive an error. If the destination node group or nodegroup is a name that is not valid, you will receive an error. Nodes and node groups that do not currently exist can be added to the destination node group. When the node group is resolved by the ngresolve command, nonexistent members are ignored.
Exit Values
Security
You must have root privilege to modify named node groups.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
Refer to the "Managing Node Groups" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional node grouping information.
Location
/usr/lpp/ssp/bin/ngaddto
Related Information
Commands: ngcreate, ngdelete, ngdelfrom, ngfind, nglist, ngnew, ngresolve
Examples
ngaddto nga 1 3 ngb
ngaddto -G g1 1 16 g2
Purpose
ngclean - Cleans up a node group, removing references to nodes and node groups that are not in the current system partition. Node groups with empty definition lists will be deleted.
Syntax
ngclean [-h] | [-G] [-r] {-a | nodegroup [nodegroup ...]}
Flags
Operands
Description
Use this command to examine node group definition lists and to remove references to nodes and node groups that do not exist in the current system partition or the SP system if -G is supplied. Node groups with empty definition lists will be deleted. If the -r flag is specified, the nodes and node groups will not be removed, but a report will be generated.
Exit Values
Security
You must have root privilege to modify a named node group.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
Refer to the "Managing Node Groups" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional node grouping information.
Location
/usr/lpp/ssp/bin/ngclean
Related Information
Command: ngaddto, ngcreate, ngdelete, ngdelfrom, ngfind, nglist, ngnew, ngresolve
Examples
ngclean -Ga
ngclean my.ng
Purpose
ngcreate - Creates and optionally populates a named node group.
Syntax
Flags
Operands
Description
Use this command to create a node group named dest_nodegroup. The destination node group is populated based on the supplied options. Node group names must begin with a letter and can be followed by any letters or numbers, a period (.), or an underscore (_). If the destination node group already exists, you will receive an error.
Exit Values
Security
You must have root privilege to create and modify named node groups.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
Refer to the "Managing Node Groups" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional node grouping information.
Location
/usr/lpp/ssp/bin/ngcreate
Related Information
Commands: ngaddto, ngdelete, ngdelfrom, ngfind, nglist, ngnew, ngresolve
Examples
To create a node group called sample_ng that contains all the nodes in the current system partition except for k22n01, enter:
ngcreate -ae k22n01 sample_ng
Purpose
ngdelete - Removes node groups from persistent storage.
Syntax
ngdelete [-h] | [-u] [-G] nodegroup [nodegroup ...]
Flags
Operands
Description
Use this command to remove node groups from persistent storage. By default, the node group is removed from any node group that contains it. If the -u flag is specified, references to this deleted node group will remain in containing node groups.
Exit Values
Security
You must have root privilege to delete node groups.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
Refer to the "Managing Node Groups" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional node grouping information.
Location
/usr/lpp/ssp/bin/ngdelete
Related Information
Commands: ngaddto, ngcreate, ngdelfrom, ngfind, nglist, ngnew, ngresolve
Examples
To delete nodegroups ngc and ngd, enter:
ngdelete ngc ngd
Purpose
ngdelfrom - Deletes nodes and node groups from the definition list of the destination node group.
Syntax
Flags
Operands
Note: | Nodes numbers and node group names being removed can be intermixed. |
Description
Use this command to remove nodes and node groups from the definition list of the destination node group. If the -G flag is specified, the dest_nodegroup must be global. If the -G flag is not specified, the dest_nodegroup must belong to the current system partition.
Exit Values
Security
You must have root privilege to modify named node groups.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
Refer to the "Managing Node Groups" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional node grouping information.
Location
/usr/lpp/ssp/bin/ngdelfrom
Related Information
Commands: ngaddto, ngcreate, ngdelete, ngfind, nglist, ngnew, ngresolve
Examples
To remove node 5 and node group ngc from nga, enter:
ngdelfrom nga 5 ngc
Purpose
ngfind - Returns a list of all node groups whose definition list contains the specified node or node group.
Syntax
ngfind [-h] | [-G] nodegroup | node
Flags
Operands
Description
Use this command to list all node groups that contain the specified node or node group in their definition list. If the specified node group does not exist, you will receive an error. Use this command to determine what other node groups would be affected by changes to the specified node group.
Exit Values
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
Refer to the "Managing Node Groups" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional node grouping information.
Location
/usr/lpp/ssp/bin/ngfind
Related Information
Commands: ngaddto, ngcreate, ngdelete, ngdelfrom, nglist, ngnew, ngresolve
Examples
To display a list of all node groups that contain node group test_B, enter:
> ngfind test_B test_A test_D
Purpose
nglist - Returns a list of all node groups in the current system partition.
Syntax
nglist [-h] | [-G]
Flags
Operands
None.
Description
Use this command to list all node groups in the current system partition to standard output. If the -G flag is specified, it will list all system node groups.
Standard Output
A list of node groups is written to standard output, one node group per line.
Exit Values
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
Refer to the "Managing Node Groups" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional node grouping information.
Location
/usr/lpp/ssp/bin/nglist
Related Information
Commands: ngaddto, ngcreate, ngdelete, ngdelfrom, ngfind, ngnew, ngresolve
Examples
> nglist nga ngb sampleng test_A
> nglist -G g1 g2 g3 test_A
Note: | The global node group test_A is not the same as node group test_A in the current system partition. The global scope and system partition dependent scope are independent name spaces and are stored in separate classes in the System Data Repository (SDR). |
Purpose
ngnew - Creates but does not populate new node groups in persistent storage.
Syntax
ngnew [-h] | [-G] nodegroup [nodegroup ...]
Flags
Operands
Description
Use this command to create new node groups. If the nodegroup already exists, you will receive an error. A valid node group name must begin with a letter. If the nodegroup is not a valid name, you will receive an error. If a node group in the list cannot be successfully created, it will not affect the creation of the other supplied node groups. A nonzero return code is returned.
Exit Values
Security
You must have root privilege to create persistent node groups.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
Refer to the "Managing Node Groups" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional node grouping information.
Location
/usr/lpp/ssp/bin/ngnew
Related Information
Commands: ngaddto, ngcreate, ngdelete, ngdelfrom, ngfind, nglist, ngresolve
Examples
To create node groups called nga, ngb, and ngc, enter:
ngnew nga ngb ngc
Purpose
ngresolve - Returns a list of hosts in the specified node group.
Syntax
ngresolve [-h] | [-u | -n | -w | -d] [-G] nodegroup [nodegroup ...]
Flags
Operands
Description
Use this command to resolve the supplied named node groups into their constituent nodes. Nodes and node groups that are in the supplied node group but do not currently exist, will resolve to an empty list. If the -u flag is specified, these nonexistent nodes and node groups will be displayed.
Standard Output
A resolved list of nodes is written to standard output, one node per line.
Exit Values
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
Refer to the "Managing Node Groups" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional node grouping information.
Location
/usr/lpp/ssp/bin/ngresolve
Related Information
Commands: ngaddto, ngcreate, ngdelete, ngdelfrom, ngfind nglist, ngnew
Examples
> ngresolve -u nga 1 3 ngb
> ngresolve nga 1 3 6 8
> ngresolve -w nga k22n01.ppd.pok.ibm.com k22n03.ppd.pok.ibm.com k22n06.ppd.pok.ibm.com k22n08.ppd.pok.ibm.com
> ngresolve -d nga 129.40.157.65 129.40.157.67 129.40.157.70 129.40.157.72
Purpose
nodecond - Conditions an SP processing node.
Syntax
nodecond [-G] [-n] frame_ID slot_ID
Flags
Operands
Description
Node conditioning is the administrative procedure used to obtain the Ethernet hardware address of an SP processing node or to initiate a network boot of an SP processing node. The Ethernet hardware address is required by SP System Management for the proper configuration of the system. A network boot of the node is required by the System Management installation procedures.
By default, the nodecond command initiates a network boot of the node specified by the frame_ID and slot_ID operands. The specified node must be in the current system partition unless the -G flag is also specified. The frame ID is any configured frame number and the slot ID is taken from the set 1 through 16. The command completes when the node has booted to the point of configuring its console. Optionally, the nodecond command obtains the Ethernet hardware address of the processing node, specified by the frame_ID and slot_ID operands. The hardware address is written to standard output and the node is left powered off with the keylock in the Normal position.
As the command executes, it writes status information indicating its progress to /var/adm/SPlogs/spmon/nc/nc.frame_ID.slot_ID.
This command uses the SP Hardware Monitor. Therefore, the user must be authorized to access the Hardware Monitor subsystem and, for the frame specified to the command, the user must be granted Virtual Front Operator Panel (VFOP) and S1 (serial port on the node that you can access via the s1term command) permission. Since the Hardware Monitor subsystem uses SP authentication services, the user must execute the kinit command prior to executing this command. Alternatively, site-specific procedures can be used to obtain the tokens that are otherwise obtained by kinit.
Files
Related Information
Commands: hmcmds, hmmon, s1term
Examples
nodecond -n 5 1 > eth_adrr.5.1
nodecond 7 16
Purpose
nrunacct - Runs on each node every night to merge raw accounting data from the login, fee, disk, print, and process subsystems.
Syntax
Flags
Operands
Description
The nrunacct command is the main daily accounting shell procedure, for each individual node. Normally initiated by the cron daemon, the nrunacct command merges the day's raw connect, fee, disk, queuing system (printer), and process accounting data files for the node.
This command has two parameters that must be entered from the keyboard should you need to restart the nrunacct procedure. The date parameter, YYYYMMDD enables you to specify the date for which you want to rerun the node accounting. The state parameter enables a user with administrative authority to restart the nrunacct procedure at any of its states. For more information on restarting nrunacct procedures and on recovering from failures, see "Restart Procedure."
The nrunacct command protects active accounting files and summary files in the event of runtime errors, and records its progress by writing descriptive messages into the /var/adm/acct/nite/activeYYYYMMDD file. When the nrunacct procedure encounters an error, it sends mail to users root and adm, and writes standard errors to /var/adm/acct/nite/accterr.
The nrunacct procedure also creates two temporary files, lock and lock1, in the directory /var/adm/acct/nite, which it uses to prevent two simultaneous calls to the nrunacct procedure. It uses the lastdate file (in the same directory) to prevent more than one invocation per day.
The nrunacct command breaks its processing into separate, restartable states. As it completes each state, it writes the name of the next state in the /var/adm/acct/nite/stateYYYYMMDD file.
Restart Procedure
To restart the nrunacct command after a failure, first check the /var/adm/acct/nite/activeYYYYMMDD file for diagnostic messages, then fix any damaged data files, such as pacct or wtmp. Remove the lock files and lastdate file (all in the /var/adm/acct/nite directory, before restarting the nrunacct command. You must specify the YYYYMMDD parameter if you are restarting the nrunacct command. It specifies the date for which the nrunacct command is to rerun accounting. The nrunacct procedure determines the entry point for processing by reading the /var/adm/acct/nite/statefileYYYYMMDD file. To override this default action, specify the desired state on the nrunacct command line.
It is not usually a good idea to restart the nrunacct command in the SETUP state. Instead, perform the setup actions manually and restart accounting with the WTMPFIX state, as follows:
/usr/lpp/ssp/bin/nrunacct YYYYMMDD WTMPFIXIf the nrunacct command fails in the PROCESS state, remove the last ptacct file, because it is incomplete.
Files
Restrictions
Access Control: This command should grant execute (x) access only to members of the adm group.
Related Information
Commands: acctcms, acctcom, acctcon1, acctcon2, acctmerg, accton, acctprc1, acctprc2, crontab, fwtmp, nrunacct,
Daemon: cron
Subroutine: acct
File format: acct, failedlogin, tacct, wtmp
The System Accounting information found in AIX Version 4 System Management Guide
Examples
nohup /usr/lpp/ssp/bin/nrunacct 19950601 2>> \ /var/adm/acct/nite/accterr &This example restarts nrunacct for the day of June 1 (0601), 1995. The nrunacct command reads the file /var/adm/acct/nite/statefile19950601 to find out the state with which to begin. The nrunacct command runs in the background (&), ignoring all INTERRUPT and QUIT signals (nohup). Standard error output (2) is added to the end (>>) of the /var/adm/acct/nite/accterr file.
nohup /usr/lpp/ssp/bin/nrunacct 19950601 FEES 2>> \ /var/adm/acct/nite/accterr &This example restarts the nrunacct command for the day of June 1 (0601), 1995, starting with the FEES state. The nrunacct command runs in the background (&), ignoring all INTERRUPT and QUIT signals (the nohup command). Standard error output (2) is added to the end (>>) of the /var/adm/acct/nite/accterr file.
Purpose
p_cat - Issues a parallel cat of files.
Syntax
p_cat [-w - | noderange | 'hostlist args'] file_name file_name ...
Parameters
The p_cat command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Description
The p_cat command issues the AIX cat command on multiple hosts. p_cat uses dsh to execute the cat command on multiple hosts. The output of the cat command is written to standard output.
Since p_cat uses dsh, proper authentication and authorization to issue these commands is necessary.
Files
Related Information
SP commands: dsh, pexec
AIX command: cat
Examples
To copy /.rhosts from each host1, host2, and host3 to the local /.rhosts file (described previously), enter:
p_cat -w host1,host2,host3 /.rhosts >> /.rhosts
Purpose
pcp - Specifies a parallel copy of local files and directories to other hosts.
Syntax
Parameters
The pcp command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Description
The pcp command copies files from the local host to one or more others in parallel. pcp is similar to rcp and in fact uses SP rcp via dsh. The -r and -p flags are passed to SP rcp.
Since pcp uses dsh and SP rcp, proper authentication and authorization to issue these commands is necessary.
Note: |
Since the pcp command uses the secure version of rcp,
your .klogin or .rhosts files need to be set up to
authorize you on each of the nodes to which you are copying a file. Refer to
the chapter on security in IBM Parallel System Support Programs for
AIX: Administration Guide. Otherwise, you see:
Permission denied messages from the nodes for which you are not authorized. |
Related Information
Commands: dsh, hostlist
SP Command: rcp
Examples
pcp -w host1,host2 /etc/fileq /etc/filen
pcp '-a' sysctl.acl /etc
hostlist -av -e badnode | pcp -w - -r /etc/new /etc/new
pcp "-av -e badnode" -r /etc/new /etc/new
Purpose
pdf - Displays file system statistics on multiple nodes in parallel.
Syntax
pdf [-w - | noderange | 'hostlist args'] [file system1 [file system2 ... ]]
Parameters
The pdf command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Acceptable parameters to the pdf command are zero or more blank-separated file system names to be displayed. If no parameters are supplied, all local file systems (on each specified node) are displayed.
Description
The pdf command displays file systems and their usage statistics on one or more nodes in parallel. pdf is similar to the df command, but does not use it. Differences are:
Since pdf uses sysctl, proper authentication and authorization to issue these commands is necessary.
Related Information
Commands: hostlist, sysctl
Examples
pdf '-a'
pdf -w node1,node2 /tmp
Purpose
penotify - Adds, removes, or shows AIX error log notification objects.
Syntax
Flags
Operands
None.
Description
Use this command to add, remove, or show notification objects in the errnotify Object Data Management (ODM) class. The AIX errdemon matches logged errors to objects in this class to execute a method defined in the class object. The error class, error label, error type, alert, resource name, resource class, and resource type parameters are used for matching to logged errors. Refer to the AIX Version 4 General Programming Concepts: Writing and Debugging Programs for descriptions of error notification object class fields.
When a match occurs, the errdemon executes the notify method passing up to nine ($1-$9) parameters related to the logged error.
If the -w parameter begins with a slash (/), it is interpreted as a file containing a list of nodes to execute the command on; otherwise, it can be a comma-delimited list of host names or a single-quoted, blank-delimited list of host names. If neither the -a nor -w parameters are used, the command defaults to the local node.
Security
You must have a Kerberos principal defined in the /etc/logmgt.acl file to run this command.
Related Information
The AIX Version 4 General Programming Concepts: Writing and Debugging Programs
The IBM Parallel System Support Programs for AIX: Administration Guide
Examples
penotify -w k47n10,k47n12,k47n15 -f show
penotify -a -f remove -n HDISK0_ERR
penotify -w /tmp/nodelist -f add -n PEND_ERR -P \ -m'/spdata/sys1/EN_meth/EN_pend $1' -t PEND -c S
This adds a notification object named PEND_ERR to all nodes in the /tmp/nodelist file. The object will persist when the system is restarted, and will match error records of type PEND and class S. The method that is executed by errdemon when a matching error occurs will be /spdata/sys1/EN_meth/EN_pend, and it will be passed the $1 parameter (sequence number). The notification method must be accessible to each node.
Purpose
perspectives - Invokes the launch pad of the SP Perspectives graphical user interface (GUI).
Syntax
Flags
Note: | Most flags accepted by X will also be recognized. For example, -display displayname. |
Operands
None.
Description
Use this command to invoke the SP Perspectives Launch Pad. The Launch Pad is a small, customizable GUI from which the user can start (or launch) executables associated with maintaining and monitoring an IBM RS/6000 SP.
The main window shows an icon for each executable that can be launched. Double-clicking on an icon launches the associated executable.
Preferences that define the look and layout of the Perspectives window are prioritized in the following order:
Files
The User's Preferences are read from and saved to $HOME/.perspectives(User Profile Name).
The System Preferences are read from and saved to /usr/lpp/ssp/perspectives/profiles/.perspectives(System Profile name).
Restrictions
Any user may run perspectives. Launching certain executables may require root privilege to run.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
For information on using the perspectives GUI, please see the online help and the "Using the SP Perspectives GUI" chapter in the IBM Parallel System Support Programs for AIX: Administration Guide.
Location
/usr/lpp/ssp/bin/perspectives
Related Information
Specific perspective windows may be brought up directly by invoking the following commands: spevent, sphardware, spperfmon, spsyspar, and spvsd.
Examples
perspectives
perspectives -fontSize 14
Purpose
pexec - Specifies the parallel execution of a command.
Syntax
pexec [-w - | noderange | 'hostlist args'] command command_args
Parameters
The pexec commands requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Description
The pexec command issues a command on multiple hosts in parallel. The output is formatted so that distinct output is displayed only once. The pexec command uses dsh to execute the specified command on multiple hosts. The output of the ls command is written to standard output and formatted. The pls, prm, pmv, pfind, and pps commands are simply links to pexec.
Since pexec uses dsh, proper authentication and authorization to issue these commands is necessary.
Note: | If any of the pls, prm, pfind, pps, pmv are renamed, they do not work properly. |
Files
Related Information
Commands: dsh, dshbak, hostlist
Examples
pexec -w host1,host2,host3 ls /
hostlist -a -v -e badnode | pexec -w -cp -r /etc/new /etc/new
pexec "-a -v -e badnode" cp -r /etc/new /etc/new
Purpose
pexscr - Runs local and remote programs in parallel.
Syntax
pexscr
Flags
None.
Operands
None.
Description
The pexscr command executes particular commands on particular processors in parallel. pexscr reads lines of the following format from standard input:
host_name: arbitrary_commandand executes each arbitrary_command on the specified host. All commands are run in parallel. SP rsh is used to run the remote commands, and local commands are run directly. Host names can include any parameter that may be specified on the rsh command.
Since pexscr uses dsh, proper authentication and authorization to issue these commands is necessary.
Related Information
Commands: dsh, rsh
Examples
To remove a file on host1 and rename a file on host2 simultaneously, enter:
pexscr <<END host1: rm /tmp/shnozzola host2: mv /tmp/shnozzola /tmp/bozo END
Purpose
pfck - Displays file system statistics on multiple hosts in parallel based on usage criteria.
Syntax
Parameters
The pfck command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Subsequent flags specify file system usage statistics criteria to be applied in searching for file systems to display. At least one of the following flags must be specified. Multiple flags are allowed.
File system usage statistics criteria are logically ORed together when comparing against actual usage information. That is, if a file system meets any of the search criteria, then it is displayed.
Description
The pfck command displays file systems and their usage statistics from one or more nodes in parallel based on usage criteria.
Since pfck uses sysctl, proper authentication and authorization to issue these commands is necessary.
Related Information
Commands: hostlist, sysctl
Examples
pfck -a -pf 20
pfck -w node1,node2,node4 -pu 98
Purpose
pfind - Specifies a parallel find of files with a matching expression.
Syntax
pfind [-w - | noderange | 'hostlist args'] find_args
Parameters
The pfind command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Description
The pfind command issues the AIX find command on multiple hosts. The output is formatted so that distinct output is displayed only once. The pfind command uses dsh to execute the find command on multiple hosts. The output of the ls commands is written to standard output and formatted. The pfind command is identical to pexec find.
Since pfind uses dsh, proper authentication and authorization to issue these commands is necessary.
Files
Related Information
Commands: dsh, find, hostlist, pexec
Examples
To find out if the file elvis is contained in /usr/bin on any host1, host2, and host3 (described previously), enter:
pfind -w host1,host2,host3 /usr/bin -print -name "elvis"
Purpose
pfps - Finds and performs operations on processes on multiple hosts in parallel based on the value of an expression.
Syntax
pfps [-w - | noderange | 'hostlist args'] operation [expression]
Parameters
The pfps command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Description
The pfps command performs operations on processes on one or more hosts in parallel. These operations include printing information about processes (-print), sending a signal to the process (-kill), and changing the priority of the process (-nice).
Authorization is via an Access Control List (ACL) on each node and is required when users try to kill a process that they do not own or nice a process to a higher priority. ACLs for pfs are contained in the /etc/sysctl.pfps.acl file.
An expression can also be specified using the preceding flags to select processes for when the expression evaluates to true. Flags are ANDed together unless the -or flag is used.
Parentheses can be used to group flags, but parenthesis must be separated from flags by a space. Also, parenthesis or any special shell character should be escaped with a backslash (\).
Since pfps uses sysctl, proper authentication and authorization to issue these commands is necessary.
Files
Related Information
Commands: hostlist, kill, nice, ps, renice, sysctl
Examples
pfps '-a' -print
pfps -w host1,host2 -rtime 24:00 -tn daemond -kill HUP
WCOLL=./wcollective pfps '' \( -cpu 10 -or -mem 10 \) -o root -print
Purpose
pls - Specifies a parallel list of files and directories.
Syntax
pls [-w - | noderange | 'hostlist args'] ls_args
Parameters
The pls command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Description
The pls command issues the AIX ls command on multiple hosts. The output is formatted so that duplicate output is displayed only once. The pls command uses dsh to execute the ls command on multiple hosts. The output of the ls commands is written to standard output and formatted. The pls command is identical to pexec ls.
Since pls uses dsh, proper authentication and authorization to issue these commands is necessary.
Files
Related Information
Commands: dsh, ls, pexec
Examples
To list the contents of /usr from each host1, host2, and host3 (described previously), enter:
pls -w host1,host2,host3 /usr
Purpose
pmanctrl - Controls the Problem Management subsystem.
Syntax
pmanctrl {-a | -s | -k | -d | -c | -t | -o | -r | -h}
Flags
Operands
None.
Description
Problem Management is a general purpose facility for monitoring and reacting to specific event occurrences within the SP system. The pmanctrl command controls the operations of the subsystems that are required for Problem Management. These subsystems are under the control of the System Resource Controller (SRC) and belong to a subsystem group called pman. Associated with each subsystem is a daemon.
An instance of the Problem Management subsystem executes on the control workstation and on every node of a system partition. Because Problem Management provides its services within the scope of a system partition, its subsystems are said to be system partition sensitive. For this reason, the pmanctrl command is normally invoked by the syspar_ctrl command during installation of the system, boot or reboot of individual nodes, and partitioning or repartitioning of the system.
From an operational point of view, the Problem Management subsystem group is organized as follows:
The pman subsystem is associated with the pmand daemon. The pmanrm subsystem is associated with the pmanrmd daemon.
The subsystem names on the nodes are pman and pmanrm. There is one of each subsystem per node and it is associated with the system partition to which the node belongs.
On the control workstation, there are multiple instances of each subsystem, one for each system partition. Accordingly, the subsystem names on the control workstation have the system partition name appended to them. For example, for system partitions named sp_prod and sp_test, the subsystems on the control workstation are named pman.sp_prod, pman.sp_test, pmanrm.sp_prod, and pmanrm.sp_test.
The pmand daemon provides the majority of Problem Management functions.
The pmanrmd daemon provides command-based resource monitor data to the pmand daemon.
The pmanctrl command provides a variety of controls for operating the Problem Management subsystems:
Unless the -c flag is used, the pmanctrl command only operates within a single partition. On a node, the pmanctrl command operates within the system partition to which the node belongs. On the control workstation, the pmanctrl command operates within any single partition, which can be chosen by setting the SP_NAME environment variable.
When the pmanctrl command is called with the -a flag, it uses the mkssys command to add the subsystems to the SRC, and it takes the necessary steps to make sure that the subsystems are automatically started when the node is booted.
When the pmanctrl command is called with the -s flag, it uses the startsrc command to start the pman and pmanrm subsystems.
When the pmanctrl command is called with the -k flag, it uses the stopsrc command to stop the pman and pmanrm subsystems.
When the pmanctrl command is called with the -d flag, it uses the rmssys command to delete the subsystems from the SRC, and if there are no more Problem Management subsystems remaining, it makes sure there is no /etc/inittab entry for the Problem Management subsystem.
When the pmanctrl command is called with the -c flag, it stops all running Problem Management subsystems, removes them all from the SRC, and makes sure there is no /etc/inittab entry for the Problem Management subsystem.
When the pmanctrl command is called with the -t flag, it uses the traceson command to turn on tracing in the pman subsystem. Tracing is not available for the pmanrmd subsystem.
When the pmanctrl command is called with the -o flag, it uses the tracesoff command to turn off tracing in the pman subsystem. Tracing is not available for the pmanrmd subsystem.
While they are running, the Problem Management daemons provide information about their operation and errors by writing entries in a log file that is located in the /var/adm/SPlogs/pman directory. On the control workstation, the pmand daemon writes to a log file named pmand.syspar_name.log, and the pmanrmd daemon writes to a log file named pmanrmd.syspar_name.log, where syspar_name is the name of the system partition. On the nodes, the pmand daemon writes to a log file named pmand.log and the pmanrmd daemon writes to a log file named pmanrmd.log.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
The "Using the Problem Management Subsystem" chapter in IBM Parallel System Support Programs for AIX: Administration Guide
IBM AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in IBM AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/pmanctrl
Related Information
SP Command: syspar_ctrl
AIX Commands: mkssys, rmssys, startsrc, stopsrc
Examples
pmanctrl -a
pmanctrl -s
pmanctrl -k
pmanctrl -d
pmanctrl -c
pmanctrl -t
pmanctrl -o
Purpose
pmandef - Defines events and resulting actions to the Problem Management subsystem.
Syntax
To subscribe to an event and associate a set of actions with that event, use the following:
To deactivate or activate a Problem Management subscription, use the following:
To query or remove a Problem Management subscription, use the following:
Flags
Flags that specify the type of request follow:
Flags that specify the hosts to be affected by the request follow:
Flags that define a Problem Management subscription follow:
echo this is a test >/tmp/event.outis allowed.
echo this is a test >/tmp/rearm.outis allowed.
Operands
None.
Description
The Problem Management subsystem allows you to specify that an action takes place as the result of a specific event occurrence within the SP system. The Problem Management subsystem registers for the event with the Event Management subsystem. When the Event Management subsystem reports the occurrence of the event back to Problem Management, the Problem Management subsystem performs the requested action on your behalf. The actions that are performed as the result of an event can be any or all of the following:
The pmandef command is the user interface for making such requests to the Problem Management subsystem. For example, running the following command on node 5:
pmandef -s Program_Monitor \ -e 'IBM.PSSP.Prog.pcount:NodeNum=12;ProgName=mycmd;UserName=bob:X@0==0' \ -r "X@0==1" -c "echo program has stopped >/tmp/myevent.out" \ -C "echo program has restarted >/tmp/myrearm.out"
causes the command echo program has stopped >/tmp/myevent.out to run on node 5 whenever the number of processes named "mycmd" and owned by user "bob" on node 12 becomes 0. When this number increases back to 1, the command echo program has restarted >/tmp/myrearm.out runs on node 5.
If you do not want the command to run on the same node from which the pmandef command was issued, then use one of the -h, -N, or -n flags.
The following example causes the commands to run on both k21n01 and k21n02 whenever bob's program dies or gets restarted on either nodes 12 or 13.
pmandef -s Program_Monitor \ -e 'IBM.PSSP.Prog.pcount:NodeNum=12-13;ProgName=mycmd;UserName=bob:X@0==0' \ -r "X@0==1" -c /usr/local/bin/start_recovery \ -C /usr/local/bin/stop_recovery -h k21n01,k21n02
The following example causes the commands to run on nodes 1, 2, 3, and 7 whenever bob's program dies or gets restarted on any of nodes 1, 2, 3, 4, 5, or 13. If bob's program dies on node 4, the command /usr/local/bin/start_recovery runs on nodes 1, 2, 3, and 7.
pmandef -s Program_Monitor \ -e 'IBM.PSSP.Prog.pcount:NodeNum=1-5,13;ProgName=mycmd;UserName=bob:X@0==0' \ -r "X@0==1" -c /usr/local/bin/start_recovery \ -C /usr/local/bin/stop_recovery -n 1-3,7
If you want to define a subscription for more than one node but you want the command to run only on the same node where the event occurs, use the -h local option. Consider the following command:
pmandef -s Filesystem_Monitor \ -e 'IBM.PSSP.aixos.FS.%totused:NodeNum=10-14;VG=myvg;LV=mylv:X>95' \ -l "filesystem is almost full" -h localWhenever the file system associated with the "mylv" logical volume and "myvg" volume group on node 11 becomes more than 95 percent full, the text filesystem is almost full gets written to the AIX error log and BSD syslog only on node 11. Whenever the same thing occurs on node 12, the same text gets written to the AIX error log and BSD syslog only on node 12. The file system filling up on node 11 is really a separate event than the file system filling up on node 12 or node 13, and the -h local option is just a convenient way to define actions for a whole a bunch of events at the same time.
Issuing the pmandef command with the -s flag to associate an action with an event creates a Problem Management "subscription." When you issue the pmandef command to create a Problem Management subscription, the definition of the subscription gets stored in the System Data Repository (SDR) so the definition becomes permanent. As soon as the subscription gets stored in the SDR, the pmandef command also requests the affected Problem Management daemons within the SP system to start acting on the new subscription. Since it is possible for some of the nodes affected by this to be down, the pmandef command is considered successful once the subscription is stored in the SDR. The failure to reach all of the affected Problem Management daemons is not considered to be an error, because they will eventually pick up the new subscription once they get restarted.
If the Event Management resource variable, instance vector, predicate, or re-arm predicate contains an error, the error will not get discovered until the Problem Management daemon registers for the event with the Event Management subsystem. When this happens, the subscription definition does not automatically get removed from the SDR. You must remove the subscription by using the pmandef command with the -u flag. The argument to the -u flag is the same name that was previously specified as the argument to the -s flag.
The pmandef command with the -u flag removes the subscription definition from the SDR and tells the appropriate Problem Management daemons to stop monitoring for the event associated with that subscription. The pmandef command with the -d can be used to turn off or "deactivate" a subscription. This also tells the appropriate Problem Management daemons to stop monitoring for the event associated with that subscription, but it does not remove the subscription definition from the SDR. The subscription remains deactivated until you call pmandef with the -a flag to "activate" the subscription. You can also use the pmandef command with the -q flag to query the appropriate Problem Management daemons for status about a subscription. This just tells you whether each daemon is currently monitoring for the event associated with that subscription.
The pmandef command is based on sysctl, which uses Kerberos for user authentication. As a result, all users of the pmandef command must have valid Kerberos credentials. In addition, the user's Kerberos principal must be listed in the /etc/sysctl.pman.acl file on the local node, in order to store the subscription in the SDR. It must also be listed on all the nodes that are affected by the new subscription, in order for the affected Problem Management daemons to be notified of the new subscription. If the user's Kerberos principal is listed only in the /etc/sysctl.pman.acl file on the local node, the subscription will be stored in the SDR, but the Problem Management daemons will not act on the new subscription until the next time they are restarted.
The user's Kerberos principal is more than just a mechanism to validate access to the Problem Management subsystem. The Kerberos principal that was in effect when the subscription was created, via the pmandef -s command, is stored as part of the subscription definition, and it is used to establish the ownership of that subscription. Modifications to the subscription, via the pmandef command with the -u, -d, and -a flags, are only allowed by a user with the same Kerberos principal that is stored in the subscription definition. The Kerberos principal that is stored in the subscription definition is also used by the Problem Management daemon to decide whether the action that is to result from the occurrence of an event should be allowed.
After the Problem Management daemon receives notification that an event has occurred, and before it performs the action for that event, the Problem Management daemon checks to see whether the Kerberos principal that is stored in the subscription definition is authorized to perform the requested action on the node where the Problem Management daemon is running. If the requested action is an entry in the AIX error log and BSD syslog or the generation of an SNMP trap, the Kerberos principal that owns the subscription must be listed in the root user's $HOME/.klogin file. If the requested action is the execution of a command, the Kerberos principal must be listed in the $HOME/.klogin file of the user that will be used to run the command. The user that will be used to run the command is by default the same user who issued the pmandef -s command to create the subscription. A different user can be specified to pmandef by using the -U flag.
Files
Exit Values
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference
"Using the Problem Management Subsystem" in IBM Parallel System Support Programs for AIX: Administration Guide
Location
/usr/lpp/ssp/bin/pmandef
Related Information
Commands: pmanquery, sysctl
Examples
In this example, the user mustard is authenticated as the Kerberos principal hotdog@XYZ.COM. If mustard issues the command:
pmandef -s Program_Monitor \ -e 'IBM.PSSP.Prog.pcount:NodeNum=12;ProgName=mycmd;UserName=relish:X@0==0' \ -r "X@0==1" -c /usr/local/bin/start_recovery \ -C /usr/local/bin/stop_recovery -n 5 -U ketchupa subscription named "Program_Monitor" and owned by Kerberos principal hotdog@XYZ.COM will be created, if the Kerberos principal hotdog@XYZ.COM is listed in the /etc/sysctl.pman.acl file on the local node. The requested actions for this subscription are to occur on node 5, so the Kerberos principal must also be listed in the /etc/sysctl.pman.acl file on node 5, in order for the Problem Management daemon on node 5 to act on this subscription immediately. Otherwise, the Problem Management daemon will not act on this subscription until the next time it gets restarted.
This subscription requests that whenever the number of processes named "mycmd" and owned by user relish on node 12 becomes 0, the command start_recovery will run on node 5. When the number of processes increases back to 1, the command stop_recovery will run on node 5. The subscription also requests that the commands start_recovery and stop_recovery are to run on node 5 as the user ketchup. The start_recovery and stop_recovery commands will not run on node 5 unless the Kerberos principal hotdog@XYZ.COM is listed in the $HOME/.klogin file of the user ketchup. If the -U flag had been omitted, this subscription would have requested the commands to run as the user mustard, and this would have required mustard's $HOME/.klogin file to list hotdog@XYZ.COM.
Once the subscription is created, any user who is authenticated as the Kerberos principal hotdog@XYZ.COM can deactivate, activate, or remove this subscription.
To deactivate:
pmandef -d Program_Monitor
To activate:
pmandef -a Program_Monitor
To remove:
pmandef -u Program_Monitor
Purpose
pmanquery - Queries the System Data Repository (SDR) for a description of a Problem Management subscription.
Syntax
Flags
Flags that specify the scope of the search follow:
Flags that format the output follow:
Operands
None.
Description
After a Problem Management subscription definition is stored in the SDR by the pmandef command, you can use the pmanquery command to retrieve the subscription definition. The pmanquery command prints the details of the subscription definition in raw format which can then be parsed by other applications.
The -n and -k flags control the scope of the search for subscriptions in the SDR. Refer to the Examples section of this command for pmanquery flag usage information.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
The "Using the Problem Management Subsystem" chapter in IBM Parallel System Support Programs for AIX: Administration Guide
Location
/usr/lpp/ssp/bin/pmanquery
Related Information
Commands: pmandef
Examples
pmanquery -n my_handle
pmanquery -n my_handle -k hotdog@XYZ.COM
pmanquery -n all -k hotdog@XYZ.COM
pmanquery -n all
pmanquery -n my_handle -k all
pmanquery -n all -k all
Purpose
pmanrmdloadSDR - Reads a pmanrmd configuration file and loads the information into the System Data Repository (SDR).
Syntax
pmanrmdloadSDR ConfigFileName
Flags
None.
Operands
Description
The Problem Management subsystem provides 16 resource variables, named IBM.PSSP.pm.User_state1 through IBM.PSSP.pm.User_state16. These are predefined resource variables that have been set aside for users to create their own resource monitors. A resource monitor that you create through Problem Management is a command that gets executed repeatedly by the pmanrmd daemon at a specific interval. The standard output from the command is supplied to the Event Management subsystem as the value for the resource variable. You can then use the pmandef command to subscribe to events for that resource variable.
The resource variable name, resource monitor command, sampling interval, and list of nodes for which the resource monitor is defined are stored in the SDR. The pmanrmdloadSDR command is used to store those definitions in the SDR.
You define your resource monitor to the pmanrmd daemon by doing the following:
Control workstation:
stopsrc -s pmanrmd.syspar_name startsrc -s pmanrmd.syspar_namewhere syspar_name is the name of the system partition.
Node:
stopsrc -s pmanrmd startsrc -s pmanrmd
For a more complete description of Problem Management resource monitors, refer to the "Using the Problem Management Subsystem" chapter in IBM Parallel System Support Programs for AIX: Administration Guide
Files
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference
The "Using the Problem Management Subsystem" chapter in IBM Parallel System Support Programs for AIX: Administration Guide
Location
/usr/lpp/ssp/bin/pmanrmdloadSDR
Related Information
Commands: pmandef
Purpose
pmv - Specifies a parallel file move.
Syntax
pmv [-w - | noderange | 'hostlist args'] mv_args
Parameters
The pmv command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Note: | The -i is not supported (the dsh command does not support standard input to remote hosts). |
Description
The pmv command issues the AIX mv command on multiple hosts. The output is formatted so that duplicate output is displayed only once. The pmv command uses dsh to execute the mv command on multiple hosts. The output of the ls commands is written to standard output and formatted. The pmv command is identical to pexec mv.
Since pmv uses dsh, proper authentication and authorization to issue these commands is necessary.
Files
Related Information
Commands: dsh, mv, pexec
Examples
To move a file from each host1, host2, and host3 to a different directory, enter:
pmv -w host1,host2,host3 /tmp/shnozzola /etc/shnozzola
Purpose
ppred - Performs a command on those hosts for which a test is satisfied.
Syntax
Parameters
The ppred command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Description
The ppred command performs a test on remote hosts in parallel. On each host where the test succeeds, a command is run. Optionally, a command can be specified that runs if the test fails.
Since ppred uses dsh, proper authentication and authorization to issue these commands is necessary.
Related Information
Commands: dsh, hostlist, test
Examples
To verify that a file exists and is a regular file on the host occupying the first slot in each of 4 frames, enter:
ppred '-s 1-4:1' '-f /etc/passwd' 'echo \'host_name\''
Purpose
pps - Specifies a parallel ps command.
Syntax
pps [-w - | noderange | 'hostlist args'] ps_args
Parameters
The pps command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Description
The pps command uses dsh to execute the ps command on multiple hosts. The output of the ls commands is written to standard output and formatted so that distinct output is presented only once. The pps command is identical to pexec ps.
Since pps uses dsh, proper authentication and authorization to issue these commands is necessary.
Files
Related Information
Commands: dsh, pexec, ps
Examples
To list processes on each host1, host2, and host3 (described previously), enter:
pps -w host1,host2,host3 -ef
Purpose
preparevsd - Makes an IBM Virtual Shared Disk available.
Syntax
preparevsd -a | vsd_name...
Flags
Operands
Description
The preparevsd command brings the specified IBM Virtual Shared Disks from the stopped state to the suspended state. The IBM Virtual Shared Disks are made available. Open and close requests are honored, while read and write requests are held until the IBM Virtual Shared Disks are brought to the active state. If they are in the suspended state, this command leaves them in the suspended state.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit vsd_mgmtand select the Prepare an IBM Virtual Shared Disk option.
Files
Security
You must have root privilege to run this command.
Restrictions
If you have the IBM Recoverable Virtual Shared Disk product installed and operational, do not use this command. The results may be unpredictable.
See IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfgvsd, ctlvsd, lsvsd, resumevsd, startvsd, stopvsd, suspendvsd, ucfgvsd
Examples
To bring the IBM Virtual Shared Disk vsd1vg1n1 from the stopped state to the suspended state, enter:
preparevsd vsd1vg1n1
Purpose
prm - Specifies a parallel file remove.
Syntax
prm [-w - | noderange | 'hostlist args'] rm_args
Parameters
The prm command requires the first flag or parameter on the command line to be a specification of the hosts on which the command is to be executed.
Note: | To use the working collective file specified by the WCOLL environment variable, you must specify a null string as the first argument. Refer to the dsh command for more information about a working collective. |
Description
The prm command issues the AIX rm command on multiple hosts. The output is formatted so that distinct output is displayed only once. The prm command uses dsh to execute the rm command on multiple hosts. The output of the ls commands is written to standard output and formatted. The prm command is identical to pexec rm.
Since prm uses dsh, proper authentication and authorization to issue these commands is necessary.
Files
Related Information
Commands: dsh, rm, pexec
Examples
To remove a file from each host1, host2, and host3 (described previously), enter:
prm -w host1,host2,host3 /tmp/shnozzola
Purpose
psyslclr - Removes entries from syslog log files on a set of nodes.
Syntax
Flags
Operands
None.
Description
Use this command to delete log entries in syslogd generated log files. Options allow for selecting the files and records that are trimmed.
The arguments to options -d, -f, -l, -n, -r, and -w can be a comma-delimited or single-quoted, blank-delimited list of values. If the -l flag is used, the command will only trim records from the specified list of log file names. If the -l flag is not passed, the command will first parse the syslog configuration file (the default is /etc/syslog.conf) to select files for trimming.
The -f and -p flags can be used to control selecting files in the configuration file. All files found in the configuration file will be trimmed if the -f and -p flags are not used.
The -d, -e, -n, -r, -s, and -y flags are used to match log entries to be deleted. A record must match a value from each of the flags that are used to be trimmed. If a flag is not passed, all records match for that field. To delete all records, use the -y flag with 0 as the argument. If the -w flag begins with a slash (/), it is interpreted as a file containing a list of nodes to execute the command on; otherwise, it can be a list as described previously. If neither the -a nor the -w flags are used, the command defaults to the local node.
Note: | The syslogd daemon does not log the year in records timestamps. The comparisons for start and end times are done on a per record basis and could cause unexpected results if the log file is allowed to span more than one year. The syslogd daemon is stopped during this process so trimming activity should be planned accordingly. It is then restarted using the default or alternate syslog configuration file. |
Files
Security
You must have a Kerberos principal defined in the /etc/logmgt.acl file to run this command.
Related Information
Command: psyslrpt
Daemon: syslogd
Examples
psyslclr -a -y 30
psyslclr -w k47n10 -s 04110000 -e 07230000 -r ftp,snmpd
psyslclr -w /tmp/nodelist -f mail,user -p error -y 0
Purpose
psyslrpt - Generates reports of records in syslog log files on a set of nodes.
Syntax
Flags
Operands
None.
Description
Use this command to generate reports of log entries in syslogd generated log files. Options allow for selecting the files and records that are reported. The arguments to options -d, -f, -l, -n, -r, and -w can be a comma-delimited or single-quoted, blank-delimited list of values. If the -l flag is used, the command will report records from the specified list of log file names. If the -l flag is not passed, the command will first parse the syslog configuration file (the default is /etc/syslog.conf) to select files for reporting.
The -f and -p options can be used to control the selecting of files in the configuration file. All files found in the configuration file are reported on if the -f and -p flags are not used.
The -d, -e, -n, -r, and -s options are used to match log entries to be reported. A record must match a value from each of these flags that are used to be reported. If a flag is not passed, all records match for that field. If the -w argument begins with slash (/), it is interpreted as a file containing a list of nodes to execute the command on; otherwise, it can be a list as described previously. If neither the -a nor -w flags are used, the command defaults to the local node.
Files
Security
You must be Kerberos authenticated to run this command.
Related Information
Command: psyslclr
Daemon: syslogd
The IBM Parallel System Support Programs for AIX: Administration Guide
Examples
psyslrpt -a -s 03030000
psyslrpt -w k47n10 -s 04110000 -e 07230000 -r ftp,snmp
psyslrpt -w k47n12,k47n15 -d'10479 1157' -l /var/adm/SPlogs/SPdaemon.log
Note: | syslogd does not log the year in record timestamps. The comparisons for start and end times are done on a per record basis and could cause unexpected results if the log file is allowed to span more than one year. |
Purpose
rcmdtgt - Obtains and caches a ticket-granting-ticket for the local realm, with a maximum allowed lifetime, using the service key for the instance of rcmd on the local host.
Syntax
rcmdtgt
Flags
None.
Operands
None.
Description
Use this command to obtain a ticket-granting-ticket with a maximum allowed lifetime, using the service key for rcmd.localhost found in the service key file at /etc/krb-srvtab. When using SP authentication services, these tickets have an unlimited lifetime. When using AFS authentication services, a maximum of 30 days is enforced.
This command must be run as root and is intended for use in shell scripts and other batch-type facilities. The rcmdtgt command retrieves your initial ticket and puts it in the ticket file specified by your KRBTKFILE environment variable. If the KRBTKFILE variable is undefined, your ticket is stored in the /tmp/tktuid file, where uid specifies your user identification number.
Note: | These tickets are shared by all processes running under the user's IDs. The KRBTKFILE environment variable can be set to change the location of the ticket cache file. |
Files
Related Information
Commands: kdestroy, kinit,
File: krb.conf
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Examples
The following example, excerpted from the sample script.cust file, shows how rcmdtgt can be used in a shell script to perform the authentication required to use the SP rcp command:
# set the host name from which you will copy the file. SERVER='cat /etc/ssp/server_host_name | cut -d" " -f1' # Define a temporary ticket cache file, then get a ticket export KRBTKFILE=/tmp/tkt.$$ /usr/lpp/ssp/rcmd/bin/rcmdtgt # # Perform kerberos-authenticated rcp /usr/lpp/ssp/rcmd/bin/rcp $SERVER:/etc/resolv.conf /etc/resolv.conf # Remove the ticket cache file /bin/rm $KRBTKFILE
Purpose
/usr/lpp/ssp/rcmd/bin/rcp - Transfers files between a local and a remote host.
Note: | The remote-to-remote copy is restricted at this time. |
Syntax
Flags
When this flag is not used, the umask being honored is the value stored in the appropriate database. It is not the value that is set by issuing the umask command. The permission and ownership values that result from the umask command do not affect those stored in the database.
Operands
Note: | Because the rcp command assumes that a colon (:) terminates a host name, you must insert a backslash (\) before any colons that are embedded in the local file and directory names. |
Note: | Because the rcp command assumes that a colon (:) terminates a host name, you must insert a backslash (\) before any colons that are embedded in the local file and directory names. |
Description
The rcp command is used to copy one or more files between the local host and a remote host or between files at the same host.
Remote destination files and directories require a specified Host: parameter. If a remote host name is not specified for either the source or the destination, the rcp command is equivalent to the cp command. Local file and directory names do not require a Host: parameter.
Note: | Because the rcp command assumes that a colon (:) terminates a host name, you must insert a backslash (\) before any colons that are embedded in the local file and directory names. |
If a Host is not prefixed by a User@ parameter, the local user name is used at the remote host. If a User@ parameter is entered, that name is used.
If the path for a file or directory on a remote host is not specified or is not fully qualified, the path is interpreted as beginning at the home directory for the remote user account. Additionally, any metacharacters that must be interpreted at a remote host must be quoted using a backslash (\), a double quotation mark ('), or a single quotation mark (').
If the /usr/kerberos/bin/rcp path does not exist, a link is created from /usr/lpp/ssp/rcmd/bin/rcp to /usr/kerberos/bin/rcp. The SP rcp command uses this path when initially starting an rcp process on a remote host for compatibility purposes.
File Permissions and Ownership
By default, the permissions mode and ownership of an existing destination file are preserved. Normally, if a destination file does not exist, the permissions mode of the destination file is equal to the permissions mode of the source file as modified by the umask command (a special command in the Korn shell) at the destination host. If the rcp command -p flag is set, the modification time and mode of source files are preserved at the destination host.
The user name entered for the remote host determines the file access privileges the rcp command uses at that host. Additionally, the user name given to a destination host determines the ownership and access modes of the resulting destination file or files. The remote host allows access if one of the following conditions is satisfied:
For security reasons, any $HOME/.klogin file must be owned by either the remote user or root, and only the owner should have read and write access.
Should user authentication fail or should the authentication subroutine call fail, the SP rcp command passes the arguments to the AIX rcp command. In this case, the user needs normal AIX rcp access to a host for the command to be successful or the permission denied message is returned. The SP rcp command issues several authentication error messages before passing arguments to the AIX rcp command. The messages are also visible from system management applications that use the SP rcp command.
This command does not support encryption by users.
Files
Prerequisite Information
Refer to AIX Version 4 System Management Guide: Communications and Networks for a network overview.
Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for an overview.
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Related Information
SP Commands: kinit, kshd, rsh
AIX Commands: cp, ftp, rlogin, rsh, rshd, tftp, umask
Examples
In the following examples, the local user is listed in the authentication database and a ticket was obtained by issuing a kinit and the password. In the following examples, host1 is the local host and host2 is the remote host:
/usr/lpp/ssp/rcmd/bin/rcp localfile host2:/home/eng/jane
The file localfile from the local host is copied to the remote host host2.
/usr/lpp/ssp/rcmd/bin/rcp host2:/home/eng/jane/newplan/home/eng/mary
The file /home/eng/jane/newplan is copied from remote host host1 to local host host2.
/usr/lpp.ssp/rcmd/bin/rcp -p -r report jane@host2:report
The directory subtree report is copied from the local host to the home directory of user jane at remote host host2 and all modes and modification times are preserved. The remote file /home/jane/.rhosts includes an entry specifying the local host and user name.
/usr/lpp/ssp/rcmd/bin/rsh r05n07 'export KRBTKTFILE=/tmp/rcmdtkt$$; \ /usr/lpp/ssp/rcmd/bin/rcmdtgt; \ /usr/lpp/ssp/rcmd/bin/rcp /tmp/stuff r05n05:/tmp/stuff;'
The root user sets the KRBTKTFILE environment variable to the name of a temporary ticket-cache file and then obtains a service ticket by issuing the rcmdtgt command. The rcp command uses the service ticket to authenticate from host r05n07 to host r05n05.
Purpose
removehsd - Removes one or more data striping devices, the IBM Virtual Shared Disks associated with them, and the System Data Repository (SDR) information for IBM Virtual Shared Disks on the associated nodes. If the only volume groups defined on the associated nodes are the ones used by the data striping devices that are being removed, those volume groups are also removed.
Syntax
Flags
Operands
None.
Description
Use this command to remove the logical volumes associated with IBM Virtual Shared Disks in the set of HSDs. The order in which the IBM Virtual Shared Disks that make up the HSDs and the HSDs themselves are removed is the reverse of the order in which they were created.
If the IBM Virtual Shared Disk or HSD is configured on any of the nodes on the system partition, this command fails, unless the -f flag is specified.
Files
Security
You must have sysctl and sysctl.vsd access and authorization from your system administrator to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: createhsd, removevsd
Examples
To unconfigure and remove the IBM Virtual Shared Disks associated with the HSD DATA and remove the HSD as well, type:
removehsd -d DATA -f
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit delete_vsdand select the Remove a Hashed Shared Disk option.
Purpose
removevsd - Removes a set of IBM Virtual Shared Disks that are not part of any data striping device (HSD).
Syntax
Flags
Operands
None.
Description
Use this command to remove the logical volumes associated with the IBM Virtual Shared Disks and update the backup nodes' Object Data Managers (ODMs), if any exist. The IBM Virtual Shared Disk information will be deleted from the System Data Repository (SDR). The removal of the IBM Virtual Shared Disks is done in the reverse of the order in which they were created. Volume groups are not removed with this command.
If the IBM Virtual Shared Disk is configured on any of the nodes on the system partition, this command fails, unless the -f flag is specified.
Note: | This command fails if one of the IBM Virtual Shared Disks named in vsd_names belongs to an HSD. To remove IBM Virtual Shared Disks that belong to an HSD, use removehsd. |
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit delete_vsdand select the Remove an IBM Virtual Shared Disk option.
Files
Security
You must have sysctl and sysctl.vsd access and authorization from your system administrator to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: createvsd, removehsd
Examples
To unconfigure and remove all defined IBM Virtual Shared Disks in a system or system partition, enter:
removevsd -a -f
Purpose
resumevsd - Activates an available IBM Virtual Shared Disk.
Syntax
resumevsd [-p | -b] -a | vsd_name ...
Flags
Note: | This flag is used only by the IBM Recoverable Virtual Shared Disk product. |
Operands
Description
The resumevsd command brings the specified IBM Virtual Shared Disks from the suspended state to the active state. The IBM Virtual Shared Disks remains available. Read and write requests which had been held while the IBM Virtual Shared Disk was in the suspended state are resumed.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit vsd_mgmtand select the Resume an IBM Virtual Shared Disk option.
Files
Security
You must have root privilege to run this command.
Restrictions
See IBM Parallel System Support Programs for AIX: Managing Shared Disks
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfgvsd, ctlvsd, lsvsd, preparevsd, startvsd, stopvsd, suspendvsd, ucfgvsd
Examples
To bring the IBM Virtual Shared Disk vsd1vg1n1 from the suspended state to the active state, enter:
resumevsd vsd1vg1n1
Purpose
rmkp - Removes Kerberos principals.
Syntax
rmkp -h
rmkp [-n] [-v] {name[.instance]|name.|.instance} ...
Flags
Operands
Note: | The name must be followed by a period and the instance must be preceded by a period. |
Description
Use this command to remove principals from the local Kerberos database. You will be prompted to confirm each deletion prior to its execution. This command will not remove any of the four principals that were predefined by Kerberos when the database was created. Deleted entries are saved in the /var/kerberos/database/rmkp.save.<PID> file, in the readable ASCII format produced by the kdb_util dump command. The rmkp command should normally be used only on the primary server. If there are secondary authentication servers, the push-kprop command is invoked to propagate the change to the other servers. The command can be used to update a secondary server's database, but the changes may be negated by a subsequent update from the primary.
Files
Standard Output
When the -v option is omitted, only the prompt for confirmation is written to standard output. When the -v flag is specified, the disposition of each selected principal is indicated by a message, and the name of the file containing the removed entries is printed. The -v flag has no effect on error messages written to standard error.
Exit Values
Security
The rmkp command can be run by the root user logged in on a Kerberos server host. It can be invoked indirectly as a Sysctl procedure by a Kerberos database administrator who has a valid ticket and is listed in the admin_acl.add file.
Restrictions
When you execute the rmkp command through the Sysctl procedure of the same name, the -n flag is added to your command invocation. This is required because Sysctl does not provide an interactive environment that supports prompting for confirmation. Suppressing confirmation increases the risk of unintentionally removing the wrong principal. In this mode, each principal to be removed must be named explicitly; selection of multiple principals by name or instance alone is not allowed. Since nonroot Kerberos administrators can execute this command only through Sysctl, you must be root on the server to use the special notation for selecting multiple principals.
Location
/usr/kerberos/etc/rmkp
Related Information
Commands: chkp, kadmin, kdb_util, lskp, mkkp, sysctl
Examples
rmkp tempuser
You should receive a prompt similar to the following:
Confirm removal of principal tempuser? (y or n): y
rmkp -v joe. frank rcmd.node25tr
You should receive prompts similar to the following:
Confirm removal of principal joe? (y or n): n joe was not removed Confirm removal of principal joe.admin? (y or n): y joe.admin was removed Confirm removal of principal frank? (y or n): y frank was removed Confirm removal of principal rcmd.node25tr? (y or n): y rcmd.node25tr was removed Removed entries were saved in /var/kerberos/database/rmkp.save.7942
Purpose
/usr/lpp/ssp/rcmd/bin/rsh - Executes the specified command at the remote host.
Syntax
{rsh | remsh} RemoteHost [-n] [-l user] [-k Kerberos_realm] command
Flags
Operands
Description
The rsh command executes command at RemoteHost. If command is not specified, the usage message is displayed. The rsh command sends standard input from the local command line to the remote command and receives standard output and standard error from the remote command.
Note: | Since any input to the remote command must be specified on the local command line, you cannot use the rsh command to execute an interactive command on a remote host. If you need to execute an interactive command on a remote host, use the AIX rlogin command. If you do not specify a command, the rsh command displays the usage message and exits. |
Remote Command Execution
While the remote command is executing, pressing the INTERRUPT, TERMINATE, or QUIT key sequences sends the corresponding signal to the remote process. However, pressing the STOP key sequence stops only the local process. Normally, when the remote command terminates, the local rsh process terminates.
To have shell metacharacters interpreted on the remote host, place the metacharacters inside double quotes (' '). Otherwise, the metacharacters are interpreted by the local shell.
When using the rsh command, you can create a link to a path (that you have permission to write) using a host name as the link name. For example,
ln -s /usr/lpp/ssp/rcmd/bin/rsh host_nameAt the prompt, if you enter the host name and a parameter (command), the rsh command is used to remotely execute the command specified on the command line on the remote host called host name. For example, if you are linked to remote host opus and want to perform the date command, you would enter:
opus dateAt the prompt, if you enter the host name with no parameters, rsh displays the usage message and exits.
Files
Security
If you do not specify the -l flag, the local user name is used at the remote host. If -l user is entered, the specified user name is used at the remote host. In either case, the remote host allows access only if at least one of the following conditions is satisfied:
For security reasons, any $HOME/.klogin file must be owned by either the remote user or root, and only the owner should have read and write access.
Should user authentication fail or should the authentication subroutine call fail, SP rsh passes the arguments to the AIX rsh command. In this case, the user needs normal AIX rsh access to a host for the command to be successful or the permission denied message is returned. The SP rsh command issues several authentication error messages before passing arguments to the AIX rsh command. The messages are also visible from system management applications that use the SP rsh command.
This command does not support encryption by users.
Prerequisite Information
Refer to AIX Version 4 System Management Guide: Communications and Networks for a network overview.
Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for an overview.
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Related Information
SP Commands: kinit, kshd, rcp
AIX Commands: rcp, rlogin, rshd, telnet
Examples
In the following examples, the local user is listed in the authentication database and has obtained a ticket by issuing a kinit and the password.
/usr/lpp/ssp/rcmd/bin/rsh host2 dfThe amount of free disk space on host2 is displayed on the local system.
/usr/lpp/ssp/rcmd/bin/rsh host2 cat test1 ">>" test2The file test1 is appended to test2 on remote host host2.
/usr/lpp/ssp/rcmd/bin/rsh host2 cat test2 >> test3The remote file test2 on host2 is appended to the local file test3.
/usr/lpp/ssp/rcmd/bin/rsh host2 -l jane cat test4 >> test5The remote file test4 is appended to the local file test5 using user jane's permissions at the remote host.
This example shows how the root user can issue an rcp on a remote host. The root user must be in the authentication database and must have already issued kinit on the local host. The command is issued at the local host to copy the file, stuff, from node r05n07 to node r05n05.
/usr/lpp/ssp/rcmd/bin/rsh r05n07 'export KRBTKTFILE=/tmp/rcmdtkt$$; \ /usr/lpp/ssp/rcmd/bin/rcmdtgt; \ /usr/lpp/ssp/rcmd/bin/rcp /tmp/stuff r05n05:/tmp/stuff;'
The root user sets the KRBTKTFILE environment variable to the name of a temporary ticket-cache file and then obtains a service ticket by issuing the rcmdtgt command. The rcp uses the service ticket to authenticate from host r05n07 to host r05n05.
Purpose
SDR_test - Verifies that the installation and configuration of the System Data Repository (SDR) component of the SP system completed successfully.
Syntax
SDR_test [-q] [-l log_file]
Flags
Operands
None.
Description
This command verifies that the SDR is functioning properly. After clearing out a class name, it creates the class, performs various SDR tasks, and removes the class when done.
A return code of 0 indicates that the test completed as expected; otherwise it returns the number of failures. If you do not specify the -q flag, a message is displayed on standard output that indicates the success or failure of the tests. In either case, the command returns 0 if successful, 1 if not. If errors are detected, more detailed information is recorded in the log file. If you do not specify the -l flag, error messages are recorded in /var/adm/SPlogs/SDR_test.log.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit SP_verify
and select the System Data Repository option.
Files
Related Information
Commands: CSS_test, jm_install_verify, jm_verify, SYSMAN_test, spmon_ctest, spmon_itest
Examples
To verify the System Data Repository following installation, saving error messages in sdr_err in the current working directory, enter:
SDR_test -l sdr_err
Purpose
SDRAddSyspar - Creates a new daemon using the System Resource Controller (SRC). The new daemon creates a subdirectory under the /spdata/sys1/sdr/partitions directory.
Syntax
SDRAddSyspar IP_address
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command creates a new instance of the SDR daemon and passes it the IP address of the system partition. It does not perform all of the system management tasks involved in creating a system partition.
Purpose
SDRArchive - Archives the entire contents of the System Data Repository (SDR), except for the archives directory, for later retrieval.
Syntax
SDRArchive [append_string]
Flags
None.
Operands
Description
Use this command to tar the contents of the SDR and put the file in the /spdata/sys1/sdr/archives subdirectory. You might want to mount this directory from another machine or physical disk drive to protect against a failure of the drive that the SDR exists on. The file name is backup.JULIANdate.HHMM.append_string, where JULIANdate.HHMM is a number or string uniquely identifying the date and time of the archive and append_string is the argument entered in the command invocation, if specified.
Purpose
SDRChangeAttrValues - Changes attribute values of one or more objects.
Syntax
SDRChangeAttrValues class_name [attr==value ... ] attr=value ...
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command changes one or more attribute values in a specified object with certain other attribute values.
Purpose
SDRClearLock - Unlocks a System Data Repository (SDR) class.
Syntax
SDRClearLock class_name
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
Use this command when a process that obtained a lock ends abnormally and fails to unlock the class.
Purpose
SDRCreateAttrs - Creates new attributes for a System Data Repository (SDR) class.
Syntax
SDRCreateAttrs class_name attr=datatype ...
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command creates one or more new attributes for a target class.
Purpose
SDRCreateClass - Creates a partitioned class.
Syntax
SDRCreateClass class_name attr=datatype ...
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command creates a partitioned class and defines its attributes.
Purpose
SDRCreateFile - Reads the specified AIX file and puts it in the System Data Repository (SDR) under the specified SDR file name.
Syntax
SDRCreateFile AIX_filename SDR_filename
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command creates a partitioned SDR file from an AIX file. Use SDRCreateSystemFile to create a system file. Use SDRRetrieveFile to retrieve the file.
Purpose
SDRCreateObjects - Creates new objects in a system class or a partitioned class.
Syntax
SDRCreateObjects class_name attr=value ...
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command creates one or more new objects. Not all attributes for an object need to be specified in this call; however, a subset of the attributes that uniquely identify this object must be entered at this time.
Purpose
SDRCreateSystemClass - Creates a system class.
Syntax
SDRCreateSystemClass class_name attr=datatype ...
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command creates a system class and defines its attributes.
Purpose
SDRCreateSystemFile - Creates a file that can be retrieved from any system partition.
Syntax
SDRCreateSystemFile AIX_filename SDR_filename
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command reads the AIX file and puts it in the repository under the SDR file name. Note that only ASCII files can be saved. Results are unpredictable if binary files are used with this command. Clients connected to any system partition can read this file.
Use SDRRetrieveFile to retrieve this file. If a system file and a partitioned file exist with the same name, the partitioned file will be returned from SDRRetrieveFile.
Purpose
SDRDeleteFile - Deletes the specified System Data Repository (SDR) file.
Syntax
SDRDeleteFile SDR_filename
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command deletes the partitioned file SDR_filename, if it exists. If the SDR_filename partitioned file does not exist, it will delete the SDR_filename system file. This command will not delete both the partitioned file and the system file.
Purpose
SDRDeleteObjects - Deletes objects from the System Data Repository (SDR).
Syntax
SDRDeleteObjects class_name [attr==value ... ]
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command deletes one or more objects. All objects in the specified class with attribute values matching those specified are deleted. If no attr==value pairs are specified, this command will match all objects in the class and all objects will be deleted.
Purpose
SDRGetObjects - Sends contents of attributes in specified object to standard output.
Syntax
Flags
Operands
Description
This command retrieves and sends to standard output attribute values in the specified objects.
Examples
SDRGetObjects -G Adapter adapter_type==css0 node_number netaddr
You should receive output similar to the following:
node_number netaddr 1 129.40.102.129 3 129.40.102.131 5 129.40.102.133 6 129.40.102.134 7 129.40.102.135 8 129.40.102.136 9 129.40.102.137 10 129.40.102.138 11 129.40.102.139 12 129.40.102.140 13 129.40.102.141 14 129.40.102.142 15 129.40.102.143 16 129.40.102.144
SDRGetObjects -G Node reliable_hostname switch_node_number \ switch_chip_port
You should receive output similar to the following:
reliable_hostname switch_node_number switch_chip_port k3n01.hpssl.kgn.ibm.com 0 3 k3n03.hpssl.kgn.ibm.com 2 0 k3n05.hpssl.kgn.ibm.com 4 1 k3n06.hpssl.kgn.ibm.com 5 0 k3n07.hpssl.kgn.ibm.com 6 2 k3n08.hpssl.kgn.ibm.com 7 3 k3n09.hpssl.kgn.ibm.com 8 3 k3n10.hpssl.kgn.ibm.com 9 2 k3n11.hpssl.kgn.ibm.com 10 0 k3n12.hpssl.kgn.ibm.com 11 1 k3n13.hpssl.kgn.ibm.com 12 1 k3n14.hpssl.kgn.ibm.com 13 0 k3n15.hpssl.kgn.ibm.com 14 2 k3n16.hpssl.kgn.ibm.com 15 3
SDRGetObjects -G -x Node node_number hdw_enet_addr > /etc/bootptab.info
You should receive output similar to the following:
1 02608C2D58D2 3 10005AFA2375 4 10005AFA22CE 5 10005AFA22B2 6 10005AFA2410 7 10005AFA223F 8 10005AFA2417 9 02608C2DA0C7 11 02608C2D9F62 13 02608C2D9E75 15 10005AFA1B03 16 10005AFA2B9B
Purpose
SDRListClasses - Lists the class names in the System Data Repository (SDR).
Syntax
SDRListClasses
Flags
None.
Operands
None.
Description
This command outputs all of the class names (system and partitioned) currently defined in the SDR to standard output.
Purpose
SDRListFiles - Lists all of the files in the system file area first, then lists all of the files in the system partition area.
Syntax
SDRListFiles
Flags
None.
Operands
None.
Description
This command outputs all the system partition files first, then the system files.
Purpose
SDRMoveObjects - Moves objects from one system partition to another.
Syntax
SDRMoveObjects source_syspar target_syspar class_name [attr==value ...]
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command moves any objects in class_name that match all of the attr==value pairs from the source_syspar to the target_syspar.
Purpose
SDRRemoveSyspar - Removes all of the partitioned classes in the System Data Repository (SDR) associated with the system partition whose address is IP_address. It removes the daemon that serves this system partition using the System Resource Controller (SRC).
Syntax
SDRRemoveSyspar IP_address
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command deletes a system partition in the SDR. It does not perform all of the system management tasks involved in deleting a system partition.
Purpose
SDRReplaceFile - Replaces the specified System Data Repository (SDR) file with the specified AIX file.
Syntax
SDRReplaceFile AIX_filename SDR_filename
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
This command searches first for a partitioned file, then for a system file, and replaces the first one found.
Purpose
SDRRestore - Extracts the contents of the archived System Data Repository (SDR).
Syntax
SDRRestore archive_file
Flags
None.
Operands
Description
Attention |
---|
The System Data Repository (SDR) commands are to be used by the IBM Parallel System Support Programs for AIX (PSSP) system management software. Use of these commands by a user can cause corruption of system configuration data. Exceptions are: SDRArchive, SDRGetObjects, SDRListClasses, SDRListFiles, SDRRetrieveFile, SDR_test, and SDRWhoHasLock. |
Use this command to remove the contents of the SDR and retrieve the archived contents of the archive_file. The archive_file must be in the /spdata/sys1/sdr/archives directory. Any new SDR daemons that represent partitions in the restored SDR are then started and any daemons that are not in the new SDR are stopped.
Related Information
Purpose
SDRRetrieveFile - Retrieves the specified System Data Repository (SDR) file into an AIX file.
Syntax
SDRRetrieveFile SDR_filename AIX_filename
Flags
None.
Operands
Description
This command searches first for a partitioned file, then for a system file if a partitioned file was not found.
Purpose
SDRWhoHasLock - Returns transaction ID of lock on specified class.
Syntax
SDRWhoHasLock class_name
Flags
None.
Operands
Description
The lock transaction ID returned from this command takes the form host_name:pid:session, where host_name is the long name of the machine running the process with the lock, pid is the process ID of the process that has the lock, and session is the number of the client's session with the System Data Repository (SDR).
Purpose
seqfile - Creates node sequence files for system startup and shutdown using information in the System Data Repository (SDR).
Syntax
Flags
Operands
None.
Description
seqfile uses information in the SDR to determine dependencies of SP nodes on /usr and, optionally, and boot/install servers, and to write the dependencies to standard output in the format of the node sequence files /etc/cstartSeq and /etc/cshutSeq.
/usr servers must shut down after and start up before their clients. Boot-install servers must start up before their clients. The node sequence files, /etc/cshutSeq and /etc/cstartSeq, have lines that describe these dependencies. The seqfile command eliminates the need for you to create these files from scratch. If the nodes in your system have sequencing dependencies in addition to those related to boot/install and /usr servers and clients, you can edit the output of seqfile to define those relationships.
seqfile defines only the nodes that have dependencies; if there are no /usr or boot/install dependencies, seqfile generates no output.
If you do not have a /etc/cstartSeq or /etc/cshutSeq file, the cstartup and cshutdown commands use seqfile to determine the default startup or shutdown sequence.
Files
The following files reside on the control workstation:
Related Information
Commands: cshutdown, cstartup
Examples
seqfile -b > /etc/cstartSeq
seqfile > /etc/cshutSeq
seqfile | more
Purpose
services_config - Configures designated services on nodes or the control workstation.
Syntax
services_config
Flags
None.
Operands
None.
Description
Use this command to configure SP services on the node or control workstation.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/install/bin/services_config
Related Information
Commands: setup_server, spbootins
Examples
To configure the SP services on a node or the control workstation, enter:
services_config
Purpose
sethacws - Sets the state of the control workstation.
Syntax
sethacws state
Flags
None.
Operands
Description
Use this command to set the current state of the control workstation. It is valid only when issued on the control workstation. When the command is executed and the calling process is not on a control workstation, an error occurs.
Note: | The High Availability Cluster Multiprocessing (HACMP) event scripts and installation scripts supplied with the High Availability Control Workstation (HACWS) option of the IBM Parallel System Support Programs for AIX (PSSP), set the control workstation state. The state is changed during fail over or reintegration in the HACWS supplied pre- and post-event scripts for HACMP. The administrator should not normally have to set the control workstation state. |
Exit Values
The following are the valid state values and their defined control workstation state:
Security
You must have root privilege to run this command.
Prerequisite Information
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for information on the HACWS option.
Location
/usr/bin/sethacws
Related Information
Command: lshacws
Subroutines: hacws_set, hacws_stat
Examples
sethacws 32
sethacws 16
sethacws 2
sethacws 1
sethacws 0
Purpose
setup_authent - Sets up a workstation to use SP authentication services.
Syntax
setup_authent
Flags
None.
Operands
None.
Description
The setup_authent command configures SP authentication services during SP installation on the control workstation and on other IBM RS/6000 workstations connected to an SP system. It is not executed on SP nodes, where authenticated client services are automatically installed. Executing this command invokes an interactive dialog, in which instructions for the various steps are displayed and various utility programs are invoked to accomplish the configuration.
There are several ways that setup_authent can configure these services. The method chosen is based on runtime choice, the combination of SP options installed on the workstation, and the contents of any predefined authentication configuration file that you have supplied.
Primary Server: When the local system is to be configured as the primary server, both ssp.clients and ssp.authent SP options must have been installed. You may supply the configuration files, /etc/krb.conf and /etc/krb.realms, or let the system create one that lists the local system as the sole authentication server in the local realm. This command creates the files used by the authentication and ticket granting services. These include the configuration files, the authentication database files, and the master key cache file. The server daemon, kerberos, is added to the inittab and started.
The administration of the authentication database is handled by the kadmind daemon, which is also added to inittab and started. The setup_authent command requires you to define the initial principal who administers the database. Access control list files are created containing this name, to be used by kadmind for authorization.
This command invokes kinit to log you in as this administrator, in order to define additional principals used by the SP authenticated services for monitoring and system administration. A server key file is created for use by the monitor commands, SP remote commands, and Sysctl remote command execution facility.
Backup Server: When the local workstation is to be configured as a secondary server, ssp.clients and ssp.authent must be installed. You must supply the configuration files, listing the local host as a slave server and some other workstation as the primary authentication server. The primary server must be configured and running and be available by standard TCP/IP connection to the local host.
You are required to authenticate your identity as the administrative user that you defined when you configured the primary server. The service principals for the local host are added to the primary database, and the server key file is created for them. Then the kpropd daemon is used in conjunction with the kprop command (executed remotely on the primary server) to copy the master database onto the local system. The server daemon, kerberos, is then added to the inittab and started.
Authentication Server: When the local host is to be configured only to provide authentication client services, just ssp.clients needs to be installed. As in the case of the slave server, you must supply the configuration files. In this case, however, the local host is not listed as a server. setup_authent simply requires the information to know how to get to the primary authentication server (already configured and accessible).
You are required to authenticate your identity as the administrative user that you defined when you configured the primary server. The service principals for the local host are added to the primary database, and the server key file is created for them.
Using AFS Authentication Services: When AFS authentication is to be configured, the local host must have already been established as either an AFS server or an AFS client. The CellServDB and ThisCell files are expected to exist in the /usr/vice/etc directory (or linked to that path). ssp.clients is the only required SP authentication option. When setup_authent finds these AFS configuration files on the local system, it allows you the choice of whether to use AFS authentication. If you choose not to use AFS, processing follows one of the other three variations described previously. When using AFS, you must supply an AFS user name and password that is a valid administrative ID in the local AFS cell. setup_authent creates the local service principals in the AFS database and creates a server key file for the SP authenticated services to use on the local host.
If you choose AFS authentication, you must do so for all workstations you configure with setup_authent, including the control workstation for your SP system.
You can reexecute setup_authent to change the configuration of your authentication services, but you add varying degrees of risk to system operations depending on how far you have progressed in the installation of the control workstation and nodes. Running it again on the control workstation prior to executing install_cw is not a problem. Reconfiguring a client workstation has little risk of disruption. A slave can be reconfigured provided the primary server is available. If the primary server must be reconfigured, all slave and client systems have to be reconfigured after the new primary server is up. If the control workstation is an authentication server, you have to recustomize any SP nodes previously booted, after running setup_authent.
Files
Related Information
Commands: add_principal, ext_srvtab, kadmind, kdb_edit, kdb_init, kdb_util, kerberos, kinit, klist, krb_conf, krb_realms, ksrvutil, kstash
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Purpose
setup_CWS - Updates control workstation files and directories for installation tasks.
Syntax
setup_CWS [-h]
Flags
Operands
None.
Description
Use this command to update control workstation files and directories for installation tasks. This includes control workstation-specific Kerberos and other files. This command can only be run on the control workstation.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/setup_CWS
Related Information
Commands: setup_server
Examples
To update the control workstation environment for installation, enter:
setup_CWS
Purpose
setup_logd - Sets up the logging daemon (splogd). This is called by installation scripts when the IBM RS/6000 control workstation is installed. It can also be run by root on a different workstation to have splogd spawned by the System Resource Controller (SRC).
Syntax
setup_logd
Flags
None.
Operands
None.
Description
To run the splogd logging daemon on a workstation other than the control workstation, install the ssp.clients option on that workstation and run setup_logd. You may want to do this in order to do the following:
By default the /spdata/sys1/spmon/hwevents file is set up to do error logging and state change logging for all frames. If you are installing splogd on a workstation besides the control workstation in order to call your own script, you should edit the /spdata/sys1/spmon/hwevents file, removing the entries for SP_STATE_LOG and SP_ERROR_LOG and add a call for your own script. Refer to the splogd command for instructions.
The setup_logd command performs the following steps:
If you do not want to perform any of the preceding steps on your workstation, do not run setup_logd. If you are only using splogd to call your own script, you might only want to do step 4 and step 5 (add splogd to SRC and /etc/inittab).
To run the logging daemon on a separate workstation, you must add the following to the /etc/environment file:
SP_NAME={control_workstation}To move a subset of error logging off of the control workstation, edit /spdata/sys1/spmon/hwevents on the control workstation to define the subset that you want to monitor. Then stopsrc and startsrc the logging daemon on the control workstation to reread the hwevents file.
Starting and Stopping the splogd Daemon
The splogd daemon is under System Resource Controller (SRC) control. It uses the signal method of communication in SRC. The splogd daemon is a single subsystem and not associated with any SRC group. The subsystem name is splogd. In order to start the splogd daemon, use the startsrc -s splogd command. This starts the daemon with the default arguments and SRC options. The splogd daemon is setup to be respawnable and be the only instance of the splogd daemon running on a particular node or control workstation. Do not start the splogd daemon from the command line without using the startsrc command to start it.
To stop the splogd daemon, use the stopsrc -s splogd command. This stops the daemon and does not allow it to respawn.
To display the status of the splogd daemon, use the lssrc -s splogd command.
If the default startup arguments need to be changed, use the chssys command to change the startup arguments or the SRC options. Refer to AIX Version 4 Commands Reference and AIX Version 4 General Programming Concepts: Writing and Debugging Programs for more information about daemons under SRC control and how to modify daemon arguments when under SRC.
To view the current SRC options and daemon arguments, use the odmget -q "subsysname=splogd" SRCsubsys command.
Files
Related Information
Daemon: splogd
Refer to IBM Parallel System Support Programs for AIX: Installation and Migration Guide for more information on setting up Hardware Monitor clients on separate workstations and the System Resource Controller.
Examples
startsrc -s splogd
stopsrc -s splogd
lssrc -s splogd
lssrc -a
odmget -q "subsysname=splogd" SRCsubsys
Purpose
setup_server - Configures a node or control workstation as a boot/install server.
Syntax
setup_server [-h]
Flags
Operands
None.
Description
Use this command to set up the node on which it is run as a boot/install server for client nodes as defined in the System Data Repository (SDR).
On a boot/install server, this command:
You must have a ticket-granting-ticket to run this command. Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for additional information on ticket-granting-tickets.
If you do not have a ticket-granting-ticket, you must run kinit.
Creation of the Network Installation Management (NIM) lppsource resource on a boot/install server will result in setup_server creating a lock in the lppsource directory on the control workstation. The setup_server command calls mknimres which creates the lock.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/setup_server
Related Information
Commands: allnimres, create_krb_files, delnimclient, delnimmast, export_clients, mkconfig, mkinstall, mknimclient, mknimint, mknimmast, mknimres, setup_CWS, unallnimres
Examples
To prepare a boot/install server node, enter the following on that node:
setup_server
Purpose
sp_configd - Starts the Simple Network Management Protocol (SNMP) SP Proxy daemon as a background process.
Syntax
sp_configd [-T] [-t secs] [-s secs] [-e secs]
Flags
Operands
None.
Description
The sp_configd daemon is internally configured as an SNMP Multiplexing Protocol (SMUX) peer, or proxy agent, of the snmpd daemon on the control workstation and on each node of the SP. For more information, refer to the "Managing SP System Events in a Network Environment" chapter in IBM Parallel System Support Programs for AIX: Administration Guide.
The sp_configd daemon provides the following functions:
The snmpd daemon should be active before the sp_configd daemon is started. The following command activates the snmpd daemon:
startsrc -s snmpd
The snmpd daemon is controlled by the System Resource Controller (SRC) and activated whenever the system is initialized.
The sp_configd daemon has several sessions with the EM. These sessions are used to maintain SP EM variable instance data and information from the last trap issued associated with an SP EM event. See the haem command for information on starting the SP Event Manager.
The sp_configd daemon should be controlled using the SRC. IBM suggests that you do not enter sp_configd at the command line.
Manipulating the sp_configd Daemon with the System Resource Controller
The sp_configd daemon is a subsystem controlled by the SRC. Use the following SRC commands to manipulate the sp_configd daemon:
Files
smux 199/tcp
Notes:
smux 1.3.6.1.4.1.2.6.117 sp_configd_pw # sp_configdThese entries are created when the SP is installed.
sp_configd 1.3.6.1.4.1.2.6.117 sp_configd_pwThese entries are created when the SP is installed.
Security
You must have root privilege to run this command or be a member of the system group.
Related Information
See the "Understanding the SNMP Daemon" and "Problem Determination for the SNMP Daemon" chapters in AIX Version 3.2 System Management Guide: Communications and Networks.
See the "Understanding the Simple Network Management Protocol (SNMP)", "Understanding the Management Information Base (MIB)", and "xgmon Overview for Programmers" chapters in AIX Version 3.2 Communication Programming Concepts.
See the "Managing SP System Events in a Network Environment" chapter in IBM Parallel System Support Programs for AIX: Administration Guide.
Examples
startsrc -s sp_configd -a '-T'
This command starts the sp_configd daemon and logs information to the /var/tmp/sp_configd.log file.
stopsrc -s sp_configd
This command stops the daemon. The -s flag specifies the subsystem that follows to be stopped.
lssrc -s sp_configd
This command returns the daemon name, process ID, and state (active or inactive).
Purpose
sp_configdctrl - A control script that is used to manage the installation of the SP Simple Network Management Protocol (SNMP) Proxy Agent subsystem.
Syntax
sp_configdctrl { -a | -s | -k | -d | -c | -t | -o | -r | -h }
Flags
Operands
None.
Description
Use this command to install or remove the SP SNMP Proxy Agent daemon. This command can be issued only by a user with root privileges or by a member of the system group.
The sp_configdctrl control script controls the operation of the SP SNMP Proxy Agent subsystem. The subsystem is under the control of the System Resource Controller (SRC). The subsystem is called sp_configd.
An instance of the SP SNMP Proxy Agent subsystem executes on the control workstation and on every node of a system partition. Because the information about SP nodes and Event Manager (EM) variables exists in system partitions, it is said to be system partition-sensitive. This control script operates in a manner similar to the control scripts of other system partition-sensitive subsystems. It can be issued from either the control workstation or any of the system partition's nodes.
From an operational point of view, the SP SNMP Proxy Agent subsystem group is organized as follows:
The sp_configd subsystem is associated with the sp_configd daemon.
The subsystem name on the nodes and the control workstation is sp_configd. There is one daemon per node and control workstation.
The sp_configd daemon provides the SP SNMP Proxy Agent function.
The sp_configdctrl script is not normally executed from the command line. It is normally called by the syspar_ctrl command during installation of the system, and partitioning or repartitioning of the system.
The sp_configdctrl script provides a variety of controls for operating the SP SNMP Proxy Agent subsystem:
Adding the Subsystem
When the -a flag is specified, the control script uses the mkssys command to add the SP SNMP Proxy Agent subsystem to the SRC. The control script operates as follows:
Starting the Subsystem
When the -s flag is specified, the control script uses the startsrc command to start the SP SNMP Proxy Agent subsystem, sp_configd.
Stopping the Subsystem
When the -k flag is specified, the control script uses the stopsrc command to stop the SP SNMP Proxy Agent subsystem, sp_configd.
Deleting the Subsystem
When the -d flag is specified, the control script uses the rmssys command to remove the SP SNMP Proxy Agent subsystem from the SRC. The control script operates as follows:
Cleaning Up the Subsystem
When the -c flag is specified, the control script stops and removes the SP SNMP Proxy Agent subsystem from the SRC. The control script operates as follows:
Turning Tracing On
When the -t flag is specified, the control script turns tracing on for the sp_configd daemon, by stopping the daemon and restarting it with the -T option.
Turning Tracing Off
When the -o flag is specified, the control script turns tracing off for the sp_configd daemon, by stopping the daemon and restarting it without the -T option.
Refreshing the Subsystem
The -r flag has no effect for this subsystem.
Files
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must be running with an effective user ID of root.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
AIX Version 4 Commands Reference
Information about the System Resource Controller (SRC) in AIX Version 4 General Programming Concepts: Writing and Debugging Programs
Location
/usr/lpp/ssp/bin/sp_configdctrl
Related Information
Commands: sp_configd
Examples
sp_configdctrl -a
sp_configdctrl -s
sp_configdctrl -k
sp_configdctrl -d
sp_configdctrl -c
sp_configdctrl -t
sp_configdctrl -o
lssrc -s sp_configd
Purpose
spacctnd - Enters accounting data into the System Data Repository for a node or group of nodes.
Syntax
Flags
Operands
Note: | The start_frame and start_slot must resolve to a node in the current system partition. |
Note: | The node_count is considered to be within the current system partition. |
Description
Run this command during installation of the SP or later to set the accounting class identifier, the accounting enabled attribute, job charge value or the exclusive use accounting enabled attribute of a node or set of nodes.
You can use the System Management Interface Tool (SMIT) to run the spacctnd command. To use SMIT, enter:
smit node_dataand select the Accounting Information option.
Note: | This command should be run only on the control workstation. You must be logged into the control workstation as root to execute this command. |
Examples
The following example adds accounting SDR information for a system with 2 frames and 32 nodes. Accounting and exclusive use accounting is to be enabled for each node and 60 seconds of exclusive use by a user is to constitute one charge fee unit.
spacctnd -e true -j 60.0 -x true 1 1 32
Purpose
spacs_cntrl - Controls interactive access to SP nodes.
Syntax
Flags
Operands
Description
Use caution when issuing this command while the Resource Manager is running. If the Resource Manager is configured to use Login Control, you may cause loss of user state information. For more Login Control information, refer to IBM Parallel System Support Programs for AIX: Administration Guide.
The following types of access can be disallowed when spacs_cntrl block or deny is used:
The spacs_cntrl command does not allow individual types of access to be disallowed.
Duplicate user names are removed from the user list whether entered from the command line or in a file.
If you add a new user to a node for which all users are denied, you must reissue spacs_cntrl to deny the new user as well.
Flags and Logging
Flags specified in combination have the following results:
Use of the verbose flag (-v) causes the command to run longer due to extra processing required to find state information.
Examples
spacs_cntrl block betty
dsh -a rcp root@mynode:/tmp/usr.input /tmp/usr.input
dsh -a spacs_cntrl -f /tmp/usr.input block
Purpose
spadaptrs - Enters configuration data for an additional adapter for a node or series of nodes in the System Data Repository (SDR).
Syntax
Flags
If -s is not specified, the default is no.
Operands
Note: | The start_frame and start_slot must resolve to a node in the current system partition. |
Note: | The node_count is considered to be within the current system partition. |
Each IP address used in the operation must be resolved by the host command on the control workstation.
Description
Execute this command during installation of the SP to identify the IP addresses, netmask, and default route associated with node adapters other than en0. If all your IP addresses are in the same block, run this command once. If you have "holes" in your IP addressing scheme, run this command once for each block of addresses you wish to assign. Special considerations apply to how IP addressing is done for the High Performance Switch. Refer to IBM Parallel System Support Programs for AIX: Administration Guide for more High Performance Switch information.
You must have a ticket-granting-ticket to run this command. Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for additional information on ticket-granting-tickets.
If you do not have a ticket-granting-ticket, you must run kinit.
You can use the System Management Interface Tool (SMIT) to run the spadaptrs command. To use SMIT, enter:
smit node_data
and select the Additional Adapter Information option.
Notes:
Related Information
Commands: syspar_ctrl
Examples
If you specify the -s flag to skip IP addresses when setting the css0 switch addresses, you must also specify -n no to not use switch numbers for an IP address assignment. You must specify -a yes to use ARP.
spadaptrs -s yes -n no -a yes 1 1 30 css0 129.33.34.1 255.255.255.0
Purpose
spapply_config - Applies a system partition configuration to the SP system.
Syntax
spapply_config [-h] [-v] [-A] [-F] [-q] config_dir/layout_dir
Flags
Operands
Description
This command functions in two phases: verification and application. Before applying a new system partition configuration, the administrator should back up the SP system SDR. This can be accomplished by using either the SDRArchive command or by using the -A flag on spapply_config. If your system has an SP switch, the Eunpartition command must be run before applying a new system partition configuration. Failure to do this will produce unpredictable results in the new system partitions. Refer to the "Managing System Partitions" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional information.
The layout directory contains one system partition directory for each system partition in the configuration. Each partition directory contains the switch topology file and nodelist file. It also contains the custom file (created and updated by the spcustomize_syspar command). The spapply_config command verifies that these files exist. It also verifies the contents of the custom file. If an error is encountered in this verification phase, the command issues an appropriate message and terminates without attempting to apply a configuration layout that is not valid. As part of its verification phase, this command also calls the verparvsd command to determine the impact on the IBM Virtual Shared Disk subsystem of applying the specified configuration layout. If any errors or warnings are returned from verparvsd, the spapply_config command reports those messages and stops. The -F flag can be used to alter this behavior by correcting nonfatal IBM Virtual Shared Disk errors encountered in the analysis of the IBM Virtual Shared Disk subsystem.
As part of its processing, spapply_config displays to standard output the list of changed system partitions and the list of unchanged system partitions. A changed system partition is a currently-defined partition which will be changed in some way by the application of the specified configuration layout. Nodes in changed system partitions should be shutdown prior to applying that configuration. Conversely, an unchanged system partition is a currently-defined partition which will be unchanged by the application of the specified configuration layout. Nodes in unchanged system partitions can remain in operation during the application of this configuration layout. The spapply_config command issues the Eannotator, Eprimary, and Etopology commands as necessary.
The spapply_config command issues status messages which track the progress of operation to standard output. These messages along with the lists of changed and unchanged system partitions can be suppressed by the using the -q flag.
In the event that spapply_config encounters an error during the application phase, a descriptive error message is displayed and the command stops. In this case, it will be necessary to restore the SP SDR and the system partition-sensitive subsystems (for example, hats, hb, and hr) to their previous state by using the sprestore_config command.
Note: | Due to system partitioning changes, your SP_NAME environment variable may no longer be set to a valid system partition name. To get a list of valid system partition names, enter the splst_syspars -n command. Then verify that your SP_NAME environment variable is either unset or set to one of the system partition names in the list. |
Files
Related Information
Commands: Eunpartition, SDRArchive, spcustomize_syspar, spdisplay_config, sprestore_config, spverify_config, syspar_ctrl, verparvsd
Files: nodelist, topology
Examples
spapply_config config.4_12/layout.2
spapply_config -v config.8_8/layout.1
Purpose
spbootins - Enters boot/install configuration data for a node or series of nodes in the System Data Repository (SDR).
Syntax
Flags
Note: | IBM strongly suggests that you use the hardware location format. It ensures that you install on the intended disk by targeting a specific disk at a specific location. The relative location of hdisks may change depending on hardware installed or possible hardware failures. IBM strongly suggests always using the hardware location format when there are external drives present, as the manner in which the device names are defined may not be obvious. |
A node that boots in a prompted mode, comes up with the "Install/Maintenance" panel. From this panel, you can choose option 3 to start a limited function maintenance shell. You may access files in the root volume group (rootvg) by choosing the panels to mount the root volume group and enter a shell.
The diag parameter can be used only on the IBM Parallel System Support Programs for AIX (PSSP) Version 2 Release 1 and later nodes. PSSP Version 1 Release 2 nodes cannot run remote diagnostics.
You can only specify this flag for a node in an IBM Parallel System Support Programs for AIX (PSSP) Version 1 Release 2 system partition.
You can only specify this flag for a node in an IBM Parallel System Support Programs for AIX (PSSP) Version 1 Release 2 system partition.
You can only specify this flag for a node in an IBM Parallel System Support Programs for AIX (PSSP) Version 1 Release 2 system partition.
Operands
Note: | The start_frame and start_slot must resolve to a node in the current system partition. |
Note: | The node_count is considered to be within the current system partition. |
Description
Execute this command during installation of the SP, or later, to identify which node is to be the boot/install server for a node or group of nodes. You can also use this command to indicate the name of the install image to be used to install a node or group of nodes the next time they are network-installed. Each time this command is run, the setup_server command is run on each of the affected boot/install servers.
You must have a ticket-granting-ticket to run this command. Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for additional information on ticket-granting-tickets.
If you do not have a ticket-granting-ticket, you must run kinit.
You can use the System Management Interface Tool (SMIT) to run the spbootins command. To use SMIT, enter:
smit node_data
and select the Boot/Install Information option.
You cannot use SMIT if you are using AFS authentication services.
Notes:
Examples
spbootins -n 9 1 10 7
spbootins -r install 2 1 16
Purpose
spchuser - Changes the attributes of an SP user account.
Syntax
spchuser attribute=value ... name
Flags
None.
Operands
Supported Attributes and Values
Description
No flags are supported. Except for home, the rules for the supported attributes and values correspond to those enforced by the AIX chuser command.
You can only change the values of the supported attributes.
You can use the System Management Interface Tool (SMIT) to run the spchuser command. To use SMIT, enter:
smit spusers
and select the Change/Show Characteristics of a User option.
Note: | This command should be run only on the control workstation. You must be logged into the control workstation as root to execute this command. |
Examples
To change the default shell to /bin/csh, and change the secondary group membership to dev and dev2 for the user account charlie:
spchuser groups=dev,dev2 shell=/bin/csh charlie
Purpose
spcustomize_syspar - Enters or verifies customization information to be used in creating a system partition.
Syntax
Flags
Operands
Description
Use this command to customize a system partition customization file (custom) or to display the previously-entered customization information.
For a specified system partition, the customization data can be entered with the optional parameters. If the custom file does not exist, you can create one by specifying the -n and -l flags. The -d and -e flags are optional when creating a custom file. If -d and -e are not specified, the system automatically specifies default to set the default install image and primary node in the newly-created custom file. Once the custom file is created, any combination of the optional parameters can be used to update the contents of the file.
If none of the optional parameters are specified with the spcustomize_syspar command, the contents of the customization file for the specified system partition are displayed to standard output as in the spdisplay_config command with the -c flag.
Related Information
Commands: spapply_config, spdisplay_config, spverify_config
Files: nodelist, topology
Examples
spcustomize_syspar config.4_12/layout.1/syspar.1 syspar-name: my-partition-1 IP-address: 9.102.55.301 PSSP-code-level: PSSP-2.1 default-install-image: bos.4.1.2 primary-node: 9 backup-primary-node: 16
spcustomize_syspar -n my-new-partition-name -l PSSP-2.1 \ -e 7 config.4_12/layout.1/syspar.1
spcustomize_syspar -e default config.4_12/layout.1/syspar.1
Purpose
spcw_addevents - Identifies the High Availability Cluster Multiprocessing (HACMP) event scripts supplied by the High Availability Control Workstation (HACWS) to the AIX High Availability Cluster Multi-Processing (HACMP) software.
Syntax
spcw_addevents
Flags
None.
Operands
None.
Description
HACWS customizes the recovery of control workstation services by providing HACMP event scripts, which get executed by the HACMP software. The spcw_addevents command is a shell script which identifies the HACMP event scripts to HACMP, without requiring the system administrator to go through all the equivalent HACMP SMIT panels.
Exit Values
Security
You must have root privilege to run this command.
Prerequisite Information
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for additional information on the HACWS option.
Location
/usr/sbin/hacws/spcw_addevents
Purpose
spcw_apps - Starts or stops control workstation applications in a High Availability Control Workstation (HACWS) configuration.
Syntax
spcw_apps {-u | -d} [-i | -a]
Flags
Operands
None.
Description
The control workstation services are started at boot time on a regular control workstation via entries in /etc/inittab. An HACWS configuration requires the capability to stop control workstation services on one control workstation and restart them on the other. The install_hacws command removes most of the control workstation entries from /etc/inittab, and the spcw_apps command is provided as a means to stop and start control workstation services in the HACWS configuration. In addition, the spcw_apps command can be used to make the inactive control workstation act as a client of the active control workstation in order to keep the two control workstations synchronized.
Note: | The High Availability Cluster Multiprocessing (HACMP) event scripts and installation scripts supplied with the High Availability Control Workstation (HACWS) option of the IBM Parallel System Support Programs for AIX (PSSP) will start or stop the control workstation applications during a fail over or reintegration. The administrator should not normally have to start or stop the control workstation applications. |
Exit Values
Prerequisite Information
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for additional information on the HACWS option.
Location
/usr/sbin/hacws/spcw_apps
Related Information
Command: install_hacws
Examples
In the following example, assume that the primary control workstation is currently the active control workstation. This means that the primary control workstation is providing control workstation services to the SP system. When a control workstation failover occurs, the AIX High Availability Cluster Multi-Processing (HACMP) software moves the control workstation network and file system resources from the primary to the backup control workstation. In addition, control workstation applications must be stopped on the primary and restarted on the backup. HACWS provides the spcw_apps command to HACMP as the method to accomplish this. The HACMP software issues the following command on the primary:
spcw_apps -di
This command stops control workstation services on the active primary and then sets the primary to be the inactive control workstation. Next, the HACMP software issues the following command on the backup:
spcw_apps -ua
This command sets the backup to be the active control workstation and then starts the control workstation services on the backup. Finally, the HACMP software issues the following command on the primary:
spcw_apps -u
This command configures the primary to be a client of the backup (which is active) control workstation.
Purpose
spdeladap - Removes configuration data for adapters for a node or series of nodes from the System Data Repository (SDR).
Syntax
Flags
Operands
Note: | The start_frame and start_slot must resolve to a node in the current system partition. |
Note: | The node_count is considered to be within the current system partition. |
Description
Use this command to remove configuration data for adapters for a node or series of nodes from the SDR. You cannot use this command to delete data for the en0 adapter. If you want to remove configuration data for the en0 adapter, you should use the spdelnode command.
You can use the System Management Interface Tool (SMIT) to run the spdeladap command. To use SMIT, enter:
smit delete_data
and select the Delete Adapter Information option.
Notes:
Related Information
Commands: syspar_ctrl
Examples
This example deletes tr0 adapter information for the the first two nodes in frame 2:
spdeladap 2 1 2 tr0
Purpose
spdelfram - Removes configuration data for a frame or series of frames from the System Data Repository (SDR).
Syntax
spdelfram start_frame frame_count
Flags
None.
Operands
Description
Execute this command to remove configuration data for a frame or series of frames from the SDR. Any node information for nodes on the frames is also removed, as well as adapter information for any of the nodes on the frames. All the nodes on all the frames to be deleted must be shut down. A frame containing a node acting as a server for another node on a frame not being removed with this operation cannot be removed with this command. In order to remove a frame containing such a node, you must configure a different boot/install or /usr server for the client nodes on the other frames.
The spdelfram command removes all extension node and extension node adapter information for extension nodes whose node numbers are within the range of node numbers represented by a frame being deleted.
You must have a ticket-granting-ticket to run this command. Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for additional information on ticket-granting-tickets.
If you do not have a ticket-granting-ticket, you must run kinit.
You can use the System Management Interface Tool (SMIT) to run the spdelfram command. To use SMIT, enter:
smit delete_data
and select the Delete Frame Information option.
Notes:
Examples
This example deletes the last two frames in a four-frame system:
spdelfram 3 2
Purpose
spdelnode - Removes configuration data for a node or nodes from the System Data Repository (SDR).
Syntax
spdelnode {start_frame start_slot node_count | -N node_group | -l node_list}
Flags
Operands
Note: | The start_frame and start_slot must resolve to a node in the current system partition. |
Note: | The node_count is considered to be within the current system partition. |
Description
Execute this command to remove configuration data for a node or series of nodes from the SDR. Any adapter information associated with the node is also removed. All the nodes to be deleted must be shut down. A node acting as a server for another node cannot be removed with this command. In order to remove a node which is a server, you must configure a different boot/install server for the client nodes.
This command can be used to delete nodes in the current system partition. To delete nodes in another system partition, you must make it your current system partition.
You must have a ticket-granting-ticket to run this command. Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for additional information on ticket-granting-tickets.
If you do not have a ticket-granting-ticket, you must run kinit.
You can use the System Management Interface Tool (SMIT) to run the spdelnode command. To use SMIT, enter:
smit delete_data
and select the Delete Node Information option.
Notes:
Security
You must be logged into the control workstation as root to run this command.
Related Information
Commands: syspar_ctrl
Examples
spdelnode 2 1 2
spdelnode -N temp_nodes
Purpose
spdisplay_config - Displays system partition configuration information which can be used to partition an SP system.
Syntax
Flags
Operands
Description
This command displays system partition information stored in the system partition information directory structure. Depending on the option and operand specified, the information displayed is at the configuration, layout, or system partition level. The output of this command is normally restricted to the SP on which it is executed. To display information for configurations applicable to SP systems other than the one on which it is executed, a fully qualified path name must be provided. This command does not display current system partition information (that function is provided with the splstdata command). If the command is issued without specifying a directory path name, the list of all valid configuration directories for the SP on which the command is running is displayed. If none of the file option flags (-c, -d, -n) are entered, the names of the files and subdirectories located on (and all levels below if -R is specified) the specified directory are displayed. If any of the file option flags are entered, the contents of the requested files are displayed instead.
Files
Related Information
Commands: spapply_config, spcustomize_syspar, splstdata, spverify_config
Files: nodelist, topology
Examples
spdisplay_config config.2_14 config.4_12 config.8_8
spdisplay_config config.4_12 layout.1 layout.2 layout.3
spdisplay_config config.4_12/layout.2 layout.desc syspar.1 syspar.2
spdisplay_config config.4_12/layout.2/syspar.1 custom nodelist topology
If the -c flag is also supplied, only the customization information is displayed for that system partition. For example:
spdisplay_config -c config.4_12/layout.2/syspar.1 custom: syspar-name: my-partition-1 IP-address: 9.102.55.301 primary-node: 9 default-install-image: bos.4.1.2 PSSP-code-level: PSSP-2.1
If the -n flag is supplied, only the list of nodes is displayed for that system partition. For example:
spdisplay_config -n config.4_12/layout.2/syspar.1 nodelist: switch node numbers: 4 5 6 7 8 9 10 11 12 13 14 15 node numbers: 5 6 7 8 9 10 11 12 13 14 15 16
All of the commands can be issued with the -R flag to recursively display the information on the specified directory and all the levels below it.
To display the entire system partition information directory structure for the config.4_12 configuration on this SP, enter:
spdisplay_config -R config.4_12 layout.1: layout.1/layout.desc layout.1/syspar.1: layout.1/syspar.1/custom layout.1/syspar.1/nodelist layout.1/syspar.1/topology layout.1/syspar.2: layout.1/syspar.2/custom layout.1/syspar.2/nodelist layout.1/syspar.2/topology layout.2: layout.2/layout.desc layout.2/syspar.1: layout.2/syspar.1/custom layout.2/syspar.1/nodelist layout.2/syspar.1/topology layout.2/syspar.2: layout.2/syspar.2/custom layout.2/syspar.2/nodelist layout.2/syspar.2/topology
Another example to recursively display all of the customization information for the config.4_12 on this SP follows:
spdisplay_config -R -c config.4_12 layout.1/syspar.1/custom: syspar-name: my-partition-1 IP-address: 9.102.55.301 primary-node: 9 default-install-image: bos.4.1.2 PSSP-code-level: PSSP-2.1 layout.1/syspar.2/custom: syspar-name: my-partition-2 IP-address: 9.102.55.302 primary-node: 9 default-install-image: bos.4.1.2 PSSP-code-level: PSSP-2.1 layout.2/syspar.1/custom: syspar-name: my-partition-1 IP-address: 9.102.55.501 primary-node: 9 default-install-image: bos.4.1.2 PSSP-code-level: PSSP-2.1 layout.2/syspar.2/custom: syspar-name: my-partition-2 IP-address: 9.102.55.502 primary-node: 9 default-install-image: bos.4.1.2 PSSP-code-level: PSSP-2.1
spdisplay_config /spdata/sys1/syspar_configs/1nsb0isb config.16 config.4_12 config.4_4_4_4 config.4_4_8 config.8_8
Purpose
spethernt - Enters configuration data for a node or series of nodes in the System Data Repository (SDR).
Syntax
Flags
If you are specifying the nodes to be used in this operation with a node_list via the -l flag, you must not specify -s yes.
Operands
Note: | The start_frame and start_slot must resolve to a node in the current system partition. |
Note: | The node_count is considered to be within the current system partition. |
Ensure that the combination of the starting IP address, the node count operand, and the -s flag do not result in the incrementing of the IP address to an invalid IP address.
Each IP address used in the operation must be resolved by the host command on the control workstation.
Description
Execute this command during installation of the SP to identify the IP addresses, netmask, and default route associated with the en0 adapters of the nodes. If all your IP addresses are in the same block, run this command once. If you have "holes" in your IP addressing scheme, run this command once for each block of addresses you wish to assign.
You must have a ticket-granting-ticket to run this command. Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for additional information on ticket-granting-tickets.
If you do not have a ticket-granting-ticket, you must run kinit.
You can use the System Management Interface Tool (SMIT) to run the spethernt command. To use SMIT, enter:
smit node_data
and select the SP Ethernet Information option.
Notes:
Related Information
Commands: syspar_ctrl
Examples
The following example adds SDR information for an en0 network of 15 nodes (frame 1 slot 1 to frame 1 slot 16 with the first node being a wide node and the rest of the nodes thin nodes all in a single system partition) with IP addresses from 129.33.32.1 to 129.33.32.15, a netmask of 255.255.255.0, and a default route of 129.33.32.200. The addresses are to be assigned to correspond with the nodes in the frame; for example, they do not increment the IP address of a wide node by 2 before assigning the IP address of the next node.
spethernt -s no 1 1 15 129.33.32.1 255.255.255.0 129.33.32.200
Purpose
spevent - Directly invokes the Event Management window of the SP Perspectives graphical user interface (GUI).
Syntax
Flags
Note: | Most flags accepted by X will also be recognized. For example, -display displayname. |
Operands
None.
Description
Use this command to launch the Event window of the SP Perspectives GUI. The Event Perspective is a graphic vehicle for managing events and event definitions in the system.
This perspective allows the user to interact with the Event Manager and the Problem Manager. The user can create event definitions, register or unregister event definitions, and delete event definitions. A notebook is provided for viewing and editing the condition, the resource identifier, and other attributes of an event's definition. The notebook also provides users with the capability for creating new conditions. Event definitions are viewed and manipulated within system partition boundaries.
By default, the Event Definition pane displays all the event definitions with the Kerberos principal of the user within the current system partition. The user can use the filter function provided on the pane to expand or further confine the event definitions being displayed.
When the command is invoked, preferences which define the look and layout of the spevent window are prioritized in the following order:
Files
The Users Preferences are read from and saved to $HOME/.spevent(User Profile Name).
The System Preferences are read from and saved to /usr/lpp/ssp/perspectives/profiles/.spevent(System Profile name).
Restrictions
Any user can run the spevent command. Many actions require root privilege to run.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
For information on using the Event window and the SP Perspectives GUI, please see the online help and the "Using the SP Perspectives GUI" chapter in the IBM Parallel System Support Programs for AIX: Administration Guide.
For information on the Event Management services, refer to "The Event Management Subsystem" and "Using the Problem Management Subsystem" chapters in IBM Parallel System Support Programs for AIX: Administration Guide.
Location
/usr/lpp/ssp/bin/spevent
Related Information
You can also access the Event window by using the SP Perspectives Launch Pad. The perspectives command invokes the launch pad. Other Perspectives windows may be launched by invoking the following commands: sphardware, spsyspar, spvsd, and spperfmon.
Examples
spevent
spevent -noProfile
Purpose
spframe - Enters configuration data for a frame or series of frames and, optionally, set up the initial System Data Repository (SDR).
Syntax
Flags
The default is -r no.
Operands
Description
Execute this command during installation of the SP to identify the tty ports to which your frames are connected. If the tty special files for your frames are consecutively numbered, run this command once during installation, specifying -r yes. If you have "holes" in your tty special file numbering scheme, run this command once for each block of ttys you wish to assign, specifying -r yes only on the last invocation of the command.
You must have a ticket-granting-ticket to run this command. Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for additional information on ticket-granting-tickets.
You can use the System Management Interface Tool (SMIT) to run the spframe command. To use SMIT, enter:
smit enter_data
and select the Frame Information option.
Note: | This command should be run only on the control workstation. You must be logged into the control workstation as root to execute this command. |
Examples
The following example enters information for four frames (frame 1 to frame 4) and indicates that frame 1 is connected to /dev/tty1, frame 2 to /dev/tty2, and so on. The System Data Repository is to be initialized with this invocation of spframe.
spframe -r yes 1 4 /dev/tty1
Purpose
spget_syspar - Returns the IP address or name of the current system partition.
Syntax
spget_syspar [-n] [-d]
Flags
Operands
None.
Description
Use this command to display to standard output the IP address (in dotted decimal format) of the current or default system partition. The current system partition indicates the system partition to which System Data Repository (SDR) client requests are directed. The result is displayed in dotted decimal format unless -n is specified.
Restrictions
The -d flag will not work if the command is not issued on a control workstation or SP node.
Examples
spget_syspar
You should receive output similar to the following:
129.40.127.122
spget_syspar -n
You should receive output similar to the following:
k47sp1
Purpose
spgetdesc - Obtains the description information from the nodes specified and, optionally, enters it into the SDR.
Syntax
Flags
"node_number:hostname:description"for all nodes that successfully obtained the description information.
One of the following flags must be specified:
Operands
None.
Description
This command will obtain the description information from the nodes specified and, optionally, update the description attribute in the SDR Node class. Unless the -f flag is specified, the node's host_responds will be checked before it will attempt to dsh to the nodes and obtain their description information. This command requires that the user be authenticated to Kerberos using the kinit command. This command is primarily intended as a migration tool for obtaining description information from existing nodes. The description information will be obtained from new nodes when they are installed or customized.
Security
Since this command uses dsh, it requires that the user issuing the command obtain a valid Kerberos ticket.
Standard Output
The model information that is obtained will be printed to standard output as well as placed in the SDR.
Standard Error
This command writes error messages (as necessary) to standard error. Errors will be printed if any attempt to get the description information fails on a node.
Exit Values
Consequences of Error
If the command does not run successfully it terminates with an error message and a nonzero return code. Messages will inform the user of the cause of the error. For a terminal error, no description information will be obtained. For a nonterminal error, no description information will be obtained for the node that had the error.
Examples
To obtain description information for all existing nodes enter:
spgetdesc -a
To obtain description information for nodes 3 & 4 and update the SDR, enter:
spgetdesc -ul 3,4
To obtain description information from all nodes in a colon delimited format, enter:
spgetdesc -ac
Implementation Specifics
This command is part of the Parallel System Support Program (PSSP) LPP.
Location
/usr/lpp/ssp/bin/spgetdesc
Related Information
SP Commands: dsh, kinit, SDRChangeAttrValues, SDRGetObjects
AIX Commands: uname
Purpose
sphardware - Directly launches the Hardware window of the SP Perspectives graphical user interface (GUI).
Syntax
Flags
Note: | Most flags accepted by X will also be recognized. For example, -display displayname. |
Operands
None.
Description
Use this command to launch the Hardware window of the SP Perspectives GUI. From the Hardware window, the user can monitor and manipulate objects within the SP system. The SP objects included in this window are the control workstation, SP system and system partitions, nodes, node groups, and frames and switch boards.
By default, the Hardware window will display the CWS/System/Syspars Work Area and the Node Work Area. The CWS/System/Syspars pane contains all system partitions and indicates the current system partition, and the Node pane contains all the nodes in the current system partition in the "frame" display control orientation.
The Node Group and Frame/Switchboard panes can be added to the Hardware window by using combo boxes at the top of the window. Likewise, any pane that is displayed can be deleted from the window by using the Delete combo box at the top of the window.
When the command is invoked, preferences which define the look and layout of the sphardware window are prioritized in the following order:
Files
The Users Preferences are read from and saved to $HOME/.sphardware(User Profile Name).
The System Preferences are read from and saved to /usr/lpp/ssp/perspectives/profiles/.sphardware(System Profile name).
Restrictions
Any user can run the sphardware command. Many actions require root privilege to run.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
For information on using the Hardware window and the SP Perspectives GUI, please see the online help and the "Using the SP Perspectives GUI" chapter in the IBM Parallel System Support Programs for AIX: Administration Guide.
Location
/usr/lpp/ssp/bin/sphardware
Related Information
The Hardware window can also be accessed by using the SP Perspectives Launch Pad. The perspectives command invokes the launch pad. Other Perspectives windows can be launched by invoking the following commands: spevent, sphardware, spperfmon, spsyspar, and spvsd.
Examples
sphardware
sphardware -foreground chartreuse
sphardware &
Purpose
sphostnam - Enters host name configuration data for a node or series of nodes in the System Data Repository.
Syntax
Flags
Operands
Note: | The start_frame and start_slot must resolve to a node in the current system partition. |
Note: | The node_count is considered to be within the current system partition. |
Description
Execute this command during installation of the SP to specify the adapter type to be used to determine the host name by which your nodes are known. You can also use this command to indicate whether you want the long (fully qualified) or short form of a host name to be used.
You can use the System Management Interface Tool (SMIT) to run the sphostnam command. To use SMIT, enter:
smit node_data
and select the Hostname Information option.
Notes:
Examples
The following example selects the css0 adapter for the host name for a system with two frames and 32 nodes. The long form of the host name is to be used.
sphostnam -a css0 1 1 32
Purpose
sphrdwrad - Obtains hardware Ethernet addresses for SP nodes so they can be written to the System Data Repository.
Syntax
Flags
Operands
Note: | The start_frame and start_slot must resolve to a node in the current system partition. |
Note: | The node_count is considered to be within the current system partition. |
Description
Execute this command only at installation or when adding new frames or nodes. The spframe command must be run before this command so that frame information is already in the System Data Repository.
If you know your hardware Ethernet addresses, you can speed this process by putting the addresses in /etc/bootptab.info, as follows:
17 02608C2E48D9 19 02608C2D6712 21 02608C2E49A4 23 02608C2E48E2
If you do not know your hardware Ethernet addresses, use the sphrdwrad command to find them.
Notes:
You can use the System Management Interface Tool (SMIT) to run the sphrdwrad command. To use SMIT, enter:
smit enter_data
and select the Get Hardware Ethernet Addresses option.
Note: | This command should be run only on the control workstation. You must be logged into the control workstation as root to execute this command. |
Examples
To obtain Ethernet addresses for a new frame containing 8 nodes (4 wide nodes and 4 thin nodes), enter:
sphrdwrad 2 1 8
You should receive output similar to the following:
Acquiring hardware Ethernet address for node 17 Acquiring hardware Ethernet address for node 19 Acquiring hardware Ethernet address for node 21 Acquiring hardware Ethernet address for node 23 Acquiring hardware Ethernet address for node 25 Acquiring hardware Ethernet address for node 26 Acquiring hardware Ethernet address for node 27 Acquiring hardware Ethernet address for node 28 Hardware ethernet address for node 17 is 02608C2D481C Hardware ethernet address for node 19 is 02608C2D78DF Hardware ethernet address for node 21 is 02608C2D93B3 Hardware ethernet address for node 23 is 02608C2D8C3C Hardware ethernet address for node 25 is 10005AFA22B9 Hardware ethernet address for node 26 is 10005AFA230A Hardware ethernet address for node 27 is 10005AFA2229 Hardware ethernet address for node 28 is 10005AFA2210
Purpose
splm - Views, gathers, archives, or collects log and system data.
Syntax
Flags
The splm command requires the -a flag and an argument to select a function to execute. It also requires a log table that contains records specifying target nodes and files or commands.
Operands
None.
Description
Use this command to execute a number of log management functions on a single host or a set of hosts in parallel. The splm functions are driven by a log table file that contains the target node designations and associated files, and the commands to execute. The table format follows:
# Comment line <target nodes>: <file or command>
The target node portion of a table stanza is a comma-delimited list with no blanks. The values in the list are interpreted as:
The -n flag ignores the target node portion of the table and only executes on the local node. The file or command portion of the stanza specifies either a command to execute that displays information to standard output, or a file that will be archived, collected, or viewed. File specification can take advantage of Tcl file globbing (similar to csh file globbing). If the file or command portion of the stanza refers to a command to be executed, it must be followed by a redirection and a path or file name. The information generated by the command will be redirected to the path or file name under the -d top level directory. Use > or >> following the command to redirect the output. The view option ignores the file or command destinations and displays the file's contents or command output to the executing node.
!,/tmp/nodelist,k47n10: /bin/errpt -a > errpt.output
k47n10,k47n15: /etc/filesystems etc/filesystems
This copies the file to the /etc subdirectory under the -d top level directory.
Note: | The -d top level directory is always appended with a subdirectory named arch_table_name for archives, or srvc_table_name for service collections. |
splm Functions
Archive: The archive function copies files and redirects command output as specified in the input table to the top level directory on each node. The -c flag then creates a compressed tar file of the data named /topdirectory/node_name.tar.Z. The -r flag removes an archive by removing all files starting from the top level directory down.
Check: The check function can be used to check a table for errors.
Gather: The gather function moves archive or service tar files to a central location on the executing node. The -r option removes the archive or service collection on each remote node only after the tar file was successfully copied to the central location. If the node.tar.Z file is not found, the gather function will attempt to create one. Gathered tar files can be mailed or copied to a tape or disk device using the -o flag. If mailed, the files are first uuencoded. The -l flag specifies the file system on the local node where the tar files are to be gathered. The -l flag must be specified if the -s stagger flag is not used. The gather function makes two passes, if necessary. On the first pass, it allows each node to take up an equal amount of the central file system. If any nodes fail, the gather function retries those nodes, one at a time, until the file system is full or all the nodes are copied. If gather fails on any node, but a node.tar.Z file exists for that node in the central location, it is moved to node.tar.Z.old, and not sent to the output location. The -s stagger flag forces the fanout to 1, gathers the tar files one at a time, attempts to send the tar to the output location, then removes it from the local node. The -r flag cannot be used with -s. The default central location directory for stagger is /tmp.
Service: The service function first calls the AIX snap command to gather system data to the top level directory if the -p flag is used. The snap command creates a set of subdirectories based on the -p arguments. The additional data defined in the table data is then collected under the "other" subdirectory created by snap. If the -p flag is not used, the data will still be collected under the "other" subdirectory. If the -c flag is used, splm uses the snap -c command to create the tar.Z file. The -r flag can be used to remove service collections. splm calls snap -r which removes the tar file and all files under each snap subdirectory.
View: The view function displays the output of the command or contents of file entries in the input table to the local host.
Files
Security
The archive, gather, and service functions of splm require that you have a Kerberos principal defined in the /etc/logmgt.acl file. The command then runs as root on all target nodes. The viewing function requires that you be Kerberos authenticated to a valid user ID on the nodes that you are executing on. The server switches IDs from root to your authenticated ID before executing.
Related Information
AIX commands: compress snap, tar, uuencode
The IBM Parallel System Support Programs for AIX: Administration Guide
Examples
splm -a archive -c -d /var/adm/archives -t /etc/tables/archive.tab
splm -a service -c -t /spdata/sys1/logtables/amd.tab -p g
splm -a gather -k service -t /spdata/sys1/logtables/amd.tab \ -l /tmp/amdproblem -o /dev/rmt0 -r
Purpose
splogd - Reports error logging, writes state changes, and calls user exits.
Syntax
splogd [-d] [-b] [-f file_name]
Flags
Operands
None.
Description
The SP logging daemon has the following functions:
The hwevents file contains state change actions that are to be performed by the splogd logging daemon. The fields are:
There are two special keywords for function. If function is SP_ERROR_LOG, error logging is performed provided that syslog is set up and AIX error logging is set up to perform SP logging. Refer to the setup_logd command for details.
If function is SP_STATE_LOG, these state changes that meet the statement's criteria are logged to /var/adm/SPlogs/spmon/splogd.state_changes.timestamp.
Note: |
To close the current state_changes.timestamp and open a
new one, send a SIGHUP signal to splogd. For example,
kill -HUP {splogd pid} |
User Exit Arguments
When a user exit is called by splogd, the following arguments are passed:
Starting and Stopping the splogd Daemon
The splogd daemon is under System Resource Controller (SRC) control. It uses the signal method of communication in SRC. The splogd daemon is a single subsystem and not associated with any SRC group. The subsystem name is splogd. In order to start the splogd daemon, use the startsrc -s splogd command. This starts the daemon with the default arguments and SRC options. The splogd daemon is setup to be respawnable and be the only instance of the splogd daemon running on a particular node or control workstation. Do not start the splogd daemon from the command line without using the startsrc command to start it.
To stop the splogd daemon, use the stopsrc -s splogd command. This stops the daemon and does not allow it to respawn.
To display the status of the splogd daemon, use the lssrc -s splogd command.
If the default startup arguments need to be changed, use the chssys command to change the startup arguments or the SRC options. Refer to AIX Version 4 Commands Reference and AIX Version 4 General Programming Concepts: Writing and Debugging Programs for more information about daemons under SRC control and how to modify daemon arguments when under SRC.
To view the current SRC options and daemon arguments, use the odmget -q 'subsysname=splogd' SRCsubsys command.
Files
Related Information
Command: setup_logd
The "System Monitor Variables, Display Types, and Attributes Appendix" in IBM Parallel System Support Programs for AIX: Administration Guide
Examples
startsrc -s splogd
stopsrc -s splogd
lssrc -s splogd
lssrc -a
odmget -q 'subsysname=splogd' SRCsubsys
Purpose
splst_syspars - Returns the list of defined system partitions.
Syntax
splst_syspars [-n]
Flags
Operands
None.
Description
This command returns the list of the system partitions. The system partition names are in dotted decimal format unless -n is specified.
Examples
splst_syspars
You should receive output similar to the following:
129.40.127.122 129.40.127.47
splst_syspars -n
You should receive output similar to the following:
k47sp1 k47s
Purpose
splst_versions - Returns information about the PSSP code version installed on nodes in the SP system.
Syntax
Flags
Operands
None.
Description
Use this command to return a list of PSSP code versions that are installed on the nodes in the current system partition. The PSSP version and release numbers are included in the output. The modification level and fix level are not returned as part of the output. Node number 0 (zero) is considered the control workstation and is not evaluated as part of any system partition. The output is sorted in ascending order by version.
If the -t flag is omitted, there will be only one record for each version present. If the -t flag is used, there will be a record for each node.
Examples
prompt> splst_versions PSSP-2.2 PSSP-2.4
prompt> splst_versions -t 1 PSSP-2.3 5 PSSP-2.3 6 PSSP-2.3 9 PSSP-2.4
prompt> splst_versions -l -e PSSP-2.2 /* this case has mixed partitions */ PSSP-2.4
The following will be the output if only PSSP-2.2 exists in the system partition:
prompt> splst_versions -l -e PSSP-2.2 /* this case has only 2.2 in partition */
Purpose
splstadapters - Use this command to list information about adapters to standard output.
Syntax
Flags
If the -t flag is not specified, the default is to consider adapters relating to both standard and dependent nodes for output.
Operands
Description
Use this command to get configuration information about any adapter from the SDR. For a complete list of adapter attributes, see the Adapter and DependentAdapter classes in "The System Data Repository" appendix in IBM Parallel System Support Programs for AIX: Administration Guide.
Not all of the attributes are applicable to each type of adapter.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit list_extadapters
Environment Variables
The environment variable SP_NAME is used (if set) to direct this command to a system partition. The default is to use the default system partition when on the control workstation and the partition of the node when on a node.
Standard Output
This command writes informational messages to standard output.
Standard Error
This command writes all error messages to standard error.
Exit Values
Implementation Specifics
You must specify an attribute in order for it to be displayed in the output. The attribute in the sort option (-s flag) and the attributes in the form attr==value must be repeated in order for them to be displayed.
Location
/usr/lpp/ssp/bin/splstadapters
Examples
splstadapters
You should receive output similar to the following:
node_number adapter_type 1 en0 1 css0 5 en0 5 css0
splstadapters -t standard -s node_number node_number netmask
You should receive output similar to the following:
1 255.255.255.192 3 255.255.255.192
splstadapters -G adapter_type==css0
You should receive output similar to the following:
node_number adapter_type 1 css0 5 css0 7 css0 9 css0 19 css0 23 css0
Purpose
splstdata - Displays configuration data from the System Data Repository (SDR) or system information for each node.
Syntax
OR
Flags
One of the following flags must be specified with each invocation of splstdata:
The following flags are optional:
Operands
Note: | The start_frame and start_slot must resolve to a node in the current system partition. |
Note: | The node_count is considered to be within the current system partition. |
Description
You can use the System Management Interface Tool (SMIT) to run the splstdata command. To use SMIT, enter:
smit list_data
and select the System Data Repository option for the information you want to see. To see system information for each node, enter:
smit config_dataand select the option for the information you want.
Note: | This command should be run only on the control workstation. You must be logged into the control workstation as root to execute this command. |
Examples
splstdata -n
You should receive output similar to the following:
List Node Configuration Information node frame slot # # # slots initial_hostname reliable_hostname default_route processor_type processors_installed description ----------------------------------------------------------------------- 1 1 1 2 k5n01.ppd.pok.ib k5n01.ppd.pok.ib 129.40.85.126 UP 1 135_MHz_P2SC_Wide 3 1 3 2 k5n03.ppd.pok.ib k5n03.ppd.pok.ib 129.40.85.126 UP 1 77_MHz_PWR2_Wide-2 5 1 5 1 k5n05.ppd.pok.ib k5n05.ppd.pok.ib 129.40.85.126 UP 1 160_MHz_P2SC_Thin 6 1 6 1 k5n06.ppd.pok.ib k5n06.ppd.pok.ib 129.40.85.126 UP 1 120_MHz_P2SC_Thin 7 1 7 1 k5n07.ppd.pok.ib k5n07.ppd.pok.ib 129.40.85.126 UP 1 66_MHz_PWR2_Thin-2 8 1 8 1 k5n08.ppd.pok.ib k5n08.ppd.pok.ib 129.40.85.126 UP 1 66_MHz_PWR2_Thin 9 1 9 4 k5n09.ppd.pok.ib k5n09.ppd.pok.ib 129.40.85.126 MP 1 112_MHz_SMP_High 13 1 13 4 k5n13.ppd.pok.ib k5n13.ppd.pok.ib 129.40.85.126 MP 1 200_MHz_SMP_High ·
splstdata -b 1 1 4
You should receive output similar to the following:
List Node Boot/Install Information node# hostname hdw_enet_addr srvr response install_disk last_install_image last_install_time next_install_image lppsrc_name pssp_ver ----------------------------------------------------------------------- 1 k55n01.ppd.pok.i 02608CE8E3C4 0 disk hdisk1 k55n01.mksysb.12139 Wed_Jan_15_08:41:22 k55n01.mksysb.12139 aix414 PSSP-2.2 5 k55n05.ppd.pok.i 02608CE8D7ED 0 disk hdisk0 bos.obj.k55n09.@ptf Wed_Jan_15_09:05:10 bos.obj.k55n09.@ptf aix414 PSSP-2.2 9 k55n09.ppd.pok.i 02608CE8DC55 0 disk hdisk0 bos.obj.k55n09.@ptf Mon_Jan_13_10:26:40 bos.obj.k55n09.@ptf aix414 PSSP-2.2 13 k55n13.ppd.pok.i 02608CE8E342 0 maint hdisk0 bos.obj.k55n09.@ptf Wed_Jan_15_08:41:29 bos.obj.k55n09.@ptf aix414 PSSP-2.2
splstdata -p
You should receive output similar to the following:
List System Partition Information System Partitions: ------------------ k55sp1 k55s Syspar: k55sp1 ------------------------------------------------------------------------ syspar_name k55sp1 ip_address 129.40.62.179 install_image default syspar_dir /spdata/sys1/syspar_configs/2nsb0isb/config.4_28/ layout.8/syspar.1 code_version PSSP-2.2 haem_cdb_version 852558375,501538560,0 Syspar: k55s ------------------------------------------------------------------------ syspar_name k55s ip_address 129.40.62.55 install_image default syspar_dir /spdata/sys1/syspar_configs/2nsb0isb/config.4_28/ layout.8/syspar.2 code_version PSSP-2.2 haem_cdb_version 852558451,833611264,0
Purpose
splstnodes - Lists to standard output information about nodes.
Syntax
Flags
If the -t flag is not specified, the default is to consider both standard and dependent nodes for output.
Operands
Description
Use this command to get configuration information about any node from the SDR. For a complete list of node attributes, see the Node and DependentNode classes in "The System Data Repository" appendix in IBM Parallel System Support Programs for AIX: Administration Guide.
Not all of the attributes are applicable to each type of node.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit list_extnodes
Environment Variables
The environment variable SP_NAME is used (if set) to direct this command to a system partition. The default is to use the default system partition when on the control workstation and the system partition of the node when on a node.
Standard Output
This command writes informational messages to standard output.
Standard Error
This command writes all error messages to standard error.
Exit Values
Implementation Specifics
You must specify an attribute in order for it to be displayed in the output. The attribute in the sort option (-s flag) and the attributes in the form attr==value must be repeated in order for them to be displayed.
Location
/usr/lpp/ssp/bin/splstnodes
Examples
splstnodes slots_used==2
You should receive results in the following output, if four wide nodes are in the system partition in slots 1, 3, 5, and 7:
node_number 1 3 5 7
splstnodes -t standard -s node_number node_number reliable_hostname
You should receive results in the following output:
node_number reliable_hostname 1 k22n1.ppd.pok.ibm.com 3 k22n3.ppd.pok.ibm.com 5 k22n5.ppd.pok.ibm.com 7 k22n7.ppd.pok.ibm.com
splstnodes -G slots_used==2
You should receive results in output similar to the following:
node_number 1 3 5 7 19 21 23
splstnodes -t dependent node_number snmp_community_name
If you have dependent nodes, you should receive output similar to the following:
node_number snmp_community_name 8 mycomm 2 yourcomm
Purpose
splsuser - Lists the attributes of an SP user account.
Syntax
splsuser [-c | -f] name
Flags
Operands
Description
You can only list the information for one SP user at a time. Unlike the AIX lsuser command, the ALL option and the -a flag are not supported for this command.
If you specify this command with no flags, the information of the user appears in a sequential display of attribute and values.
You can use the System Management Interface Tool (SMIT) to run the splsuser command. To use SMIT, enter:
smit spusers
and select the Change/Show Characteristics of a User option.
Examples
splsuser -c rob
You should receive output similar to the following:
#name:id:pgrp:groups:home:shell:gecos:login rob:16416:1::/u/rob on k46s.hpssl.kgn.ibm.com:/bin/ksh::true
splsuser -f rob
You should receive output similar to the following:
rob: id=16416 pgrp=1 groups= home=/u/rob on k46s.hpssl.kgn.ibm.com shell=/bin/ksh gecos= login=true
Purpose
spmgrd - Automates management and configuration required for extension nodes.
Syntax
spmgrd [-s | -l] -f filename -m [size | 0]
Flags
Operands
None.
Description
The spmgrd daemon is part of the spmgr subsystem and can only be controlled using the System Resource Controller (SRC). This daemon acts as an SNMP Manager monitoring SNMP trap messages received from SNMP agents supporting dependent nodes. A trap message may contain state information about an attached dependent node or may request the transfer of configuration data for a dependent node supported by the sending SNMP agent. When requested by a trap message, spmgrd transfers configuration data to the requesting SNMP agent. The data transfer is in the form of an SNMP set-request message containing the SNMP object instantiations representing configuration aspects of the dependent node and the values to which the aspects are to be set. When a trap message indicates that a dependent node previously fenced from the switch network with the "automatic rejoin" option is now active, spmgrd will automatically issue an Eunfence command to trigger the appropriate unfence processing.
The spmgrd daemon keeps log messages in a default file or in a file specified by the filename variable if the -f flag is specified. When the size of the log file exceeds an optional user-specified maximum log file size, the spmgrd daemon rotates the log file by moving the old log file to another file as follows:
* LogFile.3 is deleted. * LogFile.2 is moved to LogFile.3. * LogFile.1 is moved to LogFile.2. * LogFile.0 is moved to LogFile.1. * LogFile is moved to LogFile.0. * LogFile continues in LogFile.
The spmgrd daemon only runs on the control workstation.
The spmgrd daemon is controlled using the SRC. The spmgrd daemon is a member of the spmgr system group. The spmgrd daemon is enabled by default and can be manipulated by SRC commands. Use the following SRC commands to manipulate the spmgrd daemon:
Files
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP) ssp.spmgr file set.
Location
/usr/lpp/ssp/bin/spmgrd
Related Information
Commands: enadmin, lssrc, startsrc, stopsrc, tracesoff, traceson
Examples
startsrc -s spmgr -a'-s'
traceson -ls spmgr (to turn on long tracing) tracesoff -s spmgr (to stop tracing)
stopsrc -s spmgr
lssrc -ls spmgr
Purpose
spmkuser - Adds a new user account to the SP system.
Syntax
spmkuser [attribute=value ... ] name
Flags
None.
Operands
Supported Attributes and Values
Description
The -a flag is not supported. Except for home, the rules for the supported attributes and values correspond to those enforced by the AIX mkuser command.
All other attribute and value pairs are not supported.
The standard administrative AIX privileges do not apply to the SP users.
This command generates a random password for the user and stores it in /usr/lpp/ssp/config/admin/newpass.log. The root user has read and write permission to this file. It is the administrators responsibility to communicate this password to the new user and periodically delete the contents of this file.
You can use the System Management Interface Tool (SMIT) to run the spmkuser command. To use SMIT, enter:
smit spusers
and select the Add a User option.
Note: | The home directory must be in an exported file system before you can run this command. |
Examples
Note |
---|
The following examples assume that the SP automounter function is configured and the following defaults are specified:
|
To create a user account for baker using the defaults specified in the spmkuser.default file and the home directory specified in the SMIT site environment panel or spsitenv command:
spmkuser baker
To create a user account for charlie with a UID of 1234, a home directory of /u/charlie that is physically located at /home/charlie on hostx, the staff primary group and the dev, the test secondary groups, and the /bin/ksh default shell:
spmkuser id=1234 groups=dev,test home=hostx:/home/charlie
Purpose
spmon - Operates the system controls and monitors system activity.
Syntax
Flags
All spmon commands require a -target parameter except those with -diagnostics, -gui, and -help parameters.
Turns the power on or off for the node, frame, or switch specified as the target. For example:
spmon -G -p off frame1 spmon -p off node16 spmon -G -p off frame2/switch spmon -G -p off frame11/switch2
The mux setting must match the physical wiring of the switch clocks and requires a frame as its target. For a switch in node 17, use a frame as the target or frame/switchN for a switch in a switch-only frame.
For each switch, checks:
The tests are in dependent order. If any of these fail, the subsequent tests do not run.
The -target flag is optional. Any parameter without a flag is assumed to be the target. You can also have multiple target-flags (-t), which are optional.
Operands
None.
Description
Any unique abbreviation of flags and keywords is acceptable.
Specify target_value with the hierarchical format (or tree structure). The format is:
/SP/frame/frameN/[nodeM|switchM]/variableX/value
You can use wildcards (*) to specify more than one target node or frame for the query command.
Note: | Though they are not hardware variables, for compatibility with older systems, the variables hostResponds and switchResponds can be used as specific targets of the spmon command for both -query and -Monitor commands. However, the variable names must be entered explicitly. These two variables are not returned if the variable specified is a wildcard (*). |
You can use aliases in place of fully qualified hierarchical target values. Aliases require less typing and may be more intuitive than the fully qualified targets. Leaving the leading slash (/) off the target indicates that it is an alias.
There are two formats for aliases:
You can include a variable and attribute after the alias.
Examples
spmon -q -t /SP/frame/frame1/node1/keyModeSwitch/value 0
spmon node1/keyModeSwitch/value 0
spmon -L frame1/node1
You should receive output similar to the following:
_____________ | _ _ _ | | |_| |_| |_| | Frame 1, Node 1 | |_| |_| |_| | |_____________|
spmon -G -q -l frame*/switch*/mux/value /SP/frame/frame1/switch17/mux/value/0
spmon -M frame1/node*/powerLED/value 2 1
spmon -M -q -l frame1/node*/powerLED/value /SP/frame/frame1/node1/powerLED/value/1 /SP/frame/frame1/node3/powerLED/value/1 /SP/frame/frame1/node5/powerLED/value/1 /SP/frame/frame1/node7/powerLED/value/1 /SP/frame/frame1/node9/powerLED/value/1 /SP/frame/frame1/node10/powerLED/value/1 /SP/frame/frame1/node11/powerLED/value/1 /SP/frame/frame1/node12/powerLED/value/1 /SP/frame/frame1/node13/powerLED/value/1 /SP/frame/frame1/node14/powerLED/value/1 /SP/frame/frame1/node15/powerLED/value/1 /SP/frame/frame1/node16/powerLED/value/1 /SP/frame/frame1/node1/powerLED/value/2 /SP/frame/frame1/node1/powerLED/value/1
Note: | "node*" returns powerLED values on switches in slots 1--16 also. |
spmon -p off frame2/node3
If node3 on frame2 is outside the current system partition, enter:
spmon -G -p off frame2/node3
spmon -p off node19
spmon -k service node1
spmon -G -p off frame1
spmon -G -p off frame1/A
spmon -G -m i frame1
or
spmon -G -m i frame1/switch
spmon -G -m i frame10/switch4
Purpose
spmon_ctest - Verifies that the System Monitor component is configured correctly.
Syntax
spmon_ctest [-l log_file] [-q]
Flags
Operands
None.
Description
This command is designed to be run after installing the SP system to verify that the System Monitor is configured correctly. The test checks to make sure that the hardware is running, that it can be queried, and determines whether any node objects were created in the System Data Repository (SDR). The test also indicates whether the RS232 lines are connected properly.
A return code of zero indicates that the test completed as expected; otherwise it returns the number of failures. If you do not specify the -q flag, a message is displayed on standard output that indicates the success or failure of the tests. In either case, the command returns 0 if successful, 1 if not. If errors are detected, more detailed information is recorded in the log file. If you do not specify the -l flag, error messages are recorded in /var/adm/SPlogs/spmon_ctest.log.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit SP_verify
and select the System Monitor Configuration option.
You must run this test from a user who has monitor authority in /spdata/sys1/spmon/hmacls. The user must also have a nonexpired authentication ticket.
Refer to Chapter 2, "RS/6000 SP Files and Other Technical Information" section of IBM Parallel System Support Programs for AIX: Command and Technical Reference for additional Kerberos information.
Files
Related Information
Commands: CSS_test, jm_install_verify, jm_verify, SDR_test, SYSMAN_test, spmon_itest
Examples
To verify installation of the SP System Monitor, saving error messages in spmon.err in the current working directory, enter:
spmon_ctest -l spmon.err
Purpose
spmon_itest - Verifies that the System Monitor is installed and operational.
Syntax
spmon_itest [-l log_file] [-q]
Flags
Operands
None.
Description
This command is designed to be run after installing the SP system to verify that the System Monitor is installed correctly.
A return code of zero indicates that the test completed as expected; otherwise it returns the number of failures. If you do not specify the -q flag, a message is displayed on standard output that indicates the success or failure of the tests. In either case, the command returns 0 if successful, 1 if not. If errors are detected, more detailed information is recorded in the log file. If you do not specify the -l flag, error messages are recorded in /var/adm/SPlogs/spmon_itest.log.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit SP_verify
and select the System Monitor Installation option.
Files
Related Information
Commands: CSS_test, jm_install_verify, jm_verify, SDR_test, SYSMAN_test, spmon_ctest
Examples
To verify installation of the SP System Monitor, saving error messages in spmon.err in the current working directory, enter:
spmon_itest -l spmon.err
Purpose
spperfmon - Directly launches the Performance Monitor window of the SP Perspectives graphical user interface (GUI).
Syntax
Flags
Note: | Most flags accepted by X will also be recognized. For example, -display displayname. |
Operands
None.
Description
Use this command to launch the SP Performance Monitor window of the SP Perspectives GUI. This tool enables the user to monitor the performance of the SP in conjunction with other licensed products: Performance Toolbox for AIX (PTX), 5765-654 and Performance Toolbox Parallel Extensions for AIX (PTPE), 5765-529.
From the Performance Monitor window, you can perform most of the PTPE command set functions through point and click operations. For example, you can easily manipulate the PTPE monitoring hierarchy and save it to the System Data Repository (SDR).
The Performance Monitor perspective window uses three panes to display SP system information:
When the command is invoked, preferences that define the look and layout of the SP Performance window are prioritized in the following order:
Files
Restrictions
Any user can run the spperfmon command. To get a read/write PTPE session requires root privilege and the user must be a member of the UNIX group 'perfmon'.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP) and the IBM Performance Toolbox Parallel Extensions for AIX separately priced feature.
Prerequisite Information
For information on using spperfmon and the SP Perspectives GUI, please see the online help and the "Using the SP Perspectives GUI" chapter in the IBM Parallel System Support Programs for AIX: Administration Guide.
Location
/usr/lpp/ssp/bin/spperfmon
Related Information
You can also access the Performance Monitor window by using the SP Perspectives Launch Pad. The perspectives command invokes the launch pad. Other Perspectives windows can be launched by invoking the following commands: spevent, sphardware, spsyspar, and spvsd.
IBM Performance Toolbox Parallel Extensions for AIX: Guide and Reference
IBM Performance Toolbox 1.2 and 2.1 for AIX: Guide and Reference
Examples
spperfmon
spperfmon -noProfile
Purpose
sprestore_config - Restores the system to a given system partitioning configuration as specified in the System Data Repository (SDR) which was previously archived.
Syntax
sprestore_config archive_file [-h]
Flags
Operands
Description
Use this command to restore the SDR from an archive file that was previously created with the SDRArchive command. In addition to restoring the SDR (using the SDRRestore command), the sprestore_config command also restores system partition-sensitive subsystems (for example, hats, hb, and hr) to their previous state. This command is most useful when recovering from an attempt to partition the SP (see the spapply_config command).
You can use the System Management Interface Tool (SMIT) to run the sprestore_config command. To use SMIT, enter:
smit syspar_restoreand enter (or select from a generated list) the name of the SDR archive from which to restore.
Notes:
Exit Values
Related Information
Commands: SDRArchive, SDRRestore, spapply_config, spcustomize_syspar, spdisplay_syspar, spverify_config, syspar_ctrl
Files: nodelist, topology
Examples
To restore the SDR and the system-partition sensitive subsystems (for example, hats, hb, and hr) from the archive 'backup.95110.1620' which was previously created using the SDRArchive command, enter:
sprestore_config backup.95110.1620
Purpose
sprmuser - Removes a user account from the SP system.
Syntax
sprmuser [-i] [-p] [-r] name
Flags
Operands
None.
Description
The -i and -r options are unique to the SP system.
You can use the System Management Interface Tool (SMIT) to run the sprmuser command. To use SMIT, enter:
smit spusers
and select the Remove a User option.
Examples
sprmuser charlie
sprmuser -pr charlie
Purpose
spsitenv - Enters configuration parameters used by SP installation and system management scripts into the System Data Repository (SDR).
Syntax
Flags
Specify true if the code is to be installed. Specify false if the code is not to be installed. The initial value is true.
The initial value is the host name of the control workstation.
To use an Internet NTP time server, your control workstation must be connected to the Internet. Specify ntp_config=internet and specify the full host name of an Internet time server with the ntp_server parameter.
To cause the control workstation and file servers to generate a consensus time based on their own date settings, specify ntp_config=consensus and specify ntp_server=''.
If you do not want to run NTP on the SP, specify ntp_config=none and ntp_server=''.
The initial value of ntp_config is consensus and the initial value of ntp_server is ''. If ntp_config is specified as either timemaster or internet, the ntp_server value must be a valid host name.
This field is meaningful only if usermgmt_config=true.
Specify print_config = false if you do not want to have the existing AIX print commands saved and have the file names linked to the SP print functions.
If you want to have the print management system integrated, specify secure if you do not want users to not have rsh privileges on the print host. Specify open if the users are to have rsh privileges on the print host.
The initial value of print_config is false.
The initial value is ''. If the value remains at '' and the value of print_config is secure, the installation script uses a print management ID of prtid as a default.
The SP Print Management System was removed from PSSP 2.3. That is, the SP Print Management System cannot be configured on nodes running PSSP 2.3 or later. We suggest the use of Printing Systems Manager (PSM) for AIX as a more general solution to managing printing on the SP system.
However, if you are running earlier versions of PSSP on some of your nodes, the SP Print Management System is still supported on those nodes. The print_config routine running on the control workstation will configure the SP Print Management System on nodes running versions of PSSP earlier than PSSP 2.3.
If you are running mixed levels of PSSP in a system partition, be sure to maintain and refer to the appropriate documentation for whatever versions of PSSP you are running.
Specify remove_image=true if the images are to be removed.
Specify remove_image=false if the images are not to be removed.
The initial value is false.
The initial value is false.
Specify usermgmt=true if you want to have the SP User Management scripts in the Security & Users SMIT menu. Specify usermgmt=false to remove the scripts from the SMIT menu.
The initial value is true.
Operands
None.
Description
Use this command during installation of the SP or at a later time to identify SP configuration parameters in use at your location.
You must have a ticket-granting-ticket to run this command. Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for additional information on ticket-granting-tickets.
If you do not have a ticket-granting-ticket, you must run kinit.
You can use the System Management Interface Tool (SMIT) to run the spsitenv command. To use SMIT, enter:
smit enter_data
and select the Site Environment Information option.
You cannot use SMIT if you are using AFS authentication services.
Notes:
Examples
The following example enters site environment parameters into the System Data Repository. The NTP configuration is consensus and the file collection management code is to be installed:
spsitenv ntp_config=consensus filecoll_config=true
Purpose
spsvrmgr - SP Supervisor Manager.
Syntax
Flags
Action checks include:
If rc is specified with the -q flag, the command will issue a return code indicating whether any of the hardware requires action. A return code of 0 indicates that no action is required. A return code of 1 indicates that at least one supervisor was found that required action.
If msg is specified with the -q flag, the command will issue a message indicating whether any of the hardware requires action. In this case, a return code of 0 is issued unless an error condition occurs.
If status is specified with the -r flag, the status is listed for all of the installed supervisors that support microcode download.
If action is specified with the -r flag, the status is listed for all of the installed supervisors that support microcode download and that also require an action.
In both cases, Status includes:
If status is specified with the -m flag, the status is listed for all of the installed supervisors that support microcode download.
If action is specified with the -m flag, the status is listed for all of the installed supervisors that support microcode download and that also require an action.
In both cases, Status includes:
Note: | This flag starts an hmcmds process to perform the actual update. Refer to the hmcmds command specifically the basecode, microcode, and the boot_supervisor command options. |
Attention |
---|
In most cases, the -u flag started processes powers off the target slots during the duration of the update. |
Operands
Description
The design of the SP supervisor control system divides the microcode used in the frame supervisor, node supervisor, and switch supervisor into the following two types:
The spsvrmgr command controls the software level and state of the supervisor applications that reside on the SP supervisor hardware.
Normally, commands are only sent to the hardware components in the current system partition. A system partition contains only processing nodes. The switches and the frames themselves are not contained in any system partition. To access hardware components not in the current system partition or to any frame or switch, use the -G flag.
The slot_spec option is interpreted as slot ID specifications. A slot ID specification names one or more slots in one or more SP frames and has either of two forms:
fidlist:sidlist or nodlist
where:
The first form specifies frame numbers and slot numbers. The second form specifies node numbers. An fval is a frame number or a range of frame numbers of the form a-b. An sval is a slot number from the set 0 through 17 or a range of slot numbers of the form a-b. An nval is a node number or a range of node numbers of the form a-b.
The relationship of node numbers to frame and slot numbers is shown in the following formula:
node_number = ((frame_number - 1) × 16) + slot_number
Note: | Node numbers can only be used to specify slots 1 through 16 of any frame. |
Refer to the hmcmds command for examples of the slot_spec.
Optionally, slot ID specifications can be provided in a file rather than as command flags. The file must contain one specification per line. The command requires that slot ID specifications be provided. If the command is to be sent to all SP hardware, the keyword all must be provided in lieu of the slot_spec option. However, the all keyword can only be specified if the -G flag is specified.
Files
The media that is the repository for the application microcode files is the /spdata/sys1/ucode directory structure.
Exit Values
Restrictions
IBM suggests that you use this command through the RS/6000 SP Supervisor Manager option of the System Management Interface Tool (SMIT).
To access this command using SMIT, enter:
smitand select the RS/6000 SP System Management option, then the RS/6000 SP Supervisor Manager option.
A list of options that correspond to the spsvrmgr command flags will be presented for selection.
You can also directly access this list of options using the following SMIT fast-path command:
smit supervisor
Implementation Specifics
You must be authorized to access the Hardware Monitor subsystem to run the spsvrmgr command. In addition, for those frames specified to the command, you must have Virtual Front Operator Panel (VFOP) permission. Commands sent to frames for which you do not have VFOP permission are ignored. Since the Hardware Monitor subsystem uses SP authentication services, you must run the kinit command prior to running this command. Alternatively, site-specific procedures can be used to obtain the tokens that are otherwise obtained by kinit.
The spsvrmgr command, by design, only interacts with SP supervisor hardware that supports the ability to download application microcode. Commands sent to slots that do not support this ability are ignored.
Location
/usr/lpp/ssp/bin/spsvrmgr
Related Information
Commands: hmcmds
Refer to the "Installing and Configuring a New RS/6000 System" chapter in IBM Parallel System Support Programs for AIX: Installation and Migration Guide.
Examples
spsvrmgr -G -q msg all
You should receive output similar to the following:
spsvrmgr: At least one occurrence of supervisor hardware was found to require attention. Enter "smit supervisor" for installation options.
spsvrmgr -G -q rc all echo $?
Example usage in a script:
spsvrmgr -G -q rc all if [[ $? = 1 ]] then echo "*** Attention*** One or more supervisors require action." echo "Enter \"smit supervisor\" for installation options." fi
spsvrmgr -G -r status 2:0-17
You should receive report output similar to the following:
spsvrmgr: Frame Slot Supervisor Media Installed Required State Versions Version Action _____ ____ __________ ____________ ____________ ____________ 2 1 Active u_10.3a.0609 u_10.3a.060b None u_10.3a.060a u_10.3a.060b ____ __________ ____________ ____________ ____________ 5 Active u_10.3a.0609 u_10.3a.060b None u_10.3a.060a u_10.3a.060b ____ __________ ____________ ____________ ____________ 9 Active u_10.1a.0609 u_10.1a.060b None u_10.1a.060a u_10.1a.060b ____ __________ ____________ ____________ ____________ 13 Active u_10.3a.0609 u_10.3a.060b None u_10.3a.060a u_10.3a.060b
spsvrmgr -G -r status all
You should receive matrix output similar to the following:
spsvrmgr: Frame Slots _____ _______________________________________________ 1 00 01 05 09 13 17 (Action) - - - - - - _____ _______________________________________________ 2 01 05 09 13 (Action) + + + - Action Codes: + -- Required - -- Not Required
spsvrmgr -G -r action 1:0-17
You should receive report output similar to the following:
spsvrmgr: Frame Slot Supervisor Media Installed Required State Versions Version Action _____ ____ __________ ____________ ____________ ____________ 1 1 Active u_10.3a.0609 u_10.3a.060a Upgrade u_10.3a.060a u_10.3a.060b ____ __________ ____________ ____________ ____________ 5 Inactive u_10.3a.0609 u_10.3a.060b Reboot u_10.3a.060a u_10.3a.060b ____ __________ ____________ ____________ ____________ 9 Inactive u_10.1a.0609 u_10.1a.060b Reboot u_10.1a.060a u_10.1a.060b
spsvrmgr -u 1:1
You should receive installation output similar to the following:
spsvrmgr: Dispatched "microcode" process [24831] for frame 1 slot 1. Process will take approximately 12 minutes to complete. spsvrmgr: Process [24831] for frame 1 slot 1 completed successfully.
spsvrmgr -u 1:5,9
You should receive installation output similar to the following:
spsvrmgr: Dispatched "boot_supervisor" process [27956] for frame 1 slot 5. Process will take less than a minute to complete. spsvrmgr: Dispatched "boot_supervisor" process [23606] for frame 1 slot 9. Process will take less than a minute to complete. spsvrmgr: Process [27956] for frame 1 slot 5 completed successfully. spsvrmgr: Process [23606] for frame 1 slot 9 completed successfully.
Purpose
spsyspar - Directly invokes the System Partitioning Aid window of the SP Perspectives graphical user interface (GUI).
Syntax
Flags
Note: | Most flags accepted by X will also be recognized. For example, -display displayname. |
Operands
None.
Description
Use this command to launch the System Partitioning Aid window of the SP Perspectives GUI. The System Partitioning Aid Perspective is used to view and manage the current system partitioning configuration. This tool can also be used to generate new configurations.
When the command is invoked, preferences which define the look and layout of the System Partitioning Aid window are prioritized in the following order:
Files
The users preferences are read from and saved to $HOME/.spsyspar(User Profile Name). The System Preferences are read from and saved to /usr/lpp/ssp/perspectives/profiles/.spsyspar(System Profile name). If a new system partitioning configuration is created, the following files are created under the layout directory: layout.desc, nodes.syspar and a system partition directory for each system partition in the layout. For each system partition directory, a node list file and topology file are created.
Restrictions
Any user can run the spsyspar command. Many actions require root privilege to run.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
For information on using the System Partitioning Aid window and the SP Perspectives GUI, please see the online help and the "Using the SP Perspectives GUI" chapter in the IBM Parallel System Support Programs for AIX: Administration Guide.
Refer to the "Managing System Partitions" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional information on the System Partitioning Aid.
Location
/usr/lpp/ssp/bin/spsyspar
Related Information
You can also access the hardware window by using the SP Perspectives Launch Pad. The perspectives command invokes the launch pad. Other Perspectives windows may be launched by invoking the following commands: spevent, sphardware, spperfmon, and spvsd. The sysparaid command provides a command line interface into the System Partitioning Aid.
Examples
spsyspar
spsyspar -background hot_pink
Purpose
spverify_config - Verifies the active system partition configuration information for the SP system.
Syntax
spverify_config
Flags
None.
Operands
None.
Description
This command is run by the spapply_config command after the System Data Repository (SDR) is updated. It can also be run by an administrator to verify that the SDR information is consistent (such as, after a system outage or a problem with the SDR). (This verification is only performed on a system which was partitioned beyond the initial single partition created at initial installation.)
Exit Values
Related Information
Commands: spapply_config, spcustomize_syspar, spdisplay_config
Files: nodelist, topology
Examples
To verify that the information in the SDR matches the customization information previously supplied by the user, enter:
spverify_config
Purpose
spvsd - Directly launches the IBM Virtual Shared Disk window of the SP Perspectives graphical user interface (GUI).
Syntax
Flags
Note: | Most flags accepted by X will also be recognized. For example, -display displayname. |
Operands
None.
Description
Use this command to launch the IBM Virtual Shared Disk window of the SP Perspectives GUI. This perspective allows the user to view and control the IBM Virtual Shared Disk system.
By default, when the window is brought up, it displays:
The current system partition will be selected within the system partitions. The IBM VSD Nodes pane displays all nodes in the current system partition. The IBM VSDs and HSDs Definitions pane displays all existing IBM VSD/HSD definitions within the system partition. You can control which panes are displayed by using the Add pane and Delete pane combo boxes.
When the command is invoked, preferences that define the look and layout of the spvsd window are prioritized in the following order:
Files
The Users Preferences are read from and saved to $HOME/.spvsd(User Profile Name). The System Preferences are read from and saved to /usr/lpp/ssp/perspectives/profiles/.spvsd(System Profile name).
Restrictions
Any user can run the spvsd command. Many actions require root privilege to run.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Prerequisite Information
For information on using the IBM Virtual Shared Disk window and the SP Perspectives GUI, please see the online help and the "Using the SP Perspectives GUI" chapter in the IBM Parallel System Support Programs for AIX: Administration Guide. For information about the IBM Virtual Shared Disk, see IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Location
/usr/lpp/ssp/bin/spvsd
Related Information
You can access the IBM Virtual Shared Disk window by using the SP Perspectives Launch Pad. The perspectives command invokes the launch pad. Other Perspectives windows may be launched by invoking the following commands: spevent, sphardware, spperfmon, and spsyspar.
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Examples
spvsd
spvsd -fontBold
Purpose
startvsd - Makes an IBM Virtual Shared Disk available and activates it.
Syntax
startvsd [-p | -b] -a | vsd_name ...
Flags
This option is only used by IBM Recoverable Virtual Shared Disk product.
See IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Note: | This flag is used only by the IBM Recoverable Virtual Shared Disk product. |
Operands
Description
The startvsd command makes the specified IBM Virtual Shared Disks available and activates them. It is equivalent to running the preparevsd command followed by the resumevsd command on the specified IBM Virtual Shared Disk.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit vsd_mgmtand select the Start an IBM Virtual Shared Disk option.
Files
Security
You must have root privilege to run this command.
Restrictions
See IBM Parallel System Support Programs for AIX: Managing Shared Disks
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfgvsd, ctlvsd, lsvsd, preparevsd, resumevsd, stopvsd, suspendvsd, ucfgvsd
Examples
To make available and activate the IBM Virtual Shared Disk vsd1vg1n1, enter:
startvsd vsd1vg1n1
Purpose
statvsd - Displays IBM Virtual Shared Disk device driver statistics of a node.
Syntax
statvsd
Flags
None.
Operands
None.
Description
The statvsd command displays the level of IBM Virtual Shared Disk parallelism, several IBM Virtual Shared Disk IP device driver counters, nonzero sequence numbers and the nodes they are for, the node out cast status, and the max_IP_msg_size. If either the expected or outgoing sequence number for a node is nonzero, both are displayed along with the node number. If both the expected and outgoing sequence numbers are zero, no sequence numbers are displayed for that node.
The level of IBM Virtual Shared Disk parallelism defaults to 9 and is set via the ctlvsd -p command.
Sequence numbers are initially all zero at the first cfgvsd. They are incremented as requests are sent (outgoing) and received (expected), and reset via ctlvsd -R | -r.
The counters all start at zero at the first cfgvsd, then are incremented as the events they count occur. Use suspendvsd and stopvsd to ensure that there is no IBM Virtual Shared Disk activity; then use ctlvsd to reset the counters.
The requests queued waiting for a request block, pbuf, cache block, and buddy buffer counters indicate shortages of these resources. All four are tunable values on the vsdnode command. If a significant increase in these counters occurs during the running of a critical application, see IBM Parallel System Support Programs for AIX: Managing Shared Disks for information about tuning IBM Virtual Shared Disk performance and how to respond to resource shortages.
When a user buffer address is not on a page boundary, two IBM Virtual Shared Disks can share a page in I/O requests. Typically, when a local IBM Virtual Shared Disk server is copying data to the user buffer, the DMA hides the page. If the client receives the data from a remote IBM Virtual Shared Disk server, network protocol interrupts the local I/O; however, the page is still hidden by the DMA. Therefore, the IBM Virtual Shared Disk places the remote request on a Rework_Q, swaps control to the local I/O, and later performs rework by copying data from the network protocol mbuf to the user buffer.
A request, or its corresponding response, may be lost due to transmission failure or failure to allocate an mbuf. The current IBM Virtual Shared Disk communication protocol implements an exponential back-off retransmission strategy. A request is retransmitted to the server a fixed number of times. The IBM Virtual Shared Disk IP device driver waits about 2 seconds for a response after initially sending the request before retransmission. Thereafter, it waits about twice as long as the last time as it cycles through the fixed number of retries. If a response is not received after the timeout expires on the last retransmission attempt, the request fails with the ETIMEOUT errno value. Currently, the sum total of retransmission time is about 15 minutes. If a request is not responded to after about 10 transmissions of the request over a 15 minute period, the request is failed. statvsd displays the number of requests failed due to timeouts in the timeouts counter. The retries counters display the number of requests that have been retransmitted for each retry bucket value. The number of numbers on the retries line of statvsd output indicates the fixed number of times a request is retransmitted. The total retries is the sum of the retries bucket values, which is the total number of request retransmissions.
There is no tuning that can be performed to affect these values; the values are provided for information only. If a request to an IBM Virtual Shared Disk is being retransmitted, and the IBM Virtual Shared Disk is suspended and subsequently resumed, the request starts over with a fresh 15 minute retry period.
If a server gets heavily loaded, it may not be able to respond to a request fast enough to prevent the client from retransmitting the request. If the server responds after the client has retransmitted the request, the client rejects the response since its sequence number no longer matches the current sequence number of the request. The client records this event in the rejected response counter that statvsd displays.
If a server receives a request with an unexpected sequence number, a rejected request event is recorded in a counter that statvsd displays.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands:
Refer to IBM Parallel System Support Programs for AIX: Managing Shared Disks for information on tuning IBM Virtual Shared Disk performance and sequence numbers.
Examples
The following example displays IBM Virtual Shared Disk device driver statistics:
VSD driver: IP interface 9 vsd parallelism 24576 vsd max IP message size 0 requests queued waiting for a request block 0 requests queued waiting for a pbuf 0 requests queued waiting for a cache block 0 requests queued waiting for a buddy buffer 0.0 average buddy buffer wait_queue size 0 rejected requests 0 rejected responses 0 rejected no buddy buffer 0 requests rework 0 indirect I/O 0 64-byte unaligned reads 9 comm. buf pool shortage 0 timeouts retries: 0 0 0 0 0 0 0 0 0 0 total retries Non-zero Sequence numbers node# expected outgoing outcast? Incarnation: 0 3 Nodes Up with zero sequence numbers: 256 1000 2000
Purpose
stopvsd - Makes an IBM Virtual Shared Disk unavailable.
Syntax
stopvsd -a | vsd_name ...
Flags
Operands
Description
The stopvsd command brings the specified IBM Virtual Shared Disks from the suspended state to the stopped state. They become unavailable. All applications that have outstanding requests for the IBM Virtual Shared Disk see these requests terminate with error. Read and write requests fail with errno set to ENODEV. If the IBM Virtual Shared Disk is in the stopped state, this command leaves it in the stopped state.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit vsd_mgmtand select the Stop an IBM Virtual Shared Disk option.
Files
Security
You must have root privilege to run this command.
Restrictions
If you have the IBM Recoverable Virtual Shared Disk product installed and operational, do not use this command. The results may be unpredictable.
See IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfgvsd, ctlvsd, lsvsd, preparevsd, resumevsd, startvsd, suspendvsd, ucfgvsd
Examples
To bring the IBM Virtual Shared Disk vsd1vg1n1 from the suspended state to the stopped state, enter:
stopvsd vsd1vg1n1
Purpose
supfilesrv - Is the daemon that serves the file collections on the SP system.
Syntax
startsrc -s supfilesrv
stopsrc -s supfilesrv
Flags
There are no flags or options for this daemon. All arguments and options are defined to the System Resource Controller (SRC).
Description
This daemon executes on the control workstation and the boot and install servers. It responds to requests executed on the nodes to update collections of files configured for File Collections. The supfilesrv daemon is under control of the System Resource Controller (SRC). It is normally started as part of the SP initialization scripts. Under normal circumstances, the administrator does have to start or stop the daemon.
Exit Values
Prerequisite Information
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for information on file collections.
Location
/var/sysman/etc
Related Information
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for information on file collections, the System Resource Controller, and user management.
Examples
startsrc -s supfilesrv
stopsrc -s supfilesrv
Purpose
Manages the SP file collections.
Syntax
supper [-d] [-o options] [-v] subcommands
Flags
Operands
Subcommands
Description
You can invoke supper as an interactive session by entering the supper command without any parameters or subcommands. This allows you to enter the subcommands in an interactive dialog.
Examples
/var/sysman/supper rlog
/var/sysman/supper status
/var/sysman/supper scan user.admin
/var/sysman/supper update power_system
Purpose
suspendvsd - Deactivates an available IBM Virtual Shared Disk.
Syntax
suspendvsd -a | vsd_name...
Flags
Operands
Description
The suspendvsd command brings the specified IBM Virtual Shared Disks from the active state to the suspended state. They remain available. Read and write requests which were active while the IBM Virtual Shared Disk was active are suspended and held. Subsequent read and write operations are also held. If the IBM Virtual Shared Disk is in the suspended state, this command leaves it in the suspended state.
If you issue this command for the server node and not the client, retries occur with eventual failure. Failure occurs within 15 minutes.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit vsd_mgmtand select the Suspend an IBM Virtual Shared Disk option.
Files
Security
You must have root privilege to run this command.
Restrictions
If you have the IBM Recoverable Virtual Shared Disk product installed and operational, do not use this command. The results may be unpredictable.
See IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfgvsd, ctlvsd, lsvsd, preparevsd, resumevsd, startvsd, stopvsd, ucfgvsd
Examples
To bring the IBM Virtual Shared Disk vsd1vg1n1 from the active state to the suspended state, enter:
suspendvsd vsd1vg1n1
Purpose
sysctl - Is the command interface to Sysctl remote command execution and monitoring server.
Syntax
Flags
Using a socket also enables two-way communication between the client and a command procedure. Any commands that start a dialog requiring input from the client must be run with the -s option.
Operands
If you do not specify a command, sysctl runs in interactive mode.
Description
The sysctl program provides a simple command line interface for communicating with the Sysctl server, sysctld. Together, the Sysctl client and server provide monitoring and execution abilities needed to remotely manage workstations on all nodes on the SP system. Sysctl connects to a remote host's sysctld using TCP/IP, passes keywords and commands to the server, and writes any output returned to standard output. sysctl does not interpret the commands it passes.
The Sysctl server uses the Tcl embeddable command language as the foundation for its built-in interpreter. The server is augmented with application-specific commands that may vary between servers. The Sysctl (Tcl) expressions that are passed to the server for execution can be anything from a single command to an entire script. The server uses SP authentication services for reliable third party authentication. The Sysctl request sent from the client optionally contains an authentication ticket that uniquely identifies the user that initiated the request. The server's built-in authorization mechanism controls the set of commands available to a user.
Files
Related Information
Commands: dshbak, hostlist, sysctld
The chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide.
Examples
These examples show commands issued from the shell prompt. Unless noted otherwise, the syntax is the same (minus the word sysctl) when issuing the command from within an interactive session of sysctl:
sysctl listfs -h ceti-alpha5
sysctl -h ceti-alpha5 info commands
sysctl -h ceti-alpha5 acladd -p arielle.admin
If you do not specify a realm, the local realm is assumed. If ceti-alpha5 is in realm TESTCELL.HAL.COM, sysctl returns the following message:
-principal arielle.admin@TESTCELL.HAL.COM
sysctl -h ceti-alpha5 acladd -f /test/data/mount.acl \ -p arielle.admin
sysctl -h ceti-alpha5 aclcheck -f /test/data/mount.acl arielle
The server returns 1 if arielle is in the ACL, 0 if it is not.
sysctl -h ceti-alpha5 checkauth -cmd test:mount_if
The server returns 1 if you are authorized, 0 if not.
Purpose
sysctld - Contains the remote execution and monitoring server.
Syntax
sysctld [-sd] [-a acl] [-f file] [-k keyfile] [-l log_file] [-P port]
Flags
Operands
None.
Description
The sysctld daemon is the server component for the Sysctl remote command execution facility. Security and performance characteristics of sysctld make it an ideal mechanism for managing a large, distributed computing environment. Typically, one instance of sysctld runs on every workstation. Commands are sent to a sysctld server via the sysctl client program. When a command request is received, sysctld parses and executes the request using an embedded Tcl command interpreter. Authorization to execute commands is determined by authorization callbacks that are attached to each command. These callbacks are pieces of Tcl code that implement the security policy for the sysctld server.
Security and Access Control
Sysctl uses the SP authentication service for reliable third-party authentication. Command requests sent by the client include an authentication ticket giving the identity of the client. As the server executes the commands sent by the client, the client's access to commands and variables is determined dynamically via authorization callbacks. In a typical command language, a procedure has a name, a set of arguments, a set of commands which form the body of the procedure, and a return value. With Sysctl procedures, an additional attribute is added, a policy for determining who is able to run the command. These policies are implemented using authorization callbacks. An authorization callback is a piece of Tcl code that is attached to a command and determines the access policy for that command. The authorization callback for a command is invoked whenever a client attempts to execute the command. If the callback returns a Tcl error, the command is not executed and the following message is returned to the caller:
Authorization Denied
If the callback returns a normal Tcl result, the command executes normally.
Sysctl variables also have an authorization callback attached to them which determines the read-access policy for the variable. Therefore, it is possible to create a "private" variable in which you restrict the set of clients who have access to its value.
The sysctld server defines a set of commands which are designed to be used as authorization callbacks. These commands provide a simple authorization policy; if more complex authorizations are required, you have the ability to code your own authorization callback procedures.
Bypassing Authorization Callbacks
Under certain circumstances, the authorization callbacks are bypassed by the server. While the checks are bypassed, the user has access to all commands (and read/write access to variables) defined in the server. The authorization callbacks are bypassed while the server is:
Since authorization checks are bypassed during procedure execution, users are able to execute procedures that contain commands they are not authorized to run directly. For example, if a variable has the SYSTEM callback attached to it, a procedure for which a user is authorized can reference or modify the variable even though that user cannot modify or reference it.
Authorization Variables
Several variables, accessible as read-only variables in the interpreters or as environment variables, are set by the server prior to executing the client request. These variables provide a mechanism for external commands and procedures to perform their own authorization checking independent of the server's standard authorization checks. They are:
Determining Connection Authorization Policy
Whenever a client connects to the server, the command svcconnect is invoked. If the user is not authorized to run this command or if the command returns a Tcl error, the result of the command is returned to the user and the connection is broken. Therefore, the svcconnect command determines the connection policy for the server. By default, svcconnect simply returns a normal result and its authorization callback is AUTH. This implies that any authenticated user can connect. This policy can be altered by changing the authorization for the svcconnect callback (via the setauth command), or by redefining the svcconnect procedure (via the create proc command).
Server Configuration
At start up, the server reads a configuration file. By default, this file is named /etc/sysctl.conf. The -f flag is used to specify an alternate configuration file. The server interprets the contents of the configuration file as Tcl commands. Typically, additional commands and variables are defined in this file. Also, commands are available that instruct the server to read additional configuration files or dynamically load in shared libraries. In this way, the set of commands available to a sysctld server is extendable. Refer to the sysctl.conf file for more details.
Signals
Sending a SIGHUP signal to the server causes it to close the log file, delete and re-create all interpreters, reread the configuration files, and reinitialize the log file. The svcrestart command performs the same function.
Logging
The sysctld default log file is /var/adm/SPlogs/sysctl/sysctld.log. This default can be overridden with the -l command line option, or by by setting the LOG variable in the Sysctl configuration file. Each time a request is received, the svclogevent Sysctl command is invoked. By default, it writes a record to the log file giving the identity of the user who sent the request or "unknown" if the user is not authenticated. A different logging policy can be achieved by redefining the svclogevent procedure in the server's configuration file.
While running in security audit mode, each line written to the log file is tagged with a Connection ID field which is used to filter the audit trail for a particular connection in cases where multiple connections are processed simultaneously.
Starting and Stopping the sysctld Daemon
The sysctld daemon is under System Resource Controller (SRC) control. It uses the signal method of communication in SRC. The sysctld daemon is a single subsystem and not associated with any SRC group. The subsystem name is sysctld. In order to start the sysctld daemon, use the startsrc -s sysctld command. This starts the daemon with the default arguments and SRC options. The sysctld daemon is setup to be respawnable and be the only instance of the sysctld daemon running on a particular node or control workstation. Do not start the sysctld daemon from the command line without using the startsrc command to start it.
To stop the sysctld daemon, use the stopsrc -s sysctld command. This stops the daemon and does not allow it to respawn.
To display the status of the sysctld daemon, use the lssrc -s sysctld command.
If the default startup arguments need to be changed, use the chssys command to change the startup arguments or the SRC options. Refer to AIX Version 4 Commands Reference and AIX Version 4 General Programming Concepts: Writing and Debugging Programs for more information about daemons under SRC control and how to modify daemon arguments when under SRC.
To view the current SRC options and daemon arguments, use the odmget -q 'subsysname=sysctld' SRCsubsys command.
Files
Related Information
Command: sysctl
Files: sysctl.acl, sysctl.conf
Examples
startsrc -s sysctld
stopsrc -s sysctld
lssrc -s sysctld
lssrc -a
odmget -q 'subsysname=sysctld' SRCsubsys
Purpose
SYSMAN_test - Verifies that the installation and customization of the Systems Management components of the SP system completed successfully.
Syntax
SYSMAN_test [-q | -v] [-l log_file]
Flags
Operands
None.
Description
The SYSMAN_test command performs various tests to determine whether the systems management components of the SP system are completely installed and customized properly.
A return code of 0 indicates that the test completed as expected; otherwise it returns the number of failures. If you do not specify the -q flag, a message is displayed on standard output that indicates the success or failure of the tests. In either case, the command returns 0 if successful, 1 if not. If errors are detected, more detailed information is recorded in the log file. If you do not specify the -l flag, error messages are recorded in /var/adm/SPlogs/SYSMAN_test.log.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit SP_verify
and select the RS/6000 SP System Management option.
Files
Related Information
Commands: CSS_test, jm_install_verify, jm_verify, SDR_test, spmon_ctest, spmon_itest
Examples
To verify systems management following customization, saving error messages in sm.errors in the current working directory, enter:
SYSMAN_test -l sm.errors
Purpose
syspar_ctrl - Starts, stops, adds, deletes, and refreshes the system partition-sensitive subsystems installed on your SP system.
Syntax
Flags
Note: | IBM suggests only turning on a particular subsystem's trace by providing a subsystem name. If the trace is turned on for all subsystems, the volume of data produced may quickly fill up /var. |
Operands
syspar_ctrl option haem
Description
This command acts as an interface to the system partition-sensitive subsystems supporting the functions that are shared by all subsystems. This command is also referred to as the Syspar Controller. It can be used to add or delete, start or stop, refresh or restore the subsystems, and various other functions. When used on the control workstation, it works with the subsystems on the control workstation. When used on the nodes, it works with the subsystems on the nodes. The refresh option is an exception. In order to refresh some subsystems, the subsystem must be refreshed on both the control workstation and on the nodes. In this case, the refresh on the control workstation will dsh an appropriate refresh command from the control workstation to the appropriate nodes.
This command supports two types of options: primitive options and macro options. Primitive options are passed directly to the underlying control scripts, for example, -a (add), -d (delete), -r (refresh). Macro options conveniently group a commonly used set of primitive options into one option, for example, -R (restore). All of the subsystems and each subsystem's control script that are managed by the Syspar Controller are listed in the Syspar Controller subsystems file. By default, all of the control scripts listed in the Syspar Controller subsystems file will be called unless a subsystem_name is provided. In that case, the control script for just the specified subsystem will be called.
This command is automatically called when the system is partitioned (spapply_config) to first stop and delete the system partition-sensitive subsystems from system partitions that are being removed, and then to add and start the system partition-sensitive subsystems (for example, hats, hb, and hr) in new system partitions.
The Syspar Controller is also called when restoring the SDR with sprestore_config to first clean up and then add and start the system partition-sensitive subsystems (for example, hats, hb and hr) in each system partition.
The Syspar Controller also needs to be called with refresh flag (-r) by the System Administrator using the command line whenever a node is added or deleted from the system, or a node is migrated to a new level of PSSP.
Files
Security
You must be running with an effective user ID of root.
Environment Variables
Exit Values
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/syspar_ctrl
Related Information
Commands: emonctrl, hatsctrl, hbctrl, hrctrl, haemctrl, hagsctrl, pmanctrl, sp_configdctrl, spapply_config, spcw_apps, sprestore_config
Examples
syspar_ctrl -G -A
syspar_ctrl -G -D
syspar_ctrl -r
syspar_ctrl -R
syspar_ctrl -k
syspar_ctrl -h haem
syspar_ctrl -E
lssrc -a | grep spp1 lssrc -a | grep sp_configd
Note: | The SDR is not managed by the System Controller. |
Purpose
sysparaid - Creates a layout for a new system partition configuration of an SP system.
Syntax
Flags
Operands
Description
Use this command to invoke the System Partitioning Aid, a tool for generating new system partition configuration layouts. When invoked with the -i flag, it creates a switch-map file that will help the user to generate the input_file for creating a layout for a desired system partition configuration. When invoked with no flags or with the -s or -t flags, it attempts to partition the system according to the input requirement. If the attempt is unsuccessful, it will output appropriate error messages to the log and exit. If the attempt is successful and the -s flag is specified, the newly created layout will be saved at the desired location specified by the flag argument.
Standard Input
This command requires an input file when invoked with no flag or the -s flag. The template for the input file can be found in /spdata/sys1/syspar_configs/bin.
Standard Output
Informational messages are written to standard output.
Standard Error
Error messages are written to standard error.
Output Files
This command creates spa.snapshot and spa.metrics under tmpdir (if specified) or under the current working directory. If the -s flag is specified and the attempt is successful, it creates the following under the layout directory:
When invoked with the -i flag, the command creates spa.sysinfo under tmpdir (if specified), or under the current working directory.
Extended Description
The sysparaid command uses a set of built-in rules to create a layout for a desired system partition configuration. The following startup guidelines will help to generate an acceptable input to the command:
sysparaid -i -t spa_dir
This command places a spa.sysinfo file in the spa_dir if the -t flag is used; otherwise, it places it in the current directory. If an SDR is not present, issue the following command:
sysparaid -i -t spa_dir topology_file
where topology_file is the name of the topology file for the system to be partitioned.
Note: | In this case, only the switch_port_numbers are provided. No node_numbers are available. |
Security
Any user can run this command. Only users authorized to write to the system partitioning directory can save a generated layout under it.
Location
/usr/lpp/ssp/bin
Related Information
The spsyspar command provides the graphical user interface (GUI) for the System Partitioning Aid.
Examples
Number of Nodes in System: 32 Number of Frames in System: 2 Frame Type: tall Switch Type: HiPS Number of Switches in Node Frames: 2 Number of Switches in Switch Only Frames: 0 Node Numbering Scheme: switch_port_number Number of Partitions: 3 Partition Name: part1 Number of Nodes in Partition: 8 0 - 7 Partition Name: part2 Number of Nodes in Partition: 8 8 - 15 Partition Name: part3 Number of Nodes in Partition: 16 remaining_nodes
To use /tmp as the working directory, enter:
sysparaid -t /tmp inpfile
You should receive a message similar to the following:
A layout, for the desired system partition configuration or an equivalent, can be created. To save this layout, invoke the command again with -s option.
To save the layout for this configuration under /spdata/sys1/syspar_configs/2nsb0isb/config.8_8_16/layout.myconfig, enter:
sysparaid -s myconfig inpfile
To save the layout for this configuration under /tmp/custom/config1, enter:
sysparaid -s /tmp/custom/config1 inpfile
Number of Nodes in System: 87 Number of Frames in System: 6 Frame Type: tall Switch Type: SP Number of Switches in Node Frames: 6 Number of Switches in Switch Only Frames: 4 Node Numbering Scheme: switch_port_number Number of Partitions: 2 Partition Name: ProductionPartition Number of Nodes in Partition: 82 0 4 16 - 95 Partition Name: TestPartition Number of Nodes in Partition: 5 2 6 8 10 12
If you enter the sysparaid -s myconfig inpfile command, this configuration will be saved under /spdata/sys1/syspar_configs/6nsb4isb/config.12_84/layout.myconfig. Note that the nine unspecified switch port numbers have been allocated to one of the two system partitions.
Number of Nodes in System: 8 Number of Frames in System: 2 Frame Type: tall Switch Type: SP Number of Switches in Node Frames: 1 Number of Switches in Switch Only Frames: 0 Node Numbering Scheme: node_number Number of Partitions: 3 Partition Name: part1 Number of Nodes in Partition: 2 25 29 Partition Name: part2 Number of Nodes in Partition: 4 1 5 17 21 Partition Name: part3 Number of Nodes in Partition: 2 3 7
This input file for a particular SP system returned the location of an existing layout:
The layout for the desired/equivalent system partition configuration is under /spdata/sys1/syspar_configs/1nsb0isb/config.4_4_8/layout.2
switch_number switch_chip switch_port_number node_number 1 4 9 25 1 4 13 29 1 5 0 1 1 5 1 17 1 5 4 5 1 5 5 21 1 6 2 3 1 6 6 7
Number of Nodes in System: 32 Number of Frames in System: 2 Frame Type: tall Switch Type: NA Number of Switches in Node Frames: 0 Number of Switches in Switch Only Frames: 0 Node Numbering Scheme: switch_port_number Number of Partitions: 2 Partition Name: part1 Number of Nodes in Partition: 14 2 - 5 10 11 13 15 19 24 - 25 29 - 31 Partition Name: partition2 Number of Nodes in Partition: 18 remaining_nodes
To save the layout for this configuration under /spdata/sys1/syspar_configs/2nsb0isb/config.14_18/layout.myconfig, enter:
sysparaid -s myconfig inpfile
Purpose
s1term - Opens a connection to an SP node's S1 serial port.
Syntax
s1term [-G] [-w] frame_ID slot_ID
Flags
Operands
Description
Use this command to open a connection to the S1 serial port of the SP node contained in the slot specified by the frame_ID and slot_ID operands. The specified node must be in the current system partition unless the -G flag is also specified. By default, the connection is read only. As data arrives from the serial port, it is written to standard output. When the connection is read/write and standard input is a terminal, the terminal is placed in raw mode, that is, canonical processing is turned off in the terminal driver. As data is read from standard input, it is sent to the S1 serial port. Standard input and output can be files or pipes.
When the connection is read only, the command terminates upon receipt of a signal, usually generated by the terminal Interrupt key. When in read/write mode, the command terminates when either the termination character or End-of-File is read from standard input. The termination character is Ctrl-x by default. Another termination character can be used by setting the S1TERMESC environment variable to the octal (denoted by leading 0), decimal or hexadecimal (denoted by leading 0x) value of the desired termination character.
Note: | The termination character must only be one byte. |
To execute this command, the user must be authorized to access the Hardware Monitor subsystem and, for the frame specified to the command, must be granted S1 permission. Since the Hardware Monitor subsystem uses SP authentication services, the user must execute the kinit command prior to executing this command. Alternatively, site-specific procedures can be used to obtain the tokens that are otherwise obtained by kinit.
Files
Related Information
Commands: hmcmds, hmmon
Examples
s1term -w 12 8
s1term 9 2 > s1term.output
Purpose
ucfghsd - Makes a data striping device (HSD) for the IBM Virtual Shared Disks unavailable.
Syntax
ucfghsd -a | hsd_name...
Flags
Operands
Description
This command unconfigures the already defined data striping devices . This command does not change the definition of the HSDs; it makes the HSDs unavailable on one node.
Files
Security
You must have root privilege to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfghsd, defhsd, hsdatalst, lshsd, undefvsd
Examples
To unconfigure the data striping device hsd1, enter:
ucfghsd hsd1
Purpose
ucfghsdvsd - Makes a data striping device (HSD) and the IBM Virtual Shared Disks that comprise it unavailable on one node.
Syntax
ucfghsdvsd -a | {hsd_name...}
Flags
Operands
None.
Description
Use this command to unconfigure HSDs and their underlying IBM Virtual Shared Disks. This command does not change the definition of the HSDs and IBM Virtual Shared Disks; it just makes them unavailable. The underlying IBM Virtual Shared Disks do not have to be in the stopped state for this command to work. The IBM Virtual Shared Disks will be stopped and then unconfigured.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit hsd_mgmtand select the Unconfigure an HSD and its Underlying IBM Virtual Shared Disks option.
Files
Security
You must have root privilege or sysctl and sysctl.vsd access and authorization from your system administrator to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfghsdvsd, ucfghsd, ucfgvsd
Examples
To unconfigure the data striping device hsd1 and the IBM Virtual Shared Disks that comprise it, enter:
ucfghsdvsd hsd1
Purpose
ucfgvsd - Makes an IBM Virtual Shared Disk unavailable.
Syntax
ucfgvsd -a | vsd_name ...
Flags
Operands
Description
The ucfgvsd command unconfigures the specified IBM Virtual Shared Disks. This command does not change any IBM Virtual Shared Disk definitions. It moves IBM Virtual Shared Disks from the stopped state to the defined state.
If a configured HSD is using this IBM Virtual Shared Disk, you must first unconfigure the HSD before you unconfigure the IBM Virtual Shared Disk.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit vsd_mgmtand select the Unconfigure an IBM Virtual Shared Disk option.
Files
Security
You must have root privilege to run this command.
Restrictions
If you have the IBM Recoverable Virtual Shared Disk product installed and operational, do not use this command. The results may be unpredictable.
See IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: cfgvsd, ctlvsd, lsvsd, preparevsd, resumevsd, startvsd, stopvsd, suspendvsd
Examples
To unconfigure the IBM Virtual Shared Disk vsd1vg1n1 in the stopped state, enter:
ucfgvsd vsd1vg1n1
Purpose
unallnimres - Deallocates Network Installation Management (NIM) resources from a NIM master to one or more NIM clients.
Syntax
unallnimres -h | -l node_list
Flags
Operands
None.
Description
Use this command to unallocate all NIM resources from a NIM client.
Standard Error
This command writes error messages (as necessary) to standard error.
Exit Values
Security
You must have root privilege to run this command.
Implementation Specifics
This command is part of the IBM Parallel System Support Programs (PSSP) Licensed Program Product (LPP).
Location
/usr/lpp/ssp/bin/unallnimres
Related Information
Commands: allnimres, setup_server
Examples
To unallocate boot/installation resources to boot/install client nodes 1, 3, and 5 from their respective boot/install servers, enter:
unallnimres -l 1,3,5
Purpose
undefhsd - Undefines a data striping device (HSD).
Syntax
undefhsd hsd_name...
Flags
None.
Operands
Description
This command is used to remove a data striping device (HSD) by removing its definition from the system, including the special device files in /dev. The HSDs must be unconfigured and in the defined state on all nodes in the system partition.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit delete_vsdand select the Undefine a Hashed Shared Disk option.
Files
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: defhsd, hsdatalst
Examples
To delete the HSD information associated with the data striping device hsd1 from the SDR, enter:
undefhsd hsd1
Purpose
undefvsd - Undefines an IBM Virtual Shared Disk.
Syntax
undefvsd vsd_name ...
Flags
None.
Operands
Description
This command is used to remove IBM Virtual Shared Disk definition data from the System Data Repository (SDR) and any special device files from /dev for the given vsd_names on all the IBM Virtual Shared Disk nodes. The IBM Virtual Shared Disks must be unconfigured and in the defined state on all the IBM Virtual Shared Disk nodes.
You can use the System Management Interface Tool (SMIT) to run the undefvsd command. To use SMIT, enter:
smit delete_vsdand select the Undefine a Virtual Shared Disk option.
Files
Security
You must be in the bin group to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Command: defvsd
Examples
To delete the IBM Virtual Shared Disk information associated with IBM Virtual Shared Disk vsd1vg2n1 from the SDR, enter:
undefvsd vsd1vg2n1
Purpose
unfencevsd - Gives applications running on a node or group of nodes access to an IBM Virtual Shared Disk or group of IBM Virtual Shared Disks that were previously fenced from applications running on those nodes.
Syntax
unfencevsd [-v] vsd_name_list {-n node_list [-f] | -r}
Flags
Note: | Use unfencevsd -v -n to unfence nodes. Only use -r to remove an IBM Virtual Shared Disk fence record from the SDR while no IBM Virtual Shared Disk is configured on any node. |
Operands
None.
Description
Under some circumstances, the system may believe a node has failed and may begin recovery procedures when the node is actually operational, but is cut off from communication with other nodes running the same application. In this case, the "failed" node must not be allowed to serve requests for the IBM Virtual Shared Disks it normally manages until recovery is complete and the other nodes running the application recognize the failed node as operational. The fencevsd command prevents the failed node from filling requests for its IBM Virtual Shared Disks. The unfencevsd command allows fenced nodes to regain access to the IBM Virtual Shared Disks they normally act as servers for.
This command can be run from any node.
Note: | This command will fail if you do not specify a current server (primary or backup) to an IBM Virtual Shared Disk with the -v flag. |
Note: | This command changes SDR attributes when issued with the -r flag. Specify -r only when disks have already been removed from a fenced IBM Virtual Shared Disk. |
Files
Security
You must have root privilege to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: fencevsd, lsfencevsd, lsvsd, updatevsdtab, vsdchgserver
Refer to IBM Parallel System Support Programs for AIX: Managing Shared Disks for information on how to use this command in writing applications.
Examples
unfencevsd -v vsd1,vsd2 -n 5
unfencevsd -v vsd1,vsd2 -n 7 -f
Purpose
updatehsd - Lets you change the option in the System Data Repository (SDR) that prevents overwriting the Logical Volume Control Block (LVCB) for specified data striping devices (HSDs).
Syntax
Flags
Operands
None.
Description
Use this command only on the control workstation.
Note: | This utility is very powerful. Misuse can destroy the contents of a database. Only the superuser should be allowed to run it. If a database has been loaded on a configured HSD, modifying the protect_lvcb or not_protect_lvcb option will destroy the database. |
The HSD name must be specified. You must choose either protect_lvcb or not_protect_lvcb. The data striping device must be defined in the System Data Repository (SDR).
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit set_HSDdd_parmsand select the Update Hashed Shared Disk Options. option or
smit hsd_mgmtd_parmsand select the Set/Show HSD Device Driver Operational Parameters. option or the Update Hashed Shared Disk Options option.
Files
Security
You must have sysctl and sysctl.vsd access and authorization from your system administrator to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: defhsd, dshsd, hsdatalst
Examples
To set the protect_lvcb option for hsdcont01 and hsdcont02, enter:
updatehsd -d hsdcont01,hsdcont02 -o protect_lvcb
Purpose
updatevsdnode - Changes IBM Virtual Shared Disk options in the System Data Repository (SDR).
Syntax
Flags
Operands
None.
Description
Use updatevsdnode to change the specified values in the SDR for all nodes in node_list.
Note: | This command only changes the information in the SDR. In order to effectively configure the IBM Virtual Shared Disks, you must first unconfigure all the IBM Virtual Shared Disks and then reconfigure them. |
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit vsd_mgmtand select the Set/Show IBM Virtual Shared Disk Device Driver Operational Parameters option or the Update IBM Virtual Shared Disk Device Driver Node Parameters option.
Files
Security
You must have sysctl and sysctl.vsd access and authorization from your system administrator to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Command: vsdatalst, vsdnode
Examples
To change buddy buffer options after you've installed a new SP Switch Adapter, enter:
updatevsdnode ALL -b 4096 -x 4 -s 5This command leaves min_buddy_buffer_size at 4KB, the default, lowers the max_buddy_buffer_size to 4KB as well, and allocates a maximum of 5 buddy buffers.
Purpose
updatevsdtab - Changes the IBM Virtual Shared Disk cache/nocache option in the System Data Repository (SDR).
Syntax
updatevsdtab {-v vsd_names | -a} {[-o {cache | nocache}] [-s]} [-f]
Flags
Operands
None.
Description
Use this command to update the SDR, if necessary. If the -f flag is specified, the IBM Virtual Shared Disks involved will be reconfigured (using sysctl) on all nodes that are up and initially had these IBM Virtual Shared Disks configured.
You can use the System Management Interface Tool (SMIT) to run this command. To use SMIT, enter:
smit vsd_mgmtand select the Set/Show IBM Virtual Shared Disk Device Driver Operational Parameters option or the Update IBM Virtual Shared Disk Options option.
Files
Security
You must have sysctl and sysctl.vsd access and authorization from your system administrator to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: defvsd, updatevsdnode
Examples
updatevsdtab -a -o nocache
updatevsdtab -v USER1n3 -s
Purpose
verparvsd - Verifies IBM Virtual Shared Disk system partitioning.
Syntax
verparvsd [-F] [-o output_file] layout_directory [new_partition ...]
Flags
Operands
Description
Use this command to verify that the system partition proposed in the layout_directory will work for all the existing IBM Virtual Shared Disk data. The spapply_config command invokes this command to partition the IBM Virtual Shared Disk data during a system partition operation. The verparvsd command extracts all IBM Virtual Shared Disk data from nodes involved in the system partitioning and writes SDR commands to the output file that will reload the IBM Virtual Shared Disk SDR data into the correct new system partitions. This file is executed during the system partitioning process to partition the IBM Virtual Shared Disk data.
The spapply_config command invokes this command and its output to effect IBM Virtual Shared Disk system partitioning. You can also invoke the command prior to invoking the spapply_config command to see how well suited the desired layout is for the existing IBM Virtual Shared Disk configuration as defined in the SDR.
This command only checks and processes the new system partitions listed on the command line. If some existing system partitions are to be unchanged in the system partitioning operation, do not list those system partition names on the command line. If no new system partitions are listed, the default is to process all system partitions in the layout directory.
This command checks to see if the IBM Virtual Shared Disk data can be partitioned as specified by the layout directory without any problems. The command reports any problems it identifies, as well as reports how it would fix the problem.
The verparvsd command places global volume groups (GVGs) in the system partition containing their primary server node. IBM Virtual Shared Disks are placed in the system partition of their GVG. HSDs are placed in the system partition containing their first IBM Virtual Shared Disk.
The verparvsd command looks for the following types of errors in each new system partition:
As a corollary, if an HSD with .BAD at the end of its name is found in the new system partition to have all its IBM Virtual Shared Disks in the system partition, the .BAD will be removed from its name.
Files
Exit Values
The verparvsd command looks for error types (described previously) in each new system partition and corrects them as specified:
In either case, verparvsd processes all the IBM Virtual Shared Disk data, and generates a complete list of errors on standard error, and a complete SDR command list to the output file.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: defhsd, defvsd, spapply_config, spdisplay_config, vsdnode, vsdvg
Examples
To see how well suited the configuration specified in the config.4_4_8/layout.6 layout directory is to your IBM Virtual Shared Disk configuration, enter:
verparvsd config.4_4_8/layout.6
Purpose
vhostname - Sets or displays the virtual host name.
Syntax
vhostname [-s] [host_name]
Flags
Operands
Note: | You must have root authority to use the host_name operand. |
Description
Use this command to display or set the virtual host name of the local host. Only users with root authority can set the virtual host name. The host_name is stored in the /etc/vhostname file.
If displaying the virtual host name and the virtual host name has not been set and the /etc/vhostname file does not exist, vhostname will return the real host name from the kernel variable.
When setting the virtual host name, if the /etc/vhostname file does not exist, it will be created. If it does exist, the file contents will be overwritten by the new virtual host name.
To clear the virtual host name so that the virtual host name no longer exists, remove the /etc/vhostname file.
Note: | You must have root authority to remove the /etc/vhostname file. |
The virtual host name is used in fail over situations when an application has associated the host name in the kernel of a particular machine to the service it is providing. When the application is restarted on the fail over node that has a different host name, the application may fail or act incorrectly. If the application needs to associate a host name with a particular service and it cannot handle having multiple host names, a virtual host name can be provided. The application can call vhostname instead of hostname and get the host name of the node it normally runs on. This eliminates the need to change the real host name in the kernel on the fail over node. It should be noted that changing the real host name in the kernel can cause problems with other applications that rely on the real host name in the kernel to identify the physical machine.
Note: | The High Availability Cluster Multiprocessing (HACMP) event scripts provided with the High Availability Control Workstation (HACWS) option of the IBM Parallel System Support Programs for AIX (PSSP) set and clear the virtual host name in the HACMP pre- and post-event scripts. The administrator normally should not have to set or clear the virtual host name. |
Files
Exit Values
Related Information
Subroutines: getvhostname, setvhostname
AIX command: hostname
AIX Subroutines: gethostname, sethostname
Examples
vhostname
vhostname spcw_prim
vhostname -s
A vhostname of donald prints out.
rm /etc/vhostname
Note: | You must have root authority to remove the /etc/vhostname file. |
Purpose
vsdatalst - Displays IBM Virtual Shared Disk system definition data from the System Data Repository (SDR).
Syntax
vsdatalst [-G] -g | -n | -v
Flags
Only one of the following flags can be specified with each invocation of vsdatalst:
Operands
None.
Description
Use this command to display one of several kinds of information to standard output.
You can use the System Management Interface Tool (SMIT) to run the vsdatalst command. To use SMIT, enter:
smit list_vsdand select the option for the kind of IBM Virtual Shared Disk SDR information you wish to see.
Security
You must be in the bin group to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Examples
vsdatalst -gThe system displays a message similar to the following:
Note: | backup or secondary_server_node is only enabled with IBM Recoverable Virtual Shared Disk. |
VSD Global Volume Group Information Global Volume Local Server Node Numbers: eio_ Group name VG name primary backup recovery Recovery ---------------- -------- ----------- -------- -------- -------- hunter-rileysvg rileysvg 1 0 0 0 ppstest1-rootvg rootvg 3 0 0 0 tattooine-rootvg rootvg 2 0 0 0
vsdatalst -nThe system displays a message similar to the following:
VSD Node Information Initial Maximum VSD rw Buddy Buffer: node VSD cache cache req. req. min. max. size: # # host_name adapt. buffers buffers count count size size maxbufs ---- --------- ------ ------- ------- ----- ----- ---- ---------------- 1 hunter tr0 64 256 256 48 4096 65536 4 2 tattooine tr0 64 256 256 48 4096 65536 4 3 ppstest1 tr0 64 256 256 48 4096 65536 4
vsdatalst -vThe system displays a message similar to the following:
VSD Table VSD name logical volume Global Volume Group minor# option ----------------- --------------- ----------------------- ------ ------ vsd.rlv01 rlv01 hunter-rileysvg 2 cache vsd.rlv02 rlv02 hunter-rileysvg 3 cache vsd.vsd1 vsd1 tattooine-rootvg 1 nocache vsd.vsdp1 vsd1 ppstest1-rootvg 4 nocache
Purpose
vsdchgserver - Updates the server function for a global volume group defined within the IBM Virtual Shared Disk subsystem. When the current acting server of the global volume group is changed or updated, the server function will also be moved.
Syntax
Flags
Operands
None.
Description
The vsdchgserver command allows the serving function for a global volume group defined on a primary node to be taken over by the secondary node, or to be taken over by the primary node from the secondary node. This allows an application to continue to use IBM Virtual Shared Disks in situations where the cable or adapter between the physical disks and one of the attached nodes is not working.
The IBM Recoverable Virtual Shared Disk system automatically updates the IBM Virtual Shared Disk devices if, and only if, the vsdchgserver command is used to flip the currently-defined primary node and secondary node in the global volume group specified in the -g flag.
Files
Security
You must have root privilege to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Refer to IBM Parallel System Support Programs for AIX: Managing Shared Disks for information on how to use this command in writing applications.
Examples
To change the primary server node for the global volume group node12vg to node 1 and the secondary node to node 2, with EIO recovery, enter:
vsdchgserver -g node12vg -p 1 -b 2 -o 1
Purpose
vsddiag - Displays information about the status of IBM Virtual Shared Disks.
Syntax
vsddiag
Flags
None.
Operands
None.
Description
This command displays information about IBM Virtual Shared Disks that can help you determine their status and collect information that helps IBM service representatives diagnose system problems.
Note: | The vsddiag command can only be used when no IBM Virtual Shared Disk I/O is in progress. |
Files
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: vsdatalst, vsdsklst
Examples
To display information about the IBM Virtual Shared Disks in your system or system partition, enter:
vsddiagIf all IBM Virtual Shared Disk's are created and configured correctly, the output is:
Checking server vsds Checking VSD request sequence number. Checking device drivers. end of vsdl1diag:checkvsdl1 program.
If there are no IBM Virtual Shared Disk's defined, the output is:
k5n02.ppd.pok.ibm.com VSD_ERROR:3:No IBM Virtual Shared Disks are configured on this node. k5n01.ppd.pok.ibm.com VSD_ERROR:3:No IBM Virtual Shared Disks are configured on this node. Checking server vsds Checking VSD request sequence number. Checking device drivers. end of vsdl1diag:checkvsdl1 program.
If there is something wrong with the IBM Virtual Shared Disk's, the output is:
k5n02.ppd.pok.ibm.com VSD_ERROR:3:No IBM Virtual Shared Disks are configured on this node. k5n01.ppd.pok.ibm.com VSD_ERROR:3:No IBM Virtual Shared Disks are configured on this node. Checking server vsds Checking VSD request sequence number. Checking device drivers. vsdl1diag:checkvsdl1: 0034-619 Device driver on node 14 is not at the same level as others on this SP system or system partition. vsdl1diag:checkvsdl1: 0034-620 VSD Maximum IP Message Size on node 14 is not at the same level as others on this SP system or system partition.
Purpose
vsdelnode - Removes IBM Virtual Shared Disk information for a node or series of nodes from the System Data Repository (SDR).
Syntax
vsdelnode node_number ...
Flags
None.
Operands
Description
This command is used to remove IBM Virtual Shared Disk data for a node or series of nodes from the SDR.
The vsdelnode command makes the listed nodes no longer IBM Virtual Shared Disk nodes so that no IBM Virtual Shared Disks can be accessed from them. This command fails for any nodes that are servers for any global volume groups.
You can use the System Management Interface Tool (SMIT) to run the vsdelnode command. To use SMIT, enter:
smit delete_vsdand select the Delete IBM Virtual Shared Disk Node Information option.
Security
You must be in the bin group to run this command.
Restrictions
If you have the IBM Recoverable Virtual Shared Disk product installed and operational, do not use this command. The results may be unpredictable.
See IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: vsdatalst, vsdnode
Examples
To delete IBM Virtual Shared Disk node information for nodes 3 and 6, enter:
vsdelnode 3 6
Purpose
vsdelvg - Removes IBM Virtual Shared Disk global volume group information from the System Data Repository (SDR).
Syntax
vsdelvg [-f] global_group_name ...
Flags
Operands
Description
Use this command to remove IBM Virtual Shared Disk global volume group information from the SDR. If any IBM Virtual Shared Disks are defined on a global volume group, the vsdelvg command fails unless -f is specified. If -f is specified, any such IBM Virtual Shared Disks must be unconfigured and in the defined state on all the IBM Virtual Shared Disk nodes to be deleted.
You can use the System Management Interface Tool (SMIT) to run the vsdelvg command. To use SMIT, enter:
smit delete_vsdand select the Delete IBM Virtual Shared Disk Global Volume Group Information option.
Files
Security
You must be in the bin group to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: vsdatalst, vsdvg, undefvsd
Examples
To delete the IBM Virtual Shared Disk information associated with global volume group vg1n1 from the SDR, enter:
vsdelvg vg1n1
Purpose
vsdnode - Enters IBM Virtual Shared Disk information for a node or series of nodes into the System Data Repository (SDR).
Syntax
Flags
None.
Operands
Note: | If the application issues requests larger than 64KB, the number of buddy buffers can be increased, depending on the size of the requests. |
Description
Use this command to make the specified nodes IBM Virtual Shared Disk nodes and to assign their IBM Virtual Shared Disk operational parameters. The operational parameters are: adapter name, initial cache buffer count, maximum cache buffer count, read/write request count, IBM Virtual Shared Disk request count, and buddy buffer parameters. If this information is the same for all nodes, run this command once. If the information is different for the nodes, run this command once for each block of nodes that should have the same IBM Virtual Shared Disk information.
You can use the System Management Interface Tool (SMIT) to run the vsdnode command. To use SMIT, enter:
smit vsd_dataand select the IBM Virtual Shared Disk Node Information option.
Files
Security
You must be in the bin group to run this command.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: vsdatalst, vsdelnode
Refer to IBM Parallel System Support Programs for AIX: Managing Shared Disks for defining IBM Virtual Shared Disk information in the SDR.
Examples
The following example adds SDR information for a css0 network and nodes 1 through 8 with css0 being the adapter name used for IBM Virtual Shared Disk communications. Initially, this is accomplished with 64 cache buffers and with a maximum of 256 cache buffers to be used for IBM Virtual Shared Disk support, with allowance for 48 outstanding read/write requests per IBM Virtual Shared Disk and 256 outstanding IBM Virtual Shared Disk requests per node with two 256KB max_buddy_buffers and a 4KB min_buddy_buffer_size.
vsdnode 1 2 3 4 5 6 7 8 css0 64 256 256 48 4096 65536 2 262144
Purpose
vsdsklst - Produces output that shows you the disk resources used by the IBM Virtual Shared Disk subsystem across a system or system partition.
Syntax
vsdsklst {-d | -v | -dv} {[-a] | -n node_list]} [-G]
Flags
Operands
None.
Description
Use this command to check disk utilization across a system or system partition.
Files
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Command: vsdatalst
Examples
This command:
vsdsklst -dv -adisplays the following information on a system that has volume groups and IBM Virtual Shared Disks defined on nodes 1, 3, 5, 7, 10, and 12. Node 5 is temporarily inactive.
k7n12.ppd.pok.ibm.com Node Number:12; Node Name:k7n12.ppd.pok.ibm.com Volume group:rootvg; Partition Size:4; Total:537; Free:315 Physical Disk:hdisk0; Total:537; Free:315 Volume group:vsdvg; Partition Size:4; Total:537; Free:533 Physical Disk:hdisk1; Total:537; Free:533 VSD Name:1HsD8n12{lv1HsD8n12}; Size:2 VSD Name:1HsD20n12{lv1HsD20n12}; Size:2
k7n01.ppd.pok.ibm.com Node Number:1; Node Name:k7n01.ppd.pok.ibm.com Volume group:rootvg; Partition Size:4; Total:537; Free:210 Physical Disk:hdisk0; Total:537; Free:210 Volume group:vsdvg; Partition Size:4; Total:537; Free:533 Physical Disk:hdisk1; Total:537; Free:533 VSD Name:1HsD1n1{lv1HsD1n1}; Size:2 VSD Name:1HsD13n1{lv1HsD13n1}; Size:2
k7n05.ppd.pok.ibm.com No response
k7n10.ppd.pok.ibm.com Node Number:10; Node Name:k7n10.ppd.pok.ibm.com Volume group:rootvg; Partition Size:4; Total:537; Free:303 Physical Disk:hdisk0; Total:537; Free:303 VSD Name:vsdn10v1{lvn10v1}; Size:4 VSD Name:vsdn10v2{lvn10v2}; Size:4 VSD Name:vsdn10v3{lvn10v3}; Size:4 Volume group:vsdvg; Partition Size:4; Total:537; Free:533 Physical Disk:hdisk1; Total:537; Free:533 VSD Name:1HsD6n10{lv1HsD6n10}; Size:2 VSD Name:1HsD18n10{lv1HsD18n10}; Size:2
k7n03.ppd.pok.ibm.com Node Number:3; Node Name:k7n03.ppd.pok.ibm.com Volume group:rootvg; Partition Size:4; Total:537; Free:269 Physical Disk:hdisk0; Total:537; Free:269 VSD Name:vsdn03v1{lvn03v1}; Size:4 VSD Name:vsdn03v2{lvn03v2}; Size:4 VSD Name:vsdn03v3{lvn03v3}; Size:4 Volume group:vsdvg; Partition Size:4; Total:537; Free:533 Physical Disk:hdisk1; Total:537; Free:533 VSD Name:1HsD2n3{lv1HsD2n3}; Size:2 VSD Name:1HsD14n3{lv1HsD14n3}; Size:2
k7n07.ppd.pok.ibm.com Node Number:7; Node Name:k7n07.ppd.pok.ibm.com Volume group:rootvg; Partition Size:4; Total:537; Free:300 Physical Disk:hdisk0; Total:537; Free:300 VSD Name:vsdn07v1{lvn07v1}; Size:4 VSD Name:vsdn07v2{lvn07v2}; Size:4 VSD Name:vsdn07v3{lvn07v3}; Size:4 Volume group:vsdvg; Partition Size:4; Total:537; Free:533 Physical Disk:hdisk1; Total:537; Free:533 VSD Name:1HsD4n7{lv1HsD4n7}; Size:2 VSD Name:1HsD16n7{lv1HsD16n7}; Size:2
To view the output for a specific node, type:
vsdsklst -n 12The output is:
k7n07.ppd.pok.ibm.com Node Number:7; Node Name:k7n07.ppd.pok.ibm.com Volume group:rootvg; Partition Size:4; Total:537; Free:300 Physical Disk:hdisk0; Total:537; Free:300 VSD Name:vsdn07v1{lvn07v1}; Size:4 VSD Name:vsdn07v2{lvn07v2}; Size:4 VSD Name:vsdn07v3{lvn07v3}; Size:4 Volume group:vsdvg; Partition Size:4; Total:537; Free:533 Physical Disk:hdisk1; Total:537; Free:533 VSD Name:1HsD4n7{lv1HsD4n7}; Size:2 VSD Name:1HsD16n7{lv1HsD16n7}; Size:2
If both the rootvg and testvg volume groups are varied on, the system displays output similar to the following:
Node Number:12; Node Name:k21n12.ppd.pok.ibm.com Volume group:rootvg; Partition Size:4; Total:537; Free:47 Physical Disk:hdisk0; Total:537; Free:47 VSD Name:1HsD1n12[lv1HsD1n12]; Size:5 VSD Name:1HsD2n12[lv1HsD2n12]; Size:5 VSD Name:vsd4n12[lvvsd4n12]; Size:4 VSD Name:vsd5n12[lvvsd5n12]; Size:4 VSD Name:vsd6n12[lvvsd6n12]; Size:4 Volume group:testvg; Partition Size:4; Total:537; Free:313 Physical Disk:hdisk1; Total:537; Free:313 VSD Name:vsd14n12[lvvsd14n12]; Size:4
If the testvg volume group is not varied on, the system displays output similar to the following:
Node Number:12; Node Name:k21n12.ppd.pok.ibm.com Volume group:rootvg; Partition Size:4; Total:537; Free:47 Physical Disk:hdisk0; Total:537; Free:47 VSD Name:1HsD1n12[lv1HsD1n12]; Size:5 VSD Name:1HsD2n12[lv1HsD2n12]; Size:5 VSD Name:vsd4n12[lvvsd4n12]; Size:4 VSD Name:vsd5n12[lvvsd5n12]; Size:4 VSD Name:vsd6n12[lvvsd6n12]; Size:4 Volume group:testvg is not varied on. Physical Disk:hdisk1;
Instead of issuing this command directly, you should use the appropriate SMIT panels to view it in the best format. To view information about volume groups, type:
smit lsvgTo view information about logical volumes, type:
smit lslvTo view information about physical volumes, type:
smit lspv
Purpose
vsdvg - Defines an IBM Virtual Shared Disk global volume group.
Syntax
Flags
Operands
This can be specified in four different ways:
Note: | This operand is used only by the IBM Recoverable Virtual Shared Disk product. |
Description
Use this command to define volume groups for use by IBM Virtual Shared Disk support. This is done by specifying the local volume group name, the node on which it resides, and the name by which the volume group will be known throughout the cluster.
If eio_recovery is set (to a value of 1) due to disk failure (EIO error), the IBM Recoverable Virtual Shared Disk system will perform a full recovery by flipping the current primary node and the secondary node and doing one more retry on the new primary node.
You can use the System Management Interface Tool (SMIT) to run the vsdvg command. To use SMIT, enter:
smit vsd_dataand select the IBM Virtual Shared Disk Global Volume Group Information option.
Files
Security
You must be in the bin group to run this command.
Restrictions
The secondary_server_node operand is used only by the IBM Recoverable Shared Disk product.
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Command: vsdelvg
Examples
vsdvg vg2n17 17
vsdvg vg1p3s15 3 15
Purpose
vsdvgts - Reads the timestamp from the volume group descriptor area (VGDA) of the physical disks and sets the value in the System Data Repository (SDR).
Syntax
vsdvgts [-a] [volgrp]
Flags
Operands
Description
Use this command to update the timestamp that the IBM Recoverable Virtual Shared Disk software uses to determine if a twin-tailed volume group has changed. When the software detects a change, the recovery scripts export the volume group and then import the volume group.
This command can be used to avoid exporting the volume group and then importing the volume group during recovery in situations where the export and import operations are not really necessary. This command should be used very carefully.
Exit Values
Security
You must have root privilege to issue the vsdvgts command.
Implementation Specifics
This command is part of the IBM Recoverable Virtual Shared Disk Licensed Program Product (LPP).
Prerequisite Information
See "Using the IBM Recoverable Virtual Shared Disk Software" in IBM Parallel System Support Programs for AIX: Managing Shared Disks.
Location
/usr/lpp/csd/bin/vsdvgts
Examples
To update the timestamp associated with the IBM Virtual Shared Disk volume group vsdvg1 for just this node, enter:
vsdvgts vsdvg1
Purpose
vsdvts - Verifies that the IBM Virtual Shared Disk software works.
Syntax
vsdvts [-b block_size] [-n number_of_blocks] vsd_name [file]
Flags
Operands
Description
Attention |
---|
Data on vsd_name and its underlying logical volume is overwritten and, therefore, destroyed. Use this command after you have defined an IBM Virtual Shared Disk (including its underlying logical volume), but before placing application data on it. |
Use this command to verify that the vsd_name is in the active state and then to write the specified part of file to the raw vsd_name device, /dev/rvsd_name. This command reads the data back from the IBM Virtual Shared Disk, then compares it to file. If the data is the same, the test is successful and vsdvts succeeds. Otherwise, vsdvts fails. The dd command is used for all I/O operations.
Try vsdvts on both a server and client node (for example, on both the node with a logical volume and one without it).
Prerequisite Information
IBM Parallel System Support Programs for AIX: Managing Shared Disks
Related Information
Commands: vsdnode, vsdvg, defvsd, cfgvsd, startvsd, dd
The preceding commands are listed in their order of use.
This part of the book contains the SP files, subroutines, and other technical information.
Purpose
auto.master - Specifies the master input file to the AIX automount daemon defining the file systems to be controlled and their associated map files.
Description
The auto.master file is the master map file for the AIX automount daemon. It identifies the file systems that are to be controlled by the automounter and the directory map file associated with each file system. It may also contain default mount information for specific file systems. The auto.master file may reference a Network Information Service (NIS) configuration map that is to be used by the automounter. Entries in the master map file use one of the following formats:
+NIS_map Directory_path Automount_map_name [default_mount_options]
The first form specifies an NIS configuration map that contains mount point and automounter map information. The second form directly identifies the directory path of the file system that is to be controlled by the automounter, along with the map file containing entries for the supported directories within that file system. Default options to be used when the file system is mounted can be optionally specified.
Files
Related Information
AIX Daemons: automount
The "Managing the Automounter" chapter in IBM Parallel System Support Programs for AIX: Administration Guide
The "Network File System (NFS)" and "Network Information Service (NIS)" chapters in AIX Version 4 System Management Guide: Communications and Networks
Examples
The following example is a copy of the default auto.master file shipped with the SP modified to support an NIS configuration map:
# The following entry will use the NIS configuration map if NIS is # defined for this system and the database exists +auto_master # The following entry provides automount support for users' home # directories /u /etc/auto/maps/auto.u -rw,hard,retry=3,timeo=40,rsize=4096, \ wsize=4096
Purpose
haemloadlist File - Event Management configuration data that is to be loaded into the System Data Repository (SDR)
Description
IBM Parallel System Support Programs for AIX (PSSP) includes a number of resource monitors that use the Resource Monitor Application Programming Interface (RMAPI) to supply information about system resources to the Event Management subsystem. Information about the monitors and the resources is defined in the haemloadlist file and includes:
Before the Event Management subsystem can use this information, it must first be loaded into the System Data Repository (SDR) and then compiled into a binary Event Management Configuration Database (EMCDB) file. The haemloadcfg command is used to load the data from the haemloadlist file into the SDR. The haemcfg command is then used to compile the data from the SDR into the binary EMCDB file. Both of these commands are issued automatically by the haemctrl command when it is used to start the Event Management subsystem on the control workstation.
For resource monitors other than those supplied by PSSP, IBM suggests that you create one or more separate files in load list format to contain their Event Management configuration data. You can specify the name of any file in load list format on the haemloadcfg command.
You can use SDR commands to work with objects in the Event Management classes directly. However, IBM suggests that you use load list files and the haemloadcfg command to work with objects in these classes.
You cannot use System Management Interface Tool (SMIT) panels to work with objects in these classes.
The Format of an Event Management Load List File
Source information for the EMCDB is kept in the several partitioned classes in the SDR, each of which has a set of associated attributes. The format of an Event Management load list file, which can be used as input to the haemloadcfg command, follows the structure of the Event Management data in the SDR.
This man page describes the format of the Event Management load list file. For detailed information about its content (the Event Management SDR classes, attributes, and their values), see the description of the EMCDB in IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference.
The haemloadlist file consists of stanzas, each of which describes an object in one of the Event Management SDR classes. The stanzas may appear in any order in the file.
Each stanza begins with the SDR class name to which the object belongs, followed by one or more lines that define the attributes of the object. The SDR class name must begin in column 1.
An attribute line consists of the attribute name followed by an = (equal sign) followed by a data field. The = (equal sign) may not be surrounded by blank spaces. The data field consists of a single value for the attribute. If the field contains blanks, it must be surrounded by double quotes ("). If the field contains a double quote character, it should be preceded by a backslash (\"). If the field contains a backslash character, it should be preceded by a second backslash (\\).
The data field must be on the same line as the attribute. The attribute lines for a stanza stop at either the start of a new stanza or the end-of-file character.
Comments may be present in the file. Any line in which the first nonwhite space character is a pound sign (#) is a comment. Blank lines are also considered comment lines and are ignored.
The EM_Resource_Variable Stanza
The EM_Resource_Variable SDR class contains one object for each resource variable defined in the database. Accordingly, there is one EM_Resource_Variable stanza for each resource variable that is defined.
Here is an example of a stanza that defines a resource variable of type Quantity:
EM_Resource_Variable rvName="IBM.PSSP.aixos.PagSp.totalsize" rvDescription="99" rvValue_type="Quantity" rvData_type="long" rvInitial_value="0" rvClass="IBM.PSSP.aixos.PagSp" rvPTX_name="PagSp/totalsize" rvLocator="NodeNum" rvDynamic_instance="0"
Here is an example of a stanza that defines a resource variable of type State:
EM_Resource_Variable rvName="IBM.PSSP.Prog.pcount" rvDescription="1009" rvValue_type="State" rvData_type="SBS" rvClass="IBM.PSSP.Prog" rvLocator="NodeNum" rvDynamic_instance="1"
The EM_Structured_Byte_String Stanza
The EM_Structured_Byte_String SDR class contains one object for each structured field that is defined in a structured byte string. Accordingly, there is one EM_Structured_Byte_String stanza for each structured field that is defined.
The IBM.PSSP.Prog.pcount resource variable is an SBS that has three structured fields. Here is an example of the stanzas that define the fields:
EM_Structured_Byte_String sbsVariable_name="IBM.PSSP.Prog.pcount" sbsField_name="CurPIDCount" sbsField_type="long" sbsField_SN="0" EM_Structured_Byte_String sbsVariable_name="IBM.PSSP.Prog.pcount" sbsField_name="PrevPIDCount" sbsField_type="long" sbsField_SN="1" EM_Structured_Byte_String sbsVariable_name="IBM.PSSP.Prog.pcount" sbsField_name="CurPIDList" sbsField_type="cstring" sbsField_SN="2"
The EM_Instance_Vector Stanza
The EM_Instance_Vector SDR class contains one object for each instance vector element that is defined for a resource and all of its resource variables. Accordingly, there is one EM_Instance_Vector stanza for each instance vector element that is defined.
Here are some examples of stanzas that define instance vectors. The instance vector for the resource named IBM.PSSP.aixos.PagSp contains only one instance vector element. The instance vector for the resource named IBM.PSSP.Prog contains three elements.
EM_Instance_Vector ivResource_name="IBM.PSSP.aixos.PagSp" ivElement_name="NodeNum" ivElement_description="701"
EM_Instance_Vector ivResource_name="IBM.PSSP.Prog" ivElement_name="UserName" ivElement_description="701" EM_Instance_Vector ivResource_name="IBM.PSSP.Prog" ivElement_name="ProgName" ivElement_description="701" EM_Instance_Vector ivResource_name="IBM.PSSP.Prog" ivElement_name="NodeNum" ivElement_description="701"
The EM_Resource_Class Stanza
The EM_Resource_Class SDR class contains one object for each resource variable class that is defined in the database. Accordingly, there is one EM_Resource_Class stanza for each resource variable class that is defined.
Here are two examples of stanzas that define resource classes:
EM_Resource_Class rcClass="IBM.PSSP.aixos.PagSp" rcResource_monitor="aixos" rcObservation_interval="30" rcReporting_interval="0"
EM_Resource_Class rcClass="IBM.PSSP.Prog" rcResource_monitor="IBM.PSSP.harmpd" rcObservation_interval="0" rcReporting_interval="0"
The EM_Resource_Monitor Stanza
The EM_Resource_Monitor SDR class contains one object for each resource monitor that is defined in the database. Accordingly, there is one EM_Resource_Monitor stanza for each resource monitor that is defined.
Here are some examples of stanzas that define resource monitors:
EM_Resource_Monitor rmName="aixos" rmMessage_file="PEM.cat" rmMessage_set="3" rmConnect_type="internal"
EM_Resource_Monitor rmName="IBM.PSSP.harmpd" rmPath="/usr/lpp/ssp/bin/haemRM/harmpd" rmMessage_file="PEM.cat" rmMessage_set="2" rmConnect_type="server" rmPTX_prefix="0" rmPTX_description="0" rmPTX_asnno="0"
EM_Resource_Monitor rmName="IBM.PSSP.harmld" rmPath="/usr/lpp/ssp/bin/haemRM/harmld" rmMessage_file="harm_des.cat" rmMessage_set="1" rmConnect_type="server" rmPTX_prefix="IBM/PSSP.harmld" rmPTX_description="1,2" rmPTX_asnno="2"
Related Information
Commands: haemloadcfg, haemcfg
For a general overview of configuring Event Management, see "The Event Management Subsystem" chapter of IBM Parallel System Support Programs for AIX: Administration Guide.
For a description of the SDR classes and attributes that are related to the EMCDB, see IBM Parallel System Support Programs for AIX: Event Management Programming Guide and Reference.
Purpose
/spdata/sys1/spmon/hmacls - Defines the Access Control Lists (ACLs) used by the Hardware Monitor daemon.
Description
The /spdata/sys1/spmon/hmacls file contains permission specifications for users to execute the various Hardware Monitor operations. Each line in the file consists of three white-space separated tokens in the following format:
obj user_name permissions
The obj token is either a frame ID or the host name of the control workstation (known as the Monitor and Control Node (MACN) by the Hardware Monitor). The user_name token is the user's principal name or principal name and instance.
The permissions token specifies which operations the user specified by the user_name token can execute against the object specified by the obj token.
If the obj token is a frame ID, the permissions token is one or more characters taken from the set v, s, and m. A definition of each follows:
The VFOP permission implies the monitor permission. These permissions are described in the various commands. If the obj token is the MACN host name, the permissions token must be the character a. The character a is the administrative permission required to use the hmadm command.
Users are authorized to use the Hardware Monitor by virtue of having their names in the /spdata/sys1/spmon/hmacls file and have issued the kinit command with that name.
Files
Related Information
Commands: hmadm, hmcmds, hmmon, nodecond, s1term
Examples
The following is an example of a hmacls file:
workstn3.kgn.ibm.com john.admin a 1 john.admin vsm 2 john.admin vsm 3 john.admin vsm 1 mary m 2 mary m 3 mary m
Purpose
.klogin - Specifies remote principals that can use a local user account.
Description
The $HOME/.klogin file defines which users and services on any remote hosts (computers in a network) within an authentication realm are allowed to invoke commands on the local user account.
The format of the $HOME/.klogin file is:
principal.instance@realm
A typical .klogin file, if present, looks similar to the following:
harry@KGN.IBM.COM beverly.root@KGN.IBM.COM root.admin@KGN.IBM.COM user.wkst3@KGN.IBM.COM user1@KGN.IBM.COM rcmd.wkst3@KGN.IBM.COM
The principal name is the name of the remote user using SP authentication to execute remote commands on a local user account.
The instance is used to distinguish among variations on the principal name and may be used to indicate special privileges such as root authorization. In many environments, the instance is used with a service to indicate the workstation the service is running on. The service entries are usually found only in the root directory .klogin file.
The realm is the authentication realm. This may be different if there are several authentication realms, each with a different realm name, in your environment.
If the originating remote user is authenticated to one of the principals named in the .klogin file, access is granted to the account. The owner of the account is granted access if there is no .klogin file. If a .klogin file is present, the owner must also be listed in order to gain access to his or her account from a remote host.
For security reasons, any $HOME/.klogin file must be owned by either the local user or root, and only the owner should have read and write access.
Files
Prerequisite Information
Related Information
SP Commands: kinit, rcp, rsh
SP Daemon: kshd
Examples
The following examples assume both Jeff and Anna have principal names in the authentication database (anna@KGN.IBM.COM, jeff@KGN.IBM.COM) and have issued a kinit to be authenticated on their local host. In addition, there is one authentication realm signified by KGN.IBM.COM.
anna@KGN.IBM.COM jeff@KGN.IBM.COM
Anna's entry must be present in the .klogin file since the .klogin file exists. Jeff can now use the -l flag on the rsh or rcp to access Anna's account.
Purpose
Kerberos - Contains an introduction to SP authentication services.
Description
Kerberos authenticates individual users in a network environment. After authenticating your identity, you can use facilities such as Sysctl, the SP System Monitor, and the authenticated versions of network utilities rsh and rcp, without having to present passwords to remote hosts and without having to use .rhosts files.
Before you can use the SP authenticated commands, you must make sure that you are added to the authentication database. You can use the kinit command to find out. This command tries to authenticate your identity in the system. The kinit command prompts you for a principal name and password. Enter your principal name and password. If the command accepts your authentication information without sending you a message, you are already registered. If you enter your user name and kinit responds with the following message:
Kerberos principal unknowncontact your system administrator.
A principal identifier contains three parts. The first is the user or service name. The second is the instance, which in the case of a user is usually null. Some users can have privileged instances, however, such as root or admin. In the case of a service, the instance is the name of the machine on which it runs (for example, there can be a rcmd (kshd daemon) service running on the machine abc, which is different from the kshd service running on the machine xyz). The third part of the name is the realm. The realm corresponds to the service providing authentication for the principal. For example, computing resources within an enterprise can be partitioned into multiple administrative units for convenience or other business reasons.
When writing a principal identifier, the user or service name is separated from the instance (if not null) by a period, and the realm (if not the local realm) follows, preceded by an at sign (@).
When you authenticate your identity using the kinit command, you are given an initial ticket (which is an encrypted protocol message that provides authentication). This ticket is used to request other tickets from the authentication service for SP authenticated services such as SP rsh and SP rcp. The ticket transactions are done transparently, so you do not have to worry about their management.
Be aware that tickets expire. Some tickets, such as admin instance tickets, may expire in a few minutes, while other tickets may be good for hours, days, or weeks depending on the installation's policy. If your login session extends beyond the time limit, you have to reauthenticate your identity using the kinit command to get new tickets.
For more information about the kinit and kdestroy commands, see the kinit and kdestroy command.
Related Information
Commands: kadmin, kdestroy, kinit, klist, kpasswd
Purpose
/etc/krb.conf - Contains the SP authentication configuration file.
Description
The krb.conf file contains configuration information describing the local authentication realm and the location of authentication servers for known realms.
The first line of the krb.conf file contains the name of the local authentication realm. Additional lines specify the location of an authentication server for a realm. The krb.conf file must contain at least one entry for each realm used by the local system. Each line specifying the location of an authentication server must be of the form:
REALM_NAME host_name
or
REALM_NAME host_name admin server
When "admin server" follows the host name, it indicates that the hosts also provides an administrative database server.
Related Information
SP File: krb.realms
Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for additional Kerberos information.
Examples
The following example of an /etc/krb.conf shows a simple configuration consisting a single realm with two servers, the primary and one secondary:
EAST.COAST EAST.COAST master.authent.abc.com admin server EAST.COAST backup.authent.abc.com
Here, "admin server" identifies the system whose full host name is "master.authent.abc.com" as the primary server, responsible for administration of the master database. Note that, in this case, there would have to be information in the /etc/krb.realms file to map the two host names or the domain name authent.abc.com to the local realm name, "EAST.COAST". See the Example section of the krb.realms file.
Purpose
/etc/krb.realms - Specifies the translations from host names to authentication realms.
Description
The krb.realms file provides a translation from a host name or a network domain name to the authentication realm name for the services provided by that host. Each line of the translation file is in one of the following forms (domain names should begin with a period (.)):
host_name realm_name domain_name realm_name
If a host name exactly matches the host_name field in a line of the first form, the corresponding realm is the realm of the host. If a host name does not match any host_name in the file but its domain exactly matches a domain name, the corresponding realm is the realm of the host.
If no translation entry applies, the host's realm is considered to be the host name's domain portion converted to uppercase. If the host name does not contain a domain name, the host's realm is considered to be the host name converted to uppercase.
Related Information
SP File: krb.conf
Refer to the chapter on security in IBM Parallel System Support Programs for AIX: Administration Guide for additional Kerberos information.
Examples
The following example of an /etc/krb.realms shows the entries that could be used to map a host name or a domain name to a realm. These names correspond to those used in the krb.conf file example.
master.authent.abc.com EAST.COAST .authent.abc.com EAST.COAST
The first line maps a specific host name to the realm "EAST.COAST". If the host name were "master.east.coast", no entry would have been required. The second entry maps all host names whose domain portion is "authent.abc.com" to the same domain. The default mapping for this realm is:
.east.coast EAST.COAST
This type of mapping is always assumed, even if the /etc/krb.realms file is empty.
Purpose
SDR_dest_info - Provides connection addresses for System Data Repository (SDR) clients.
Description
The SDR_dest_info file provides connection addresses for System Data Repository (SDR) clients. It contains the following fields:
Related Information
Refer to the IBM Parallel System Support Programs for AIX: Administration Guide for additional information on the SDR_dest_info file.
Examples
This example shows the contents of an SDR_dest_info file for a node. The node is a member of system partition 129.40.127.46. The default system partition on this system is 129.40.127.47. The host name of the node's system partition is k99sp1. The host name of the default system partition is k99s.
default:129.40.127.47 primary:129.40.127.46 nameofprimary:k99sp1 nameofdefault:k99s
Purpose
sysctl.acl - Contains the Access Control List (ACL) for the Sysctl ACL authorization callback.
Description
The sysctl.acl file contains the list of users authorized to access objects which are assigned the ACL authorization callback. Each line of the file must conform to one of the following formats:
The first line of this (or any Sysctl ACL file) must begin with the line:
#acl#
Note: | Including a principal in the sysctl.acl file gives that principal the same level of authority as having the root password. |
Prerequisite Information
See the descriptions of SP authentication and Sysctl in IBM Parallel System Support Programs for AIX: Administration Guide.
Related Information
SP Command: sysctl
SP Daemon: sysctld
Examples
The following sysctl.acl file defines two principals and includes the contents of another ACL file:
#acl# # # This is the Sysctl ACL file for the system. It defines two # principals: "root.@HPSSL.KGN.IBM.COM" and "rcr.@HPSSL.KGN.IBM.COM" # It also includes the contents of file /etc/my.acl.file # _PRINCIPAL root.@HPSSL.KGN.IBM.COM _PRINCIPAL rcr.@HPSSL.KGN.IBM.COM _ACL_FILE /etc/my.acl.file
Purpose
sysctl.conf - Configures the Sysctl server (sysctld) running on the local SP node.
Description
The sysctl.conf file configures the local Sysctl server daemon by optionally creating variables, procedures and classes, setting variables, loading shared libraries, and executing Sysctl commands. These items are used by the server in the processing of requests from local and remote Sysctl clients.
The default location of this file is /etc/sysctl.conf. An alternate configuration file location can be specified by using the -f flag when starting the server (see the sysctld command).
The /etc/sysctl.conf file contains Sysctl commands which are read and executed by the Sysctl server during its initialization. The following commands are supported:
create var buildTop /usr/lpp/ssp AUTH
creates the variable buildTop, assigns it a value of /usr/lpp/ssp, and assigns it an authorization callback of AUTH. The variable buildTop can be referenced within Sysctl commands and procedures.
Another example:
create var STARTTIME [exec /bin/date] NONE
creates the variable STARTTIME, assigns it the value returned from the execution of the /bin/date command at server initialization, and assigns it an authorization callback of NONE. This variable contains the date and time at which the server was started on the node.
create proc mydate {} AUTH {exec /bin/date}
creates the procedure mydate which has no parameters. This procedure has an authorization callback of AUTH. The procedure is comprised of a single statement (exec /bin/date).
create class sys $buildTop/samples/sysctl/sys.cmds
creates the class sys. If $buildTop is defined as /usr/lpp/ssp, the file /usr/lpp/ssp/samples/sysctl/sys.cmds contains the definition of the sys class.
include $buildTop/samples/sysctl/pdfpfck.cmds
causes the Sysctl server to read the contents of the specified file at initialization time.
set LOG /var/sysctld.log_file
The values assigned to the ACL, LOG, and KEY variables are overridden by the optional command line arguments -a, -l, and -k.
Environment variables (such as the default PATH) can also be set within the configuration file by assigning values to the env() array. For example:
set env(PATH) /usr/bin:/bin:/usr/etc:/etc
sets the PATH for the Sysctl server. The env() array is assigned an authorization callback of SYSTEM which prevents its modification from outside the Sysctl server by a request sent by a Sysctl client.
library name init function ------------ ------------- libxxx.sl xxx_Init() libxxx.a xxx_Init()
The Sysctl server exports an API which the library uses to define commands, variables, authorization callbacks, and interpreter deletion callbacks. See the load() help page for details.
setauth -cmd svcconnect NONE
Prerequisite Information
See the description of Sysctl in IBM Parallel System Support Programs for AIX: Administration Guide.
Related Information
SP Command: sysctl
SP Daemon: sysctld
Purpose
tuning.commercial - Contains initial performance tuning parameters for a typical commercial SP environment.
Description
This file is a Korn shell script file containing commands to set network performance tuning parameters. It can be copied to the /tftpboot/tuning.cust file on the control workstation for propagation to the nodes.
Related Information
SP Commands: cptuning
AIX Commands: This file contains invocations of the no command.
SP Files: tuning.default, tuning.development, tuning.scientific
IBM Parallel System Support Programs for AIX: Installation and Migration Guide
Purpose
tuning.default - Contains initial (default) performance tuning parameters for an SP environment.
Description
This file is a Korn shell script file containing commands to set network performance tuning parameters. In the absence of explicit administrator action, this file is copied to the /tftpboot/tuning.cust file on the control workstation for propagation to the nodes.
Related Information
SP Commands: cptuning
AIX Commands: This file contains invocations of the no command.
SP Files: tuning.commercial, tuning.development, tuning.scientific
IBM Parallel System Support Programs for AIX: Installation and Migration Guide
Purpose
tuning.development - Contains initial performance tuning parameters for a typical software development/interactive SP environment.
Description
This file is a Korn shell script file containing commands to set network performance tuning parameters. It can be copied to the /tftpboot/tuning.cust file on the control workstation for propagation to the nodes.
Related Information
SP Commands: cptuning
AIX Commands: This file contains invocations of the no command.
SP Files: tuning.commercial, tuning.default, tuning.scientific
IBM Parallel System Support Programs for AIX: Installation and Migration Guide
Purpose
tuning.scientific - Contains initial performance tuning parameters for a typical engineering or scientific SP environment.
Description
This file is a Korn shell script file containing commands to set network performance tuning parameters. It can be copied to the /tftpboot/tuning.cust file on the control workstation for propagation to the nodes.
Related Information
Commands: cptuning
AIX Commands: This file contains invocations of the no command.
SP Files: tuning.commercial, tuning.default, tuning.development
IBM Parallel System Support Programs for AIX: Installation and Migration Guide
Purpose
getvhostname - Gets the virtual host name.
Library
Availability Library (libavail.a)
Syntax
#include <vhost.h> int getvhostname (name, name_length); char *name; int name_length;
Description
Use this subroutine to retrieve the virtual host name of a host machine. This routine is similar to the gethostname system call with the exception that it retrieves the virtual host name from the /etc/vhostname file instead of using a kernel variable. The getvhostname subroutine is a library call and gethostname is a system call.
The name is retrieved from the /etc/vhostname file. If the file does not exist, the gethostname system call is used and the real host name is returned. If excess space is provided, the returned name parameter is null terminated. If insufficient space is provided, the returned name parameter is truncated to fit in the given space. Virtual host names are limited to MAX_VHOSTNAME_LEN bytes (255), not including the terminating null character. The MAX_VHOSTNAME_LEN macro is defined in the vhost.h header file. In order to guarantee sufficient buffer space to hold the virtual host name, the name_length parameter should be MAX_VHOSTNAME_LEN + 1 or 256.
To clear the virtual host name so that the virtual host name no longer exists, remove the /etc/vhostname file.
Note: | You must have root authority to remove the /etc/vhostname file. |
The virtual host name is used in fail over situations when an application has associated the host name in the kernel of a particular machine to the service it is providing. When the application is restarted on the fail over node that has a different host name, the application may fail or act incorrectly. If the application needs to associate a host name with a particular service and it cannot handle having multiple host names, a virtual host name can be provided. The application can call getvhostname instead of gethostname and get the host name of the node it normally runs on. This eliminates the need to change the real host name in the kernel on the backup node. It should be noted that changing the real host name in the kernel can cause problems with other applications that rely on the real host name to identify the physical machine.
Note: | The High Availability Cluster Multiprocessing (HACMP) event scripts supplied with the High Availability Control Workstation (HACWS) option of the IBM Parallel System Support Programs for AIX (PSSP), set and clear the virtual host name in the supplied HACMP pre- and post-event scripts. The administrator should not normally have to set the virtual host name. |
Return Values
Upon successful completion, the getvhostname subroutine returns a value of 0. Otherwise, a value of -1 is returned, the global variable errno is set to identify the error, and the contents of the buffer pointed to by the name parameter are indeterminate.
Error Values
The getvhostname subroutine is unsuccessful if the following error occurs:
If one of the system calls used to retrieve the virtual host name from the /etc/vhostname file fails (for example, open or read), errno is set by the system call that failed.
Examples
rm /etc/vhostname
Note: | You must have root authority to remove the /etc/vhostname file. |
#include <vhost.h> main ( ) { char name [MAX_VHOSTNAME_LEN + 1]; getvhostname (name, (MAX_VHOSTNAME_LEN + 1)); }
Related Information
Commands: vhostname
Subroutines: setvhostname
AIX Commands: hostname
AIX Subroutines: gethostname, sethostname
Header Files: vhost.h
Purpose
hacws_set - Sets the state of the control workstation.
Library
Availability Library (libavail.a)
Location
Syntax
#include <hacws.h> int hacws_set (state); int state;
Parameters
Description
Use this subroutine to set the current state of the control workstation. It is valid only when issued on the control workstation. When the subroutine is called and the calling process is not on a control workstation, an error occurs.
Note: | The High Availability Cluster Multiprocessing (HACMP) event scripts and installation scripts supplied with the High Availability Control Workstation (HACWS) option of the IBM Parallel System Support Programs for AIX (PSSP), set the control workstation state. The state is changed during fail over or reintegration in the HACWS supplied pre- and post-event scripts for HACMP. The administrator should not normally have to set the control workstation state. |
Return Values
Upon successful completion, the hacws_set subroutine returns a value of 0. Otherwise, a value of -1 is returned and the global variable errno is set to identify the error.
Error Values
The hacws_set subroutine is unsuccessful if any of the following errors occur:
If one of the system calls used to store the HACWS state value into the /etc/hacws.state file fails (for example, open, write, rename), errno is set by the system call that failed.
Macros
The /usr/include/hacws.h header file defines the following macros as valid input values for the hacws_set subroutine:
Prerequisite Information
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for information on the HACWS option.
Related Information
Commands: lshacws, sethacws
Subroutines: hacws_set
Header Files: hacws.h
Purpose
hacws_stat - Gets the state of the control workstation.
Library
Availability Library (libavail.a)
Location
Syntax
#include <hacws.h> int hacws_stat (void);
Description
Use this subroutine to return the current state of the control workstation. It returns an integer that indicates the state of the primary or backup control workstation and specifies whether the control workstation is a high availability configuration. This subroutine is valid only when issued on the control workstation. When the subroutine is called and the calling process is not on a control workstation, an error occurs.
Note: | The High Availability Cluster Multiprocessing (HACMP) event scripts and installation scripts supplied with the High Availability Control Workstation (HACWS) option of the IBM Parallel System Support Programs for AIX (PSSP), set the control workstation state. The state is changed during fail over or reintegration in the HACWS supplied pre- and post-event scripts for HACMP. The administrator should not normally have to set the control workstation state. |
Return Values
Upon successful completion, the hacws_stat subroutine returns a nonnegative value. If the hacws_stat subroutine is unsuccessful, a value of -1 is returned and the global variable errno is set to identify the error.
Error Values
The hacws_stat subroutine is unsuccessful if any of the following errors occur:
If one of the system calls used to retrieve the HACWS state value from the /etc/hacws.state file fails (for example, open or read), errno is set by the system call that failed.
Macros
The /usr/include/hacws.h header file defines the following macros for the nonnegative return values for the hacws_stat subroutine:
Prerequisite Information
Refer to IBM Parallel System Support Programs for AIX: Administration Guide for information on the HACWS option.
Related Information
Commands: lshacws, sethacws
Subroutines: hacws_set
Header Files: hacws.h
Purpose
LAPI_Address - Gets an unsigned integer value for a specified address.
C Syntax
#include <lapi.h> int LAPI_Address(my_addr, ret_addr) void *my_addr; uint *ret_addr;
FORTRAN Syntax
include 'lapif.h' LAPI_ADDRESS(my_addr, ret_addr, ierror) INTEGER my_addr; INTEGER ret_addr; INTEGER ierror;
Parameters
Description
Use this subroutine in FORTRAN programs when specified addresses need to be stored in an array. In FORTRAN, the concept of address ('&') does not exist as it does in C. This function gives that ability to FORTRAN.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Purpose
LAPI_Address_init - Exchanges virtual addresses for non-Single Program Multiple Data (SPMD) programs and for dynamically allocated data.
C Syntax
#include <lapi.h> int LAPI_Address_init(hndl, my_addr, add_tab) lapi_handle_t hndl; void *my_addr; void *add_tab[];
FORTRAN Syntax
include 'lapif.h' LAPI_ADDRESS_INIT(hndl, my_addr, add_tab, ierror) INTEGER hndl; <type> my_addr(*); INTEGER add_tab(*); INTEGER ierror;
Parameters
Description
Use this subroutine to exchange virtual and dynamically allocated addresses. add_tab is an array of pointers of size >= LAPI_Qenv(,NUM_TASKS,). This function is a collective call over the LAPI context hndl which fills the table add_tab with the entries supplied by each task. Upon completion of this call, add_tab[i] will contain the entry provided by task i.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Purpose
LAPI_Amsend - Invokes a user-provided Active Message (AM) handler to run on a remote (target) process.
C Syntax
#include <lapi.h> typedef void (compl_hndlr_t)(hndl, user_info); lapi_handle_t *hndl; LAPI context passed in from LAPI_Amsend. void * user_info; Buffer (user_info) pointer passed in from header handler (void * (hnd_hndlr_t)). typedef void * (hdr_hndlr_t)(hndl, uhdr, uhdr_len, msg_len, comp_h, user_info); lapi_handle_t *hndl; LAPI context passed in from LAPI_Amsend. void * uhdr; uhdr passed in from LAPI_AMsend. uint *uhdr_len; uhdr_len passed in from LAPI_Amsend. uint * msg_len; udata_len passed in from LAPI_Amsend. compl_hndlr_t ** comp_h; Function address of completion handler (void (compl_hndlr_t)) that needs to be filled out by this header handler function. void ** user_info; Buffer pointer (user_info) that is provided by this head handler function to pass to the completion handler. int LAPI_Amsend(hndl, tgt, hdr_hdl, uhdr, uhdr_len, udata, udata_len, tgt_cntr, org_cntr, cmpl_cntr) lapi_handle_t hndl; uint tgt; void *hdr_hdl; void *uhdr; uint uhdr_len; void *udata; uint udata_len; lapi_cntr_t *tgt_cntr; lapi_cntr_t *org_cntr; lapi_cntr_t *cmpl_cntr;
FORTRAN Syntax
include 'lapif.h' COMPL_H(hndl, user_info); INTEGER hndl; INTEGER user_info; INTEGER FUNCTION HDR_HDL(hndl, uhdr, uhdr_len, msg_len, comp_h, user_info) INTEGER hndl; INTEGER uhdr; INTEGER uhdr_len; INTEGER msg_len; INTEGER comp_h; INTEGER user_info; LAPI_AMSEND(hndl, tgt, hdr_hdl, uhdr, uhdr_len, udata, udata_len, tgt_cntr org_cntr, cmpl_cntr, ierror) INTEGER hndl; INTEGER tgt; <type> hdr_hdl(*); INTEGER uhdr; INTEGER uhdr_len; INTEGER udata; INTEGER udata_len; <type> tgt_cntr(*); INTEGER org_cntr; INTEGER cmpl_cntr; INTEGER ierror;
Parameters
Description
Use this subroutine to transfer the hdr_hdl function pointer along with the contents of uhdr and udata from the origin to the tgt target process. When the message arrives at the target process, the hdr_hdl header handler is invoked at the tgt target with the pointer to uhdr as one of the parameters.
The user-supplied header handler is expected to return a buffer pointer (user_info as the return value) in which udata is to be copied. The header handler is also expected to save any information that will be required later by the completion handler. The header handler returns (through reference parameters) the completion handler and a pointer to the saved information (user_info).
Note: | The header handler should be nonblocking because no progress on the messages associated with hndl can be made until control is returned to the communications library from the header handler. |
After the header handler returns, the udata (if any) is copied into the user-specified buffer (user_info). When all of the udata is copied into the user buffer, the completion handler you specified through the header handler is enqueued.
After the parameters (including the contents of uhdr and udata) are copied out of the memory at the origin, the org_cntr is incremented. After the completion handler finishes running at the tgt, the tgt_cntr is incremented. If the completion handler specified is NULL, the tgt_cntr is incremented after all of the udata is copied into the user-specified buffers. If the user-specified buffer is NULL and the completion handler is also NULL, the tgt_cntr will be incremented in some implementation-specific manner. Either counter addresses may be NULL.
This is a nonblocking call. The calling process cannot change the uhdr origin header and udata data until completion at the origin is signaled by the org_cntr being incremented. Similarly, you can assume that the specified AM handler has run at the tgt only after the tgt_cntr target counter is incremented. The cmpl_cntr and tgt_cntr counters are incremented after the AM handler has completed execution at the target. When the AM handler has both a hdr_hdl header handler and a comp_h completion handler, the cmpl_cntr and tgt_cntr counters are incremented after the completion handler has completed execution. If the AM handler has only a hdr_hdl header handler, the cmpl_cntr and tgt_cntr counters will be incremented after the header handler has completed execution. This call can be made synchronous if the origin waits for the cmpl_cntr update to complete.
The length (uhdr_len) of the user-specified header is constrained by an implementation specified maximum value (LAPI_Qenv(,MAX_UHDR_SZ,)). In the current implementation, the amount of udata sent per packet is LAPI_Qenv(,MAX_UHDR_SZ,) - uhdr_len. To get the best bandwidth, uhdr_len should be as small as possible.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Fence, LAPI_Getcntr, LAPI_Qenv, LAPI_Waitcntr
Purpose
LAPI_Fence - Enforces order on Low-Level Application Programming Interface (LAPI) calls.
C Syntax
#include <lapi.h> int LAPI_Fence(hndl) lapi_handle_t hndl;
FORTRAN Syntax
include 'lapif.h' LAPI_FENCE(hndl, ierror) INTEGER hndl; INTEGER ierror;
Parameters
Description
Use this subroutine to enforce order on LAPI calls. If a process calls LAPI_Fence(), all the LAPI operations that were initiated by that process, before the fence using the LAPI context hndl, are guaranteed to complete at the target processes. This occurs before any of its communication operations using hndl, initiated after the fence, complete at the target processes. This is a data fence which means that the data movement is complete. This is not an operation fence which would need to include Active Message completion handlers completing on the target.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Amsend, LAPI_Get, LAPI_Gfence, LAPI_Put, LAPI_Rmw
Purpose
LAPI_Get - Copies data from a remote process to the local address on a local process.
C Syntax
#include <lapi.h> int LAPI_Get(hndl, tgt, len, tgt_addr, org_addr, tgt_cntr, org_cntr) lapi_handle_t hndl; uint tgt; uint len; void *tgt_addr; void *org_addr; lapi_cntr_t *tgt_cntr; lapi_cntr_t *org_cntr;
FORTRAN Syntax
include 'lapif.h' LAPI_GET(hndl, tgt, len, tgt_addr, org_addr, tgt_cntr, org_cntr, ierror) INTEGER hndl; INTEGER tgt; INTEGER len; <type> tgt_addr(*); INTEGER org_addr; <type> tgt_cntr(*); INTEGER org_cntr; INTEGER ierror;
Parameters
Description
Use this subroutine to transfer the len number of bytes from the tgt_addr address at the target process to the org_addr virtual address at the origin process over the port identified by hndl. After the data is copied out of the memory at the tgt_addr, the tgt_cntr is incremented. After the data arrives at the origin, the org_cntr is incremented. If either counter address is NULL, the data transfer occurs, but the corresponding counter increment does not take place.
This is a nonblocking call in that the calling program cannot assume that the target buffer can be changed, nor that the contents of the memory pointed to by the org_addr on the origin is ready for use. However, after the origin waits for the org_cntr update to complete, the origin can use the org_addr data. Similarly, the target can reuse the target buffer tgt_addr only after it has waited for the tgt_cntr update to complete at the target.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Fence, LAPI_Getcntr, LAPI_Gfence, LAPI_Put, LAPI_Qenv, LAPI_Waitcntr
Purpose
LAPI_Getcntr - Gets the integer value of the counter.
C Syntax
#include <lapi.h> int LAPI_Getcntr(hndl, cntr, val) lapi_handle_t hndl; lapi_cntr_t *cntr; int *val;
FORTRAN Syntax
include 'lapif.h' LAPI_GETCNTR(hndl, cntr, val, ierror) INTEGER hndl; INTEGER cntr; INTEGER val; INTEGER ierror;
Parameters
Description
Use this subroutine to get the integer value of cntr. It can be used to find how much progress is being made in LAPI context hndl. In conjunction, LAPI_Probe() can be used to make progress in LAPI context hndl if LAPI_Getcntr() is called inside a loop.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Amsend, LAPI_Get, LAPI_Probe, LAPI_Put, LAPI_Setcntr, LAPI_Waitcntr
Purpose
LAPI_Gfence - Enforces order on Low-Level Application Programming Interface (LAPI) calls on all processes.
C Syntax
#include <lapi.h> int LAPI_Gfence(hndl) lapi_handle_t hndl;
FORTRAN Syntax
include 'lapif.h' LAPI_GFENCE(hndl, ierror) INTEGER hndl; INTEGER ierror;
Parameters
Description
This is a collective call. On completion of this call, it is assumed that all LAPI communications associated with hndl from all processes has quiesced.
Note: | Although hndl is a local variable, it has a set of nodes that were associated with it at LAPI_Init all of which have to participate in this operation in order for it to complete. |
This is a data fence which means that the data movement is complete. This is not an operation fence which would need to include Active Message completion handlers completing on the target.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Fence
Purpose
LAPI_Init - Initializes the Low-Level Application Programming Interface (LAPI) subsystem.
C Syntax
#include <lapi.h> int LAPI_Init(hndl, lapi_info) lapi_handle_t *hndl; lapi_info_t *lapi_info;
FORTRAN Syntax
include 'lapif.h' LAPI_INIT(hndl, lapi_info, ierror) INTEGER hndl; INTEGER lapi_info(10); INTEGER ierror;
Parameters
Description
Use this subroutine to instantiate a new context of the LAPI subsystem and to initialize it. A handle to the newly-created LAPI context is returned in hndl. All subsequent LAPI calls can use hndl to specify the context of the LAPI operation. The lapi_info structure (lapi_info_t) needs to be filled in:
typedef struct { /* Not in use currently */ lapi_dev_t protocol; /* OUT - Which protocol is initialized */ int info2 /* Future support */ int info3; /* Future support */ int info4; /* Future support */ int info5; /* Future support */ int info6; /* Future support */ LAPI_err_hndlr *err_hndlr; /* IN - User registered error handler */ void *info_info2; /* Future support */ void *info_info3; /* Future support */ void *info_info4; /* Future support */ } lapi_info_t;
lapi_dev_t is defined as follows:
typedef enum {NULL_DEV=0, TB2_DEV, TB3_DEV, UDP_DEV, VIRTUAL_DEV, LAST_DEV} lapi_dev_t;
Note: | Only the TB3_DEV lapi_dev_t type is supported at this time. |
You can register an error handler through the lapi_info structure.
To create a function, you need the following parameters:
void (User func name) (lapi_handle_t *hndl, /* LAPI handle */ int *error_code, /* Error code */ lapi_err_t *err_type, /* GET/PUT/RMW/AM/ INTERNAL */ int *task_id, /* Current node */ int *src); /* Source node */
The error code (*error_code) of LAPI_ERR_TIMEOUT is a recoverable error if you choose to ignore it in your error handler. All other error codes are currently terminal and you should do clean-up processing of your process and terminate the process (for example, exit()).
An error occurs if any LAPI calls are made before calling LAPI_Init(), except for LAPI_Address() and LAPI_Msg_string().
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Term
Purpose
LAPI_Msg_string - Gets the Low-Level Application Programming Interface (LAPI) and system message string.
C Syntax
#include <lapi.h> LAPI_Msg_string(error_code, buf) int error_code; void * buf;
FORTRAN Syntax
include 'lapif.h' LAPI_MSG_STRING(error_code, buf, ierror) INTEGER error_code; INTEGER buf(40); INTEGER ierror;
Parameters
Description
Use this subroutine to return the message string representation of the return value for a specific LAPI call.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Init, LAPI_Term
Purpose
LAPI_Probe - Transfers control to the communications subsystem to check for arriving messages and to make progress in polling mode.
C Syntax
#include <lapi.h> int LAPI_Probe(hndl) lapi_handle_t hndl;
FORTRAN Syntax
include 'lapif.h' LAPI_PROBE(hndl, ierror) INTEGER hndl; INTEGER ierror;
Parameters
Description
Use this subroutine to transfer control to the communications subsystem in order to make progress on messages associated with the context hndl.
Note: | There is no guarantee about receipt of messages on the return from this function. |
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Getcntr, LAPI_Setcntr, LAPI_Waitcntr
Purpose
LAPI_Put - Puts data into the target address on a target process.
C Syntax
#include <lapi.h> int LAPI_Put(hndl, tgt, len, tgt_addr, org_addr, tgt_cntr, org_cntr, cmpl_cntr) lapi_handle_t hndl; uint tgt; uint len; void *tgt_addr; void *org_addr; lapi_cntr_t *tgt_cntr; lapi_cntr_t *org_cntr; lapi_cntr_t *cmpl_cntr;
FORTRAN Syntax
include 'lapif.h' int LAPI_PUT(hndl, tgt, len, tgt_addr, org_addr, tgt_cntr, org_cntr, cmpl_cntr, ierror) INTEGER hndl; INTEGER tgt; INTEGER len; <type> tgt_addr(*); INTEGER org_addr; <type> tgt_cntr(*); INTEGER org_cntr; INTEGER cmpl_cntr; INTEGER ierror;
Parameters
Description
Use this subroutine to transfer the len number of bytes from the org_addr virtual address on the origin to the tgt target process at the tgt_address address over the port identified by hndl. After the data is copied out of the memory at org_addr, the org_cntr is incremented. After the data arrives at the tgt, the tgt_cntr is incremented. If either counter address is NULL, the data transfer occurs, but the corresponding counter increment does not take place.
This is a nonblocking call in that the calling program cannot assume that the origin buffer can be changed, nor that the contents of the memory pointed to by tgt_addr on tgt is ready for use. However, after the origin waits for the org_cntr update to complete, the origin can modify the org_addr origin buffer. Similarly, the target can modify the data in the tgt_addr target buffer after it has waited for the tgt_cntr update to complete on the target. This call can be made synchronous if the origin waits for the cmpl_cntr update to complete.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Fence, LAPI_Get, LAPI_Getcntr, LAPI_Gfence, LAPI_Qenv, LAPI_Waitcntr
Purpose
LAPI_Qenv - Queries the Low-Level Application Programming Interface (LAPI) interface for parallel job information.
C Syntax
#include <lapi.h> int LAPI_Qenv(hndl, query, ret_val) lapi_handle_t hndl; lapi_query_t query; int *ret_val;
FORTRAN Syntax
include 'lapif.h' LAPI_QENV(hndl, query, ret_val, ierror) INTEGER hndl; INTEGER query; INTEGER ret_val; INTEGER ierror;
Parameters
Description
Use this subroutine to query the LAPI interface for information about a specific LAPI instance. lapi_query_t defines the types of LAPI queries available.
typedef enum { TASK_ID=0, /* Query task id of current task in job */ NUM_TASKS, /* Query number of tasks in job */ MAX_UHDR_SZ, /* Query max. user header size for AM */ MAX_DATA_SZ, /* Query max. data length that can be sent */ ERROR_CHK, /* Query & Set parameter checking on(1)/off(0) */ TIMEOUT, /* Query & Set current comm. timeout setting in seconds */ MIN_TIMEOUT, /* Query minimum comm. timeout setting */ MAX_TIMEOUT, /* Query maximum comm. timeout setting */ INTERRUPT_SET, /* Query & Set interrupt on(1)/off(0) */ MAX_PORTS, /* Query max. available comm. ports */ MAX_PKT_SZ, /* This is the payload size of 1 packet */ NUM_REX_BUFS, /* Number of retransmission buffers */ REX_BUF_SZ, /* Size of Each retransmission buffer in bytes */ LAST_QUERY} lapi_query_t;
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Amsend, LAPI_Get, LAPI_Put, LAPI_Senv
Purpose
LAPI_Rmw - Provides the synchronization primitives.
C Syntax
#include <lapi.h> int LAPI_Rmw(hndl, op, tgt, tgt_var, in_val, prev_tgt_val, org_cntr) lapi_handle_t hndl; RMW_ops_t op; uint tgt; int *tgt_var; int *in_val; int *prev_tgt_val; lapi_cntr_t *org_cntr;
FORTRAN Syntax
include 'lapif.h' LAPI_RMW(hndl, op, tgt, tgt_var, in_val, prev_tgt_val, org_cntr, ierror) INTEGER hndl; INTEGER op; INTEGER tgt; <type> tgt_var(*); INTEGER in_val; INTEGER prev_tgt_val; INTEGER org_cntr; INTEGER ierror;
Parameters
Description
Use this subroutine to synchronize two independent operations, such as two processes sharing a common data structure. The operation is performed at the tgt target process and is atomic. The operation takes an in_val from the origin and performs one of four selected op operations on a tgt_var variable at the tgt target, and then replaces the tgt_var target variable with the results of the op operation. The prev_tgt_val original value of the tgt_var target variable is returned to the origin.
The valid operations for op are:
The operations are performed over the context referred to by hndl. The outcome of the execution of these calls is as if the following code was executed atomically:
*prev_tgt_val = *tgt_var; *tgt_var = f(*tgt_var, *in_val);
where:
f(a,b) = a + b for FETCH_AND_ADD
f(a,b) = a | b for FETCH_AND_OR (bitwise or)
f(a,b) = b for SWAP
For COMPARE_AND_SWAP, in_val is treated as a pointer to an array of two integers, and the op is the following atomic operation:
if(*tgt_var == in_val[0]) { *prev_tgt_val = TRUE; *tgt_var = in_val[1]; } else { *prev_tgt_val = FALSE; }
All the calls are nonblocking. To test for completion, use the LAPI_Getcntr and LAPI_Waitcntr functions. There is no tgt_cntr on RMW calls and they do not provide any indication of completion on the tgt process.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Getcntr, LAPI_Probe, LAPI_Qenv, LAPI_Setcntr, LAPI_Waitcntr
Purpose
LAPI_Senv - Sets the Low-level Application Programming Interface (LAPI) environment for the specified context.
C Syntax
#include <lapi.h> int LAPI_Senv(hndl, query, set_val) lapi_handle_t hndl; lapi_query_t query; int set_val;
FORTRAN Syntax
include 'lapif.h' LAPI_SENV(hndl, query, set_val, ierror) INTEGER hndl; INTEGER query; INTEGER set_val; INTEGER ierror;
Parameters
Description
Use this subroutine to set the LAPI environment for a specific LAPI instance. lapi_query_t defines the types of LAPI set environment variables.
typedef enum { ... ERROR_CHK, /* Query & Set parameter checking on(1)/off(0) */ TIMEOUT, /* Query & Set current communications timeout setting in seconds */ INTERRUPT_SET, /* Query & Set interrupt on(1)/off(0) */ ... } lapi_query_t;
To obtain the default values of the settings, use the LAPI_Qenv function.
Note: | If ERROR_CHK is set to 0 for all LAPI calls, parameter error checking is ignored (for example, LAPI_ERR_BAD_PARAMETER is not returned). |
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Qenv
Purpose
LAPI_Setcntr - Sets a counter to a specified value.
C Syntax
#include <lapi.h> int LAPI_Setcntr(hndl, cntr, val) lapi_handle_t hndl; lapi_cntr_t *cntr; int val;
FORTRAN Syntax
include 'lapif.h' LAPI_SETCNTR(hndl, cntr, val, ierror) INTEGER hndl; INTEGER cntr; INTEGER val; INTEGER ierror;
Parameters
Description
Use this subroutine to set the cntr to the appropriate value. The LAPI context associated with hndl may or may not be polled for incoming messages.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Getcntr, LAPI_Probe, LAPI_Waitcntr
Purpose
LAPI_Term - Terminates and cleans up the Low-Level Application Programming Interface (LAPI) subsystem.
C Syntax
#include <lapi.h> int LAPI_Term(hndl) lapi_handle_t hndl;
FORTRAN Syntax
include 'lapif.h' LAPI_TERM(hndl, ierror) INTEGER hndl; INTEGER ierror;
Parameters
Description
Use this subroutine to terminate the LAPI context specified by hndl. Any LAPI notification threads associated with this context are terminated. An error occurs when any LAPI calls are made using hndl after LAPI_Term() is called, except for LAPI_Msg_string() and LAPI_Address().
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Init
Purpose
LAPI_Waitcntr - Waits until a specified counter reaches the value specified.
C Syntax
#include <lapi.h> int LAPI_Waitcntr(hndl, cntr, val, cur_cntr_val) lapi_handle_t hndl; lapi_cntr_t *cntr; int val; int *cur_cntr_val;
FORTRAN Syntax
include 'lapif.h' LAPI_WAITCNTR(hndl, cntr, val, cur_cntr_val, ierror) INTEGER hndl; INTEGER cntr; INTEGER val; INTEGER cur_cntr_val; INTEGER ierror;
Parameters
Description
Use this subroutine to wait until the cntr reaches or exceeds the specified val. Once the cntr reaches the val, the cntr is decremented by that value. (We say decremented rather than set to zero since the cntr could have had a value greater than the specified val when the call was made.) This call may or may not check for message arrivals over the LAPI context hndl; LAPI_Probe will always check for message arrivals.
Return Values
The following is returned when an error occurs:
Prerequisite Information
Refer to the "Understanding and Using the Communications Low-Level Application Programming Interface" chapter in IBM Parallel System Support Programs for AIX: Administration Guide for additional LAPI information.
Related Information
Subroutines: LAPI_Amsend, LAPI_Get, LAPI_Getcntr, LAPI_Probe, LAPI_Put, LAPI_Rmw, LAPI_Setcntr
Purpose
setvhostname - Sets the virtual host name.
Library
Availability Library (libavail.a)
Syntax
#include <vhost.h>
int setvhostname (name, name_length); char *name; int name_length;
Parameters
Description
Use this subroutine to set the virtual host name of a host machine. Only programs with a root user ID can use this subroutine. This routine is similar to the sethostname system call with the exception that it stores the virtual host name in the /etc/vhostname file instead of using a kernel variable. The setvhostname subroutine is a library call and sethostname is a system call.
The name is stored in the /etc/vhostname file. If the file does not exist, it will be created. If it does exist, the file contents will be overwritten by the new virtual host name. Virtual host names are limited to MAX_VHOSTNAME_LEN bytes (255), not including the terminating null character. The MAX_VHOSTNAME_LEN macro is defined in the vhost.h header file. The name_length parameter does not have to allow for the terminating null character, therefore, the largest allowable value for name_length is MAX_VHOSTNAME_LEN.
To clear the virtual host name so that the virtual host name no longer exists, remove the /etc/vhostname file.
Note: | You must have root authority to remove the /etc/vhostname file. |
The virtual host name is used in fail over situations when an application has associated the host name in the kernel of a particular machine to the service it is providing. When the application is restarted on the fail over node that has a different host name, the application may fail or act incorrectly. If the application needs to associate a host name with a particular service and it cannot handle having multiple host names, a virtual host name can be provided. The application can call getvhostname instead of gethostname and get the host name of the node it normally runs on. This eliminates the need to change the real host name in the kernel on the backup node. It should be noted that changing the real host name in the kernel can cause problems with other applications that rely on the real host name to identify the physical machine.
Note: | The High Availability Cluster Multiprocessing (HACMP) event scripts supplied with the High Availability Control Workstation (HACWS) option of the IBM Parallel System Support Programs for AIX (PSSP), set and clear the virtual host name in the HACMP pre- and post-event scripts. The administrator should not normally have to set the virtual host name. |
Return Values
Upon successful completion, the setvhostname subroutine returns a value of 0. Otherwise, a value of -1 is returned and the global variable errno is set to identify the error.
Error Values
The setvhostname subroutine is unsuccessful if the following error occurs:
If one of the system calls used to store the virtual host name into the /etc/vhostname file fails (for example, open, write, rename), errno is set by the system call that failed.
Examples
rm /etc/vhostname
Note: | You must have root authority to remove the /etc/vhostname file. |
#include <string.h> #include <vhost.h> main ( ) { char name[]='spcw_prim'; setvhostname(name, strlen(name)); }
Related Information
Commands: vhostname
Subroutines: getvhostname
AIX Commands: hostname
AIX Subroutines: gethostname, sethostname
Purpose
swclockGetIncrement - Returns the hertz frequency at which the switch clock operates.
Library
Switch Clock Library (libswclock.a)
Location
/usr/lib/libswclock.a
Syntax
#include <swclock.h> int swclockGetIncrement(swclock_handle_t swclock_handle);
Parameters
Description
Use this thread-safe subroutine to obtain the hertz frequency at which the switch clock operates. Switch clock frequency can be used to convert the switch clock value returned by the swclockRead subroutine.
Return Values
Upon successful completion, the swclockGetIncrement subroutine returns the hertz frequency at which the switch clock operates. Otherwise, a value of -1 is returned and the global variable errno is set to identify the error.
Error Values
Related Information
Subroutines: swclockInit, swclockRead, swclockReadSec, swclockTerm
Header File: swclock.h
Purpose
swclockInit - Initializes the switch clock read interface for a thread.
Library
Switch Clock Library (libswclock.a)
Location
/usr/lib/libswclock.a
Syntax
#include <swclock.h> swclock_handle_t swclockInit(void);
Description
Use this thread-safe subroutine to initialize the switch clock read interface for the current thread. It returns a handle which must be passed as an input parameter to all other switch clock library subroutines.
Usage Notes:
Return Values
Upon successful completion, the swclockInit subroutine returns a handle that must be passed as input to all other switch clock library subroutines. Otherwise, a value of 0 is returned and the global variable errno is set to identify the error.
Error Values
Related Information
IBM Parallel System Support Programs for AIX: Administration Guide
Subroutines: swclockGetIncrement, swclockRead, swclockReadSec, swclockTerm
Header File: swclock.h
Purpose
swclockRead - Returns the current switch clock value.
Library
Switch Clock Library (libswclock.a)
Location
/usr/lib/libswclock.a
Syntax
#include <swclock.h> long64_t swclockRead(swclock_handle_t swclock_handle);
Parameters
Description
Use this thread-safe subroutine to read the switch clock. The switch clock value can be converted using the frequency returned by the swclockGetIncrement subroutine. The swclockRead subroutine can be called as many times as needed once the switch clock read interface is initialized for the current thread.
The switch clock is synchronous across all nodes active on a switch. Its value is set to zero when the primary node powers on, and can be reset during switch operation and management.
Usage Notes:
Return Values
Upon successful completion, the swclockRead subroutine returns the current value of the switch clock. Otherwise, a value of -1 is returned and the global variable errno is set to identify the error.
Error Values
Related Information
IBM Parallel System Support Programs for AIX: Diagnosis and Messages Guide
IBM Parallel System Support Programs for AIX: Administration Guide
Subroutines: swclockGetIncrement, swclockInit, swclockReadSec, swclockTerm
Header File: swclock.h
Purpose
swclockReadSec - Returns the current switch clock value in seconds.
Library
Switch Clock Library (libswclock.a)
Location
/usr/lib/libswclock.a
Syntax
#include <swclock.h> double swclockReadSec(swclock_handle_t swclock_handle);
Parameters
Description
Use this thread-safe subroutine to read the switch clock. The swclockReadSec subroutine returns switch clock value converted to seconds. It can be called as many times as needed to read the switch clock once the switch clock read interface is initialized for the current thread.
The switch clock is synchronous across all nodes active on a switch. Its value is set to zero when the primary node powers on, and can be reset during switch operation and management.
Usage Notes:
Return Values
Upon successful completion, the swclockReadSec subroutine returns the current value of the switch clock converted to seconds. Otherwise, a value of -1 is returned and the global variable errno is set to identify the error.
Error Values
Related Information
IBM Parallel System Support Programs for AIX: Diagnosis and Messages Guide
IBM Parallel System Support Programs for AIX: Administration Guide
Subroutines: swclockGetIncrement, swclockInit, swclockRead, swclockTerm
Header File: swclock.h
Purpose
swclockTerm - Terminates the switch clock read interface for a thread.
Library
Switch Clock Library (libswclock.a)
Location
/usr/lib/libswclock.a
Syntax
#include <swclock.h> int swclockTerm(swclock_handle_t swclock_handle);
Parameters
Description
Use this thread-safe subroutine to terminate the switch clock read interface for the current thread. Switch clock library subroutines called subsequent to the swclockTerm subroutine will fail unless the thread reinitializes the interface. If the swclockTerm subroutine is not called, the switch clock read interface will be terminated when the thread itself terminates.
Return Values
Upon successful completion, the swclockTerm subroutine returns a value of 0. Otherwise, a value of -1 is returned and the global variable errno is set to identify the error.
Error Values
Related Information
Subroutines: swclockGetIncrement, swclockInit, swclockRead, swclockReadSec
Header File: swclock.h
Note: | Colors may vary depending on the type o/f display you are using. |
The following list contains valid color names that can be supplied as
optional arguments to the -backgroundColor and
-foregroundColor flags:
Color | Red | Green | Blue |
---|---|---|---|
alice blue | 240 | 248 | 255 |
antique white | 250 | 235 | 215 |
aquamarine | 127 | 255 | 212 |
azure | 240 | 255 | 255 |
beige | 245 | 245 | 220 |
bisque | 255 | 228 | 196 |
black | 0 | 0 | 0 |
blanched almond | 255 | 235 | 205 |
blue | 0 | 0 | 255 |
blue violet | 138 | 43 | 226 |
brown | 165 | 42 | 42 |
burlywood | 222 | 184 | 135 |
cadet blue | 95 | 158 | 160 |
chartreuse | 127 | 255 | 0 |
chocolate | 210 | 105 | 30 |
coral | 255 | 127 | 80 |
cornflower blue | 100 | 149 | 237 |
cornsilk | 255 | 248 | 220 |
cyan | 0 | 255 | 255 |
dark goldenrod | 184 | 134 | 11 |
dark green | 0 | 100 | 0 |
dark khaki | 189 | 183 | 107 |
dark olive green | 85 | 107 | 47 |
dark orange | 255 | 140 | 0 |
dark orchid | 153 | 50 | 204 |
dark salmon | 233 | 150 | 122 |
dark sea green | 143 | 188 | 143 |
dark slate blue | 72 | 61 | 139 |
dark turquoise | 0 | 206 | 209 |
dark violet | 148 | 0 | 211 |
deep pink | 255 | 20 | 147 |
deep sky blue | 0 | 191 | 255 |
dim gray | 105 | 105 | 105 |
dodger blue | 30 | 144 | 255 |
firebrick | 178 | 34 | 34 |
floral white | 255 | 250 | 240 |
forest green | 34 | 139 | 34 |
ghost white | 248 | 248 | 255 |
gold | 255 | 215 | 0 |
goldenrod | 218 | 165 | 32 |
gray | 190 | 190 | 190 |
green | 0 | 255 | 0 |
green yellow | 173 | 255 | 47 |
honeydew | 240 | 255 | 240 |
hot pink | 255 | 105 | 180 |
indian red | 205 | 92 | 92 |
ivory | 255 | 255 | 240 |
khaki | 240 | 230 | 140 |
lavender | 230 | 230 | 250 |
lavender blush | 255 | 240 | 245 |
lawn green | 124 | 252 | 0 |
lemon chiffon | 255 | 250 | 205 |
light blue | 173 | 216 | 230 |
light coral | 240 | 128 | 128 |
light cyan | 224 | 255 | 255 |
light goldenrod | 238 | 221 | 130 |
light gray | 211 | 211 | 211 |
light pink | 255 | 182 | 193 |
light salmon | 255 | 160 | 122 |
light sea green | 32 | 178 | 170 |
light sky blue | 135 | 206 | 250 |
light slate blue | 132 | 112 | 255 |
light slate gray | 119 | 136 | 153 |
light steel blue | 176 | 196 | 222 |
light yellow | 255 | 255 | 224 |
lime green | 50 | 205 | 50 |
linen | 250 | 240 | 230 |
magenta | 255 | 0 | 255 |
maroon | 176 | 48 | 96 |
medium aquamarine | 102 | 205 | 170 |
medium blue | 0 | 0 | 205 |
medium orchid | 186 | 85 | 211 |
medium purple | 147 | 112 | 219 |
medium sea green | 60 | 179 | 113 |
medium slate blue | 123 | 104 | 238 |
medium spring green | 0 | 250 | 154 |
medium turquoise | 72 | 209 | 204 |
medium violet red | 199 | 21 | 133 |
midnight blue | 25 | 25 | 112 |
mint cream | 245 | 255 | 250 |
misty rose | 255 | 228 | 225 |
moccasin | 255 | 228 | 181 |
navajo white | 255 | 222 | 173 |
navy blue | 0 | 0 | 128 |
old lace | 253 | 245 | 230 |
oldlace | 253 | 245 | 230 |
olive drab | 107 | 142 | 35 |
orange | 255 | 165 | 0 |
orange red | 255 | 69 | 0 |
orchid | 218 | 112 | 214 |
pale goldenrod | 238 | 232 | 170 |
pale green | 152 | 251 | 152 |
pale turquoise | 175 | 238 | 238 |
pale violet red | 219 | 112 | 147 |
papaya whip | 255 | 239 | 213 |
peach puff | 255 | 218 | 185 |
peru | 205 | 133 | 63 |
pink | 255 | 192 | 203 |
plum | 221 | 160 | 221 |
powder blue | 176 | 224 | 230 |
purple | 160 | 32 | 240 |
red | 255 | 0 | 0 |
rosy brown | 188 | 143 | 143 |
royal blue | 65 | 105 | 225 |
saddle brown | 139 | 69 | 19 |
salmon | 250 | 128 | 114 |
sandy brown | 244 | 164 | 96 |
sea green | 46 | 139 | 87 |
seashell | 255 | 245 | 238 |
sienna | 160 | 82 | 45 |
sky blue | 135 | 206 | 235 |
slate blue | 106 | 90 | 205 |
slate gray | 112 | 128 | 144 |
snow | 255 | 250 | 250 |
spring green | 0 | 255 | 127 |
steel blue | 70 | 130 | 180 |
tan | 210 | 180 | 140 |
thistle | 216 | 191 | 216 |
tomato | 255 | 99 | 71 |
turquoise | 64 | 224 | 208 |
violet | 238 | 130 | 238 |
violet red | 208 | 32 | 144 |
wheat | 245 | 222 | 179 |
white | 255 | 255 | 255 |
white smoke | 245 | 245 | 245 |
yellow | 255 | 255 | 0 |
yellow green | 154 | 205 | 50 |
Note: | Fonts will vary depending on the type of Xmachine or Xstation you are using. |
The following list contains font names that can be supplied as optional arguments to the -fontFamily flag:
This glossary includes terms and definitions from:
The following cross-references are used in this glossary:
This section contains some of the terms that are commonly used in the SP publications.
IBM is grateful to the American National Standards Institute (ANSI) for permission to reprint its definitions from the American National Standard Vocabulary for Information Processing (Copyright 1970 by American National Standards Institute, Incorporated), which was prepared by Subcommittee X3K5 on Terminology and Glossary of the American National Standards Committee X3. ANSI definitions are preceded by an asterisk (*).
Other definitions in this glossary are taken from IBM Vocabulary for Data Processing, Telecommunications, and Office Systems (SC20-1699) and IBM DATABASE 2 Application Programming Guide for TSO Users (SC26-4081).