User Tools

Site Tools


gpfs:gpfs_cmd

GPFS commands

mmuserauth

mmuserauth service check [--data-access-method {file|object|all}] [-r|--rectify]
          [-N|--nodes {node-list|cesNodes}][--server-reachability]
          
--server-reachability: correct config errors
          

mmchnode

--gateway | --nogateway
    Specifies whether the node is to be designated as a gateway node or not.
--manager | --nomanager
    Designates the node as part of the pool of nodes from which file system managers and token managers are selected. 
--nonquorum
    Designates the node as a non-quorum node. If two or more quorum nodes are downgraded at the same time, GPFS must be stopped on all nodes in the cluster. GPFS does not have to be stopped if the nodes are downgraded one at a time.
--perfmon | --noperfmon
    Specifies whether the node is to be designated as a perfmon node or not.
--nosnmp-agent
    Stops the SNMP subagent and specifies that the node should no longer serve as an SNMP collector node. For more information, see GPFS SNMP support.
--quorum
    Designates the node as a quorum node.
    Note: If you are designating a node as a quorum node, and adminMode central is in effect for the cluster, you must ensure that GPFS is up and running on that node (mmgetstate reports the state of the node as active).
--snmp-agent
    Designates the node as an SNMP collector node. If the GPFS daemon is active on this node, the SNMP subagent will be started as well. For more information, GPFS SNMP support.
    
--ces-enable
    Enables Cluster Export Services (CES) on the node.
--ces-disable
    Disables CES on the node. 
--ces-group=Group[,Group...]
    Adds one or more groups to the specified nodes. Each group that is listed is added to all the specified nodes. 
--noces-group=Group[,Group...]
    Removes one or more groups from the specified nodes.
[root@prscale-a-02 ~]# mmcloudgateway node list
 Node  Cloud node name                  Cloud Node Class
--------------------------------------------------------
    1  prscale-a-01                     tct_group1
[root@prscale-a-02 ~]# mmchnode --cloud-gateway-disable -N prscale-a-01 --cloud-gateway-nodeclass tct_group1
[root@prscale-a-01 ~]# mmcloudgateway service status

=============================
Cloud Node Class:  tct_group1
=============================

 Cloud Service                Status      Reason
------------------------------------------------
 tct_tiering1-vault_backup_01  ENABLED

                                           Server     Account /   Container /
 Node  Daemon node name                    Status     CSAP        File System/Set               Status    Reasons
-----------------------------------------------------------------------------------------------------------------
    1  prscale-a-01                        STARTED    vault_backup_01
                                                      vault_backup_01                                ONLINE
[root@prscale-a-01 ~]# mmcloudgateway cloudStorageAccessPoint list

Configured cloudStorageAccessPoint options from node class tct_group1:
----------------------------------------------------------------------
  csapName                        :  vault_backup_01
  accountName                     :  vault_backup_01
  url                             :  http://10.0.0.1
  mpuPartsSize                    :  134217728
  sliceSize                       :  524288
  proxyIp                         :
  proxyPort                       :
  region                          :
[root@prscale-a-01 ~]# mmcloudgateway account list

Configured account options from node class tct_group1:
------------------------------------------------------
  accountName                     :  vault_backup_01
  cloudType                       :  s3
  userName                        :  s4KhHAJtjQ2xxxxxxxx
  tenantId                        :
[root@prscale-a-01 ~]# mmcloudgateway service status

=============================
Cloud Node Class:  tct_group1
=============================

 Cloud Service                Status      Reason
------------------------------------------------
 tct_tiering1-vault_backup_01  ENABLED

                                           Server     Account /   Container /
 Node  Daemon node name                    Status     CSAP        File System/Set               Status    Reasons
-----------------------------------------------------------------------------------------------------------------
    1  prscale-a-01                        STARTED    vault_backup_01
                                                      vault_backup_01                                ONLINE
[root@prscale-a-01 ~]# mmcloudgateway cloudService delete    --cloud-nodeclass tct_group1 --cloud-service-name tct_tiering1-vault_backup_01
mmcloudgateway: Sending the command to the first successful node starting with prscale-a-01
mmcloudgateway: This may take a while...
mmcloudgateway: Command completed successfully on prscale-a-01.
mmcloudgateway: Command completed.

[root@prscale-a-02 ~]# mmdelnode -N prscale-a-01 mmdelnode: Node prscale-a-01 still appears in the following callbacks:

      CLOUDGWnotify_1

mmdelnode: Command failed. Examine previous error messages to determine cause.

[root@prscale-a-02 ~]# mmlscallback -Y mmlscallback::HEADER:version:reserved:reserved:identifier:command:priority:sync:timeout:event:node:parms:onError:object: mmlscallback::0:1:::GUI_CCR_CHANGE:/usr/lpp/mmfs/gui/callbacks/global/ccrChangedCallback_421.sh::false::ccrFileChange:GUI_MGMT_SERVERS:root %25eventName %25ccrObjectName %25ccrObjectVersion::gui_master_node,_gui.user.repo,_gui.snapshots,_gui.notification,_gui.settings,_gui.dashboards,_gui.keystore,_gui.keystore_settings,_gui.ldap_settings,_gui.policysettings,gui_jobs,mmsdrfs,spectrum-scale-object-policies.conf: mmlscallback::0:1:::GUI_CM_TAKEOVER:/usr/lpp/mmfs/gui/callbacks/global/cmTakeOverCallback_421.sh::false::clusterManagerTakeover:GUI_MGMT_SERVERS:root %25eventName %25eventNode %25clusterManager::: mmlscallback::0:1:::GPFS_STARTUP_SHUTDOWN:/usr/lpp/mmfs/gui/callbacks/global/startupShutdownCallback_421.sh::false::shutdown,startup:GUI_MGMT_SERVERS:root %25eventName %25eventNode::: mmlscallback::0:1:::GUI_NODES:/usr/lpp/mmfs/gui/callbacks/global/NodeQuorumCallback_421.sh::false::nodeJoin,nodeLeave,quorumLoss,quorumNodeJoin,quorumNodeLeave,quorumReached:GUI_MGMT_SERVERS:root %25eventName %25eventNode %25quorumNodes %25downNodes %25upNodes %25clusterName::: mmlscallback::0:1:::GUI_DISK_SPACE:/usr/lpp/mmfs/lib/gui_enablement/diskSpaceCallback::false::lowDiskSpace,noDiskSpace:GUI_SERVERS:root %25eventName %25eventNode %25reason %25fsName %25filesetName %25storagePool::: mmlscallback::0:1:::GUI_MOUNT_ACTION:/usr/lpp/mmfs/lib/gui_enablement/mountActionCallback::false::mount,unmount:GUI_SERVERS:root %25eventName %25eventNode %25fsName::: mmlscallback::0:1:::GUI_THRESHOLD_MIGRATION:/usr/lpp/mmfs/lib/gui_enablement/thresholdMigration::false::lowDiskSpace,noDiskSpace:GUI_SERVERS:root %25fsName %25storagePool %25reason %25eventNode::: mmlscallback::0:1:::CLOUDGWnotify_1:/opt/ibm/MCStore/scripts/mcstorenotify::true::startup,preShutdown,quorumLoss:10.0.0.1:%25eventName -N prscale-a-01 tct_group1:::

[root@prscale-a-02 ~]# mmdelcallback CLOUDGWnotify_1 mmdelcallback: Propagating the cluster configuration data to all

affected nodes.  This is an asynchronous process.

[root@prscale-a-02 ~]#

mmperfmon

mmperfmon config show

Do not edit the config file, automatically generated

/opt/IBM/zimon/ZIMonCollector.cfg
[root@prscale-a-01 ~]# systemctl restart pmcollector.service
[root@prscale-a-01 ~]# mmperfmon config show --config-file OutputFile.txt

Edit the output file to remove or add parameters or nodes

[root@prscale-a-01 ~]# mmperfmon config update --config-file OutputFile.txt
[root@prscale-a-01 ~]# mmdelnode -N prscale-a-02
Verifying GPFS is stopped on all affected nodes ...
mmdelnode: Removing GPFS system files on all deleted nodes ...
mmdelnode: Command successfully completed
mmdelnode: Propagating the cluster configuration data to all
  affected nodes.  This is an asynchronous process.
gpfs/gpfs_cmd.txt · Last modified: 2023/03/14 22:10 by manu