====== GPFS health ======
mmhealth thresholds (+GUI +REST) monitors the usage:
$ mmhealth thresholds list
### Threshold Rules ###
rule_name metric error warn direction filterBy groupBy sensitivity
----------------------------------------------------------------------------------------------------------------------------
InodeCapUtil_Rule Fileset_inode 90.0 80.0 high gpfs_cluster_name,gpfs_fs_name,gpfs_fset_name 300
DataCapUtil_Rule DataPool_capUtil 97.0 90.0 high gpfs_cluster_name,gpfs_fs_name,gpfs_diskpool_name 300
..
List all events
[root@prscale-a-01 ~]# mmhealth node eventlog
2021-06-09 14:47:43.228189 CEST unmounted_fs_check WARNING The filesystem gpfs02lv is probably needed, but not mounted
2021-06-09 14:47:48.256313 CEST disk_found INFO The disk GPFS_NSD_A_D1_0001 was found
2021-06-09 14:52:28.475693 CEST fs_remount_mount INFO The filesystem gpfs02lv was mounted normal
Clear the events:
[root@prscale-a-01 ~]# mmhealth node eventlog --clear
Clear ALL messages on the Web (GUI) interface
[root@prscale-a-01 ~]# /usr/lpp/mmfs/gui/cli/lshealth --reset
Already resolved errors that continue to be displayed in mmhealth and the GUI:
How to remove them (and this annoying TIPS):
mmdsh -N mmsysmonc clearDB
mmdsh -N mmsysmoncontrol restart
mmhealth event hide
Check cluster health
[root@prscale-a-01 ~]# mmhealth cluster show
Component Total Failed Degraded Healthy Other
----------------------------------------------------------------------------------------------------------------
NODE 3 0 0 0 3
GPFS 3 0 0 0 3
NETWORK 3 0 0 3 0
FILESYSTEM 3 0 0 3 0
DISK 21 0 0 21 0
CES 2 0 0 2 0
CESIP 1 0 0 1 0
FILESYSMGR 2 0 0 2 0
GUI 2 0 0 2 0
PERFMON 3 0 0 3 0
THRESHOLD 3 0 0 3 0
More compact
[root@prscale-a-01 ~]# mmhealth cluster show -Y
mmhealth:Summary:HEADER:version:reserved:reserved:component:entityname:total:failed:degraded:healthy:other:
mmhealth:Summary:0:1:::NODE:NODE:3:0:0:0:3:
mmhealth:Summary:0:1:::GPFS:GPFS:3:0:0:0:3:
mmhealth:Summary:0:1:::NETWORK:NETWORK:3:0:0:3:0:
mmhealth:Summary:0:1:::FILESYSTEM:FILESYSTEM:3:0:0:3:0:
mmhealth:Summary:0:1:::DISK:DISK:43:0:0:43:0:
mmhealth:Summary:0:1:::CES:CES:2:0:0:2:0:
mmhealth:Summary:0:1:::CESIP:CESIP:1:0:0:1:0:
mmhealth:Summary:0:1:::CLOUDGATEWAY:CLOUDGATEWAY:2:0:0:2:0:
mmhealth:Summary:0:1:::FILESYSMGR:FILESYSMGR:2:0:0:2:0:
mmhealth:Summary:0:1:::GUI:GUI:2:0:0:2:0:
mmhealth:Summary:0:1:::PERFMON:PERFMON:3:0:0:3:0:
mmhealth:Summary:0:1:::THRESHOLD:THRESHOLD:3:0:0:3:0:
Per node
[root@prscale-a-01 ~]# mmhealth node show
Node name: prscale-a-01
Node status: TIPS
Status Change: 2 days ago
Component Status Status Change Reasons
----------------------------------------------------------------------------------------------------------------------------------
GPFS TIPS 2 days ago callhome_not_enabled, gpfs_maxfilestocache_small, total_memory_small
NETWORK HEALTHY 4 days ago -
FILESYSTEM HEALTHY 4 days ago -
DISK HEALTHY 4 days ago -
CES HEALTHY 4 days ago -
CESIP HEALTHY 4 days ago -
CLOUDGATEWAY HEALTHY 4 days ago -
FILESYSMGR HEALTHY 4 days ago -
GUI HEALTHY 4 days ago -
PERFMON HEALTHY 4 days ago -
THRESHOLD HEALTHY 4 days ago -
[root@prscale-a-01 ~]# mmhealth node show -y
Option with missing argument
Additional messages:
invalid option: -y
[root@prscale-a-01 ~]# mmhealth node show -Y
mmhealth:Event:HEADER:version:reserved:reserved:node:component:entityname:entitytype:event:arguments:activesince:identifier:ishidden:
mmhealth:State:HEADER:version:reserved:reserved:node:component:entityname:entitytype:status:laststatuschange:
mmhealth:State:0:1:::prscale-a-01:NODE:prscale-a-01:NODE:TIPS:2021-10-03 03%3A55%3A43.152676 CEST:
mmhealth:State:0:1:::prscale-a-01:CES:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A50%3A34.180949 CEST:
mmhealth:State:0:1:::prscale-a-01:BLOCK:prscale-a-01:NODE:DISABLED:2021-10-01 09%3A23%3A59.486199 CEST:
mmhealth:State:0:1:::prscale-a-01:NFS:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A50%3A34.180949 CEST:
mmhealth:State:0:1:::prscale-a-01:AUTH_OBJ:prscale-a-01:NODE:DISABLED:2021-10-01 09%3A24%3A17.174618 CEST:
mmhealth:State:0:1:::prscale-a-01:CESNETWORK:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A49%3A03.938605 CEST:
mmhealth:State:0:1:::prscale-a-01:CESNETWORK:ens192:NIC:HEALTHY:2021-10-01 09%3A24%3A05.346180 CEST:
mmhealth:State:0:1:::prscale-a-01:OBJECT:prscale-a-01:NODE:DISABLED:2021-10-01 09%3A24%3A05.421280 CEST:
mmhealth:State:0:1:::prscale-a-01:SMB:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A48%3A18.924751 CEST:
mmhealth:State:0:1:::prscale-a-01:AUTH:prscale-a-01:NODE:DISABLED:2021-10-01 09%3A24%3A02.017860 CEST:
mmhealth:State:0:1:::prscale-a-01:HDFS_NAMENODE:prscale-a-01:NODE:DISABLED:2021-10-01 09%3A24%3A17.295829 CEST:
mmhealth:State:0:1:::prscale-a-01:CLOUDGATEWAY:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A25%3A08.337661 CEST:
mmhealth:State:0:1:::prscale-a-01:CLOUDGATEWAY:tct_tiering1-vault_backup_01:TCT_SERVICE:HEALTHY:2021-10-01 09%3A24%3A33.013963 CEST:
mmhealth:State:0:1:::prscale-a-01:CLOUDGATEWAY:vault_backup_01/:TCT_CSAP:HEALTHY:2021-10-01 09%3A24%3A33.023072 CEST:
mmhealth:State:0:1:::prscale-a-01:CESIP:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A35%3A32.969821 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A52%3A49.707704 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_A_CES_0001:NSD:HEALTHY:2021-10-01 09%3A52%3A49.715357 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_M_A_0001:NSD:HEALTHY:2021-10-01 09%3A24%3A14.576509 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_M_A_0002:NSD:HEALTHY:2021-10-01 09%3A24%3A14.596580 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_M_A_0003:NSD:HEALTHY:2021-10-01 09%3A24%3A14.618590 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0001:NSD:HEALTHY:2021-10-01 09%3A24%3A14.637984 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_M_A_0004:NSD:HEALTHY:2021-10-01 09%3A24%3A14.656978 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0002:NSD:HEALTHY:2021-10-01 09%3A24%3A14.673522 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0003:NSD:HEALTHY:2021-10-01 09%3A24%3A14.692509 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0004:NSD:HEALTHY:2021-10-01 09%3A24%3A14.711926 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0005:NSD:HEALTHY:2021-10-01 09%3A24%3A14.742578 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0006:NSD:HEALTHY:2021-10-01 09%3A24%3A14.761914 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0007:NSD:HEALTHY:2021-10-01 09%3A24%3A14.788854 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0008:NSD:HEALTHY:2021-10-01 09%3A24%3A14.808564 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0009:NSD:HEALTHY:2021-10-01 09%3A24%3A14.830882 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0010:NSD:HEALTHY:2021-10-01 09%3A24%3A14.852833 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0011:NSD:HEALTHY:2021-10-01 09%3A24%3A14.876191 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_D_A_0012:NSD:HEALTHY:2021-10-01 09%3A24%3A14.888040 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS03_NSD_M_A_0005:NSD:HEALTHY:2021-10-01 09%3A24%3A14.915370 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS03_NSD_M_A_0006:NSD:HEALTHY:2021-10-01 09%3A24%3A14.931229 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS03_NSD_M_A_0007:NSD:HEALTHY:2021-10-01 09%3A24%3A14.942819 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS03_NSD_M_A_0008:NSD:HEALTHY:2021-10-01 09%3A24%3A14.956576 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS03_NSD_M_A_0009:NSD:HEALTHY:2021-10-01 09%3A24%3A14.970095 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS03_NSD_M_A_0010:NSD:HEALTHY:2021-10-01 09%3A24%3A14.993296 CEST:
mmhealth:State:0:1:::prscale-a-01:DISK:GPFS_NSD_A_D1_0001:NSD:HEALTHY:2021-10-01 09%3A24%3A15.019350 CEST:
mmhealth:State:0:1:::prscale-a-01:GUI:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A53%3A32.099884 CEST:
mmhealth:State:0:1:::prscale-a-01:THRESHOLD:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A24%3A01.090924 CEST:
mmhealth:State:0:1:::prscale-a-01:THRESHOLD:MemFree_Rule:THRESHOLD_RULE:HEALTHY:2021-10-01 09%3A24%3A32.160416 CEST:
mmhealth:State:0:1:::prscale-a-01:THRESHOLD:SMBConnPerNode_Rule:THRESHOLD_RULE:HEALTHY:2021-10-01 09%3A29%3A32.474178 CEST:
mmhealth:State:0:1:::prscale-a-01:THRESHOLD:active_thresh_monitor:THRESHOLD_MONITOR:HEALTHY:2021-10-01 09%3A35%3A32.982353 CEST:
mmhealth:State:0:1:::prscale-a-01:THRESHOLD:SMBConnTotal_Rule:THRESHOLD_RULE:HEALTHY:2021-10-01 09%3A38%3A48.166426 CEST:
mmhealth:State:0:1:::prscale-a-01:PERFMON:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A24%3A17.345148 CEST:
mmhealth:State:0:1:::prscale-a-01:FILESYSTEM:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A48%3A17.866719 CEST:
mmhealth:State:0:1:::prscale-a-01:FILESYSTEM:cesSharedRootlv:FILESYSTEM:HEALTHY:2021-10-01 09%3A48%3A17.879658 CEST:
mmhealth:State:0:1:::prscale-a-01:FILESYSTEM:gpfs01lv:FILESYSTEM:HEALTHY:2021-10-01 09%3A48%3A17.893977 CEST:
mmhealth:State:0:1:::prscale-a-01:FILESYSTEM:gpfs02lv:FILESYSTEM:HEALTHY:2021-10-01 09%3A24%3A17.503234 CEST:
mmhealth:State:0:1:::prscale-a-01:GPFS:prscale-a-01:NODE:TIPS:2021-10-03 03%3A55%3A43.146453 CEST:
mmhealth:Event:0:1:::prscale-a-01:GPFS:prscale-a-01:NODE:callhome_not_enabled::2021-10-01 09%3A24%3A16.102766 CEST::no:
mmhealth:Event:0:1:::prscale-a-01:GPFS:prscale-a-01:NODE:gpfs_maxfilestocache_small::2021-10-01 09%3A24%3A16.139998 CEST::no:
mmhealth:Event:0:1:::prscale-a-01:GPFS:prscale-a-01:NODE:total_memory_small::2021-10-01 09%3A24%3A16.169878 CEST::no:
mmhealth:State:0:1:::prscale-a-01:NETWORK:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A24%3A17.308930 CEST:
mmhealth:State:0:1:::prscale-a-01:NETWORK:ens192:NIC:HEALTHY:2021-10-01 09%3A24%3A17.323203 CEST:
mmhealth:State:0:1:::prscale-a-01:FILESYSMGR:prscale-a-01:NODE:HEALTHY:2021-10-01 09%3A24%3A19.114252 CEST:
mmhealth:State:0:1:::prscale-a-01:FILESYSMGR:gpfs02lv:FILESYSTEMMGMT:HEALTHY:2021-10-01 09%3A24%3A19.137516 CEST:
Check Protocols components
[root@prscale-a-01 ~]# mmces state show -a -Y
mmces:stateShow:HEADER:version:reserved:reserved:NODE:AUTH:BLOCK:NETWORK:HDFS_NAMENODE:AUTH_OBJ:NFS:OBJ:SMB:CES:
mmces:stateShow:0:1:::prscale-a-01:DISABLED:DISABLED:HEALTHY:DISABLED:DISABLED:HEALTHY:DISABLED:HEALTHY:HEALTHY:
mmces:stateShow:0:1:::prscale-b-01:DISABLED:DISABLED:HEALTHY:DISABLED:DISABLED:HEALTHY:DISABLED:HEALTHY:HEALTHY:
Check particular event
[root@prscale-a-02 ~]# mmhealth cluster show
Component Total Failed Degraded Healthy Other
-------------------------------------------------------------------------------------
NODE 3 0 0 0 3
GPFS 3 0 0 0 3
NETWORK 3 0 0 3 0
FILESYSTEM 3 0 0 3 0
DISK 31 0 0 31 0
CES 2 0 0 2 0
CESIP 1 0 0 1 0
FILESYSMGR 1 0 0 1 0
GUI 3 0 1 2 0
PERFMON 3 0 0 3 0
THRESHOLD 3 0 0 3 0
mmhealth cluster show [ NODE | GPFS | NETWORK [ UserDefinedSubComponent ]
| FILESYSTEM [UserDefinedSubComponent ]| DISK [UserDefinedSubComponent ]
| CES |AUTH | AUTH_OBJ | BLOCK | CESNETWORK | NFS | OBJECT | SMB
| HADOOP |CLOUDGATEWAY | GUI | PERFMON | THRESHOLD
| AFM [UserDefinedSubComponent] ]
[-Y] [--verbose]
[root@prscale-a-02 ~]# mmhealth cluster show GUI
Component Node Status Reasons
------------------------------------------------------------------------------------------
GUI prscale-q-b-01 HEALTHY -
GUI prscale-a-02 HEALTHY -
GUI prscale-a-01 DEGRADED gui_refresh_task_failed
[root@prscale-a-01 ~]# /usr/lpp/mmfs/gui/cli/runtask help --debug
[AFM_FILESET_STATE, AFM_NODE_MAPPING, ALTER_HOST_NAME, CALLBACK, CALLHOME, CALLHOME_STATUS, CAPACITY_LICENSE, CES_ADDRESS, CES_STATE, CES_SERVICE_STATE, CES_USER_AUTH_SERVICE, CLUSTER_CONFIG, CONNECTION_STATUS, DAEMON_CONFIGURATION, DF, DISK_USAGE, DISKS, FILESETS, FILESYSTEM_MOUNT, FILESYSTEMS, FILE_AUDIT_LOG_CONFIG, GUI_CONFIG_CHECK, GPFS_JOBS, DIGEST_NOTIFICATION_TASK, HEALTH_STATES, HEALTH_TRIGGERED, HOST_STATES, HOST_STATES_CLIENTS, INODES, KEYSTORE, LOG_REMOVER, MASTER_GUI_ELECTION, MOUNT_CONFIG, NFS_EXPORTS, NFS_EXPORTS_DEFAULTS, NFS_SERVICE, NODE_LICENSE, NODECLASS, OBJECT_STORAGE_POLICY, OS_DETECT, PM_MONITOR, PM_SENSORS, PM_TOPOLOGY, POLICIES, QUOTA, QUOTA_DEFAULTS, QUOTA_ID_RESOLVE, QUOTA_MAIL, RDMA_INTERFACES, REMOTE_CONFIG, REMOTE_CLUSTER, REMOTE_FILESETS, REMOTE_GPFS_CONFIG, REMOTE_HEALTH_STATES, SMB_GLOBALS, SMB_SHARES, SNAPSHOTS, SNAPSHOTS_FS_USAGE, SNAPSHOT_MANAGER, SQL_STATISTICS, STATE_MAIL, STORAGE_POOL, SYSTEMUTIL_DF, TCT_ACCOUNT, TCT_CLOUD_SERVICE, TCT_NODECLASS, THRESHOLDS, WATCHFOLDER, WATCHFOLDER_STATUS, TASK_CHAIN]
[root@prscale-a-01 ~]# /usr/lpp/mmfs/gui/cli/runtask CLUSTER_CONFIG --debug
debug: locale=en_US
debug: Running 'mmsdrquery 'sdrq_cluster_info' all ' on node localhost
debug: Running 'mmsdrquery 'sdrq_nsd_info' all ' on node localhost
debug: Running 'mmlscluster -Y ' on node localhost
debug: Running 'mmsdrquery 'sdrq_node_info' all ' on node localhost
debug: Running 'mmlsnodeclass 'GUI_MGMT_SERVERS' -Y ' on node localhost
debug: Running 'mmlsnodeclass 'GUI_SERVERS' -Y ' on node localhost
EFSSG1000I The command completed successfully.
[root@prscale-a-01 ~]# mmhealth event show gui_refresh_task_failed
Event Name: gui_refresh_task_failed
Event ID: 998254
Description: One or more GUI refresh tasks failed. This could mean that data in the GUI is outdated.
Cause: There can be several reasons.
User Action: 1.) Check if there is additional information available by executing '/usr/lpp/mmfs/gui/cli/lstasklog [taskname]'. 2.) Run the specified task manually on the CLI by executing '/usr/lpp/mmfs/gui/cli/runtask [taskname] --debug'. 3.) Check the GUI logs under /var/log/cnlog/mgtsrv. 4.) Contact IBM Support if this error persists or occurs more often.
Severity: WARNING
State: DEGRADED
[root@prscale-a-01 ~]# mmhealth event resolve 998254
The specified event gui_refresh_task_failed is not manually resolvable.
A. Objectif
Contrôler l’état d’un filesystem GPFS et rectifier les anomalies
B. Principales entités concernées
Système
C. Description générale et flux des données
D. Définition des termes
Node names
Correspondance GPFS-AIX
GPFS AIX
p5-gpfs-h oragpfh
p5-gpfs-r oragpfr
p5-gpfs-k oragpfk
E. Instructions spécifiques
1) Affichez les informations
mmlsnsd affiche les filesystemes GPFS connues sur la machine. (peut s’exécuter sur les 3 machines)
root@oragpfh:/home/root>mmlsnsd
File system Disk name Primary node Backup node
---------------------------------------------------------------------------
orafs DiskR p5-gpfs-r p5-gpfs-h
orafs DiskH p5-gpfs-r p5-gpfs-h
orafs DiskK gpfs-k
orafs2 DiskR2 p5-gpfs-r p5-gpfs-h
orafs2 DiskH2 p5-gpfs-r p5-gpfs-h
orafs2 DiskK2 gpfs-k
root@oragpfh:/home/root>mmlsnsd -L
L’option –L affiche les NSD volume ID
File system Disk name NSD volume ID Primary node Backup node
--------------------------------------------------------------------------------------------
orafs DiskR 0A040120452A20B4 p5-gpfs-r p5-gpfs-h
orafs DiskH 0A040120452A20B6 p5-gpfs-r p5-gpfs-h
orafs DiskK AC13131D48008499 gpfs-k
orafs2 DiskR2 0A04012046827A9E p5-gpfs-r p5-gpfs-h
orafs2 DiskH2 0A04012046827AA0 p5-gpfs-r p5-gpfs-h
orafs2 DiskK2 AC13131D480084BD gpfs-k
root@oragpfh:/home/root>mmlsnsd -M
L’option –L affiche la correspondance disque GPFS – disque OS/machine
ex : DiskR = hdisk0 sur p5-gpfs-r (primary node)= hdisk3 sur p5-gpfs-h (backup node)
Disk name NSD volume ID Device Node name Remarks
---------------------------------------------------------------------------------------
DiskR 0A040120452A20B4 /dev/hdisk0 p5-gpfs-r primary node
DiskR 0A040120452A20B4 /dev/hdisk3 p5-gpfs-h backup node
DiskH 0A040120452A20B6 /dev/hdisk3 p5-gpfs-r primary node
DiskH 0A040120452A20B6 /dev/hdisk0 p5-gpfs-h backup node
DiskK AC13131D48008499 /dev/descgpfslv gpfs-k primary node
DiskR2 0A04012046827A9E /dev/hdisk4 p5-gpfs-r primary node
DiskR2 0A04012046827A9E /dev/hdisk5 p5-gpfs-h backup node
DiskH2 0A04012046827AA0 /dev/hdisk5 p5-gpfs-r primary node
DiskH2 0A04012046827AA0 /dev/hdisk4 p5-gpfs-h backup node
DiskK2 AC13131D480084BD /dev/descgpfslv2 gpfs-k primary node
mmlsdisk permet d’afficher les disques d’un filesysteme GPFS. (peut s’exécuter sur les 3 machines)
root@oragpfh:/home/root>mmlsdisk orafs
disk driver sector failure holds holds
name type size group metadata data status availability
------------ -------- ------ ------- -------- ----- ------------- ------------
DiskR nsd 512 1 yes yes ready up
DiskH nsd 512 2 yes yes ready up
DiskK nsd 512 3 no no ready up
Le statu normal des disques est ready, la disponibilité est up
En cas d’incohérences, un message d’alerte est affiché après les informations disques !
ex : Attention : Due to an earlier configuration change, the file system
may contain data that is at risk of being lost
Autes commandes :
mmlscluster
root@oragpfh:/home/root>mmlscluster
GPFS cluster information
========================
GPFS cluster name: p5-gpfs-r
GPFS cluster id: 12399281700916488274
GPFS UID domain: p5-gpfs-r
Remote shell command: /usr/bin/ssh
Remote file copy command: /usr/bin/scp
GPFS cluster configuration servers:
-----------------------------------
Primary server: p5-gpfs-r
Secondary server: p5-gpfs-h
Node number Node name IP address Full node name Remarks
-----------------------------------------------------------------------------------
1 p5-gpfs-h 10.11.10.2 p5-gpfs-h quorum node
2 p5-gpfs-r 10.11.10.3 p5-gpfs-r quorum node
3 gpfs-k 10.11.10.4 gpfs-k quorum node
mmlsfs « Fs-name » (Fs-name ={orafs|orafs2}
root@oragpfh:/home/root>mmlsfs orafs
flag value description
---- -------------- -----------------------------------------------------
-s roundRobin Stripe method
-f 8192 Minimum fragment size in bytes
-i 512 Inode size in bytes
-I 16384 Indirect block size in bytes
-m 2 Default number of metadata replicas
-M 2 Maximum number of metadata replicas
-r 2 Default number of data replicas
-R 2 Maximum number of data replicas
-j cluster Block allocation type
-D posix File locking semantics in effect
-k posix ACL semantics in effect
-a 1048576 Estimated average file size
-n 32 Estimated number of nodes that will mount file system
-B 262144 Block size
-Q user;group Quotas enforced
none Default quotas enabled
-F 185344 Maximum number of inodes
-V 8.01 File system version. Highest supported version: 8.02
-u yes Support for large LUNs?
-z no Is DMAPI enabled?
-E yes Exact mtime mount option
-S no Suppress atime mount option
-d DiskR;DiskH;DiskK Disks in file system
-A yes Automatic mount option
-o none Additional mount options
-T /kora Default mount point
2) Réactiver / resynchroniser un disque
Si l’état d’un disque est différent de ready et up, il faut utiliser la commande mmchdisk pour activer le disque et mmrestripefs pour resynchroniser le disque.
Ces commandes mettent un certain temps (jusqu’à 30 minutes). Il est fortement conseiller d’attendre la fin d’une commande avant de lancer la suivante.
Si le disque DiskH est down mmchdisk orafs start -d DiskH
Si le disque DiskH est suspended mmchdisk orafs resume –d DiskH
Si la commande mmlsdisk renseigne des alertes :
Attention : Due to an earlier configuration change, the file system
may contain data that is at risk of being lost
==> mmrestripefs “Fs-name” –r –N mount (Fs-name ={orafs|orafs2})
relancer la commande mmlsdisk “Fs-name” pour conrôler le résultat
Attention : Due to an earlier configuration change, the file system
is no longer properly balanced.
==> mmrestripefs “Fs-name” -b -N mount (Fs-name ={orafs|orafs2})
relancer la commande mmlsdisk “Fs-name” pour conrôler le résultat