5 vi /etc/hosts 6 mount /dev/sr0 /mnt 7 mv /etc/yum.repos.d/CentOS-Base.repo /etc/yum.repos.d/CentOS-Base.repo.old 8 yum -y install ntp 9 yum -y install gcc 10 vi /etc/selinux/config 11 systemctl stop firewalld 12 systemctl disable firewalld 20 df 21 bootlist -m normal -o 22 multipath -ll 23 bootlist -m normal sda sdg 25 systemctl start ntpd 26 systemctl enable ntpd 27 systemctl disable firewalld 34 cd rsct/ 35 ll 36 rpm -Uhv src-3.2.1.3-16106.ppc64.rpm 40 yum -y install ksh 41 yum -y install sg3_utils 42 yum -y install nfs-utils 43 yum -y install lshw 44 yum -y install net-tools 45 yum -y install telnet 46 yum -y install psmisc 48 rpm -Uhv rsct.core-3.2.1.3-16106.ppc64.rpm rsct.core.utils-3.2.1.3-16106.ppc64.rpm 49 rpm -Uhv src-3.2.1.3-16106.ppc64.rpm 50 rpm -Uhv rsct.core-3.2.1.3-16106.ppc64.rpm rsct.core.utils-3.2.1.3-16106.ppc64.rpm 51 rpm -Uhv devices.chrp.base.ServiceRM-2.5.1.1-16106.ppc64.rpm DynamicRM-2.0.3-1.ppc64.rpm 52 rpm -Uhv nmon-16e-0.el7.ppc64.rpm 53 rpm -Uhv ppc64-diag-2.6.10-1.el7.ppc64.rpm lsvpd-1.7.7-1.el7.ppc64.rpm libvpd-2.2.5-1.el7.ppc64.rpm 54 bootlist -m normal -o 55 lssrc -a 56 ntpq -p 63 cd 4.2.1 65 ./Spectrum_Scale_Protocols_Standard-4.2.1.1-ppc64-Linux-install 66 cd /usr/lpp/mmfs/4.2.1.1/gpfs_rpms 74 rpm -Uhv gpfs.base-4.2.1-1.ppc64.rpm gpfs.docs-4.2.1-1.noarch.rpm gpfs.ext-4.2.1-1.ppc64.rpm gpfs.gpl-4.2.1-1.noarch.rpm gpfs.gskit-8.0.50-57.ppc64.rpm gpfs.msg.en_US-4.2.1-1.noarch.rpm 75 yum install m4 76 rpm -Uhv gpfs.base-4.2.1-1.ppc64.rpm gpfs.docs-4.2.1-1.noarch.rpm gpfs.ext-4.2.1-1.ppc64.rpm gpfs.gpl-4.2.1-1.noarch.rpm gpfs.gskit-8.0.50-57.ppc64.rpm gpfs.msg.en_US-4.2.1-1.noarch.rpm 77 ls 78 rpm -Uhv gpfs.java-4.2.1-1.ppc64.rpm gpfs.gui-4.2.1-1.noarch.rpm 79 yum -y install nc bzip2 postgresql-server 80 rpm -Uhv gpfs.java-4.2.1-1.ppc64.rpm 81 cd .. 82 ll 83 cd zimon_rpms 84 ll 85 rpm -Uhv gpfs.gss.pmsensors-4.2.1-1.el7.ppc64.rpm gpfs.gss.pmcollector-4.2.1-1.el7.ppc64.rpm 86 yum -y install boost-regex [root@rhlabh1 zimon_rpms]# rpm -Uhv gpfs.gss.pmcollector-4.2.1-1.el7.ppc64.rpm gpfs.gss.pmsensors-4.2.1-1.el7.ppc64.rpm [root@rhlabh1 gpfs_rpms]# rpm -Uhv gpfs.gui-4.2.1-1.noarch.rpm [root@rhlabh1 gpfs_rpms]# export LINUX_DISTRIBUTION=REDHAT_AS_LINUX [root@rhlabh1 gpfs_rpms]# mmbuildgpl --build-package yum -y install kernel-devel gcc-c++ rpm-build [root@rhlabh1 gpfs_rpms]# rpm -Uhv /root/rpmbuild/RPMS/ppc64/gpfs.gplbin-3.10.0-327.el7.ppc64-4.2.1-1.ppc64.rpm # yum -y install libarchive rpm -Uhv smb_rpms/gpfs.smb-4.3.11_gpfs_21-3.el7.ppc64.rpm # yum -y install libwbclient cd /usr/lpp/mmfs/4.2.1.1/ganesha_rpms rpm -Uhv nfs-ganesha-gpfs-2.3.2-0.ibm24.el7.ppc64.rpm nfs-ganesha-2.3.2-0.ibm24.el7.ppc64.rpm nfs-ganesha-utils-2.3.2-0.ibm24.el7.ppc64.rpm # yum -y install gpfs.protocols-support [root@rhlabh1 yum.repos.d]# cat cdrom.repo gpfs.repo [cdrom] name=CDROM Repo baseurl=file:///mnt enabled=1 gpgcheck=0 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-redhat-release [root@rhlabh1 yum.repos.d]# cat gpfs.repo [gpfs_base] name=GPFS_base baseurl=file:/usr/lpp/mmfs/4.2.1.1/gpfs_rpms/ enabled=1 gpgcheck=0 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-redhat-release [gpfs_obj] name=GPFS_obj baseurl=file:/usr/lpp/mmfs/4.2.1.1/object_rpms/ enabled=1 gpgcheck=0 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-redhat-release [gpfs_nfs] name=GPFS_nfs baseurl=file:/usr/lpp/mmfs/4.2.1.1/ganesha_rpms/ enabled=1 gpgcheck=0 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-redhat-release [gpfs_smb] name=GPFS_smb baseurl=file:/usr/lpp/mmfs/4.2.1.1/smb_rpms/ enabled=1 gpgcheck=0 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-redhat-release [gpfs_zimon] name=GPFS_zimon baseurl=file:/usr/lpp/mmfs/4.2.1.1/zimon_rpms/ enabled=1 gpgcheck=0 gpgkey=file:///etc/pki/rpm-gpg/RPM-GPG-KEY-redhat-release # yum -y install createrepo
[root@rhlabh1 ~]# cat nsd_disk.lst %nsd: device=/dev/dm-11 nsd=nsd_t1_h servers=rhlabh1,rhlabr1 usage=dataAndMetadata failureGroup=1 pool=system %nsd: device=/dev/dm-10 nsd=nsd_t2_h servers=rhlabh1,rhlabr1 usage=dataOnly failureGroup=1 pool=slow %nsd: device=/dev/dm-12 nsd=nsd_t1_r servers=rhlabh1,rhlabr1 usage=dataAndMetadata failureGroup=2 pool=system %nsd: device=/dev/dm-13 nsd=nsd_t2_r servers=rhlabh1,rhlabr1 usage=dataOnly failureGroup=2 pool=slow [root@rhlabh1 ~]# mmcrfs gpfs1 -F nsd_disk.lst -k nfs4 -T /gpfs1 -m 2 -M2 -r2 -R2 The following disks of gpfs1 will be formatted on node rhlabr1: nsd_t1_h: size 30720 MB nsd_t2_h: size 30720 MB nsd_t1_r: size 30720 MB nsd_t2_r: size 30720 MB Formatting file system ... Disks up to size 280 GB can be added to storage pool system. Disks up to size 280 GB can be added to storage pool slow. Creating Inode File Creating Allocation Maps Creating Log Files Clearing Inode Allocation Map Clearing Block Allocation Map Formatting Allocation Map for storage pool system Formatting Allocation Map for storage pool slow Completed creation of file system /dev/gpfs1. mmcrfs: Propagating the cluster configuration data to all affected nodes. This is an asynchronous process. [root@rhlabh1 ~]# mmlsfs all File system attributes for /dev/gpfs1: ====================================== flag value description ------------------- ------------------------ ----------------------------------- -f 8192 Minimum fragment size in bytes -i 4096 Inode size in bytes -I 16384 Indirect block size in bytes -m 2 Default number of metadata replicas -M 2 Maximum number of metadata replicas -r 2 Default number of data replicas -R 2 Maximum number of data replicas -j cluster Block allocation type -D nfs4 File locking semantics in effect -k nfs4 ACL semantics in effect -n 32 Estimated number of nodes that will mount file system -B 262144 Block size -Q none Quotas accounting enabled none Quotas enforced none Default quotas enabled --perfileset-quota No Per-fileset quota enforcement --filesetdf No Fileset df enabled? -V 15.01 (4.2.0.0) File system version --create-time Fri Sep 30 16:11:20 2016 File system creation time -z No Is DMAPI enabled? -L 33554432 Logfile size -E Yes Exact mtime mount option -S No Suppress atime mount option -K whenpossible Strict replica allocation option --fastea Yes Fast external attributes enabled? --encryption No Encryption enabled? --inode-limit 123008 Maximum number of inodes --log-replicas 0 Number of log replicas --is4KAligned Yes is4KAligned? --rapid-repair Yes rapidRepair enabled? --write-cache-threshold 0 HAWC Threshold (max 65536) -P system;slow Disk storage pools in file system -d nsd_t1_h;nsd_t2_h;nsd_t1_r;nsd_t2_r Disks in file system -A yes Automatic mount option -o none Additional mount options -T /gpfs1 Default mount point --mount-priority 0 Mount priority [root@rhlabh1 ~]# mmlsfs all -d File system attributes for /dev/gpfs1: ====================================== flag value description ------------------- ------------------------ ----------------------------------- -d nsd_t1_h;nsd_t2_h;nsd_t1_r;nsd_t2_r Disks in file system [root@rhlabh1 ~]# mmmount all Fri Sep 30 16:16:04 CEST 2016: mmmount: Mounting file systems ... [root@rhlabh1 ~]# mmlsdisk /dev/gpfs1 disk driver sector failure holds holds storage name type size group metadata data status availability pool ------------ -------- ------ ----------- -------- ----- ------------- ------------ ------------ nsd_t1_h nsd 512 1 Yes Yes ready up system nsd_t2_h nsd 512 1 No Yes ready up slow nsd_t1_r nsd 512 2 Yes Yes ready up system nsd_t2_r nsd 512 2 No Yes ready up slow [root@rhlabh1 ~]# mmlsdisk /dev/gpfs1 -M Disk name IO performed on node Device Availability ------------ ----------------------- ----------------- ------------ nsd_t1_h localhost /dev/dm-11 up nsd_t2_h localhost /dev/dm-10 up nsd_t1_r localhost /dev/dm-12 up nsd_t2_r localhost /dev/dm-13 up [root@rhlabh1 ~]# df Filesystem 1K-blocks Used Available Use% Mounted on /dev/mapper/ca_rhlabh1-root 23089088 4019816 19069272 18% / devtmpfs 1778752 0 1778752 0% /dev tmpfs 1815168 0 1815168 0% /dev/shm tmpfs 1815168 81344 1733824 5% /run tmpfs 1815168 0 1815168 0% /sys/fs/cgroup /dev/mapper/mpatha2 508588 173280 335308 35% /boot tmpfs 363072 0 363072 0% /run/user/0 gpfs1 125829120 3387392 122441728 3% /gpfs1 [root@rhlabh1 ~]# mmdf /dev/gpfs1 disk disk size failure holds holds free KB free KB name in KB group metadata data in full blocks in fragments --------------- ------------- -------- -------- ----- -------------------- ------------------- Disks in storage pool: system (Maximum disk size allowed is 261 GB) nsd_t1_h 31457280 1 Yes Yes 29829632 ( 95%) 584 ( 0%) nsd_t1_r 31457280 2 Yes Yes 29829632 ( 95%) 584 ( 0%) ------------- -------------------- ------------------- (pool total) 62914560 59659264 ( 95%) 1168 ( 0%) Disks in storage pool: slow (Maximum disk size allowed is 261 GB) nsd_t2_h 31457280 1 No Yes 31391232 (100%) 376 ( 0%) nsd_t2_r 31457280 2 No Yes 31391232 (100%) 376 ( 0%) ------------- -------------------- ------------------- (pool total) 62914560 62782464 (100%) 752 ( 0%) ============= ==================== =================== (data) 125829120 122441728 ( 97%) 1920 ( 0%) (metadata) 62914560 59659264 ( 95%) 1168 ( 0%) ============= ==================== =================== (total) 125829120 122441728 ( 97%) 1920 ( 0%) Inode Information ----------------- Number of used inodes: 4038 Number of free inodes: 118970 Number of allocated inodes: 123008 Maximum number of inodes: 123008 [root@rhlabh1 gpfs]# mmlsconfig Configuration data for cluster gpfs_test.rhlabh1: ----------------------------------------------------- clusterName gpfs_test.rhlabh1 clusterId 9668046452208786064 autoload yes profile gpfsProtocolDefaults dmapiFileHandleSize 32 minReleaseLevel 4.2.1.0 ccrEnabled yes cipherList AUTHONLY maxblocksize 16M [cesNodes] maxMBpS 5000 numaMemoryInterleave yes enforceFilesetQuotaOnRoot yes workerThreads 512 [common] tscCmdPortRange 60000-61000 adminMode central File systems in cluster gpfs_test.rhlabh1: ---------------------------------------------- /dev/gpfs1
Creation of a file system or fileset or path for a CES shared root, and creation of an object fileset The installation toolkit uses a shared root storage area to install the protocols on each node. This storage is also used by NFS and object protocols to maintain system data associated with the cluster integration we provide. This storage can be a subdirectory in an existing file system or it can be a file system on its own. Once this option is set, changing it will requires a restart of GPFS. You can use the installation toolkit to set up this CES shared root storage area if you use the toolkit for GPFS installation and file system creation. For more information, see “Using the installation toolkit to perform installation tasks: Explanations and examples” on page 271. However, if you want to set up shared root before launching the installation toolkit, the following steps can be used: 1. Create a file system or fileset for shared root. Size must be at least 4 GB. 2. Use the following command: mmchconfig cesSharedRoot=path_to_the_filesystem/fileset_created_in_step_1 For Object, the installation toolkit creates an independent fileset in the GPFS file system that you name.
[root@rhlabh1 gpfs]# mmchconfig cesSharedRoot=/gpfs1 mmchconfig: Command successfully completed mmchconfig: Propagating the cluster configuration data to all affected nodes. This is an asynchronous process. [root@rhlabh1 gpfs]# mmlsconfig Configuration data for cluster gpfs_test.rhlabh1: ----------------------------------------------------- clusterName gpfs_test.rhlabh1 clusterId 9668046452208786064 autoload yes profile gpfsProtocolDefaults dmapiFileHandleSize 32 minReleaseLevel 4.2.1.0 ccrEnabled yes cipherList AUTHONLY maxblocksize 16M [cesNodes] maxMBpS 5000 numaMemoryInterleave yes enforceFilesetQuotaOnRoot yes workerThreads 512 [common] tscCmdPortRange 60000-61000 cesSharedRoot /gpfs1 adminMode central File systems in cluster gpfs_test.rhlabh1: ---------------------------------------------- /dev/gpfs1 [root@rhlabh1 gpfs]# mmchnode --ces-enable -N rhlabh1,rhlabr1 Fri Sep 30 17:12:30 CEST 2016: mmchnode: Processing node rhlabr1 Fri Sep 30 17:12:50 CEST 2016: mmchnode: Processing node rhlabh1 mmchnode: Propagating the cluster configuration data to all affected nodes. This is an asynchronous process. [root@rhlabh1 gpfs]# mmlscluster GPFS cluster information ======================== GPFS cluster name: gpfs_test.rhlabh1 GPFS cluster id: 9668046452208786064 GPFS UID domain: gpfs_test.rhlabh1 Remote shell command: /usr/bin/ssh Remote file copy command: /usr/bin/scp Repository type: CCR Node Daemon node name IP address Admin node name Designation --------------------------------------------------------------------- 1 rhlabh1 10.10.10.103 rhlabh1 quorum-manager-perfmon 2 rhlabr1 10.10.10.104 rhlabr1 quorum-manager-perfmon [root@rhlabr1 /]# mmlscluster --ces GPFS cluster information ======================== GPFS cluster name: gpfs_test.rhlabh1 GPFS cluster id: 9668046452208786064 Cluster Export Services global parameters ----------------------------------------- Shared root directory: /gpfs1 Enabled Services: None Log level: 0 Address distribution policy: even-coverage Node Daemon node name IP address CES IP address list ----------------------------------------------------------------------- 1 rhlabh1 10.10.10.103 None 2 rhlabr1 10.10.10.104 None [root@rhlabh1 gpfs]# mmces service enable NFS rhlabr1: Redirecting to /bin/systemctl start rpcbind.service rhlabr1: Redirecting to /bin/systemctl start nfs-ganesha.service rhlabh1: Redirecting to /bin/systemctl start rpcbind.service rhlabh1: Redirecting to /bin/systemctl start nfs-ganesha.service mmchconfig: Command successfully completed mmchconfig: Propagating the cluster configuration data to all affected nodes. This is an asynchronous process. [root@rhlabh1 gpfs]# mmces service enable SMB rhlabr1: Redirecting to /bin/systemctl start gpfs-ctdb.service rhlabh1: Redirecting to /bin/systemctl start gpfs-ctdb.service rhlabr1: Wait for ctdb to become ready. State=STARTUP rhlabh1: Wait for ctdb to become ready. State=STARTUP rhlabr1: Wait for ctdb to become ready. State=STARTUP rhlabh1: Wait for ctdb to become ready. State=STARTUP rhlabr1: mmchconfig: Command successfully completed rhlabh1: mmchconfig: Unable to obtain the GPFS configuration file lock. Retrying ... rhlabr1: mmchconfig: Propagating the cluster configuration data to all rhlabr1: affected nodes. This is an asynchronous process. rhlabr1: Redirecting to /bin/systemctl start gpfs-smb.service rhlabh1: mmchconfig: Lock creation successful. rhlabh1: mmchconfig: Command successfully completed rhlabh1: mmchconfig: Propagating the cluster configuration data to all rhlabh1: affected nodes. This is an asynchronous process. rhlabh1: Redirecting to /bin/systemctl start gpfs-smb.service mmchconfig: Command successfully completed mmchconfig: Propagating the cluster configuration data to all affected nodes. This is an asynchronous process. [root@rhlabr1 /]# mmlscluster --ces GPFS cluster information ======================== GPFS cluster name: gpfs_test.rhlabh1 GPFS cluster id: 9668046452208786064 Cluster Export Services global parameters ----------------------------------------- Shared root directory: /gpfs1 Enabled Services: NFS SMB Log level: 0 Address distribution policy: even-coverage Node Daemon node name IP address CES IP address list ----------------------------------------------------------------------- 1 rhlabh1 10.10.10.103 None 2 rhlabr1 10.10.10.104 None [root@rhlabh1 gpfs]# mmces service list --all Enabled services: NFS SMB rhlabr1: NFS is running, SMB is running rhlabh1: NFS is running, SMB is running [root@rhlabh1 gpfs]# mmnfs export add /gpfs1 mmnfs export list mmcesfuncs.sh: Current authentication: none is invalid. This operation can not be completed without correct Authentication configuration. Configure authentication using: mmuserauth mmnfs export add: Command failed. Examine previous error messages to determine cause. [root@rhlabh1 gpfs]# mmnfs export list mmnfs export list [E]: Cannot list any exports. There are no exports to show. [root@rhlabh1 gpfs]# mmuserauth Usage: mmuserauth service Configure the system authentication service [root@rhlabh1 gpfs]# mmuserauth service Usage: mmuserauth service create Create new authentication configuration mmuserauth service check Check authentication configuration and correct if needed mmuserauth service list List current authentication configuration mmuserauth service remove Remove current authentication configuration [root@rhlabh1 gpfs]# mmuserauth service list FILE access not configured PARAMETERS VALUES ------------------------------------------------- OBJECT access not configured PARAMETERS VALUES ------------------------------------------------- [root@rhlabh1 gpfs]# mmuserauth service create --data-access-method file --type userdefined File authentication configuration completed successfully. [root@rhlabh1 gpfs]# FILE access configuration : USERDEFINED PARAMETERS VALUES ------------------------------------------------- OBJECT access not configured PARAMETERS VALUES ------------------------------------------------- [root@rhlabh1 gpfs]# mmnfs export add '/gpfs01/backupdb' -c '10.1.0.0/16(Access_Type=RW,squash=root_squash,protocols=3:4)' [root@gpfs01 scripts]# mmchdisk gpfs01lv change -d "GPFS_NSD_META01:::metadataOnly" Attention: Disk parameters were changed. Use the mmrestripefs command with the -r option to relocate data and metadata. Verifying file system configuration information ... mmchdisk: Propagating the cluster configuration data to all affected nodes. This is an asynchronous process. [root@gpfs01 scripts]# mmrestripefs gpfs01lv -r Scanning file system metadata, phase 1 ... Scan completed successfully. Scanning file system metadata, phase 2 ... Scan completed successfully. Scanning file system metadata, phase 3 ... Scan completed successfully. Scanning file system metadata, phase 4 ... Scan completed successfully. Scanning user file metadata ... 100.00 % complete on Thu Apr 26 15:02:24 2018 ( 606208 inodes with total 3879 MB data processed) Scan completed successfully. [root@gpfs01 scripts]# mmlsdisk gpfs01lv disk driver sector failure holds holds storage name type size group metadata data status availability pool ------------ -------- ------ ----------- -------- ----- ------------- ------------ ------------ GPFS_NSD_DATA01 nsd 512 -1 No Yes ready up system GPFS_NSD_DATA02 nsd 512 -1 No Yes ready up system GPFS_NSD_DATA03 nsd 512 -1 No Yes ready up system GPFS_NSD_DATA04 nsd 512 -1 No Yes ready up system GPFS_NSD_DATA05 nsd 512 -1 No Yes ready up system GPFS_NSD_DATA06 nsd 512 -1 No Yes ready up system GPFS_NSD_META01 nsd 512 -1 Yes No ready up system GPFS_NSD_META02 nsd 512 -1 Yes No ready up system GPFS_NSD_META03 nsd 512 -1 Yes No ready up system GPFS_NSD_META04 nsd 512 -1 Yes No ready up system [root@gpfs01 scripts]#