This shows you the differences between two versions of the page.
| Both sides previous revision Previous revision Next revision | Previous revision | ||
|
storage:svc_cmd [2021/01/26 16:32] manu |
storage:svc_cmd [2025/11/10 11:38] (current) manu [Replace SVC node] |
||
|---|---|---|---|
| Line 12: | Line 12: | ||
| There are several CLI usability features that I routinely find users are not aware of, so I thought I would share some of them here: | There are several CLI usability features that I routinely find users are not aware of, so I thought I would share some of them here: | ||
| + | |||
| + | ==== Hidden commands ==== | ||
| + | |||
| + | Get the real time of a node: | ||
| + | <cli prompt='>'> | ||
| + | IBM_FlashSystem:V5100-02:superuser>svqueryclock | ||
| + | Fri Nov 5 14:53:21 CET 2021 | ||
| + | </cli> | ||
| __Reference__ | __Reference__ | ||
| Line 315: | Line 323: | ||
| //Error// using replication metro mirror (PPRC) or global mirror --> change the system layout on all devices to have the same virtualization layer | //Error// using replication metro mirror (PPRC) or global mirror --> change the system layout on all devices to have the same virtualization layer | ||
| - | ==== Enable SVC NPIV ==== | + | ==== Enable SVC NPIV (required in v8.7) ==== |
| Here we talk about SVC FC port virtualization, host with NPIV is supported for a long time. | Here we talk about SVC FC port virtualization, host with NPIV is supported for a long time. | ||
| + | |||
| + | === Change from Not virtualized to Transitional === | ||
| Enabling NPIV target port functionality on a new cluster | Enabling NPIV target port functionality on a new cluster | ||
| After creating the cluster, but before starting to add hosts, issue | After creating the cluster, but before starting to add hosts, issue | ||
| - | chiogrp fctargetportmode transitional <IO group> | + | <cli> |
| + | chiogrp -fctargetportmode transitional <IO group> | ||
| + | </cli> | ||
| on each populated I/O group. | on each populated I/O group. | ||
| - | Once that is done, issue | + | |
| - | chiogrp fctargetportmode enabled <IO group> | + | === Update the zoning for ports connected to hosts === |
| + | |||
| + | Ensure that the hosts are using the NPIV ports for host I/O. To verify that you are logged in to these hosts with these NPIV ports, enter the following command | ||
| + | <cli> | ||
| + | lsfabric -host host_id_or_name | ||
| + | </cli> | ||
| + | Wait a minimum of 15 minutes before you move from transitional to enabled state. | ||
| + | |||
| + | === Change from Transitional to NPIV === | ||
| + | |||
| + | To change the NPIV setting to enabled, enter the following command: | ||
| + | <cli> | ||
| + | chiogrp -fctargetportmode enabled | ||
| + | </cli> | ||
| on each populated I/O group. | on each populated I/O group. | ||
| If using WWPN based zoning, issue the command to view the set of FC ports. | If using WWPN based zoning, issue the command to view the set of FC ports. | ||
| + | <cli> | ||
| lstargetportfc | lstargetportfc | ||
| + | </cli> | ||
| All ports that display virtualized=yes need to be zoned into hosts as described above. | All ports that display virtualized=yes need to be zoned into hosts as described above. | ||
| Line 451: | Line 479: | ||
| </cli> | </cli> | ||
| - | ==== Manualy upgrade controler firmware/software ==== | + | ==== Manualy upgrade controler firmware/software/reset drives status ==== |
| + | If you have pending upgrade on drives, for a long time and never end | ||
| + | <cli prompt='>'> | ||
| + | IBM_Storwize:v3700-Quorum:superuser>lsdriveupgradeprogress | ||
| + | id status estimated_completion_time | ||
| + | 0 not_scheduled | ||
| + | 7 not_scheduled | ||
| + | 51 completed 221013120044 | ||
| + | 52 completed 221013120550 | ||
| + | 53 completed 221013121056 | ||
| + | 54 completed 221013121602 | ||
| + | </cli> | ||
| + | |||
| + | Reset the status to **not_scheduled** | ||
| + | |||
| + | First copy firmware and test upgrade tool into /home/admin/upgrade | ||
| <cli prompt='>'> | <cli prompt='>'> | ||
| IBM_Storwize:v3700-Quorum:superuser>lsdumps -prefix /home/admin/upgrade | IBM_Storwize:v3700-Quorum:superuser>lsdumps -prefix /home/admin/upgrade | ||
| Line 458: | Line 501: | ||
| 0 IBM_StorwizeV3000_INSTALL_upgradetest_30.6 | 0 IBM_StorwizeV3000_INSTALL_upgradetest_30.6 | ||
| 1 IBM2072_INSTALL_7.8.1.11 | 1 IBM2072_INSTALL_7.8.1.11 | ||
| - | IBM_Storwize:v3700-Quorum:superuser>applysoftware -file IBM2072_INSTALL_7.8.1.11 | + | </cli> |
| - | CMMVC8542E The System Update Checker has not been run prior to updating. | + | |
| + | Second install the upgrade test else firmware can't be installed | ||
| + | <cli prompt='>'> | ||
| + | IBM_Storwize:v3700-Quorum:superuser>applysoftware -file IBM_StorwizeV3000_INSTALL_upgradetest_30.6 | ||
| + | </cli> | ||
| + | |||
| + | Third start test utility | ||
| + | <cli prompt='>'> | ||
| IBM_Storwize:v3700-Quorum:superuser>svcupgradetest -v 7.8.1.11 | IBM_Storwize:v3700-Quorum:superuser>svcupgradetest -v 7.8.1.11 | ||
| svcupgradetest version 30.6 | svcupgradetest version 30.6 | ||
| Line 484: | Line 534: | ||
| clear_dump(nodes[node]._id, dump['filename']) | clear_dump(nodes[node]._id, dump['filename']) | ||
| File "/home/utilities/svcupgradetest/delete_old_dumps.py", line 75, in clear_dump | File "/home/utilities/svcupgradetest/delete_old_dumps.py", line 75, in clear_dump | ||
| - | command = ['/compass/bin/svctask', 'cleardumps', '-prefix', '{}/{}'.format(DUMPS_DIR, dump_name), node_id] | + | command = ['/compass/bin/svctask', 'cleardumps', '-prefix', '{}/{}'.format(DUMPS_DIR, dump_name), |
| + | node_id] | ||
| ValueError: zero length field name in format | ValueError: zero length field name in format | ||
| https://www.ibm.com/support/knowledgecenter/en/STLM5A/com.ibm.storwize.v3700.710.doc/v3500_qi_installing_memory_module.html | https://www.ibm.com/support/knowledgecenter/en/STLM5A/com.ibm.storwize.v3700.710.doc/v3500_qi_installing_memory_module.html | ||
| Line 498: | Line 549: | ||
| One issue will prevent the system being | One issue will prevent the system being | ||
| updated to 7.8.1.11 until it has been resolved | updated to 7.8.1.11 until it has been resolved | ||
| + | </cli> | ||
| - | IBM_Storwize:v3700-Quorum:superuser> | + | If it keeps schedules pending, then restart **tomcat** |
| + | <cli prompt='>'> | ||
| + | IBM_Storwize:v3700:superuser> satask restartservice -service tomcat | ||
| + | IBM_Storwize:v3700:superuser> satask restartservice -service cimom | ||
| + | </cli> | ||
| + | Wait about 5 min and then retry the upgrade | ||
| + | Now you are ready to upgrade the new firmware | ||
| + | <cli prompt='>'> | ||
| IBM_Storwize:v3700-Quorum:superuser>applysoftware -file IBM2072_INSTALL_7.8.1.11 | IBM_Storwize:v3700-Quorum:superuser>applysoftware -file IBM2072_INSTALL_7.8.1.11 | ||
| - | CMMVC8542E The System Update Checker has not been run prior to updating. | ||
| - | IBM_Storwize:v3700-Quorum:superuser>applysoftware -file IBM_StorwizeV3000_INSTALL_upgradetest_30.6 | ||
| </cli> | </cli> | ||
| - | ====== SVC changing topology for enhanced strech cluster ====== | + | |
| + | Use the following command to check the upgrade progress, take care it will stay on 0, after it will switch to 50% after reboot of first controller... | ||
| + | <cli prompt='>'> | ||
| + | IBM_Storwize:v3700-Quorum:superuser> lsupdate | ||
| + | </cli> | ||
| + | |||
| + | ====== SVC changing topology for enhanced stretch cluster ====== | ||
| == Configuration steps == | == Configuration steps == | ||
| Line 591: | Line 654: | ||
| mkvdisk -mdiskgrp pool1 -iogrp io_grp0 -size 1099511627776 -unit b -name DCAV-PRTSM02-99 | mkvdisk -mdiskgrp pool1 -iogrp io_grp0 -size 1099511627776 -unit b -name DCAV-PRTSM02-99 | ||
| </cli> | </cli> | ||
| + | |||
| + | ===== Users management ===== | ||
| + | |||
| + | ==== Groups ==== | ||
| + | |||
| + | List groups: | ||
| + | <cli prompt='>'> | ||
| + | IBM_FlashSystem:fs7300:user01> lsusergrp | ||
| + | id name role remote owner_id owner_name | ||
| + | 0 SecurityAdmin SecurityAdmin no | ||
| + | 1 Administrator Administrator no | ||
| + | 2 CopyOperator CopyOperator no | ||
| + | 3 Service Service no | ||
| + | 4 Monitor Monitor no | ||
| + | 5 RestrictedAdmin RestrictedAdmin no | ||
| + | </cli> | ||
| + | |||
| + | ==== Users ==== | ||
| + | |||
| + | Create a new user with a password: | ||
| + | <cli prompt='>'> | ||
| + | IBM_FlashSystem:fs7300:user01> mkuser -name user02 -usergrp SecurityAdmin -password xxxxxx | ||
| + | </cli> | ||
| + | |||
| + | List all users | ||
| + | <cli prompt='>'> | ||
| + | IBM_FlashSystem:fs7300:user01> lsuser | ||
| + | id name password ssh_key remote usergrp_id usergrp_name | ||
| + | 0 superuser yes yes no 0 SecurityAdmin | ||
| + | 1 user02 yes no no 0 SecurityAdmin | ||
| + | 3 report no yes no 4 Monitor | ||
| + | </cli> | ||
| + | |||
| + | Create a user with ssh key axchange | ||
| + | |||
| + | * Copy the public key from you user to **/home/admin/upgrade** | ||
| + | <cli prompt='#'> | ||
| + | [root@lnx01]/storage/.ssh # scp /storage/.ssh/id_ecdsa.pub user01@fs7300:/home/admin/upgrade | ||
| + | Password: | ||
| + | id_ecdsa.pub 100% 284 411.6KB/s 00:00 | ||
| + | </cli> | ||
| + | |||
| + | * Create a user and import the ssh key | ||
| + | <cli prompt='>'> | ||
| + | IBM_FlashSystem:fs7300:user01>mkuser -name report -usergrp Monitor -keyfile /home/admin/upgrade/id_ecdsa.pub | ||
| + | User, id [8], successfully created | ||
| + | </cli> | ||
| + | |||
| + | |||
| + | |||