This shows you the differences between two versions of the page.
Next revision | Previous revision | ||
zfs [2016/05/09 19:06] k2patel created |
zfs [2019/02/13 20:46] k2patel [Solaris] |
||
---|---|---|---|
Line 4: | Line 4: | ||
- | To check zpool status | + | === To check zpool status === |
<code bash> | <code bash> | ||
zpool status [<volume>] | zpool status [<volume>] | ||
</code> | </code> | ||
- | To clear error on drive, which you think reported wrongly. | + | === To clear error on drive, which you think reported wrongly. === |
<code bash> | <code bash> | ||
zpool clear <volume name> | zpool clear <volume name> | ||
</code> | </code> | ||
- | To add Zeus Drive | + | === To add Zeus Drive === |
<code bash> | <code bash> | ||
zpool add performance log <drive> | zpool add performance log <drive> | ||
</code> | </code> | ||
- | Adding spare to volume | + | === Adding spare to volume === |
<code bash> | <code bash> | ||
zpool add performance spare <drive> | zpool add performance spare <drive> | ||
</code> | </code> | ||
- | Remove attached spare drive to volume | + | === Remove attached spare drive to volume === |
<code bash> | <code bash> | ||
zpool detach <volume> <Drive> | zpool detach <volume> <Drive> | ||
</code> | </code> | ||
- | Removing drives from volume (require to remove spare drives after raid rebuilt) | + | === Removing drives from volume (require to remove spare drives after raid rebuilt) === |
<code bash> | <code bash> | ||
zpool remove <Volume> <Drive> | zpool remove <Volume> <Drive> | ||
</code> | </code> | ||
+ | |||
+ | === List all devices / drive === | ||
+ | |||
+ | <code bash> | ||
+ | sas2ircu 0 DISPLAY | ||
+ | </code> | ||
+ | |||
+ | === ZFS Display drive information by serial number === | ||
+ | |||
+ | <code bash> | ||
+ | sas2ircu 0 DISPLAY | grep -B 9 -A 4 <Serial Number> | ||
+ | </code> | ||
+ | |||
+ | === Freenas Blink Drives === | ||
+ | |||
+ | <code bash> | ||
+ | sas2ircu 0 locate 3:7 ON | ||
+ | </code> | ||
+ | |||
+ | === Offline Drive === | ||
+ | |||
+ | <code bash> | ||
+ | zpool offline hermes 15935140517898495532 | ||
+ | </code> | ||
+ | |||
+ | === Replace disk in ZFS Pool === | ||
+ | |||
+ | <code bash> | ||
+ | zpool replace hermes 15935140517898495532 /dev/disk/by-id/ata-ST3500320AS_9QM03ATQ | ||
+ | </code> | ||
+ | |||
+ | ==== Nexenta ==== | ||
+ | === Disk / Lun === | ||
+ | To blink drive use following command in NMC | ||
+ | <code bash> | ||
+ | show lun c8t5000C5005785138Bd0 blink -y | ||
+ | </code> | ||
+ | === Nexenta Mgmt. === | ||
+ | |||
+ | To create nexenta collector report, Run following command from root prompt | ||
+ | <code bash> | ||
+ | nexenta-collector --no-upload | ||
+ | </code> | ||
+ | Printing Each JBOD Slotmap | ||
+ | <code bash> | ||
+ | nmc -c "show jbod jbod:1 slotmap" | less | ||
+ | </code> | ||
+ | Check HA Status from command line | ||
+ | <code bash> | ||
+ | /opt/HAC/RSF-1/bin/rsfcli status | ||
+ | </code> | ||
+ | Synchronize Disk Location | ||
+ | <code bash> | ||
+ | nmc -c "lunsync -r -y" | ||
+ | </code> | ||
+ | Migrate Pool to another server | ||
+ | <code bash> | ||
+ | /opt/HAC/RSF-1/bin/rsfcli -i0 move <volume> <hostname> | ||
+ | </code> | ||
+ | ==== Solaris ==== | ||
+ | |||
+ | List all faults on system | ||
+ | <code bash> | ||
+ | fmadm faulty | ||
+ | </code> | ||
+ | Clearing Faults | ||
+ | <code bash> | ||
+ | fmadm repair <faulty_id> | ||
+ | </code> | ||
+ | Checking Service status | ||
+ | <code bash> | ||
+ | svcs nm{s,v,cd} dbus rmvolmgr nmdtrace | ||
+ | </code> | ||
+ | Restarting Services | ||
+ | <code bash> | ||
+ | svcadm restart nms | ||
+ | </code> | ||
+ | Loading unloading zfs-diagnosis module | ||
+ | <code bash> | ||
+ | fmadm load /usr/lib/fm/fmd/plugins/zfs-diagnosis.so | ||
+ | </code> | ||
+ | Clearning dangling dev link | ||
+ | <code bash> | ||
+ | devfsadm -Cv | ||
+ | </code> | ||
+ | |||
+ | ==== Check status of ARC and L2 ==== | ||
+ | <code bash> | ||
+ | kstat -p zfs:0:arcstats | ||
+ | </code> | ||
+ | |||
+ |