NCDA NSO-163 revision notes
Commands
Sysconfig –r shows disk raid groups
Sysconfig –v shows storage system disk firmware
Sysconfig –a shows storage systems ip information
Disk –o hostname
Disk –s match disk id
Disk –n show unowned disks
Disk –v will list all disks
Disk –a any storage
Disk assign 0ad.2 fas2020 assigns the disk 0ad.2 to fas2020
Sync mirror
Cf giveback fails back over to th failed system once it is
back up
Cf giveback –f forces the give back but terminates dumps,
cifs sessions, parity checks raid reconstructions and raid scrubs
Cf forcegiveback will force a giveback but could cause loss
of data use only as a last resort
Cf giveback will fail if ifconfig emulates failed filer
Halt –f will halt the filer without takeover
Cf takeover will takeover a filer
Cf forcetaker use this if the cf takeover fails
Sysconfig –v show disk shelf and ontap versions
Sync mirror – mirrors volumes and aggregates disk need to be
evenly divided between the two plexes. Disk need to come from different pools.
Needs the same number of disks in each pool. Disks are slected on bytes per
sector then on disk size. If there is no equivalent size disk the filer will
grab a larger disk and down size it.
Sysconfig –r will determine if the plexes are configured
correctly (will show the plexes online normal
Aggr create aggr1 –m 6 will create a new mirrored aggregate
Sysconfig –r
Vol status –r
Aggr status –r
Vol create volx –m 6 will create a new mirrored volume
-n will preview the output and show what disks ontap has
chosen
Fcal (fibre card slots) slots 1-7 pool 0 slots 8-11 pool 1
0a 0b pool0 0c 0d pool 1
Sync mirror needs to be licensed on both hosts
Split mirror command will split the mirrored volumes
creating 2 new unmirrored volumes
Aggr split aggr1\plexname aggr_split
Vol split vol1\plexname aggr1_split
Mirror needs to be online operational before you do the
split
To resync the mirror
Vol mirror vol1 –v vol2 sure yes
Sysconfig –r checks spare disks are in the correct pools
Stretched metro
cluster
Fabric attached metro cluster can be up to 30km apart
Stretch metro cluster license needs cluster cluster_remote
syncmirror_local needs 6.4.2 or higher
Max stretch cluster 500m with om3 cables 300m with fc cables
all 2gb
270m om3 cables at 4gb
Fabric metro cluster max 100km 2gb
With 4gb 55km
Fabric have 2 banks a 4 quadrants in each fibre switch if
host is plugged into bank 1 it owns disks plugged into bank 2
Remote plexes use disk pool 1
Hardware needed for syncmirror disks shelves, adapters
cabling
Cluster config checker will check license mismatch options
mismatch incorrect network ontap version mismatch cf-config-checker-cgi
Other
WAFL overhead 10% (Write Anywhere File Layout)
Default vol snap reserve 20%
Default aggregate reserve 5%
Vol options vol0 nosnapdir on will deny access to snapshots
to users
Options nfs.hide_snapshot on will hide the snapshots in an
nfs volume
Options cifs.show_snapshot off will hide cifs shares to
users
Vol status verify a volume is online
Snap list vol0 – lists all snapshots on vol0
Snap restore –t file /vol/vol0 –s vol0_snap /vol/vol0 – will
restore file to vol0 from snapshot vol0_snap
Snap restore –r will
change the destination directory
Snapmirror volume restore can do source volumes but cannot
do destination volumes
Run df on destination to check data size
Snapshot copies
Snapshot copies are a point in time copy of a volume or
aggregate and read only
255 snap shot copies
Snap default schedule vol0: 0 2 6 @ 8, 12, 16, 20
Default snap reserve is 20%
To check snap shot rate of change use the snap delta command
- snap delta vol0
Allow cifs snap access – options cifs.show_snapshots on
Df –h shows snap reserve
Snap sched vol0 0 0 0 no snap sched
Vol options vol0 nosnap on disables snap schedule
Snap create
Snap delete
Snap delete –a will delete all snaps in that volume
Snaprestore
Use snaprestore if files are virus infected need to revert
to a software baseline, need to restore 300mb of data and only have 10mb spare.
Snaprestore does not restore snapshot schedules, volume
options, raid group size, max files per volume
To restore a file use the following command snaprestore –t
file –s vol0_snap /vol/newvol
Cannot reverse a snaprestore
Snapmirror
It is recommended not to change the visibility_interval the
default is 3 seconds.
Edit snapmirror.conf file to change mirror from async to
sync or semi-sync
Vol0 fas2020:vol0 – sync (sync mirror)
Vol fas2020:vol0 – semi-sync (semi-sync mirror)
Vol fas2020:vol0 fas2021 – 0-55/5 (async)
Changes to the snamirror.comf file on the destination will
disrupt the mirror and fail.
Options snapmirror.log.enable on will enable snapmirror
logging
/etc/logs/snapmirror stores the snapmirror logs on the
source root volume
Fas1:vol1 fas2:vol2 – 0-55/5 * * * this is an async schedule
not supported for sync or semi-sync
Fas1:vol1 fas2:vol2 semi-sync
Snapmirror initialize –L to convert the destination volume
into a snaplock volume
When visibility_interval is reached a snap is automatically
taken
Snapmirror initialize –s source vol dest vol – starts mirror
transfer
Snapmirror status –l shows the status of the mirror
Snapmirror rsync –s dest vol source vol – will re-sync the
mirror overwriting dest vol
Snap mirror break vol0
Snap mirror quiesce vol0
Snap mirror break
Snap mirror resync vol0 (after break)
Snap mirror release vol0 (Permanent break)
Snapmirror limitations
Update will fail when limit is reached each transfer beyond
their limit will retry every minute
Destination Snapmirror commands
Snapmirror initialize
Vol restrict dest vol
Snapmirror break
Snapmirror resync
Snapmirror update
Source snapmirror commands
Snapmirror break
Snapmirror resync
Snapmirror status
Options snapmirror.access on
Throttling network
Per transfer change the kbs in snapmirror.conf
Dynamic throttling lets you change the kbs rate whilst the
mirror is active snapmirror throttle 10kbs dest host dest vol
System wide throttling
Options replication.throttle.enable on enables system wide
throttling
Options replication.throttle.incoming.max_kbs
Options replication.throttle.outgoing.max_kbs
2ms RTT for sync
5ms RTT for semi-sync
Volume Snap mirror
One snapmirror license is required for source and
destination, sync and semi-sync require an additional license snapmirror_sync
It can only replicate the same type volumes Traditional to
traditional or Flex volumes to Flex volumes. Destination and source need to
have the same version of ontap. Destination volume needs to be the same size or
larger than the source volume.
Volume snap mirror supports Async, sync and semi-sync
Can be initialized through a tape device
Source volume is online and writable Destination volume is
read only
Baseline transfer is started by creating a non root
restricted volume on the destination.
All data in the snap shot copies is replicated down to the
destination
Schedules updates the mirror current snapshot is compared
with the previous snapshot.
Qtree Snapmirror
Qtree snapmirrors can do replications from traditional
volumes to flexible volumes. And the source and destination
Supports async only, destination must have 5% extra space,
Destination qtree cannot be /etc, Deep directory structure and large number of
small files may impact performance.
Snap mirror optimization
Try to stagger update schedule
Snap mirror uses available cpu cycles
100% cpu utilisation does not mean that the system
performance is degraded
You can monitor cpu using operations manager perf advisor
and sysstat
Snapmirror may conflict with other processes and may impact
response times reduce this by:
Using flexshare for the system
Schedule snap mirror for when nfs and cifs traffic is low
Reduce update frequency
Update controller
Network distance causes latency
Limit bandwidth using throttling (options
replication.throttle.enable)
Use dedicated network for snapmirror
Use multi pathing for load balancing
Look for network prblems
Snapvault
Basic unit for snapvault is the qtree
Increase concurrent transfers by installing nearstore on the
destination
The intial transfer will setup the relation between the
source and destination
Open tcp ports 10000 and 10566
With ontap 7.3 you can install the sv_ontap_pri and
sv_ontap_sec on the same system
To start a snapvault incremental restore – snapvault restore
–r –s
Snapvault update – can only be run on the destination system
Snapvault uses retention periods to expire worm snapshots
Snapvault status –c (show qtree config parameters)
Snapvault log files are located /etc/logs/snapmirror
Snap vault start –r
Open Systems Snap
Vault
The directory is the basic unit for ossv
To install open systems snapvault post check use
svinstallcheck to install use
Snapvault snaps for snaplock vol are not auto deleted
according to retention schedule
OOSV backups are only schedueled on the secondary system.
Snapvault restore –s will restore a vm with config files.
Snaplock
2 types of snaplock snaplock compliance and snaplock
enterprise (for self regulated requirements)
You can delete snaplock enterprise volumes with unexpired
WORM (Write Once Read Many)
Protection Manager
Protection manager and provisioning manager run I nthe
netapp management console
Protection manager uses the same database as operations
manager
No comments:
Post a Comment