Clustered data ontap fundementals

advertisement
Data ontap evolution
In 1992, NetApp introduced Data ONTAP and ushered in the network-attached storage industry.
Since then, NetApp has continued to add features and solutions to its product portfolio to meet the
needs of its customers. In 2004, NetApp acquired Spinnaker Networks to fold its scalable clustered
file system technology into Data ONTAP. That plan came to fruition in 2006 as NetApp released Data
ONTAP GX, the first clustered product from NetApp. NetApp also continued to enhance and sell Data
ONTAP 7-Mode. Having two products provided a way to meet the needs of the NetApp customers
who were happy with the classic Data ONTAP, while allowing customers with certain application
requirements to use Data ONTAP GX to achieve even higher levels of performance and benefit from
the flexibility and transparency afforded by its scale-out architecture. Although the goal was always
to merge the two products into one, the migration path that would eventually enable Data ONTAP 7Mode customers to get to clustered storage required a big leap. Enter the Data ONTAP 8 operating
system. The goal of the Data ONTAP 8 operating system is to provide a single product that enables
Data ONTAP 7G customers to operate a Data ONTAP 7-Mode system in a manner to which they're
accustomed while providing a first step in the eventual move to a clustered environment. Clustered
Data ONTAP enables Data ONTAP GX customers to upgrade and to continue to operate their clusters
as they're accustomed. With the release of clustered Data ONTAP 8.1 in 2011, NetApp achieved one
more major milestone in cluster storage: unified storage support. Clustered Data ONTAP supports
NAS and SAN storage.
Similarities with data ontap 7-mode
Scalability
Clustered Data ONTAP is different from "non-clustered" Data ONTAP in many ways. The most
obvious difference is that a single file system can scale beyond one storage appliance, being
distributed across a cluster of appliances. Each appliance is now called a "node," and many nodes
can be put together to form a cluster.
This scalability is important for increasing storage capacity. It is also essential for providing the
increased horsepower needed to handle an equally scalable workload, such as compute "farms"
used for high-performance computing applications. In a NAS environment, Data ONTAP cluster size
ranges from two to 24 nodes. In a SAN environment, currently the cluster size ranges from two to
eight nodes. SAN is supported only from clustered Data ONTAP 8.1 and later.
Aggregates, volumes and namespaces
Clustered Data ONTAP continues to use aggregates and flexible volumes as its storage containers.
Unlike "non-clustered" Data ONTAP, however, the volumes on a single storage appliance are no
longer standalone file systems. Rather, they are pieces of a potentially very large file system, or
namespace. This joining together of volumes to create large, distributed namespaces that appear as
a single storage system is unique to clustered Data ONTAP.
Volume movement
Another key benefit of clustered Data ONTAP is the ability to move volumes around within a cluster,
keeping the movement completely transparent to the namespace and the clients that use the
namespace.
As workloads need to be balanced to improve overall performance, and as storage allocation
requirements change over time, volumes can be moved to accommodate those needs.
Data ontap 7 mode components
To understand the software that makes all of this possible, consider Data ONTAP 7- Mode first. Data
ONTAP 7-Mode is made up of five primary components: network, protocol, WAFL, RAID, and
storage. Clustered Data ONTAP has these same components, as well as some new ones, that enable
Data ONTAP to scale beyond a single system.
-
-
-
-
-
Network
o The network interface module receives data from the client and delivers it to the
physical RAM
Protocols
o The protocol module determines the protocol used to transfer the data, such as CIFS
and NFS in a NAS environment, then sends the pertinent information on to WAFL
Wafl
o The WAFL (write anywhere file layout) module receives the raw data and places a
copy into NVRAM (non volatile random access memory)
Raid
o WAFL sends the data to RAID (redundant array of independent disk), which
calculates parity to protect the data.
Storage
o Raid sends the data and parity information to the storage module. The storage
module physically performs the write to disk.
Clustered data ontap components
Clustered Data ONTAP components are grouped into two logical layers: access protocols and
storage. Clustered Data ONTAP combines the network and protocol functionality into the access
protocols layer. Likewise, WAFL, RAID, and storage are grouped into the storage layer.
Clustering
The next step is to see how two nodes are made into a cluster. When two or more nodes are
clustered through a dedicated IP network, remote protocol access requests go directly from the
network access protocol layer of one node to the storage layer of the node that owns the remote
data. This decoupling of the network and protocol functionality from the actual data storage allows
an entire cluster of nodes to work together to make a distributed group of nodes look and act as one
single storage system.
Scaling a cluster
As the cluster grows to contain more and more nodes, the cluster's combined processing power and
capacity scale along with its increasing size, while maintaining its singlesystem image to its client
machines.
Disk types
A disk is the basic unit of storage for storage systems running Data ONTAP. Understanding how Data
ONTAP uses and classifies disks will help you manage your storage more effectively. Data ONTAP
supports five disk types: Fibre Channel (abbreviated as FC), ATA; SATA, SAS, and solid-state drive
(SSD). For a specific configuration, the disk types that are supported depend on the storage system
model, the disk shelf type, and the I/O modules installed in the system. Administrators can always
look up specific information when they need it on the NetApp Support site. Data ONTAP supports
two disk connection architectures: FC-AL and SAS. FC and ATA disks use the FC-AL disk connection
architecture. SAS, SATA, and SSDs use SAS diskconnection architecture.
Disk names: HBA port ID
Names for direct-attached disks (that is, disks that are attached directly to the storage system rather
than through a switch) are based on the host adapter or port number, followed by the disk ID, in the
following format: portID.diskID portID. IDs in this format refer to the host adapter or port number as
follows:
 Onboard expansion slots use the port number that is cabled to the disk shelf that contains the disk.
 Host adapters use the expansion slot number on the storage system where the host adapter is
attached. The port number of the host adapter is appended to the expansion slot number.
Disk names: Disk ID
This example of disk numbering uses a DS14 FC disk shelf. The disk ID is an integer from 16 to 125.
The disk ID corresponds to the disk shelf number and the bay in which the disk is installed. The
lowest disk ID is always in the far right bay of the first disk shelf. The next higher disk ID is in the next
bay to the left, and so on. The table shows the numbering system for FC loop IDs. The IDs that are
reserved and are not used are displayed here. The table can be summarized by the following
formula: DS14 Disk/Loop ID = DS14 Shelf ID * 16 + Bay Number.
Two examples of names of direct attached disks are:
-
A disk with an id of 33, which is connected by way of fc-al to the b port of an adapter
installed in expansion slot 4 would have the following name: 4b.33
A disk with an id of 22, which is connected by way of fc-a to onboard port 0a, would have
the following name: 0a.22
Disk ownership
Disk ownership determines which node owns a disk and which pool a disk is associated with.
Understanding disk ownership enables you to maximize storage redundancy and manage your hot
spares effectively. Disk ownership, which is software-based, is stored on the disk. Ownership is not
determined by the topology of the storage system's physical connections as it might have been in
previous versions of Data ONTAP. Software-based disk ownership gives you greater flexibility and
control over disk use.
Raid disk types
Data ONTAP classifies disks as one of four types for RAID: data, hot spare, parity, or double-parity.
The RAID disk type is determined by how RAID is using a disk. Later, you will use this information to
make decisions about increasing the size of an aggregate or about how many disks to use when
creating an aggregate. Point to each disk type to view its description. Data disk: A data disk is part of
a RAID group and stores data on behalf of the client. Hot spare disk: A hot spare disk does not hold
usable data, but it is available to be added to a RAID group in an aggregate. Any functioning disk that
is not assigned to an aggregate, but is assigned to a system, functions as a hot spare disk. Parity disk:
A parity disk stores data reconstruction within a RAID group. Double-Parity disk: If NetApp RAID-DP
software is enabled, a double-parity disk stores double-parity information within RAID groups.
-
-
Data disk
o A data disk is part of a raid group and stores data on behalf of the client
Hot spare disk
o Hot spare disk does not hold usable data but is available to be added to a raid group
in an aggregate, but is assigned to a system, functions as a hot spare disk
Parity disk
o A parity disk stores data reconstruction within a raid group
Double-parity disk
o If Netapp raid-dp software is enabled, a double-parity disk stores double-parity
information within raid groups
Raid groups
All Data ONTAP disks are organized into RAID groups. RAID groups provide parity protection against
data loss. Each RAID group consists of data disks, a parity disk (RAID 4), and a double-parity disk
(RAID-DP). A double-parity RAID group must have at least three disks: one or more data disks, a
parity disk, and a double-parity disk. You can add disks to your RAID groups to increase usable disk
space; however, you cannot remove disks to reduce disk space. Note that any new RAID-DP
aggregate created after upgrading to clustered Data ONTAP 8.2 must have at least five disks for
increased reliability and data protection
Raid groups: disk failure
If a data disk failure occurs in a RAID group, Data ONTAP replaces the failed disk with a spare disk.
Data ONTAP automatically uses parity data to reconstruct the failed disk's data on the replacement
disk. Meanwhile, Data ONTAP continues to serve data to clients by reconstructing data from the
parity disk while the new data disk is under construction. If a parity or double-parity disk failure
occurs in a RAID group, Data ONTAP replaces the failed disk with a spare disk and reconstructs parity
for the new disk.
Understanding aggregates
An aggregate is a logical container that encompasses the physical aspects of storage, such as disks
and RAID groups. Aggregates provide the storage for volumes, and volumes provide support for the
differing security, backup, performance, and datasharing needs of your users. Each aggregate has its
own RAID configuration, RAID groups, and set of assigned disks. When you create an aggregate, Data
ONTAP assigns the data disks and parity disks to RAID groups that are used to create the aggregate.
You can increase the size of your aggregate, either by adding disks to an existing RAID group or by
adding new RAID groups; however, you cannot remove a disk to reduce storage space. You can use
the aggregate to hold one or more FlexVol volumes. FlexVol volumes are logical file systems that
share the physical storage resources and RAID configuration of the containing aggregate.
Remote write process
The clustered Data ONTAP write process delivers high performance. When a client writes to a file,
the client operating system sends the write request to the clustered Data ONTAP operating system.
The write operation is processed by the cluster in this order: First, when a client sends a write
request to Data ONTAP, the request goes to the network module. The network module checks the
volume location database to find the aggregate that is hosting the volume and the node that is
hosting the aggregate. The network module passes the data to the cluster session manager
(abbreviated as CSM), which controls traffic through the cluster-interconnect. The CSM sends the
data through the cluster-interconnect to the target node's CSM. The storage module, which includes
the WAFL file system and the RAID layer, receives the raw data. The primary job of the WAFL file
system is to determine how the data will be written when the write is performed. Meanwhile, copies
of all write requests are stored in NVRAM as a backup that is used for emergencies. Because NVRAM
is backed by a battery, the write request will survive, even if power is lost. The WAFL file system also
mirrors NVRAM contents over the HA interconnect to the node's HA partner, in case of a storage
failover event. After the data is written to NVRAM, the WAFL file system sends an acknowledgement
back to the client that the write operation was successful. The WAFL file system continues to take
incoming data and decides how to write data on disk until a consistency point (a CP) occurs. This
typically happens either every 10 seconds or when NVRAM is half full. When a CP occurs, the WAFL
file system locks the half of NVRAM that contains a backup of write requests that are in the process
of being written. The other half of NVRAM is used for incoming requests. When the CP is finished,
the locked portion of NVRAM is flushed and made ready for use. When the CP occurs, the WAFL file
system passes data to the RAID module. The RAID module calculates parity and adds parity
information to the data that will be sent to disk. RAID then notifies the storage module that the data
is ready. The storage module physically performs the write request.
Module two
DATA ontap 7-mode terminology
The term "cluster" has historically been used in Data ONTAP 7-Mode to indicate what is now
referred to as a high-availability, or HA, pair. That is, two storage systems connected together via an
HA interconnect to enable the "take over" of storage and network interfaces of its partner when
necessary. The storage and network interfaces owned by the system that is down are still served by
the healthy partner system. This was called cluster failover (CFO).
Clustered data ontap 8 cluster
In clustered Data ONTAP, HA-pair relationships continue to exist; however, the term "cluster" means
something very different. A Data ONTAP cluster is a physical interconnectivity of storage systems,
which are called "nodes." A cluster can include from 2 to 24 nodes in a NAS environment and 2 to 8
nodes in a SAN environment. Nodes are connected to each other by a private, nonroutable 10-Gb
Ethernet interconnect. Each node has an HA partner node for storage failover (abbreviated as SFO).
Both nodes are also peer nodes within the cluster. The storage of a node can failover to its HA
partner, and its logical interfaces can failover to any node within the cluster. Starting with clustered
Data ONTAP 8.2, two new options are available for the configuration of small clusters. For the first
time, clustered Data ONTAP supports single-node clusters and switchless two-node clusters. For
more information on single-node clusters, two-node switchless clusters, and other advanced topics,
take the Clustered Data ONTAP 8.2 Administration instructor-led course.
Cluster components
A cluster, which is a physical entity, is made up of other physical and logical pieces. For example, a
cluster is made up of nodes, and each node is made up of a controller, disks, disk shelves, NVRAM,
and so on. On the disks are RAID groups and aggregates. Also, each node has a certain number of
physical network ports, each with its own MAC address.
Data virtual storage servers
Data virtual storage servers, or data Vservers, can use any or all of the physical cluster resources to
provide a file system, or "namespace." Multiple data Vservers can exist in a single cluster, and each
namespace is disjoint from the namespaces of any other data virtual servers in the cluster. The same
cluster resources can be used simultaneously by multiple data Vservers, and at the same time. Data
Vservers do not have node boundaries; they are bound only by the physical cluster on which they
are built
Data virtual storage server components
Data Vservers have their own components, all of which are also virtual. Every volume is associated
with exactly one data Vserver. All the Snapshot copies of that volume are also associated with that
data Vserver, as are all the mirrors of that volume. In this example, each circle represents a single
volume. The circles that are "stacked" behind a volume are the Snapshot copies of the volumes, and
the circles labeled with prime symbols (') after the letters are mirrors of the volumes with the
identical letters. Each data Vserver can also have its own client-facing logical interfaces, one NFS
service, one CIFS service for NAS, and one FC service and iSCSI service for SAN.
Test
Module two
Cluster-wide management
Clustered Data ONTAP provides two interfaces for managing nodes and clusters: the browser-based
NetApp OnCommand System Manager and the cluster shell commandline interface. Although you
connect to an individual node of a cluster with these user interfaces, you can see and manage the
entire cluster from any node. System Manager supports only clustered Data ONTAP 8.1 or later. For
support of all new clustered Data ONTAP features, go to the NetApp Support site to find the most
recent version of OnCommand System Manager.
Installing oncommand system manager
You can configure, manage, and monitor your cluster by using OnCommand System Manager.
System Manager is a web-based graphical management interface for managing common storagesystem functions from a web browser. System Manager enables you to configure and manage
storage in a Windows or Linux environment. System Manager is free software that you can
download directly from the NetApp Support site. Use the intuitive installer to install OnCommand
System Manager in only a few clicks. To learn more about System Manager, review the web-based
course Technical Overview of OnCommand System Manager.
Cluster shell
The cluster shell commands are organized into a command hierarchy that groups associated
commands in branches of the hierarchy. These branches are made up of command directories and
commands, similar to the directories and files within a file system. Each fully-qualified cluster shell
command contains the complete "path" of the command.
The command hierarchy
After you log in to the cluster shell of a node, you are placed at the top of the commandline
hierarchy. At any location in the hierarchy, a question mark shows you the command directories and
commands that are available at that location. Many of these are not commands at all, but rather
command directories. Command directories have a greaterthan sign (>) immediately after them. This
is a visual clue that there is "more," and that you can navigate further to get to more subdirectories
or commands.
Command navigation
For example, the volume branch of the hierarchy contains everything pertaining to volumes, such as
creating, deleting, modifying, and showing volumes. In addition to the commands for working
directly with volumes, this directory also contains commands for working with the Snapshot copies
of volumes. You can type two dots to move backward through the command hierarchy a directory at
a time. Enter the top command to go straight to the top of the command hierarchy
Command usage
To see the syntax of a command, type the command followed by a question mark. The syntax
indicates required parameters by showing them within single brackets. Optional parameters are
enclosed within double brackets. The brackets themselves are not part of the commands. The order
in which the parameters are shown is important if you want to use positional parameters rather
than named parameters. For example, to create a volume using only positional parameters, you
would type volume create <vserver<, without using the vserver, volume, and aggregate parameters.
Tab completion
In the cluster shell, tab completion works while you are typing part of a command directory or
command by finishing what you are typing and by prompting you for what's next. Tab completion
also works when you type the parameters of a particular command. Not only does it step you
through the parameters that you need for a command, it can also help you fill in the values for those
parameters.
Tab completion example
For example, you can enter the storage aggregate rename command by typing only a few
keystrokes. If you type the letter "s" and then press the Tab key, command directories that begin
with "s" are displayed. Type "st" to narrow the options down to two directories, and "sto" to narrow
the options down to one directory. Then "sto" is autocompleted to "storage."
Tab completion in the command hierarchy
After you expand storage, press the Tab key again to display the command directories within that
directory. The same tab completion functionality exists here, and you can use it to select the next
command directory or command, also with minimal keystrokes. The tab completion functionality can
also display the commands that are available for aggregates
Tab completion of parameters
After you choose a command, tab completion can step you through the parameters and values for
the command. For example, because aggregate is the first parameter, if you press the Tab key, the
aggregate parameter appears, followed by a list of aggregates in the cluster that are the possible
values for this parameter. You can select one of these aggregates and then continue filling in the
remaining parameters for this command manually or by using tab completion.
Command abbreviations
In the cluster shell, you can abbreviate any term (command directory, command, or parameter) by
typing the fewest letters required to make the term unique within its context, as you say for the
storage command. From the top level of the command hierarchy, the shortest unambiguous
abbreviation for the "storage" command directory is "sto." When you type this abbreviation, no tab
completion is needed.
Abbreviation strings
The aggregate command directory within the storage directory can be unambiguously abbreviated
as "agg," and the rename command can be abbreviated as "ren." Therefore, the storage aggregate
rename command can be abbreviated as "sto agg ren." You can follow these same rules to
abbreviate the parameters. But even better than that, you can eliminate the named parameters
altogether if you specify the parameter values in the exact canonical order shown in the command
syntax.
Module three
Raid technology in clustered data ontap
Aggregates in clustered Data ONTAP are formed by using either RAID 4 or RAID 6 technology to
group disks together. RAID 4 groups consist of data disks and one parity disk and can withstand a
single disk failure. RAID 6 groups, on the other hand, are made up of data disks and two parity disks
and can withstand two disk failures. RAID-DP technology is the NetApp implementation of RAID 6.
Aggregates
An aggregate is a logical container of one or more RAID groups. Each aggregate has its own RAID
configuration, which may be different from other aggregates on the node and in the cluster. The size
of the aggregate can be increased by adding disks to it, but it cannot be reduced. An aggregate is
"owned" by a particular node in the cluster, based on the ownership of the disks that make up the
RAID groups of the aggregate.
Flash pools
Traditional aggregates are built with matching disk types. Aggregates are constructed from RAID
groups of SATA, FC, SAS, or SSD disks. Hard disk types cannot be mixed within an aggregate. Flash
Pools introduce a high-speed SSD tier into a standard aggregate. The SSD tier acts as cache for data
and metadata within the aggregate. The benefits of Flash Pools include improved cost and
performance with fewer spindles, less rack space, and lower power and cooling requirements. Flash
Pools provide highly available storage with a simple administrative model. The cost-to-performance
and cost-to-capacity ratios are better than an SSD and SATA combination with pure FC or SAS
solutions. Flash Pools ensure predictable and improved operation when running in degraded mode,
during controller failures, and during HA takeover and giveback. They provide automatic, dynamic,
and policy-based placement of data on appropriate storage tiers at WAFL block granularity for data
and metadata
Flash pools: how to create a flash pool
You can enable Flash Pools on new or existing aggregates in three steps. No additional license is
required to create a Flash Pool. First, select a new or existing aggregate. The selected aggregate
must be a 64-bit aggregate. Then turn on the hybrid_enabled option on the aggregate. Finally, add a
new SSD RAID group to the aggregate. These steps convert the aggregate to a Flash Pool and
activate the storage tiers. The capacity of the SSD tier is not reflected in the total aggregate size. For
example, if the original aggregate has a 10-TB capacity and you add an SSD raid group with a 1-TB
capacity, the amount of capacity in the aggregate that can be provisioned is still 10 TB.
Module four
Volumes
An aggregate by itself isn't very useful, and multiple aggregates are equally boring. Things begin to
get interesting when volumes are created on aggregates. A volume is a logical piece of an aggregate
that contains files and directories arranged in a single file system "tree." Volumes can be resized
dynamically, without the addition or removal of physical resources. Data ONTAP 7-Mode has two
types of volumes: traditional and flexible. Clustered Data ONTAP has one type of volume: flexible.
Flexible volumes are the same in both products
Volumes are the context within which data is managed for a cluster. For example, volumes are the
units of data for which Snapshot copies, mirrors, and tape backups are created. Volumes can be
moved around in a cluster, and those moves are transparent to any clients that use them.
Namespaces in clustered data ontap
One of the key features of clustered Data ONTAP is the ability to "join" volumes together to create
distributed global namespaces. If a client machine mounts or maps the "root" volume of a
namespace, it can navigate to every volume in that namespace, regardless of which nodes in the
cluster are hosting the volumes.
Junctions
The magic that joins volumes together is called a junction. A junction is conceptually like a mount
point of a Unix file system. Rather than connecting one Unix partition to another, a junction
connects one volume to another volume. From a client machine, the junction simply looks like a
directory. Every volume in a namespace can be connected to every other volume in that namespace.
A junction can live at any directory or qtree level within the parent volume. A volume can be
"unmounted" from its parent and "mounted" to a new parent at any time.
Creating a volume
You create a volume within a data Vserver by using either the cluster shell CLI or OnCommand
System Manager. From the cluster shell, you create a volume by using the volume create command.
Volume attributes for the Vserver, volume name, and aggregate are required. For other attributes,
you can use the default value unless otherwise specified. In most cases, you will specify at least a
volume size and a junction path. In OnCommand System Manager, you navigate to the Vserver that
will host the volume and click the Create button. From there, you specify the attributes of the
volume. After a volume is created, the new volume is included in the list of volumes for your
selected Vserver. To mount the volume into the Vserver's namespace, navigate to the namespace in
the left pane of System Manager. When the namespace diagram appears, click the Mount button. In
the Mount Volume dialog, you can specify the junction name and the path to the new junction.
Infinite volume definition
Clustered Data ONTAP 8.1.1 introduced infinite volumes. Infinite volumes are boundless, easily
manageable, and scalable containers that exceed the current Data ONTAP limits for FlexVol capacity.
Clustered Data ONTAP 8.2 improves infinite volumes in many ways. Infinite volumes now support
NFSv3, CIFS, and NFSv4.1, including pNFS. Data ONTAP 8.2 also introduced data constituent groups,
which can be used for storage-class tiers
Infinite volume features
Infinite volumes can now coexist with FlexVol volumes on aggregates and Vservers that are enabled
for infinite volumes can coexist with Vservers that serve FlexVol volumes. Infinite volumes use a
unified security style, which enables all clients to view and set file permissions, regardless of the file
permissions that are currently in effect on a given file or directory. A unified security style also
provides unified access control lists, which facilitate access checks using both Windows and UNIX
credentials. Clustered Data ONTAP 8.2 allows you to perform disaster recovery of a namespace
constituent that has been permanently lost, from namespace mirror constituent. Infinite volumes
can have mirror relationships with infinite volumes in other clusters as well as fan-out and
bidirectional mirror relationships. Infinite volumes are also now supported on all current NetApp
platforms, except the FAS2000 series.
Not supported with infinite volume
Infinite volumes in clustered Data ONTAP 8.2 do not support some features, including single-node
clusters, qtrees and quotas, FlexCache software, Cascading mirror copies, and SnapVault software
Infinite volumes: examples
Creating infinite volumes and the aggregates that host them is very similar to creating FlexVol
volumes. You create aggregates throughout the cluster to host constituent volumes. Create a
Vserver that is capable of serving infinite volumes by using the -isrepository switch. Then create the
infinite volume to fit the capacity of the constituent aggregates. The volume show command shows
you the infinite volume. The volume show command with the -is-constituent true switch displays the
list of constituent volumes.
Module five
Nfsv4 and nfsv4.1
The clustered Data ONTAP 8.1 operating system introduced support for the NFS version 4 (NFSv4)
protocol specification, as well as elements of NFSv4.1. Clustered Data ONTAP continues to fully
support NFSv3. Starting with clustered Data ONTAP 8.2, NFSv2 is no longer supported. The key
feature of NFSv4 is referrals, which are discussed in this module. NFSv4.1 is a minor revision and
extension of version 4.0, not a modification, so it is fully compliant with the NFSv4 specification. It
extends delegations beyond files to directories and symlinks, introduces NFS sessions for enhanced
efficiency and reliability, provides parallel NFS (pNFS), which is also discussed in the module, and
fixes some problems with NFSv4.
Nfsv4.1:pnfs examples
Remote file access is defined as file access in which a client that is connected through a logical
interface (abbreviated as LIF) that goes through a physical port on one controller accesses a file that
is hosted on a different controller in the same cluster. Remote file access has traditionally been a
performance concern for clients that use clustered Data ONTAP. In this example, a client is mounted
to a data LIF that is hosted on node 2. This client has a file operation with a destination in a volume
on node 5. The request is serviced by the node 2 protocol stack. The node 2 protocol stack looks up
the location of the volume and directs the operation to node 5, which hosts the target volume. The
request traverses the cluster-interconnect, and the result is returned to the client along the same
path. With pNFS, when a file is opened by an NFS client, the mounted data LIF on node 2 serves as
the metadata path because this is the path that is used to discover the target volume's location. If
the data is hosted by node 2, the operation is managed locally. But in this case, the local node
discovers that the data is on node 5. Based on the pNFS protocol, the client is redirected to a LIF that
is hosted on node 5. This request, as well as subsequent requests to the volume, are serviced locally,
bypassing the cluster network. When a volume is moved to an aggregate on a different node, the
pNFS client data path is redirected to a data LIF that is hosted on the destination node.
Nfsv4.1 pnfs features
You can enable pNFS from the command line with the vserver nfs modify command. Before enabling
pNFS, ensure that all NFS clients are compatible with and configured to support NFSv4.1 and pNFS.
Red Hat Enterprise Linux 6.2 and Fedora 14 are targeted for full support. Before you implement
pNFS or NFSv4 referrals, you should decide which fits better on your environment because they do
not work together. By keeping network access and data local, pNFS reduces the amount of traffic
that traverses the cluster network. Unlike the NFS referral process, pNFS works seamlessly with the
client; it does not require a file system remount to ensure an optimized path. With pNFS, because
the network redirect does not happen at mount time, the file handle is not left stale when a volume
is moved to an aggregate on a different node.
SMB 2.0 and SMB 2.1
In addition to the SMB 1.0 protocol, clustered Data ONTAP supports SMB 2.0 and SMB 2.1. SMB 2.0
is a major revision of the SMB 1.0 protocol, including a complete reworking of the packet format.
SMB 2.0 also introduced several performance improvements relative to previous versions.
Enhancements include more efficient network utilization, along with request compounding, which
stacks multiple SMBs into a single network packet. Larger read and write sizes enable it to exploit
faster networks, and file and directory property caching are provided. Durable file handles allow an
SMB 2 connection to transparently reconnect to the server if a temporary disconnection occurs, such
as over a wireless connection. In addition, improved message signing with improved configuration
and interoperability with HMAC SHA-256 replace MD5 as a hashing algorithm. SMB 2.1 provides
important performance enhancements, including the client opportunistic lock (oplock) leasing
model, large maximum transmission unit support, and improved energy efficiency for client
computers. Support is provided for previous versions of SMB. The SMB 2.1 protocol also provides
several minor enhancements to the SMB 2.0 specification. The clustered Data ONTAP 8.1 operating
system supports most of the SMB 2.1 features. Starting with clustered Data ONTAP 8.2, support is
provided for resilient file handles and branch cache. Support for SMB 2.1 is automatically enabled
when you enable the SMB 2.0 protocol on a virtual storage server (Vserver). You can use this
command to enable SMB 2.0 for a Vserver
SMB 2.1 leases
One of the most important features in the SMB 2.1 protocol is the oplock leasing model. Leasing
allows a client to hold oplocks over a wider range of scenarios. The feature offers enhanced file
caching and metadata caching opportunities for the SMB client, and provides major performance
benefits by limiting the amount of data that needs to be transferred between the client computer
and the server. This enhancement particularly benefits networks with high latency. Additionally, as
the number of operations that must be directed toward an SMB file server is reduced, the SMB file
server scalability is increased. The new leasing model in SMB 2.1 allows greater file and handle
caching opportunities for an SMB 2.1 client computer, while preserving data integrity and requiring
no current application changes to take advantage of this capability. New caching levels allow more
flexibility for clients to request caching mechanism. Multiple applications can retain the cached data
even after the file handle is closed. Full data caching is supported even with multiple handles, as long
as those handles are opened on the same client.
Branch cache
BranchCache is a CIFS feature that caches content on computers local to requesting clients, either in
distributed or hosted cache mode. In distributed cache mode, branch office client computers
download content from content servers in the main office and then cache the content for other
computers in the same branch office. Distributed cache mode does not require any infrastructure or
services in the branch office beyond client computers that run Windows 7. In hosted cache mode,
branch office client computers download content from the content servers in the main office, and a
hosted cache server retrieves the content from the clients. The hosted cache server then caches the
content for other client computers. The hosted cache server at the branch office must run Windows
Server 2008 R2 or Windows Server 2012. Clustered Data ONTAP 8.2 and later versions support both
BranchCache 1 and BranchCache 2. For more information about configuring and managing
BranchCache, see the Clustered Data ONTAP 8.2 File Access and Protocols Management Guide, and
the CIFS Administration on Data ONTAP instructor-led training for Data ONTAP 8.2.
SMB 3.0
Clustered Data ONTAP 8.2 introduces support for SMB 3.0 enhancements that provide BranchCache
version 2, witness protocol, remote VSS for SMB shares, persistent file handles, ODX copy offload,
and continuously available shares.
The Data ONTAP implementation of BranchCache reduces WAN utilization and provides improved
access response time for branch office clients who are using the SMB protocol.
The Data ONTAP implementation of BranchCache reduces WAN utilization and provides improved
access response time for branch office clients who are using the SMB protocol.
Microsoft Remote VSS extensions enable Remote VSS-enabled backup services, such as
SnapManager for Hyper-V, to create application-consistent shadow copies for virtual machines that
store data and configuration files on shares.
Persistent file handles allow an SMB connection to survive a brief network outage without the need
to construct a new session.
Offloaded Data Transfer, or ODX, enables direct data transfers within or between compatible
storage devices without transferring the data through the host computer.Clustered Data ONTAP also
supports ODX for SAN protocols.
The continuously available share property enables clients to survive disruptive events such as
failover and giveback.
Hyper –v over smb
SMB 3.0 enables you to use continuously available SMB file shares to store Hyper-V virtual machine
files on volumes, providing nondisruptive operations for both planned and unplanned events. With
Data ONTAP Hyper-V over SMB, clients that connect through continuously available SMB 3.0 shares
can survive disruptive events such as takeover and giveback. The witness protocol provides
enhanced client failover capabilities for SMB 3 shares. The witness protocol facilitates faster failover
by bypassing the LIF failover recovery period. It notifies Hyper-V servers when a node is unavailable
without needing to wait for the SMB connection to time out. It provides nondisruptive operations
for events such as planned and unplanned takeover and giveback, relocation of aggregates, and
upgrades.
Additional nas protocol enhancements
Clustered Data ONTAP 8.2 adds support for NAS protocol features like NAS auditing, local users and
groups, offline files and roaming profiles. For more information about these features, see the
Clustered Data ONTAP 8.2 File Access and Protocols Management Guide, and the CIFS and NFS
Administration on Data ONTAP instructorled training for Data ONTAP 8.2.
Multiple networks needed
Clustered Data ONTAP uses three distinct physical networks. You have already experienced the
management network, which is used when connecting to the UI of a particular node. You have also
seen the use of the cluster-interconnect, which is the private network that provides for intracluster
communication among all the nodes. Another network is the "data" network. This is the one through
which NAS and SAN clients connect to the cluster. NetApp recommends that data and management
interfaces reside on separate physical networks.
NAS: management and cluster lifs
Each node connects to each of these networks through a unique LIF and network port. The
management, cluster, and data LIFs map IP addresses to network interface cards (abbreviated as
NICs) on particular nodes. For example, each node has a NIC that is designated as a management
port. A management LIF, then, is an association between an IP address on the management or data
network and a management port on a node. Likewise, each node contains two cluster ports, and
those ports are associated with IP addresses on the cluster-interconnect by means of cluster LIFs.
NAS: DATA LIFS
Data LIFs are also associations between IP addresses on the data network and network ports on
nodes. They differ from the node management and cluster LIFs, however, in that node management
and cluster LIFs are tied permanently to nodes, but data LIFs are tied to Vservers. Their association
with a node is utilitarian; for example, they use data network ports to create connections to nodes,
but those LIFs provide access to the protocol services of a Vserver, not of a node.
Cluster lif failover and migration
Each management LIF is bound to one network port of one node, but cluster LIFs can migrate among
the cluster network ports within a node. If a node in a cluster has two or more cluster ports, and a
problem is detected with a NIC, a network cable, or the cluster switch, the cluster LIF can
transparently fail over to another cluster port on the same node. This is for high availability of the
cluster itself.
Scalable san
Clustered Data ONTAP delivers a SAN target device that is scalable and provides location
transparency and single-image, cluster-wide administration to the NetApp block access portfolio.
The new SCSI target implements the SCSI-3 protocol engine for access to LUNs in clustered WAFL
(Write Anywhere File Layout) FlexVol volumes. On the front end, clustered Data ONTAP works with
the FC protocol and iSCSI transport modules to provide SCSI target services. On the back-end, it
maps SCSI operations into data packets that can traverse the cluster network to be processed by the
host storage subsystem, or for coordination among different SCSI target instances on each node.
Scalable SAN introduces new data LIF types. FC LIFs are assigned to host bus adapters. Like
traditional NAS data LIFs, iSCSI LIFs are assigned to Ethernet ports. Unlike NAS data LIFs, FC LIFs are
assigned worldwide port names instead of IP addresses. iSCSI LIFs are assigned IP addresses but
cannot be used by NAS clients; likewise, NAS LIFs cannot be used by iSCSI clients. NAS and iSCSI LIFs
can co-exist on a single Ethernet port. In clustered Data ONTAP 8.2, SAN is scalable up to eight-node
clusters.
You must configure SAN clients to use multipath I/O to access LUNs, and Asymmetric Logical Unit
Access (abbreviated as ALUA) to determine the state of a given data path to the LUNs. The direct
path to a LUN refers to the path for which the LIF and LUN are hosted by the same node. The
indirect path represents the path for which the LIF and LUN are hosted on separate nodes. Unlike
NAS LIFs, SAN LIFs do not migrate among interfaces or nodes. Therefore, the client host uses ALUA
to determine the most efficient path (or paths) to communicate to the LUN. The direct path
becomes the primary path for data transfer between the host and the LUN. When a volume that
hosts a LUN is moved to an aggregate on a different node, the Vserver updates the path status, and
the client polls the Vserver for the change. In this way, the new direct and indirect paths are chosen,
and the client selects the best possible paths. When a node goes down and storage fails over to the
partner node, the node's paths also go offline. If an appropriately zoned SAN LIF is available on the
partner node, the path to the takeover node becomes the active path until the aggregate is returned
to its home node. If the paths to a node become unavailable so that only indirect paths remain, but
the storage doesn't fail over, the client chooses an indirect path, and the data traverses the cluster
network until an optimized path is restored.
File access
The NFS, CIFS, FC, and iSCSI servers for clustered Data ONTAP are not configured per node, nor are
they configured per cluster. Rather, they are configured per cluster per Vserver. For example, for
NFS and CIFS, an NFS and a CIFS server run on each node, but they work together to appear as one
NFS and one CIFS server for the entire cluster Vserver. The FC and iSCSI servers are also configured
per cluster per Vserver. The most basic configuration for NFS, CIFS, FC, and iSCSI access for a Vserver
is simple to perform; however, many other steps are required, depending on your needs and
preferences, such as Network Information Service configuration and routing group configuration
Module six
Distributed namespaces
A namespace can and should be distributed throughout the nodes and aggregates of a cluster in
order to spread the workload across the resources of the cluster. The most obvious and simple way
of doing this is to add volumes to nodes and aggregates in a round robin fashion. As more and more
volumes are created and added to a namespace, the distribution will tend to balance out.
Flexibility with volumes
When you first begin setting up a particular virtual server and its namespace, it might be difficult to
predict how it will grow in its consumption of capacity, bandwidth, and CPU utilization. With the
flexibility of NetApp clusters, you're never locked in to a particular configuration. As volumes grow
and shrink, they might need to be moved to balance space utilization. Volumes might also need to
be moved to alleviate "hot spots" that occur due to some volumes, aggregates, or nodes being
utilized more than others. Volume movement is mostly transparent to any clients and applications
that use the volume that is being moved. The exceptions to that rule are CIFS clients. Because NFS
clients are "connectionless," whereas CIFS clients are not, CIFS clients might experience an
interruption if they are in the midst of accessing a volume that is being moved.
Flexibility with logical interfaces
Likewise, network ports on the nodes can also become hot spots, as multiple logical interfaces that
share the same node, or even the same port, begin handling more traffic. Logical interfaces can be
easily migrated to a different port, node, or port and node to balance the network load within the
cluster.
Load-sharing mirrors
In addition to mirroring data in order to protect it, clustered Data ONTAP provides mirroring for load
balancing. "Copies" of read/write volumes, which are called loadsharing mirrors (or LS mirrors), can
be used to offload read requests from their read/write volumes. Also, when a number of LS mirrors
are created for a single read/write volume, the likelihood of a read request being served locally,
rather than traversing the cluster network, is greatly increased, resulting in better read performance.
Load-sharing mirrors that are out of sync
An LS mirror is "in" the namespace at the same point as its read/write volume. So, if a volume has
any LS mirrors, all client requests are sent, transparently to the clients, to an N-Blade or SCSI-Bladeselected LS mirror, rather than to the read/write volume. If the LS mirrors become out-of-sync with
their read/write volumes, a client read request will get out-of-date information
Volume move
T Clustered Data ONTAP enables you to move a volume from one aggregate or node to another
within the same Vserver for capacity utilization, improved performance, and to satisfy SLAs. The
volume move is a nondisruptive operation. During the volume movement process, the original
volume is intact and available for clients to access. You can move a FlexVol volume to a different
aggregate, node, or both within the same Vserver.
Typical nfs mount
The typical way of mounting the root volume of a namespace on an NFS client is by using the mount
command keyword followed by the IP address of the data LIF, followed by a colon and the mount
point. This mount accesses the read/write volume except unless the volume has an LS mirror. When
a volume has an LS mirror and has replicated to it, all read and write access to that volume via that
NFS mount is sent to one of the LS mirrors rather than to the read/write volume. For any volumes
that do not have LS mirrors, their read/write volumes continue to be used.
Nfs mount for writes
To allow an NFS request to go to the read/write volume after it has been replicated to an LS mirror,
an additional mount would need to be done to use the /.admin path. For CIFS clients, an additional
step is needed within the cluster itself. You must create an additional CIFS share that uses /.admin
rather than / for its path. The clients that require read/write access must use that share.
Ls mirror selection
When multiple LS mirrors exist for a volume, the node that receives the request gives preference to
a local LS mirror. If there is no local LS mirror, Data ONTAP uses a roundrobin algorithm to choose
which "remote" LS mirror receives the request. For volumes with high read traffic, a good practice is
to have an LS mirror on every node so that all read requests are served locally. Mirroring of the root
volumes of virtual servers is highly recommended and is considered a best practice.
Dns load balancing
Clustered Data ONTAP also includes additional load-balancing functionality in the form of DNS load
balancing. The DNS load-balancing feature works in two ways. With the first method, initial client
mount requests are returned the IP address of the least-utilized node and port combination. With
the second method, automatic periodic rebalancing occurs for LIFs that host NFS connections. The
LIFs are automatically migrated throughout the cluster to achieve evenness in terms of the number
of connections, CPU utilization, and throughput. Ultimately, this feature helps to balance the overall
utilization of the cluster. It does not increase the performance of any one individual node; rather, it
ensures that each node is more evenly used. The result is better performance utilization from the
entire cluster. DNS load balancing also improves the simplicity of maintaining the cluster. Instead of
manually deciding which LIFs are used when mounting a particular global namespace, the
administrator can let the system dynamically determine which LIF is the most appropriate.
Module 7: load balancing
Distributed namespaces
A namespace can and should be distributed throughout the nodes and aggregates of a cluster in
order to spread the workload across the resources of the cluster. The most obvious and simple way
of doing this is to add volumes to nodes and aggregates in a round robin fashion. As more and more
volumes are created and added to a namespace, the distribution will tend to balance out.
Flexibility with volumes
When you first begin setting up a particular virtual server and its namespace, it might be difficult to
predict how it will grow in its consumption of capacity, bandwidth, and CPU utilization. With the
flexibility of NetApp clusters, you're never locked in to a particular configuration. As volumes grow
and shrink, they might need to be moved to balance space utilization. Volumes might also need to
be moved to alleviate "hot spots" that occur due to some volumes, aggregates, or nodes being
utilized more than others. Volume movement is mostly transparent to any clients and applications
that use the volume that is being moved. The exceptions to that rule are CIFS clients. Because NFS
clients are "connectionless," whereas CIFS clients are not, CIFS clients might experience an
interruption if they are in the midst of accessing a volume that is being moved.
Flexibility with logical interfaces
Likewise, network ports on the nodes can also become hot spots, as multiple logical interfaces that
share the same node, or even the same port, begin handling more traffic. Logical interfaces can be
easily migrated to a different port, node, or port and node to balance the network load within the
cluster.
Load-sharing mirrors
An LS mirror is "in" the namespace at the same point as its read/write volume. So, if a volume has
any LS mirrors, all client requests are sent, transparently to the clients, to an N-Blade or SCSI-Bladeselected LS mirror, rather than to the read/write volume. If the LS mirrors become out-of-sync with
their read/write volumes, a client read request will get out-of-date information. LS mirrors are ideal
for volumes that are read frequently and written infrequently
Load-sharing mirrors and read/write volumes
LS mirrors are separate flexible volumes that have a special relationship with one read/write volume.
Mirroring in clustered Data ONTAP is asynchronous; that is, the cluster administrator is responsible
for ensuring that the read/write volumes are being replicated to their mirrors when necessary,
either manually, or by means of an automated (scheduled) replication. The read-only data is only as
up-to-date as the administrator keeps it.
Load-sharing mirrors that are out of sync
An LS mirror is "in" the namespace at the same point as its read/write volume. So, if a volume has
any LS mirrors, all client requests are sent, transparently to the clients, to an N-Blade or SCSI-Bladeselected LS mirror, rather than to the read/write volume. If the LS mirrors become out-of-sync with
their read/write volumes, a client read request will get out-of-date information. LS mirrors are ideal
for volumes that are read frequently and written infrequently.
Volume move
Clustered Data ONTAP enables you to move a volume from one aggregate or node to another within
the same Vserver for capacity utilization, improved performance, and to satisfy SLAs. The volume
move is a nondisruptive operation. During the volume movement process, the original volume is
intact and available for clients to access. You can move a FlexVol volume to a different aggregate,
node, or both within the same Vserver.
Typical NFS mount
The typical way of mounting the root volume of a namespace on an NFS client is by using the mount
command keyword followed by the IP address of the data LIF, followed by a colon and the mount
point. This mount accesses the read/write volume except unless the volume has an LS mirror. When
a volume has an LS mirror and has replicated to it, all read and write access to that volume via that
NFS mount is sent to one of the LS mirrors rather than to the read/write volume. For any volumes
that do not have LS mirrors, their read/write volumes continue to be used.
NFS mount for writes
To allow an NFS request to go to the read/write volume after it has been replicated to an LS mirror,
an additional mount would need to be done to use the /.admin path. For CIFS clients, an additional
step is needed within the cluster itself. You must create an additional CIFS share that uses /.admin
rather than / for its path. The clients that require read/write access must use that share.
LS mirror selection
When multiple LS mirrors exist for a volume, the node that receives the request gives preference to
a local LS mirror. If there is no local LS mirror, Data ONTAP uses a roundrobin algorithm to choose
which "remote" LS mirror receives the request. For volumes with high read traffic, a good practice is
to have an LS mirror on every node so that all read requests are served locally. Mirroring of the root
volumes of virtual servers is highly recommended and is considered a best practice.
DNS Load balancing
Clustered Data ONTAP also includes additional load-balancing functionality in the form of DNS load
balancing. The DNS load-balancing feature works in two ways. With the first method, initial client
mount requests are returned the IP address of the least-utilized node and port combination. With
the second method, automatic periodic rebalancing occurs for LIFs that host NFS connections. The
LIFs are automatically migrated throughout the cluster to achieve evenness in terms of the number
of connections, CPU utilization, and throughput. Ultimately, this feature helps to balance the overall
utilization of the cluster. It does not increase the performance of any one individual node; rather, it
ensures that each node is more evenly used. The result is better performance utilization from the
entire cluster. DNS load balancing also improves the simplicity of maintaining the cluster. Instead of
manually deciding which LIFs are used when mounting a particular global namespace, the
administrator can let the system dynamically determine which LIF is the most appropriate.
Module 8: data protection
Snapshot technology
A Snapshot copy consists of pointers to the blocks of storage that contain volume data. As data
changes in the volume, a new block is written and the volume points to that new block, while the
Snapshot copy continues to point to the original block of storage. As long as any Snapshot copy
points to a block of data, it will not be deleted. The space that is used for Snapshot copies resides
within the volume.
Snapshot copies of volumes can be created manually by the administrator at any time, or
they can be created automatically based on Snapshot polices. To manually create a Snapshot copy,
you need to provide the volume, the cluster virtual server associated with the volume, and a
Snapshot copy name of your choosing. The process is very fast, since no data is copied at the time of
the Snapshot copy creation.
Snapshot policies
Snapshot policies dictate when Snapshot copies are to be created and how many previous copies are
to be kept. For example, a Snapshot policy might cause Snapshot copies of a particular volume to be
created hourly, with the last six hourly Snapshot copies being kept. Snapshot copies could also be
created nightly, keeping the two previous nights’ Snapshot copies, and weekly, keeping the two
previous weeks' Snapshot copies. All told, that volume would have ten Snapshot copies at all times.
Multiple snapshot policies
Unless explicitly specified, each volume that is created will be assigned the Snapshot policy named
default which, as its name implies, exists by default. You can create additional Snapshot policies with
up to five different schedules and retention counts and assign them to volumes. Each volume can
have at most one Snapshot policy, or a volume can have no Snapshot policy at all.
Restoring a volume from a snapshot copy
A Snapshot copy also acts as an inexpensive online backup. The data in a Snapshot copy can be
accessed by NFS and CIFS clients, thus allowing individual users to restore their own data when
needed. In the advanced privilege mode, a Snapshot copy can also be restored, which means that its
contents instantly become the contents of its read/write volume. The current contents of the
volume and all Snapshot copies "later" than the restored copy are removed. Obviously, this type of
operation should not be taken lightly.
Data protection mirrors
In addition to supporting mirroring for load balancing, clustered Data ONTAP provides mirroring for
data protection or disaster recovery. Called data protection, or DP mirrors, these "copies" of
read/write volumes are disk-to-disk backup copies of their associated read/write volumes. In
substance, DP mirrors are like load sharing, or LS mirrors: they are separate read-only flexible
volumes that have associations with read/write volumes, they are mirrored asynchronously, and
they are only as up-to-date as the administrator keeps them.
DP mirrors as online backups
Unlike LS mirrors, DP mirrors are not automatically included in their namespaces and are not
accessed transparently by clients. DP mirrors are ideally suited for lower-cost, higher-capacity disks,
since they are not typically accessed by high-performance applications. Also, DP mirrors are online,
disk-based backups that can be used to easily restore files, directories, or entire volumes.
Creating and replicating a DP mirror
Like LS mirrors, DP mirrors must be created before they can be replicated. The first step in creating a
mirror is to create a volume. After you create the volume, you can create the mirror. You must
choose the name of the DP mirror and the particular aggregate in the cluster on which the DP mirror
should reside. After you create the mirror, you must initialize it to sync it up to its read/write or
source volume.
Cluster peer data protection
For greater data protection, you can create a data-protection mirror relationship between two
clusters. In this type of intercluster data protection, the source volume is on one cluster, and the
destination volume is on the other. When a disaster occurs, you can use data-protection mirror
copies in the peer cluster to recover data. To create a dataprotection mirror relationship between
cluster peers, you must first set up a cluster-peer relationship. The cluster-peering facility allows two
clusters to coordinate and share resources between themselves. Before you create a cluster peer,
you must ensure that the clocks on the clusters are synchronized by using the system node date
show command. At most, the time difference between the two clusters can be 300 seconds. You
start cluster-peer configuration by using the network interface create command to create an
intercluster logical interface (LIF). Each cluster that participates in a clusterpeer relationship must
have at least one intercluster LIF assigned to each node. You must create the intercluster LIF through
the ports on the data network. Also ensure that each node in the cluster has an intercluster LIF
assigned to each node.
Snapvault backups for clusters
NetApp’s flagship disk-to-disk backup solution is available with clustered Data ONTAP 8.2. SnapVault
software leverages block-level incremental replication for a reliable, lowoverhead backup solution. It
provides efficient data protection by copying only the data blocks that have changed since the last
backup, instead of entire files. As a result, you can back up more often while reducing your storage
footprint because no redundant data is moved or stored. With direct backups between NetApp
systems, disk-to-disk vault backups minimize the need for external infrastructure and appliances. By
default, vault transfers retain storage efficiency on disk and over the network, further reducing
network traffic. You can also configure additional deduplication, compression, or both on the
destination volume. However, if additional compression is configured on the destination volume,
storage efficiencies from source to destination are not retained over the network. The key
advantages of vault backups for clusters include: reduction of backup times from hours or days to
minutes, 100% success rates for backup reliability, reduction of disk capacity requirements by 90% or
more, simplified management across enterprise applications, and minimized network traffic. For
more information about backing up FlexVol volumes to a backup vault, see the Clustered Data
ONTAP Data Protection Guide.
No snapvault commands
Because SnapVault was added to the new SnapMirror architecture and UI, there are no SnapVault
commands. SnapVault functions are accomplished with SnapMirror commands. SnapVault is
specified with the transfer type ―XDP.‖ Architecture, UI, and various behaviors were changed to
accommodate scalability and server virtualization.
Basic snapmirror commands
In clustered Data ONTAP, SnapMirror technology is organized to include several types of replication
relationships. ―DP‖ is for asynchronous data protection mirror relationships. ―LS‖ is for loadsharing mirror relationships. ―XDP‖ is for backup vault relationships. ―TDP‖ is for transition
relationships from Data ONTAP running in 7-Mode to clustered Data ONTAP. ―RST‖ is a transient
relationship for restore operations. SnapMirror commands, with the ―–type XDP‖ option, are used
to configure SnapVault. The basic SnapMirror commands include snapmirror create, snapmirror
initialize, snapmirror modify, snapmirror policy, snapmirror show, snapmirror update, and
snapmirror restore.
SMTape to seed baselines
In clustered Data ONTAP, you can attach a tape device to the source node and use SMTape in a
process that is called ―tape seeding.‖ By using SMTape commands and a tape device, you can
establish mirror and vault relationships for large source volumes without sending the initial baseline
transfer from the source node to the destination node over the network. For more information on
using vault backups, enroll in the web-based course Technical Overview of SnapVault on Clustered
Data ONTAP.
NDMP Tape Backups
Clustered Data ONTAP supports tape backups by using NDMP. NDMP is an industry standard that
allows third-party data management applications, such as Veritas NetBackup and IBM Tivoli Storage
Manager, to control backups and restores of NetApp storage appliances. The smallest unit that can
be backed
Vserver-aware NDMP
Clustered Data ONTAP now enables NDMP to function at the virtual storage server (Vserver) level.
Resources can be backed up, restored, and scoped at the Vserver level, including FlexVol volumes.
Vserver-aware backups are critical for implementing multitenancy. For NDMP to be aware of a
Vserver, NDMP data management application software must be enabled with cluster-aware backup
extensions, and the NDMP service must be enabled on the Vserver. After the feature is enabled, you
can back up and restore all volumes that are hosted across all nodes in the Vserver. An NDMP
control connection can be established on different LIF types. An NDMP control connection can be
established on any data or intercluster LIF that is owned by a Vserver that is enabled for NDMP and
owns the target volume. If a volume and tape device share the same affinity, and if the datamanagement application supports the cluster-aware backup extensions, then the backup application
can perform a local backup or restore operation—you do not need to perform a three-way backup
or restore operation. Vserver-aware NDMP user authentication is integrated with the role-based
access control mechanism. For more information about Vserver-aware NDMP and cluster-aware
backup extensions, see the Clustered Data ONTAP Data Protection Tape Backup and Recovery Guide.
Download