Professional Documents
Culture Documents
2018 IBM Systems Technical University: 22-26 Oct Rome, Italy
2018 IBM Systems Technical University: 22-26 Oct Rome, Italy
2018 IBM Systems Technical University: 22-26 Oct Rome, Italy
Trishali Nayar
Spectrum Scale, IBM 2018 IBM Systems
Technical University
22-26 Oct Rome, Italy
Session Objectives
— How to share data between Spectrum Scale Clusters
— Understand various Use Cases
— Learn updates from the latest release (5.0.2)
— Integrated tools designed to help organizations manage petabytes of data and millions of files.
— Active File Management is a clustered file system cache, using the underlying file system.
— Moves data on demand, periodically and continuously which makes it extremely flexible.
— Home Cluster/Site
The cluster or main site where data is stored.
— Cache Cluster/Site
The cluster where data is cached.
Note:
The home and cache sites are created independent of each other in terms of storage and
network configuration. The number of nodes in each of these sites can vary based on workload.
— Application Node
An application node is any node in the cache cluster that gets I/O requests from applications.
— Asynchronous Operations
Operations done at the cache like creating directories/files, writes, renames, removes, truncates or setting permissions/attributes
etc.
Once the operation is completed on the local filesystem at the cache and queued at the GW node, the response is returned to the
application.
The GW node maintains a queue of all these asynchronous operations that need to be performed at the home cluster. These will
happen at the home cluster after some delay and this process is asynchronous, but continuous.
7 © Copyright IBM Corporation 2018
Data Flow
— Pull Data
This is used to refer to the direction of data flow, when data is pulled into the AFM cache from the home. Eg- on
demand
— Push Data
This is used to refer to the direction of data flow, when data is pushed from the AFM cache to home. Or from
primary site to secondary site, in case of Disaster Recovery scenarios.
— Revalidation
The process of comparing the metadata at cache and home to determine if the data has changed at home. And if
it has, then fetch the latest contents.
— Single-writer (SW)
When a cache is configured in this mode, the cache site can exclusively write data. All asynchronous operations at the cache get pushed to the
home site asynchronously, hiding WAN latencies. This also helps provide better performance to any applications which are run at the cache, as
write-back caching is done. When any asynchronous operation happens, an application can proceed as soon as the operation happens locally on its
filesystem at the cache. This same operation also gets queued on the AFM gateway node.
There is a 1:1 relationship between the AFM single-writer cache fileset and the home fileset. This implies that all the data is to be written at the
single cache site and the home is used only for reading. AFM cannot detect or prevent home site modification of data, the administrators need to
ensure that the data is not modified or accidently corrupted.
— Local-update (LU)
This is used to pull data from home, but any changes made at the cache are not pushed to the home. When a cache is configured in this mode, the
cached data is available for both reading and writing. But the data modified at the cache site is not sent back to the home site. So, this mode serves
as a scratch-cache. After the data is modified at cache, new updates made at home for that particular data object are not pulled into the cache.
9 © Copyright IBM Corporation 2018
Modes Available
— Independent-writer (IW)
This mode allows multiple cache filesets, located in different cache clusters to be associated with a single home fileset, hence this is an
example of N:1 mapping. But the important point to be noted is that each cache site should perform asynchronous operations (includes
writes) on different files. There is no inter-cluster locking for a file getting modified, at multiple cache clusters. Each cache makes its
updates independently and these changes in the IW caches are pushed to the home. In case multiple sites modify the same file and cause
conflicts then the last writer will win. It is administrator’s responsibility to control who has write access to files, to avoid such conflicts.
Once data is updated at home, all connected IW caches can fetch those changes on-demand based on the revalidation intervals set. So on
next data access all the IW caches will get synchronized with the home. Data can also be pre-fetched into the cache.
— Note:
As seen in the above modes, depending on where the data is created/modified sometimes the home site can be referred to as the local site
and the cache site can be referred to as the remote or edge or geographically disperse site. Eg. In RO mode, the home cluster can be
called the local site and cache cluster can be considered as the remote site.
The vice versa is also true Eg- in the SW/IW mode the cache site is where data is generated and can be considered as the local site and
the home site can be considered as the remote site. So these terms local or remote site can be applied to both the cache and home sites,
based on location of data creation and direction of data flow.
— When the cache needs to be smaller than home, you can save storage costs.
— Eviction means that data blocks of files residing in the cache are removed from the local file system,
but the metadata of these files is retained at the cache.
— Automatic Eviction: The automatic eviction is based on fileset quotas.
— Manual Eviction: can be done for specific files selected by an Information Lifecycle Management
(ILM) policy. This adds more flexibility in terms of specifying which particular files shouldn’t be eating
up your disk space.
• Healthcare
and Life
Sciences
• Data
Archives/
Libraries
• Govt.
Institutions
• Central and
Branch Offices
— Peer-snapshots.
— Cache can continue despite no connectivity with home or periods when home is inaccessible.
Note: AFM DR feature is disabled by default and customers need to review the deployment with
the Spectrum Scale Development for approval
19 © Copyright IBM Corporation 2018
Disaster Recovery
AFM
(configured as
Client switches to secondary)
secondary on failure
NAS client
AFM
(configured as primary)
— https://www.ibm.com/support/knowledgecenter/STXKQY_5.0.2/com.ibm.spectrum.scale.v5r02.doc/b
1lins_quickreference_afm.htm
— https://www.ibm.com/support/knowledgecenter/STXKQY_5.0.2/com.ibm.spectrum.scale.v5r02.doc/bl
1xx_soc.htm
Trishali Nayar
Spectrum Scale
ntrishal@in.ibm.com