Professional Documents
Culture Documents
3 Intel Optane PMem and SSDs (SXXW1212r4-C)
3 Intel Optane PMem and SSDs (SXXW1212r4-C)
Introduction
This presentation introduces Intel® Optane™ PMem, describes what it is, and discusses why it is important to
Lenovo. This course has been updated to reflect new workloads that have been qualified by Intel and software
vendors.
Intel® Optane™ SSDs - solid-state drives that use Intel® Optane™ PMem technology - are also introduced
here.
Objectives
This course introduces Intel® Optane™ Persistent Memory, abbreviated PMem. This technology is supported
in Lenovo servers with certain 2nd or 3rd generation Intel® Xeon® Scalable Family processors.
In the uppermost gap there is a real need to extend memory capacity. Intel® Optane™ Persistent Memory
increases the size of available memory and adds persistence. Persistence is important because if there is a
power loss, the data is protected and the memory reload time on restart is eliminated. There is also a
performance gain from persistence.
To fill the storage gap, Intel has introduced Intel® Optane™ SSDs. Optane-based SSDs are different than
NAND-based SSDs, and provide increased lOPs and high write performance and consistency. The need for
speed and writing/rewriting of data, like caching, logging, and journaling solutions, are the ideal scenarios for
the use of Intel® Optane™ SSDs.
Intel® Optane™ Persistent Memory fills the memory capacity gap in the data center, and in the future will do the
same for client and commercial systems.
Intel® Optane™ SSDs fill the storage performance gap to bring much lower latencies and faster access to
data sets at the top of the storage layer.
Finally, and not discussed in this course, Intel® 3D NAND SSDs fill the cost performance gap in the capacity
tier, bringing massive density at low cost. These SSDs are deployed in the warm data tier, and will move into
the cold tier of the storage hierarchy currently serviced by HDDs.
• OS Operations: Paging
Paging and context switching refer to operations used by operating systems. Paging swaps memory
pages to and from disk to make disk look like part of main memory. Disk is slow, of course, and the
paging process adds latency as well.
DMA - Direct Memory Access - enables faster I/O operations by allowing peripheral devices to access
memory directly, rather than going through the CPU.
RDMA - Remote Direct Memory Access - allows peripheral devices to access memory on other hosts
through a high-speed network connection.
Intel® Optane™ PMem technology expands server memory capadty beyond what has been possible
with traditional memory DIMMs, and costs less per gigabyte than standard memory. DRAM prices have
fallen in recent years, but Intel has committed to keeping Optane™ PMem pricing competitive with
DRAM. These modules can also add persistence capability to the memory tier, resulting in reduced
downtime due to faster reboots and application restoration times.
• Memory Mode
The first mode we are going to discuss is Memory Mode. This mode provides affordable high capacity
memory, which is volatile like traditional DIMMs.
The advantage of memory mode is that it can provide large amounts of system memory at a lower total
cost than can be achieved by a DRAM-only solution. This allows increased density in virtualized
environments, where memory capacity is often a limitation.
PMem makes higher capacities affordable, because PMem is less costly than DRAM.
Applications and the operating system perceive a pool of volatile memory, as is the case in DRAM- only
systems. In this mode, no specific PMem programming is required in the applications, and data will
not be saved in the event of a power loss.
• Memory Mode
The first mode we are going to discuss is Memory Mode. This mode provides affordable high capacity
memory, which is volatile like traditional DIMMs.
The advantage of memory mode is that it can provide large amounts of system memory at a lower total
cost than can be achieved by a DRAM-only solution. This allows increased density in virtualized
environments, where memory capacity is often a limitation.
PMem makes higher capacities affordable, because PMem is less costly than DRAM.
Applications and the operating system perceive a pool of volatile memory, as is the case in DRAM- only
systems. In this mode, no specific PMem programming is required in the applications, and data will
not be saved in the event of a power loss.
First, application aligned data management gives applications direct access to memory, bypassing the
operating system and kernel. This consistently reduces latency, which results in faster insights from
data. This greatly increases capacity for inmemory databases like SAP HAN A.
Second, PMem also provides rapid recovery of applications because application data remains in memory
even after a power failure, and does not need to be loaded from storage. This reduces reboot and
recovery times from hours or minutes to just seconds.
Finally, storing data in Intel® Optane™ PMem lowers latency by allowing the CPU access to stored data
via the memory bus instead of across the I/O bus. This enables super fast storage solutions because the
response time on the memory bus is measured in nanoseconds while the response time over the I/O bus
is typically measured in milliseconds. Loading a block device driver into the OS effectively turns the
PMem into a very fast storage device. This is sometimes referred to as 'Storage over App Direct’.
In App Direct Mode, the DRAM and the Intel® Optane™ PMem both count toward the total platform
memory seen by the OS.
Lenovo supports Intel© Optane™ PMem on ThinkSystem servers configured with supporting processors.
This includes the V2 versions of several servers, which support the 3rd generation Intel Xeon Scalable
Family processors.
Rack Servers
SR570 SR590 SR630 SR630V2 SR650 SR650 V2 SR850 SR860 SR950
Dense Platforms
SD530 SD650
Flex Nodes
SN550 SN550V2 SN850
• PMem Use Cases - General
A number of use cases have been identified for Intel© Optane™ PMem. The areas where Intel®
Optane™ PMem will show the most benefit are highlighted in the categories above. Many use cases
deal with access over high-speed links, such as RDMA access, discussed earlier, and PMoF (PMem
over Fibre).
Intel® Optane™ PMem in App Direct mode allows large SAP HANA
<1> databases to be memory- resident, and also improves startup time after an outage. Servers memory
capacity is often limited by the number of permitted memory slots. Optane™ PMem modules are up to
<2> 512 GB in size, and can expand server memory beyond the DRAM limit WSSD is a software-based
<3> storage solution. Intel® Optane™ PMem allows the storage visible to a server to be accessed
through the memory bus, vastly improving performance.
• Supported Applications
The applications shown here support the use of Intel© Optane™ PMem, or, where they’re open source,
have been tested by Intel.
Note that AI, analytics, and databases feature heavily. Theses are all areas where large amounts of data
kept in memory can improve performance significantly. The infrastructure and storage workloads and
applications benefit from large quantities of memory running at close to DRAM speeds, and which can
be used as fast storage. Note also that operating systems are well represented.
• App Direct Mode Support
In this mode, as noted previously, the application must be PMem aware. That awareness, though it
involves programming changes, provides many benefits to the application environment Note that the
applications are those that benefit from large amounts of low-latency memory, typically databases and AI
environments.
NetApp MAX Data uses PMem as a data tier, and requires no changes on the part of applications that
use the storage.
• Memory Mode
Memory Mode requires no changes to the OS or the application. The slide shows applications which
have been evaluated by Intel, in most cases with ISV partners, for use in this mode. Even in this mode,
which turns PMem into non-persistent memory, the workloads are very similar to those shown in the
other modes. Note that the large quantity of memory in infrastructure and storage environments will help
relieve pressure on the (slower) disk drives by caching data. These environments are converged or
hyperconverged; external storage systems will implement their own cache.
• Key Takeaway
Here are the Intel® Optane™ PMem key takeaways to remember from this presentation.
Intel® Optane™ PMem provides high capacity, affordable memory, with persistence.
Module performance is comparable to DRAM but costs less per gigabyte than standard memory DIMMs.
Modules are available in 128, 256, and 512 GB capacities and are DDR4 pin compatible. Servers must
be populated with both Intel® Optane™ PMem and traditional DRAM. This technology enables up to 6
terabytes per CPU memory capacity.
There are multiple operational modes available to provide flexibility in feature deployment:
Memory Mode implements high capacity, non- persistent memory with DRAM caching. No software or
application changes are necessary.
App Direct Mode implements byte addressable, low latency memory with persistence similar to
storage. This mode requires persistent memory aware OS and application code.
Storage over App Mode implements a block driver to make the PMem look like storage accessible in
sectors [blocks].
NAND media reads and writes in pages, and writes only to blank (erased) pages. The granularity for
erase operations is a block, and erasing a block is a slow process. For performance reasons, page
updates are typically written to a new unused block. As new data is written, old pages become stale.
Stale pages soon occupy significant space on an SSD, and blocks must be freed up to allow new writes
to proceed. Both sequential write and more typical random read/write scenarios follow the same rules.
When a block needs to be freed up, the garbage collection process is launched. Garbage collection is a
cumbersome 3-step process.
In the first step, current pages are copied to an empty block. The second step consists of cleaning up the
entire block by erasing both stale pages and the pages copied to the new block.
Finally, the newly emptied block is available for new writes, and the NAND program/erase cycle starts all
over again. This cumbersome process is why you'll see reputable companies (like Intel) reporting drive
performance numbers using “pre-conditioned" SSDs - meaning there is data already on the drive before
the test starts, versus a fresh-out-of-the-box drive where data can be written anywhere with no garbage
collection.
This consistent level of performance makes the Optane™ SSDs well suited for caching in high-
performance storage environments such as SDS and HCI.
Drive writes per day (DWPD) is the number of times we can write to the drive per day and read it back
with no data loss. This is a drive lifetime number, not a 24 hour limitation. Optane™ SSD achieves 60
drive writes per day, or 20x better than most NAND SSDs on the market.
When the CPU and memory are not limiting the workload, lower storage latency, higher IOPS, and better
endurance allow each server to support more virtual machines. Additionally, you no longer need 10% of
your total capacity provisioned as cache. In many cases, one or two 375 GB Optane™ SSDs will be
sufficient. The bottom line is, you don’t need to purchase many gigabytes of cache, and you can support
more virtual machines per server with Intel® Optane™ SSDs.
Course Summary
Having completed this course, you should be able to:
60
10
100
30
How does Intel® Optane™ SSD break the storage bottleneck? Choose two [2].
How can Intel® Optane™ SSDs help the performance of a VMware vSAN
system?