Professional Documents
Culture Documents
Archive Center 16.2.1 Cluster Document Ingestion Whitepaper
Archive Center 16.2.1 Cluster Document Ingestion Whitepaper
Version: 16.2.1
Task/Topic Performance
Audience: Administrators, Decision makers
Platform: RHEL 7.4, Oracle 12.1.0.2
Document ID: 900002
Updated: June 11, 2019
Contents
Audience ...................................................................................................................... 3
Disclaimer .................................................................................................................... 3
Executive summary .................................................................................................... 4
Assessment overview ................................................................................................ 5
Objectives ............................................................................................................... 5
Testing methodology .................................................................................................. 6
Test setup ............................................................................................................... 6
Test strategy ........................................................................................................... 6
Test types ............................................................................................................... 6
Test results .................................................................................................................. 7
Test with four nodes and 10/20/30/40 clients ......................................................... 7
Client-side statistics ......................................................................................... 7
Observations ............................................................................................. 7
Server metrics .................................................................................................. 8
Observations ............................................................................................. 9
Test with differently sized documents ..................................................................... 9
Client-side statistics ....................................................................................... 10
Conclusions............................................................................................................... 11
Appendices ................................................................................................................ 12
Appendix A - Test environment ...................................................................... 12
Hardware and software resources .......................................................... 13
Data set ................................................................................................... 13
Appendix B - Application and system tuning guide ........................................ 14
Appendix C - References ............................................................................... 15
Audience
The document is intended for a technical audience that is planning an implementation
of OpenText™ products. OpenText recommends consulting with OpenText
Professional Services, who can assist with the specific details of individual
implementation architectures.
Disclaimer
The tests and results described in this document apply only to the OpenText
configuration described herein. For testing or certification of other configurations,
contact OpenText for more information.
All tests described in this document were run on equipment located in the OpenText
Performance Laboratory and were performed by the OpenText Performance
Engineering Group. Note that using a configuration similar to that described in this
document, or any other certified configuration, does not guarantee the results
documented herein. There may be parameters or variables that were not contemplated
during these performance tests that could affect results in other test environments.
For any OpenText production deployment, OpenText recommends a rigorous
performance evaluation of the specific environment and applications to avoid any
configuration or custom development bottlenecks that hinder overall performance.
All results in this paper are based on server–side measurements and do not capture
browser rendering of results. Actual timings including client-side timings (for example,
from browsers) may vary significantly depending on the client machine specifications,
the client network, browser variations, and other conditions of the user’s environment.
Executive summary
This paper describes the testing efforts undertaken by OpenText to benchmark the
performance of document creation with a four-node Archive Center (AC) 16.2.1 cluster.
Tests were run with different numbers of concurrent test clients to show how
performance scaled with increasing load, and with a range of different document sizes.
This paper is specific to AC 16.2.1 deployed on RHEL 7.4 in a four-node cluster, with
Oracle 12.1.0.2.
Some key findings are summarized below:
• Using a document size of 100 KB, the chart below shows the throughput and
average response time with different numbers of concurrent clients:
Figure 1 Throughput and Avg. Response Time for tests with 10, 20, 30, 40 Clients
• With a higher load of 120 clients, the chart below shows the throughput and
average response time with different document sizes:
Figure 2 Throughput and Avg. Response Time for tests with different document sizes
Assessment overview
Objectives
This assessment strives to test the responsiveness, resource consumption, and
scalability of a four-node Archive Center 16.2.1 cluster, deployed in a virtual
environment.
The following were the specific objectives of this assessment:
Determine the performance characteristics for creating documents using a test client
with the following approaches:
a. With increasing number of clients creating documents concurrently, document
throughput as load increases.
b. With different document sizes, running at the same high load level, document the
impact of document size on throughput.
Testing methodology
This section describes the tests that were executed as part of this assessment.
Test setup
All the tests were executed with a four-node AC 16.2.1 cluster running on RHEL 7.4,
with an Oracle 12.1.0.2 database. The test environment is described in Appendix A.
Test strategy
All tests were executed using a Perl based test client with the following characteristics:
• Documents were created in AC through the libdsh API.
• The test client connected to AC through an F5 Load Balancer URL.
• The test client created a specified number of documents sequentially with no think
time between requests.
• Each test ran for a duration of roughly 8 to 10 minutes.
• A batch file was used to initiate multiple instances of the test client to run tests with
different numbers of parallel clients.
Test types
Tests were run as load tests with all test client instances created simultaneously at the
start of the test, running for an overall duration of roughly 10 minutes. The following
specific tests were executed to meet the assessment objectives:
• 10 Clients with 100 KB document size (all running on test client 1)
• 20 Clients with 100 KB document size (all running on test client 1)
• 30 Clients with 100 KB document size (all running on test client 1)
• 40 Clients with 100 KB document size (all running on test client 1)
• 120 Clients with document sizes 10 KB, 20 KB, 50 KB, 100 KB (equally split
between test client 1 and 2)
Test results
This section of the report contains detailed results for each of the test types outlined in
the Testing Methodology section.
Client-side statistics
Table 1 Client-Side Statistics for tests with 10, 20, 30, 40 Clients
Throughput Avg
Number of Duration Total Throughput
Percent Response
Clients (sec) Documents (docs/sec)
Increase Time (ms)
10 575 100000 173.5 57.7
Figure 3 Throughput and Avg. Response Time for tests with 10, 20, 30, 40 Clients
Observations
• Throughput increased as the number of concurrent clients increased, although not
at a linear rate, increasing by 31.9% with 20 clients, and then by 17% with the
increase to 30 and then 40 clients.
Server metrics
This table shows server resource usage during the four tests.
Table 2 Resource Usage for Tests with 10, 20, 30, and 40 Clients
AC Node 1 AC Node 2 AC Node 3 AC Node 4 Oracle
CPU Usage
10 Client 17.8 18.2 17.3 20.42 5.5
Avg. CPU Usage (%) 40 Client 34.6 35.8 34.35 39.70 10.60
Memory Usage
10 Client 4.8 5.1 5.6 5.6 0.30
Disk Usage
10 Client <1 <1 <1 <1
10 Client 0 0 0 0 0
Disk Usage
10 Client 7.89 7.89 3.43 4.16
Network Usage
10 Client 54.19 53.65 53.98 40.47 11.24
Observations
• Resource usage across the Oracle server and all four AC nodes was well below
saturation.
• The higher CPU usage and Avg. Bytes Sent on AC node 4 reflects that it was acting
as the master in the AC cluster.
• Tests with more clients had lower MemFree values. This was mostly due to
increased memory usage by the Linux file cache.
• The low available memory on the Oracle Linux VM was also due to high cached
memory usage. Cached memory was around 26 GB during the tests.
Client-side statistics
Table 3 Client-Side Statistics for tests with different sized documents
Avg
Number Document
Throughput Response
of Clients Size
Time (ms)
10KB 817.3 147.0
120
Figure 4 Throughput and Avg. Response Time for tests with different sizes
Conclusions
The tests that were completed lead to the following conclusions:
• Using a document size of 100 KB, the tests achieved a document creation
throughput of 173.5, 228.9, 267.9, and 313.5 documents/second when running
with 10, 20, 30, and 40 concurrent clients, respectively.
• With 120 concurrent test clients, the tests achieved a document creation
throughput of 817 with document size 10 KB, 768 with document size 20 KB, 775.9
with document size 50 KB, and 506 with document size 100 KB.
Appendices
Appendix A - Test environment
The diagram below shows the architecture used for these tests.
Client Tool
F5 LB
• Each AC node had its own 50 GB SAN partition (Hitachi AMS 2100 Storage Array),
used for a disk volume associated with the disk buffer.
• Each AC node connected to a common 100 GB NAS share, which held disk
volumes associated to each of the pools.
4 core (E5-
5 OTDS 2680 v4 32 GB Win 2012 R2 OTDS 16.2.2
2.4GHz)
4 cores (Xeon
6 Oracle 12 E5-2695 v2 32 GB RHEL 7.1 Oracle 12.1.0.2
2.4GHz)
Data set
The initial AC setup was an empty system.
Most tests used text files of size 100 KB as the source for adding content into the
system. A final set of tests used some additional source documents of size 10 KB,
20 KB, and 50 KB.
F5 LB
Virtual server was created with a new pool containing the 4 AC nodes,
Virtual Server set up for round-robin load balancing for HTTP traffic on port 8080.
Appendix C - References
• Archive Center Cluster Guide:
https://knowledge.opentext.com/knowledge/piroot/ar/v160200-00/ar-
iclu/en/html/_manual.htm
About OpenText
OpenText enables the digital world, creating a better way for organizations to work with information, on premises or in the
cloud. For more information about OpenText (NASDAQ: OTEX, TSX: OTC) visit opentext.com.
Connect with us:
16
www.opentext.com/contact
Copyright © 2019 Open Text SA or Open Text ULC (in Canada).
All rights reserved. Trademarks owned by Open Text SA or Open Text ULC (in Canada).