Learning steps to benchmark a Droonga cluster and compare it to a Groonga server.
Because Droonga has compatibility to Groonga, you’ll plan to migrate your application based on Groonga to Droonga. Before that, you should benchmark Droonga and confirm that it is better alternative for your application.
Of course you may simply hope to know the difference in performance between Groonga and Droonga. Benchmarking will make it clear.
There are two major indexes to indicate performance of a system.
Latency is the response time, actual elapsed time between two moments: when the system receives a request, and when it returns a response. In other words, for clients, it is the time to wait for each request. At this index, the smaller is the better. In general, latency becomes small for lightweight queries, small size database, or less clients.
Throughput means how many request can be processed in a time. The performance index is described as “queries per second (qps)”. For example, if a Groonga server processed 10 requests in one second, that is described as “10qps”. Possibly there are 10 users (clients), or, there are 2 users and each user opens 5 tabs in his web browser. Anyway, “10qps” means that the Groonga actually accepted and responded for 10 requests while one second is passing.
You can run benchmark with the command drnbench-request-response
, introduced by the Gem package drnbench.
It measures both latency and throughput of the target service.
drnbench-request-response
benchmarks the target service, by steps like following:
Finally, the master process reports minimum/maximum/average elapsed time, “qps”, and other extra information for each case, as a CSV file like:
n_clients,total_n_requests,queries_per_second,min_elapsed_time,max_elapsed_time,average_elapsed_time,200
1,996,33.2,0.001773766,0.238031643,0.019765581680722916,100.0
2,1973,65.76666666666667,0.001558398,0.272225481,0.020047345673086702,100.0
4,3559,118.63333333333334,0.001531184,0.39942581,0.023357554419499882,100.0
6,4540,151.33333333333334,0.001540704,0.501663069,0.042344890696916264,100.0
8,4247,141.56666666666666,0.001483995,0.577100609,0.045836844514480835,100.0
10,4466,148.86666666666667,0.001987089,0.604507078,0.06949704923846833,100.0
12,4500,150.0,0.001782343,0.612596799,0.06902839555222215,100.0
14,4183,139.43333333333334,0.001980711,0.60754769,0.1033681068718623,100.0
16,4519,150.63333333333333,0.00284654,0.653204575,0.09473386513387955,100.0
18,4362,145.4,0.002330049,0.640683693,0.12581190483929405,100.0
20,4228,140.93333333333334,0.003710795,0.662666076,0.1301649290901133,100.0
You can analyze it, draw a graph from it, and so on.
(Note: Performance results fluctuate from various factors. This is just an example on a specific version, specific environment.)
Look at the result above.
See the last columns named 200
.
It means the percentage of HTTP response statuses.
200
is “OK”, 0
is “timed out”.
If clients got 400
, 500
and other errors, they will be also reported.
These information will help you to detect unexpected slow down.
Latency is easily analyzed - the smaller is the better. The minimum and average elapsed time becomes small if any cache system is working correctly on the target. The maximum time is affected by slow queries, system’s page-in/page-out, unexpected errors, and so on.
A graph of latency also reveals the maximum number of effectively acceptable connections in same time.
This is a graph of average_elapsed_time
.
You’ll see that the time is increased for over 4 clients.
What it means?
Groonga can process multiple requests completely parallelly, until the number of available processors. When the computer has 4 processors, the system can process 4 or less requests in same time, without extra latency. And, if more requests are sent, 5th and later requests will be processed after a preceding request is processed. The graph confirms that the logical limitation is true.
A graph helps you to analyze throughput performance.
You’ll see that the “qps” stagnated around 150, for 6 or more clients. This means that the target service can process 150 requests in one second, at a maximum.
In other words, we can describe the result as: 150qps is the maximum throughput performance of this system - generic performance of hardware, software, network, size of the database, queries, and more. If the number of requests for your service is growing up and it is going to reach the limit, you have to do something about it - optimize queries, replace the computer with more powerful one, and so on.
Sending same request patterns to Groonga and Droonga, you can compare performance of each system. If Droonga has better performance, it will become good reason to migrate your service from Groogna to Droonga.
Moreover, comparing multiple results from different number of Droogna nodes, you can analyze the cost-benefit performance for newly introduced nodes.
Assume that there are four Ubuntu 15.10 servers for the new Droogna cluster and they can resolve their names each other:
192.168.100.50
, the host name is node0
192.168.100.51
, the host name is node1
192.168.100.52
, the host name is node2
192.168.100.53
, the host name is node3
One is client, others are Droonga nodes.
If you have any existing service based on Groonga, it becomes the reference. Then you just have to dump all data in your Groonga database and load them to a new Droonga cluster.
Otherwise - if you have no existing service, prepare a new reference database with much data for effective benchmark. The repository wikipedia-search includes some helper scripts to construct your Groonga server (and Droonga cluster), with Japanese Wikipedia pages.
So let’s prepare a new Groonga database including Wikipedia pages, on the node0
.
Determine the size of the database. You have to use good enough size database for benchmarking.
For example, if there are three nodes node0
(8GB RAM), node1
(8GB RAM), and node2
(6GB RAM), then the database should be smaller than 6GB.
Set up the Groonga server, as instructed on the installation guide.
(on node0)
% sudo apt-get -y install software-properties-common
% sudo add-apt-repository -y universe
% sudo add-apt-repository -y ppa:groonga/ppa
% sudo apt-get update
% sudo apt-get -y install groonga
Then the Groonga becomes available.
Download the archive of Wikipedia pages and convert it to a dump file for Groonga, with the rake task data:convert:groonga:ja
.
You can specify the number of records (pages) to be converted via the environment variable MAX_N_RECORDS
(default=5000).
(on node0)
% cd ~/
% git clone https://github.com/droonga/wikipedia-search.git
% cd wikipedia-search
% bundle install --path vendor/
% time (MAX_N_RECORDS=1500000 bundle exec rake data:convert:groonga:ja \
data/groonga/ja-pages.grn)
Because the archive is very large, downloading and data conversion may take time.
After that, a dump file ~/wikipedia-search/data/groonga/ja-pages.grn
is there.
Create a new database and load the dump file to it.
This also may take more time:
(on node0)
% mkdir -p $HOME/groonga/db/
% groonga -n $HOME/groonga/db/db quit
% time (cat ~/wikipedia-search/config/groonga/schema.grn | groonga $HOME/groonga/db/db)
% time (cat ~/wikipedia-search/config/groonga/indexes.grn | groonga $HOME/groonga/db/db)
% time (cat ~/wikipedia-search/data/groonga/ja-pages.grn | groonga $HOME/groonga/db/db)
Note: number of records affects to the database size. Just for information, my results are here:
Start the Groonga as an HTTP server.
(on node0)
% groonga -p 10041 -d --protocol http $HOME/groonga/db/db
OK, now we can use this node as the reference for benchmarking.
Install Droonga to all nodes.
Because we are benchmarking it via HTTP, you have to install both services droonga-engine
and droonga-http-server
for each node.
(on node0)
% host=node0
% curl https://raw.githubusercontent.com/droonga/droonga-engine/master/install.sh | \
sudo HOST=$host bash
% curl https://raw.githubusercontent.com/droonga/droonga-http-server/master/install.sh | \
sudo ENGINE_HOST=$host HOST=$host PORT=10042 bash
% sudo droonga-engine-catalog-generate \
--hosts=node0,node1,node2
% sudo service droonga-engine start
% sudo service droonga-http-server start
(on node1)
% host=node1
...
(on node2)
% host=node2
...
Note: to start droonga-http-server
with a port number different from Groonga, we should specify another port 10042
via the PORT
environment variable, like above.
Make sure that Droonga’s HTTP server is actualy listening the port 10042
and it is working as a cluster with three nodes:
(on node0)
% curl "http://node0:10042/droonga/system/status"
{
"nodes": {
"node0:10031/droonga": {
"live": true
},
"node1:10031/droonga": {
"live": true
},
"node2:10031/droonga": {
"live": true
}
}
}
Next, prepare the Droonga database.
You can generate messages for Droonga from Groonga’s dump result, by the grn2drn
command.
Install grn2drn
Gem package to activate the command, to the Groonga server computer.
(on node0)
% sudo gem install grn2drn
And, the grndump
command introduced as a part of rroonga
Gem package provides ability to extract all data of an existing Groonga database, flexibly.
If you are going to extract data from an existing Groonga server, you have to install rroonga
before that.
(on Ubuntu server)
% sudo apt-get -y install software-properties-common
% sudo add-apt-repository -y universe
% sudo add-apt-repository -y ppa:groonga/ppa
% sudo apt-get update
% sudo apt-get -y install libgroonga-dev
% sudo gem install rroonga
(on CentOS server)
# rpm -ivh http://packages.groonga.org/centos/groonga-release-1.1.0-1.noarch.rpm
# yum -y makecache
# yum -y ruby-devel groonga-devel
# gem install rroonga
Then dump schemas and data separately and load them to the Droonga cluster.
(on node0)
% time (grndump --no-dump-tables $HOME/groonga/db/db | \
grn2drn | \
droonga-send --server=node0 \
--report-throughput)
% time (grndump --no-dump-schema --no-dump-indexes $HOME/groonga/db/db | \
grn2drn | \
droonga-send --server=node0 \
--server=node1 \
--server=node2 \
--messages-per-second=100 \
--report-throughput)
Note that you must send requests for schema and indexes to just one endpoint. Parallel sending of schema definition requests for multiple endpoints will break the database, because Droonga cannot sort schema changing commands sent to each node in parallel.
To reduce traffic and system load, you should specify maximum number of inpouring messages per second by the --messages-per-second
option.
If too many messages rush into the Droonga cluster, they may overflow - Droonga may eat up the RAM and slow down the system.
This may take much time.
For example, with the option --messages-per-second=100
, 1500000 records will be synchronized in about 4 hours (we can estimate the required time like: 150000 / 100 / 60 / 60
).
After all, now you have two HTTP servers: Groonga HTTP server with the port 10041
, and Droonga HTTP Servers with the port 10042
.
You must install the benchmark client to the computer.
Assume that you use a computer node3
as the client:
(on node3)
% sudo apt-get update
% sudo apt-get -y upgrade
% sudo apt-get install -y ruby curl
% sudo gem install drnbench
Let’s prepare request pattern files for benchmarking.
First, you have to determine the cache hit rate.
If you have any existing service based on Groonga, you can get the actual cache hit rate of the Groonga database via status
command, like:
% curl "http://node0:10041/d/status"
[
[
0,
1412326645.19701,
3.76701354980469e-05
],
{
"max_command_version": 2,
"alloc_count": 158,
"starttime": 1412326485,
"uptime": 160,
"version": "4.0.6",
"n_queries": 1000,
"cache_hit_rate": 0.5,
"command_version": 1,
"default_command_version": 1
}
]
The cache hit rate appears as "cache_hit_rate"
.
0.5
means 50%, then a half of responses are returned from cached results.
If you have no existing service, you should assume that the cache hit rate becomes 50%.
To measure and compare performance of Groonga and Droonga properly, you should prepare request patterns for benchmarking which make the cache hit rate near the actual rate. So, how do it?
You can control the cache hit rate by the number of unique request patterns, calculated with the expression:
N = 100 / (cache hit rate)
, because Groonga and Droonga (droonga-http-server
) cache 100 results at a maximum by default.
When the expected cache hit rate is 50%, the number of unique requests is calculated as: N = 100 / 0.5 = 200
Note: if the actual rate is near zero, the number of unique requests becomes too huge! For such case you should carry up the rate to 0.01 (1%) or something.
The format of the request patterns list for drnbench-request-response
is the plain text, a list of request paths for the host.
Here is a short example of requests for Groonga’s select
command:
/d/select?command_version=2&table=Pages&limit=10&match_columns=title&output_columns=title&query=AAA
/d/select?command_version=2&table=Pages&limit=10&match_columns=title&output_columns=title&query=BBB
...
If you have any existing service based on Groonga, the list should be generated from the actual access log, query log, and so on. Patterns similar to actual requests will measure performance of your system more effectively. To generate 200 unique request patterns, you just have to collect 200 unique paths from your log.
Otherwise, you’ll have to generate list of request paths from something. See the next section.
To generate 200 unique request patterns, you have to prepare 200 terms.
Moreover, all of terms must be effective search term for the Groonga database.
If you use randomly generated terms (like P2qyNJ9L
, Hy4pLKc5
, D5eftuTp
, …), you won’t get effective benchmark result, because “not found” results will be returned for most requests.
So there is a utility command drnbench-extract-searchterms
.
It generates list of terms from Groonga’s select result, like:
% curl "http://node0:10041/d/select?command_version=2&table=Pages&limit=10&output_columns=title" | \
drnbench-extract-searchterms
title1
title2
title3
...
title10
drnbench-extract-searchterms
extracts terms from the first column of records.
To collect 200 effective search terms, you just have to give a select result with an option limit=200
.
OK, let’s generate request patterns by drnbench-extract-searchterms
, from a select result.
% n_unique_requests=200
% curl "http://node0:10041/d/select?command_version=2&table=Pages&limit=$n_unique_requests&output_columns=title" | \
drnbench-extract-searchterms --escape | \
sed -r -e "s;^;/d/select?command_version=2\&table=Pages\&limit=10\&match_columns=title,text\&output_columns=snippet_html(title),snippet_html(text),categories,_key\&query_flags=NONE\&sortby=title\&drilldown=categories\&drilldown_limit=10\&drilldown_output_columns=_id,_key,_nsubrecs\&drilldown_sortby=_nsubrecs\&query=;" \
> ./patterns.txt
Note:
&
in the sed script with prefixed backslash, like \&
.--escape
option for drnbench-extract-searchterms
.
It escapes characters unsafe for URI strings.query_flags=NONE
as a part of parameters, if you use search terms by the query
parameter.
It forces ignoring of special characters in the query
parameter, to Groonga.
Otherwise you may see some errors from invalid queries.The generated file patterns.txt
becomes like following:
/d/select?command_version=2&table=Pages&limit=10&match_columns=title,text&output_columns=snippet_html(title),snippet_html(text),categories,_key&query_flags=NONE&sortby=title&drilldown=categories&drilldown_limit=10&drilldown_output_columns=_id,_key,_nsubrecs&drilldown_sortby=_nsubrecs&query=AAA
/d/select?command_version=2&table=Pages&limit=10&match_columns=title,text&output_columns=snippet_html(title),snippet_html(text),categories,_key&query_flags=NONE&sortby=title&drilldown=categories&drilldown_limit=10&drilldown_output_columns=_id,_key,_nsubrecs&drilldown_sortby=_nsubrecs&query=BBB
...
OK, it’s ready to run. Let’s benchmark Groonga and Droonga.
First, run benchmark for Groonga as the reference. Start Groonga’s HTTP server before running, if you configured a node as a reference Groonga server and daemon is stopped.
(on node0)
% groonga -p 10041 -d --protocol http $HOME/groonga/db/db
You can run benchmark with the command drnbench-request-response
, like:
(on node3)
% drnbench-request-response \
--step=2 \
--start-n-clients=0 \
--end-n-clients=20 \
--duration=30 \
--interval=10 \
--request-patterns-file=$PWD/patterns.txt \
--default-hosts=node0 \
--default-port=10041 \
--output-path=$PWD/groonga-result.csv
Important parameters are:
--step
is the number of virtual clients increased on each progress.--start-n-clients
is the initial number of virtual clients.
Even if you specify 0
, initially one client is always generated.--end-n-clients
is the maximum number of virtual clients.
Benchmark is performed progressively until the number of clients is reached to this limit.--duration
is the duration of each benchmark.
This should be long enough to average out the result.
30
(seconds) seems good for my case.--interval
is the interval between each benchmark.
This should be long enough to finish previous benchmark.
10
(seconds) seems good for my case.--request-patterns-file
is the path to the pattern file.--default-hosts
is the list of host names of target endpoints.
By specifying multiple hosts as a comma-separated list, you can simulate load balancing.--default-port
is the port number of the target endpoint.--output-path
is the path to the result file.
Statistics of all benchmarks is saved as a file at the location.While running, you should monitor the system status of the node0
, by top
or something.
If the benchmark elicits Groonga’s performance correctly, Groonga’s process uses the CPU fully (for example, 400%
on a computer with 4 processors).
Otherwise something wrong - for example, too narrow network, too low performance client.
Then you’ll get the reference result of the Groonga.
To confirm the result is valid, check the response of the status
command:
% curl "http://node0:10041/d/status"
[
[
0,
1412326645.19701,
3.76701354980469e-05
],
{
"max_command_version": 2,
"alloc_count": 158,
"starttime": 1412326485,
"uptime": 160,
"version": "4.0.6",
"n_queries": 1000,
"cache_hit_rate": 0.49,
"command_version": 1,
"default_command_version": 1
}
]
Look at the value of "cache_hit_rate"
.
If it is far from the expected cache hit rate (ex. 0.5
), something wrong - for example, too few request patterns.
Too high cache hit rate produces too high throughput unexpectedly.
After that you should stop Groonga to release CPU and RAM resources, if it is running on a Droonga node.
(on node0)
% pkill groonga
Before benchmarking, make your cluster with only one node.
(on node1, node2)
% sudo systemctl stop droonga-engine
% sudo systemctl stop droonga-http-server
(on node0)
% sudo droonga-engine-catalog-generate \
--hosts=node0
% sudo systemctl restart droonga-engine
% sudo systemctl restart droonga-http-server
To clear effects from previous benchmark, you should restart services before each test.
After that the endpoint node0
works as a Droonga cluster with single node.
Make sure that only one node is actually detected:
(on node3)
% curl "http://node0:10042/droonga/system/status"
{
"nodes": {
"node0:10031/droonga": {
"live": true
}
}
}
Run the benchmark.
(on node3)
% drnbench-request-response \
--step=2 \
--start-n-clients=0 \
--end-n-clients=20 \
--duration=30 \
--interval=10 \
--request-patterns-file=$PWD/patterns.txt \
--default-hosts=node0 \
--default-port=10042 \
--output-path=$PWD/droonga-result-1node.csv
Note that the default port is changed from 10041
(Groonga’s HTTP server) to 10042
(Droonga).
Moreover, the path to the result file also changed.
While running, you should monitor the system status of the node0
, by top
or something.
It may help you to analyze what is the bottleneck.
And, to confirm the result is valid, you should check the actual cache hit rate:
% curl "http://node0:10042/statistics/cache"
{
"hitRatio": 49.830717830807124,
"nHits": 66968,
"nGets": 134391
}
Look at the value of "hitRatio"
.
Actual cache hit rate of the HTTP server is reported in percentage like above (the value 49.830717830807124
means 49.830717830807124%
.)
If it is far from the expected cache hit rate, something wrong.
Before benchmarking, join the second node to the cluster.
(on node0, node1)
% sudo droonga-engine-catalog-generate \
--hosts=node0,node1
% sudo systemctl restart droonga-engine
% sudo systemctl restart droonga-http-server
After that both endpoints node0
and node1
work as a Droonga cluster with two nodes.
Make sure that two nodes are actually detected:
(on node3)
% curl "http://node0:10042/droonga/system/status"
{
"nodes": {
"node0:10031/droonga": {
"live": true
},
"node1:10031/droonga": {
"live": true
}
}
}
Run the benchmark.
(on node3)
% drnbench-request-response \
--step=2 \
--start-n-clients=0 \
--end-n-clients=20 \
--duration=30 \
--interval=10 \
--request-patterns-file=$PWD/patterns.txt \
--default-hosts=node0,node1 \
--default-port=10042 \
--output-path=$PWD/droonga-result-2nodes.csv
Note that two hosts are specified via the --default-hosts
option.
If you send all requests to single endpoint, droonga-http-server
will become a bottleneck, because it works as a single process for now.
Moreover, droonga-http-server
and droonga-engine
will scramble for CPU resources.
To measure the performance of your Droonga cluster effectively, you should average out CPU load per capita.
Of course, on the production environment, it should be done by a load balancer, but It’s a hassle to set up a load balancer for just benchmarking.
Instead, you can specify multiple endpoint host names as a comma-separated list for the --default-hosts
option.
And, the path to the result file also changed.
Don’t forget to monitor system status of both nodes while benchmarking. If only one node is busy and another is idling, something wrong - for example, they are not working as a cluster. You also must check the actual cache hit rate of all nodes.
Before benchmarking, join the last node to the cluster.
(on node0, node1)
% sudo droonga-engine-catalog-generate \
--hosts=node0,node1,node2
% sudo systemctl restart droonga-engine
% sudo systemctl restart droonga-http-server
After that all endpoints node0
, node1
, and node2
work as a Droonga cluster with three nodes.
Make sure that three nodes are actually detected:
(on node3)
% curl "http://node0:10042/droonga/system/status"
{
"nodes": {
"node0:10031/droonga": {
"live": true
},
"node1:10031/droonga": {
"live": true
},
"node2:10031/droonga": {
"live": true
}
}
}
Run the benchmark.
(on node3)
% drnbench-request-response \
--step=2 \
--start-n-clients=0 \
--end-n-clients=20 \
--duration=30 \
--interval=10 \
--request-patterns-file=$PWD/patterns.txt \
--default-hosts=node0,node1,node2 \
--default-port=10042 \
--output-path=$PWD/droonga-result-3nodes.csv
Note that both --default-hosts
and --output-path
are changed again.
Monitoring of system status and checking cache hit rate of all nodes are also important.
OK, now you have four results:
groonga-result.csv
droonga-result-1node.csv
droonga-result-2nodes.csv
droonga-result-3nodes.csv
As described, you can analyze them.
For example, you can plot a graph from these results like:
You can explain this graph of latency as:
You can explain this graph of throughput as:
(Note: Performance results fluctuate from various factors. This graph is just an example on a specific version, specific environment.)
In this tutorial, you did prepare a reference Groonga server and Droonga cluster. And, you studied how to prepare request patterns, how measure your systems, and how analyze the result.