cbc-pillowfight - Man Page
Stress Test for Couchbase Client and Cluster
Synopsis
cbc-pillowfight [Options]
Description
cbc-pillowfight creates a specified number of threads each looping and performing get and set operations within the cluster.
The stress test operates in the following order
- It will pre-load the items in the cluster (set by the --num-items option)
- Once the items are all loaded into the cluster, it will access all the items (within the --num-items) specification, using a combination of storage and retrieval operations (the proportion of retrieval and storage operations are controlled via the --set-pct option).
- Operations are scheduled in batches. The batches represent a single pipeline (or network buffer) which is filled with a certain amount of operations (see the --batch-size option). These batch sizes are then sent over to the cluster and the requests are serviced by it.
Tuning
Getting the right benchmark numbers highly depends on the type of environment the client is being run in. The following provides some information about specific settings which may make pillowfight generate more operations.
- Increasing the batch size will typically speed up operations, but increasing the batch size too much will actually slow it down. Additionally, very high batch sizes will cause high memory usage.
- Adding additional threads will create additional client objects and connections, potentially increasing performance. Adding too many threads will cause local and network resource congestion.
- Decreasing the item sizes (the --min-size and --max-size options) will always yield higher performance in terms of operationd-per-second.
- Limiting the working set (i.e. --num-items) will decrease the working set within the cluster, thereby increasing the chance that a given item will be inside the server's CPU cache (which is extremely fast), rather than in main memory (slower), or disk (much slower)
The benchmark tool sets up SIGQUIT (CTRL-/) handler and dumps useful diagnostics and metrics to STDERR on this signal.
Options
Options may be read either from the command line, or from a configuration file (see cbcrc(4)):
The following options control workload generation:
- -B, --batch-size=BATCHSIZE
This controls how many commands are scheduled per cycles. To simulate one operation at a time, set this value to 1.
- -I, --num-items=NUMITEMS
Set the total number of items the workload will access within the cluster. This will also determine the working set size at the server and may affect disk latencies if set to a high number.
- -p, --key-prefix=PREFIX
Set the prefix to prepend to all keys in the cluster. Useful if you do not wish the items to conflict with existing data.
- --collection=scopeName.collectionName
Access data based on full collection name path. Multiple --collection filters can specify the different scopes with different collection names. Note that default collection will not be used if the collection was specified, to enable default collection along with the named ones, it should be specified explicitly --collection=_default._default.
- -t, --num-threads=NTHREADS
Set the number of threads (and thus the number of client instances) to run concurrently. Each thread is assigned its own client object.
- -r, --set-pct=PERCENTAGE
The percentage of operations which should be mutations. A value of 100 means only mutations while a value of 0 means only retrievals.
- -n, --no-population
By default cbc-pillowfight will load all the items (see --num-items) into the cluster and then begin performing the normal workload. Specifying this option bypasses this stage. Useful if the items have already been loaded in a previous run.
- --populate-only
Stop after population. Useful to populate buckets with large amounts of data.
- -m, --min-size=MINSIZE:
- -M, --max-size=MAXSIZE
Specify the minimum and maximum value sizes to be stored into the cluster. This is typically a range, in which case each value generated will be between --min-size and --max-size bytes.
- -E, --pause-at-end
When the workload completes, do not exit immediately, but wait for user input. This is helpful for analyzing open socket connections and state.
- -c, --num-cycles
Specify the number of times the workload should cycle. During each cycle an amount of --batch-size operations are executed. Setting this to -1 will cause the workload to run infinitely.
- --sequential
Specify that the access pattern should be done in a sequential manner. This is useful for bulk-loading many documents in a single server.
- --start-at
This specifies the starting offset for the items. The items by default are generated with the key prefix (--key-prefix) up to the number of items (--num-items). The --start-at value will increase the lower limit of the items. This is useful to resume a previously cancelled load operation.
- -T, --timings
Enabled timing recorded. Timing histogram will be dumped to STDERR on SIGQUIT (CTRL-/). When specified second time, it will dump a histogram of command timings and latencies to the screen every second.
- -e, --expiry=SECONDS
Set the expiration time on the document for SECONDS when performing each operation. Note that setting this too low may cause not-found errors to appear on the screen.
@@common-options.markdown@@
- d, --durability=LEVEL
Specify durability level for mutation operations. Known values are: "none", "majority", "majority_and_persist_to_active", "persist_to_majority".
- p, --persist-to=NUMNODES
Wait until the item has been persisted to at least NUMNODES nodes' disk. If NUMNODES is 1 then wait until only the master node has persisted the item for this key. You may not specify a number greater than the number of nodes actually in the cluster. -1 is special value, which mean to use all available nodes.
- r --replicate-to=NREPLICAS
Wait until the item has been replicated to at least NREPLICAS replica nodes. The bucket must be configured with at least one replica, and at least NREPLICAS replica nodes must be online. -1 is special value, which mean to use all available replicas.
- --lock=TIME
This will retrieve and lock an item before update, making it inaccessible for modification until the update completed, or TIME has passed.
- --json
Make pillowfight store document as JSON rather than binary. This will allow the documents to nominally be analyzed by other Couchbase services such as Query and MapReduce.
JSON documents are created by creating an empty JSON object ({}) and then repeated populating it with Field_%d property names (where %d is 1 and higher), and setting its value to a repeating asterisk * up to 16 times:
{ "Field_1": "****************", "Field_2": "****************", "Field_3": "****************", "Field_4": "****************", "Field_5": "********" }
When using document size constraints, be aware that the minimum and maximum sizes (--min-size and --max-size) are not strict limits, and that the resultant sizes may be bigger or smaller by a few bytes in order to satisfy the requirements of proper JSON syntax.
- --noop
Use couchbase NOOP operations when running the workload. This mode ignores population, and all other document operations. Useful as the most lightweight workload.
- --subdoc
Use couchbase sub-document operations when running the workload. In this mode pillowfight will use Couchbase sub-document operations http://blog.couchbase.com/2016/february/subdoc-explained to perform gets and sets of data. This option must be used with --json
- --pathcount
Specify the number of paths a single sub-document operation should contain. By default, each subdoc operation operates on only a single path within the document. You can specify multiple paths to atomically executed multiple subdoc operations within a single command.
This option does not affect the --batch-size option as a subdoc command is considered as a single command (with respect to batching) regardless of how many operations it contains.
Additional Options
The following options may be included in the connection string (via the -U option) as URI-style query params (e.g. couchbase://host/bucket?option1=value1&option2=value2) or as individual key=value pairs passed to the -D switch (e.g. -Doption1=value1 -Doption2=value). The -D will internally build the connection string, and is provided as a convenience for options to be easily passed on the command-line
@@common-additional-options.markdown@@
Examples
Connection Examples
The following examples show how to connect pillowfight to different types of cluster configurations.
Connect to a bucket (a_bucket) on a cluster on a remote host (for servers version 5.x+). It uses account 'myname' and asks password interactively:
cbc-pillowfight -U couchbase://192.168.33.101/a_bucket -u myname -P-
Run against a password-less bucket (a_bucket) on a cluster on a remote host (for servers older than 5.x):
cbc-pillowfight -U couchbase://192.168.33.101/a_bucket
Connect to an SSL cluster at secure.net. The certificate for the cluster is stored locally at /home/couchbase/couchbase_cert.pem:
cbc-pillowfight -U couchbases://secure.net/topsecret_bucket?certpath=/home/couchbase/couchbase_cert.pem
Connect to an SSL cluster at secure.net, ignoring certificate verification. This is insecure but handy for testing:
cbc-pillowfight -U couchbases://secure.net/topsecret_bucket?ssl=no_verify
Connect to a password protected bucket (protected) on a remote host (for servers older than 5.x):
cbc-pillowfight -U couchbase://remote.host.net/protected -P - Bucket password: ......
Connect to a password protected bucket (for servers older than 5.x), specifying the password on the command line (INSECURE, but useful for testing dummy environments)
cbc-pillowfight -U couchbase://remote.host.net/protected -P t0ps3cr3t
Connect to a bucket running on a cluster with a custom REST API port
cbc-pillowfight -U http://localhost:9000/default
Connec to bucket running on a cluster with a custom memcached port
cbc-pillowfight -U couchbase://localhost:12000/default
Connect to a memcached (http://memcached.org) cluster using the binary protocol. A vanilla memcached cluster is not the same as a memcached bucket residing within a couchbase cluster (use the normal couchbase:// scheme for that):
cbc-pillowfight -U memcached://host1,host2,host3,host4
Connect to an SSL cluster at secure.net:
cbc-pillowfight -U couchbases://secure.net/topsecret_bucket
Run against a bucket (a_bucket) on a cluster on a remote host:
cbc-pillowfight -U couchbase://192.168.33.101/a_bucket
Run against a bucket (a_bucket) and use collections app.users and backup.stage:
cbc-pillowfight --spec couchbase://127.0.0.1/a_bucket --username Administrator --password password \ --collection app.users --collection backup.stage
Benchmark Examples
The following examples show how to configure different types of workloads with pillowfight.
Run with 20 threads/instances, each doing one operation at a time:
cbc-pillowfight -t 20 -B 1
Run 100 iterations of 2MB item sizes, using a dataset of 50 items
cbc-pillowfight -M $(1024*1024) -m $(1024*1024) -c 100 -I 50
Use JSON documents of 100k each
cbc-pillowfight --json -m 100000 -M 100000
Stress-test sub-document mutations
cbc-pillowfight --json --subdoc --set-pct 100
Todo
Rather than spawning threads for multiple instances, offer a way to have multiple instances function cooperatively inside an event loop.
Bugs
This command's options are subject to change.
See Also
History
The cbc-pillowfight tool was first introduced in libcouchbase 2.0.7
Referenced By
cbc(1), cbc-n1qlback(1), cbcrc(4).