|bonnie++(8)||System Manager's Manual||bonnie++(8)|
Bonnie++ is a program to test hard drives and file systems for performance or the lack therof. There are a many different types of file system operations which different applications use to different degrees. Bonnie++ tests some of them and for each test gives a result of the amount of work done per second and the percentage of CPU time this took. For performance results higher numbers are better, for CPU usage and latency lower are better (NB a configuration scoring a performance result of 2000 and a CPU result of 90% is better in terms of CPU use than a configuration delivering performance of 1000 and CPU usage of 60%).
There are two sections to the program's operations. The first is to test the IO throughput in a fashion that is designed to simulate some types of database applications. The second is to test creation, reading, and deleting many small files in a fashion similar to the usage patterns of programs such as Squid or INN.
All the details of the tests performed by Bonnie++ are contained in the file /usr/share/doc/bonnie++/readme.html
- the directory to use for the tests.
- the level of concurrency (default 1). The number of copies of the test to be performed at the same time.
- the size of the file(s) for IO performance measures in megabytes. If the
size is greater than 1G then multiple files will be used to store the
data, and each file will be up to 1G in size. This parameter may include
the chunk size seperated from the size by a colon. The chunk-size is
measured in bytes and must be a power of two from 256 to 1048576, the
default is 8192. NB You can specify the size in giga-bytes or the
chunk-size in kilo-bytes if you add g or k to the end of the
number respectively. This parameter may also include the number of seeks
(default 8192) and the number of seeker processes (default 5).
If the specified size is 0 then this test will be skipped.
- the number of files for the file creation test. This is measured in
multiples of 1024 files. This is because no-one will want to test less
than 1024 files, and we need the extra space on braille displays.
If the specified number is 0 then this test will be skipped.
The default for this test is to test with 0 byte files. To use files of other sizes you can specify number:max:min:num-directories:chunk-size where max is the maximum size and min is the minimum size (both default to 0 if not specified). If minimum and maximum sizes are specified then every file will have a random size from the range min..max inclusive. If you specify a number of directories then the files will be evenly distributed amoung that many sub-directories.
If max is -1 then hard links will be created instead of files. If max is -2 then soft links will be created instead of files.
- name of the machine - for display purposes only.
- RAM size in megabytes. If you specify this the other parameters will be checked to ensure they make sense for a machine of that much RAM. You should not need to do this in general use as it should be able to discover the RAM size. NB If you specify a size of 0 then all checks will be disabled...
- number of test runs. This is useful if you want to perform more than one test. It will dump output continuously in CSV format until either the number of tests have been completed, or it is killed.
- user-id to use. When running as root specify the UID to use for the tests. It is not recommended to use root (since the occasion when a Bonnie++ bug wiped out someone's system), so if you really want to run as root then use -u root. Also if you want to specify the group to run as then use the user:group format. If you specify a user by name but no group then the primary group of that user will be chosen. If you specify a user by number and no group then the group will be nogroup.
- group-id to use. Same as using :group for the -u parameter, just a different way to specify it for compatibility with other programs.
- quiet mode. If specified then some of the extra informational messages will be suppressed. Also the csv data will be the only output on standard out and the plain text data will be on standard error. This means you can run bonnie++ -q >> file.csv to record your csv data.
- -f size-for-char-io
- fast mode control, skips per-char IO tests if no parameter, otherwise specifies the size of the tests for per-char IO tests (default 20M).
- no write buffering. fsync() after every write.
- number of processes to serve semaphores for. This is used to create the semaphores for synchronising multiple Bonnie++ processes. All the processes which are told to use the semaphore with -ys will start each test with synchronization. Use "-p -1" to delete the semaphore.
- -y s|p
- perform synchronization before each test. Option s for semaphores and option p for prompting.
- use direct IO (O_DIRECT) for the bulk IO tests
- -z seed
- random number seed to repeat the same test.
- -Z random-file
- file containing random data in network byte order.
bonnie++ -y > out1 &
bonnie++ -y > out2 &
bonnie++ -y > out3 &
The second type of output is CSV (Comma Seperated Values). This can easily be imported into any spread-sheet or database program. Also I have included the programs bon_csv2html and bon_csv2txt to convert CSV data to HTML and plain-ascii respectively.
The "Name:Size etc" field has the name, filesize, IO chunk size, concurrency, number of seeks, and number of seek processes separated by : characters.
For every test two numbers are reported, the amount of work done (higher numbers are better) and the percentage of CPU time taken to perform the work (lower numbers are better). If a test completes in less than 500ms then the output will be displayed as "++++". This is because such a test result can't be calculated accurately due to rounding errors and I would rather display no result than a wrong result.
Data volumes for the 80 column text display use "K" for KiB (1024 bytes), "M" for MiB (1048576 bytes), and "G" for GiB (1073741824 bytes). So K/sec means a multiple of 1024 bytes per second.
The documentation, the Perl scripts, and all the code for testing the creation of thousands of files was written by Russell Coker, but the entire package is under joint copyright with Tim Bray.
SIGXCPU and SIGXFSZ act like SIGINT.
See http://etbe.coker.com.au/category/benchmark for further information.