List of Benchmarks
FIO
sudo apt-get install fio
mukuls@MUKUL:~/thesis/fsbench/fio-2.1.2$ fio --filename=/tmp/smr/sfs/a
--direct=1 --rw=randrw --bs=4k --rwmixread=100 --iodepth=100 --numjobs=16
--name=4ktest --group_reporting --size=15m
4ktest: (g=0): rw=randrw, bs=4K-4K/4K-4K, ioengine=sync, iodepth=100
...
4ktest: (g=0): rw=randrw, bs=4K-4K/4K-4K, ioengine=sync, iodepth=100
2.0.8
Starting 16 processes
4ktest: Laying out IO file(s) (1 file(s) / 15MB)
4ktest: (groupid=0, jobs=16): err= 0: pid=16861
read : io=245760KB, bw=326809KB/s, iops=81702 , runt= 752msec
clat (usec): min=43 , max=2762 , avg=192.77, stdev=41.90
lat (usec): min=43 , max=2762 , avg=192.84, stdev=41.91
clat percentiles (usec):
| 1.00th=[ 173], 5.00th=[ 185], 10.00th=[ 185], 20.00th=[ 187],
| 30.00th=[ 187], 40.00th=[ 189], 50.00th=[ 189], 60.00th=[ 189],
| 70.00th=[ 191], 80.00th=[ 191], 90.00th=[ 201], 95.00th=[ 215],
| 99.00th=[ 270], 99.50th=[ 306], 99.90th=[ 668], 99.95th=[ 772],
| 99.99th=[ 2128]
bw (KB/s) : min=20112, max=20432, per=6.22%, avg=20315.00, stdev=86.35
lat (usec) : 50=0.02%, 100=0.09%, 250=98.61%, 500=1.04%, 750=0.20%
lat (usec) : 1000=0.02%
lat (msec) : 2=0.01%, 4=0.02%
cpu : usr=1.41%, sys=0.95%, ctx=61515, majf=0, minf=369
IO depths : 1=100.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=0.0%
submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
issued : total=r=61440/w=0/d=0, short=r=0/w=0/d=0
Run status group 0 (all jobs):
READ: io=245760KB, aggrb=326808KB/s, minb=326808KB/s, maxb=326808KB/s,
mint=752msec, maxt=752msec
Postmark
Installation :
sudo apt-get install postmark
mukuls@MUKUL:~/thesis/smr_main/src/test/local$ postmark
PostMark v1.51 : 8/14/01
pm>set location=/tmp/smr/sfs
pm>run
Creating files...Done
Performing transactions..........Done
Deleting files...Done
Time:
1 seconds total
1 seconds of transactions (500 per second)
Files:
764 created (764 per second)
Creation alone: 500 files (500 per second)
Mixed with transactions: 264 files (264 per second)
243 read (243 per second)
257 appended (257 per second)
764 deleted (764 per second)
Deletion alone: 528 files (528 per second)
Mixed with transactions: 236 files (236 per second)
Data:
1.36 megabytes read (1.36 megabytes per second)
4.45 megabytes written (4.45 megabytes per second)
pm>
Filebench (webserver, webproxy, fileserver, createfiles, videoserver, varmail)
webproxy
filebench> load webproxy [2/222]
22815: 17.312: Web proxy-server Version 3.0 personality successfully loaded
22815: 17.313: Usage: set $dir=<dir>
22815: 17.313: set $meanfilesize=<size> defaults to 16384
22815: 17.313: set $nfiles=<value> defaults to 10000
22815: 17.313: set $nthreads=<value> defaults to 100
22815: 17.313: set $meaniosize=<value> defaults to 16384
22815: 17.313: set $iosize=<size> defaults to 1048576
22815: 17.313: set $meandirwidth=<size> defaults to 1000000
22815: 17.313: run runtime (e.g. run 60)
filebench> set $dir=/tmp/smr/sfs
filebench> set $meandirwidth=10000
filebench> set $iosize=16k
filebench> run 10
22815: 260.246: Creating/pre-allocating files and filesets
22815: 260.266: Fileset bigfileset: 10000 files, 0 leafdirs, avg dir width = 10000, avg dir depth = 1.0, 154.045MB
22815: 260.269: Removed any existing fileset bigfileset in 1 seconds
22815: 260.269: making tree for filset /tmp/smr/sfs/bigfileset
22815: 260.270: Creating fileset bigfileset...
22815: 264.734: Preallocated 7979 of 10000 of fileset bigfileset in 5 seconds
22815: 264.734: waiting for fileset pre-allocation to finish
22883: 264.734: Starting 1 proxycache instances
22884: 264.738: Starting 100 proxycache threads
22815: 265.741: Running...
22815: 275.745: Run took 10 seconds...
22815: 275.797: Per-Operation Breakdown
limit 0ops 0ops/s 0.0mb/s 0.0ms/op 0us/op-cpu [0ms - 0ms]
closefile6 0ops 0ops/s 0.0mb/s 0.0ms/op 0us/op-cpu [0ms - 0ms]
readfile6 0ops 0ops/s 0.0mb/s 0.0ms/op 0us/op-cpu [0ms - 0ms]
openfile6 0ops 0ops/s 0.0mb/s 0.0ms/op 0us/op-cpu [0ms - 0ms]
closefile5 0ops 0ops/s 0.0mb/s 0.0ms/op 0us/op-cpu [0ms - 0ms]
readfile5 0ops 0ops/s 0.0mb/s 0.0ms/op 0us/op-cpu [0ms - 0ms]
openfile5 0ops 0ops/s 0.0mb/s 0.0ms/op 0us/op-cpu [0ms - 0ms]
closefile4 0ops 0ops/s 0.0mb/s 0.0ms/op 0us/op-cpu [0ms - 0ms]
readfile4 0ops 0ops/s 0.0mb/s 0.0ms/op 0us/op-cpu [0ms - 0ms]
openfile4 1ops 0ops/s 0.0mb/s 1.9ms/op 0us/op-cpu [0ms - 1ms]
closefile3 1ops 0ops/s 0.0mb/s 1.1ms/op 0us/op-cpu [0ms - 1ms]
readfile3 2ops 0ops/s 0.0mb/s 2.8ms/op 0us/op-cpu [0ms - 3ms]
openfile3 2ops 0ops/s 0.0mb/s 1.6ms/op 0us/op-cpu [0ms - 2ms]
closefile2 4ops 0ops/s 0.0mb/s 1.0ms/op 0us/op-cpu [0ms - 1ms]
readfile2 4ops 0ops/s 0.0mb/s 1.2ms/op 2500us/op-cpu [0ms - 1ms]
openfile2 4ops 0ops/s 0.0mb/s 1.6ms/op 0us/op-cpu [0ms - 2ms]
closefile1 4ops 0ops/s 0.0mb/s 0.5ms/op 2500us/op-cpu [0ms - 0ms]
appendfilerand1 5ops 0ops/s 0.0mb/s 0.9ms/op 2000us/op-cpu [0ms - 1ms]
createfile1 6ops 1ops/s 0.0mb/s 3.3ms/op 3333us/op-cpu [0ms - 5ms]
deletefile1 6ops 1ops/s 0.0mb/s 1.6ms/op 3333us/op-cpu [0ms - 3ms]
22815: 275.797: IO Summary: 39 ops, 3.900 ops/s, (1/0 r/w), 0.0mb/s, 76364us cpu/op, 5.7ms latency
22815: 275.797: Shutting down processes
videoserver
filebench> load videoserver
22419: 17.445: Eventgen rate taken from variable
22419: 17.445: Video Server Version 3.0 personality successfully loaded
22419: 17.445: Usage: set $dir=<dir> defaults to /tmp
22419: 17.445: set $eventrate=<value> defaults to 96
22419: 17.445: set $filesize=<size> defaults to 10737418240
22419: 17.445: set $nthreads=<value> defaults to 48
22419: 17.445: set $writeiosize=<value> defaults to 1048576
22419: 17.445: set $readiosize=<value> defaults to 262144
22419: 17.445: set $numactivevids=<value> defaults to 32
22419: 17.445: set $numpassivevids=<value> defaults to 194
22419: 17.445: run runtime (e.g. run 60)
filebench> set $dir=/tmp/smr/sfs
filebench> set $filesize=16k
filebench> run 10
22419: 92.880: Creating/pre-allocating files and filesets
22419: 92.881: Fileset passivevids: 194 files, 0 leafdirs, avg dir width = 20, avg dir depth = 1.8, 2.859MB
22419: 92.884: Removed any existing fileset passivevids in 1 seconds
22419: 92.884: making tree for filset /tmp/smr/sfs/passivevids
22419: 92.888: Creating fileset passivevids...
22419: 92.923: Preallocated 104 of 194 of fileset passivevids in 1 seconds
22419: 92.924: Fileset activevids: 32 files, 0 leafdirs, avg dir width = 4, avg dir depth = 2.5, 0.545MB
22419: 92.935: Removed any existing fileset activevids in 1 seconds
22419: 92.935: making tree for filset /tmp/smr/sfs/activevids
22419: 92.937: Creating fileset activevids...
22419: 92.940: Preallocated 32 of 32 of fileset activevids in 1 seconds
22419: 92.940: waiting for fileset pre-allocation to finish
22585: 92.945: Starting 1 vidreaders instances
22585: 92.945: Starting 1 vidwriter instances
22587: 92.949: Starting 1 vidwriter threads
22586: 92.950: Starting 48 vidreaders threads
22419: 93.951: Running...
22419: 103.952: Run took 10 seconds...
22419: 103.953: Per-Operation Breakdown
serverlimit 79110ops 7910ops/s 0.0mb/s 2.7ms/op 2063042us/op-cpu [0ms - 3999ms]
vidreader 79227ops 7922ops/s 67.1mb/s 0.0ms/op 3749us/op-cpu [0ms - 11ms]
replaceinterval 0ops 0ops/s 0.0mb/s 0.0ms/op 0us/op-cpu [0ms - 0ms]
wrtclose 1ops 0ops/s 0.0mb/s 0.2ms/op 0us/op-cpu [0ms - 0ms]
newvid 1ops 0ops/s 0.0mb/s 0.2ms/op 0us/op-cpu [0ms - 0ms]
wrtopen 1ops 0ops/s 0.0mb/s 0.7ms/op 0us/op-cpu [0ms - 0ms]
vidremover 1ops 0ops/s 0.0mb/s 0.8ms/op 0us/op-cpu [0ms - 0ms]
22419: 103.953: IO Summary: 79231 ops, 7922.517 ops/s, (7922/0 r/w), 67.1mb/s, 0us cpu/op, 0.0ms latency
22419: 103.953: Shutting down processes
createfiles
filebench> load createfiles
22149: 13.084: Createfiles Version 3.0 personality successfully loaded
22149: 13.084: Usage: set $dir=<dir> defaults to /tmp
22149: 13.084: set $meanfilesize=<size> defaults to 16384
22149: 13.084: set $iosize=<size> defaults to 1048576
22149: 13.084: set $nfiles=<value> defaults to 50000
22149: 13.084: set $nthreads=<value> defaults to 16
22149: 13.084: set $meandirwidth=<size> defaults to 100
22149: 13.084: run
filebench> set $dir=/tmp/smr/sfs
filebench> run 10
22149: 28.985: Creating/pre-allocating files and filesets
22149: 29.047: Fileset bigfileset: 50000 files, 0 leafdirs, avg dir width = 100, avg dir depth = 2.3, 778.397MB
22149: 29.068: Removed any existing fileset bigfileset in 1 seconds
22149: 29.068: Creating fileset bigfileset...
22149: 29.146: Preallocated 0 of 50000 of fileset bigfileset in 1 seconds
22149: 29.146: waiting for fileset pre-allocation to finish
22154: 29.147: Starting 1 filecreate instances
22155: 29.151: Starting 16 filecreatethread threads
22149: 30.153: Running...
22149: 68.156: Run took 38 seconds...
22149: 68.156: Per-Operation Breakdown
closefile1 49985ops 1315ops/s 0.0mb/s 0.3ms/op 81us/op-cpu [0ms - 2310ms]
writefile1 49986ops 1315ops/s 20.5mb/s 0.7ms/op 213us/op-cpu [0ms - 4403ms]
createfile1 50000ops 1316ops/s 0.0mb/s 10.9ms/op 2627us/op-cpu [0ms - 4414ms]
22149: 68.156: IO Summary: 149971 ops, 3946.296 ops/s, (0/1315 r/w), 20.5mb/s, 1043us cpu/op, 11.9ms latency
22149: 68.156: Shutting down processes
varmail
filebench> load varmail
21752: 9.711: Varmail Version 3.0 personality successfully loaded
21752: 9.711: Usage: set $dir=<dir>
21752: 9.711: set $meanfilesize=<size> defaults to 16384
21752: 9.711: set $nfiles=<value> defaults to 1000
21752: 9.711: set $nthreads=<value> defaults to 16
21752: 9.711: set $meanappendsize=<value> defaults to 16384
21752: 9.711: set $iosize=<size> defaults to 1048576
21752: 9.711: set $meandirwidth=<size> defaults to 1000000
21752: 9.711: run runtime (e.g. run 60)
filebench> set $dir=/tmp/smr/sfs
filebench> run 10
21752: 26.814: Creating/pre-allocating files and filesets
21752: 26.816: Fileset bigfileset: 1000 files, 0 leafdirs, avg dir width = 1000000, avg dir depth = 0.5, 14.959MB
21752: 26.819: Removed any existing fileset bigfileset in 1 seconds
21752: 26.819: making tree for filset /tmp/smr/sfs/bigfileset
21752: 26.820: Creating fileset bigfileset...
21752: 27.345: Preallocated 805 of 1000 of fileset bigfileset in 1 seconds
21752: 27.345: waiting for fileset pre-allocation to finish
21758: 27.345: Starting 1 filereader instances
21759: 27.346: Starting 16 filereaderthread threads
21752: 28.382: Running...
21752: 38.383: Run took 10 seconds...
21752: 38.384: Per-Operation Breakdown
closefile4 3488ops 349ops/s 0.0mb/s 0.8ms/op 444us/op-cpu [0ms - 6ms]
readfile4 3488ops 349ops/s 5.3mb/s 1.9ms/op 923us/op-cpu [0ms - 22ms]
openfile4 3488ops 349ops/s 0.0mb/s 1.5ms/op 662us/op-cpu [0ms - 10ms]
closefile3 3491ops 349ops/s 0.0mb/s 0.8ms/op 455us/op-cpu [0ms - 8ms]
fsyncfile3 3491ops 349ops/s 0.0mb/s 1.1ms/op 418us/op-cpu [0ms - 18ms]
appendfilerand3 3491ops 349ops/s 2.7mb/s 1.5ms/op 745us/op-cpu [0ms - 22ms]
readfile3 3492ops 349ops/s 5.2mb/s 1.8ms/op 899us/op-cpu [0ms - 24ms]
openfile3 3493ops 349ops/s 0.0mb/s 1.4ms/op 624us/op-cpu [0ms - 22ms]
closefile2 3494ops 349ops/s 0.0mb/s 0.8ms/op 429us/op-cpu [0ms - 18ms]
fsyncfile2 3494ops 349ops/s 0.0mb/s 1.0ms/op 415us/op-cpu [0ms - 22ms]
appendfilerand2 3494ops 349ops/s 2.7mb/s 1.5ms/op 833us/op-cpu [0ms - 22ms]
createfile2 3495ops 349ops/s 0.0mb/s 14.5ms/op 7210us/op-cpu [0ms - 44ms]
deletefile1 3500ops 350ops/s 0.0mb/s 13.4ms/op 6606us/op-cpu [0ms - 43ms]
21752: 38.385: IO Summary: 45399 ops, 4539.454 ops/s, (698/698 r/w), 16.0mb/s, 1613us cpu/op, 10.5ms latency
21752: 38.385: Shutting down processes
fileserver
filebench> load fileserver
18464: 6.044: File-server Version 3.0 personality successfully loaded
18464: 6.044: Usage: set $dir=<dir>
18464: 6.044: set $meanfilesize=<size> defaults to 131072
18464: 6.044: set $nfiles=<value> defaults to 10000
18464: 6.044: set $nthreads=<value> defaults to 50
18464: 6.044: set $meanappendsize=<value> defaults to 16384
18464: 6.044: set $iosize=<size> defaults to 1048576
18464: 6.044: set $meandirwidth=<size> defaults to 20
18464: 6.044: run runtime (e.g. run 60)
filebench> set $dir=/tmp/smr/sfs
filebench> set $nfiles=1000
filebench> set $meandirwidth=20
filebench> set $meanfilesize=16k
filebench> set $nthreads=100
filebench> set $iosize=4k
filebench> set $meanappendsize=16k
filebench> run 10
18464: 69.942: Creating/pre-allocating files and filesets
18464: 69.945: Fileset bigfileset: 1000 files, 0 leafdirs, avg dir width = 20, avg dir depth = 2.3, 14.903MB
18464: 69.948: Removed any existing fileset bigfileset in 1 seconds
18464: 69.948: making tree for filset /tmp/smr/sfs/bigfileset
18464: 69.965: Creating fileset bigfileset...
18464: 70.386: Preallocated 805 of 1000 of fileset bigfileset in 1 seconds
18464: 70.386: waiting for fileset pre-allocation to finish
18473: 70.386: Starting 1 filereader instances
18474: 70.388: Starting 100 filereaderthread threads
18464: 71.390: Running...
18464: 81.392: Run took 10 seconds...
18464: 81.404: Per-Operation Breakdown
statfile1 1524ops 152ops/s 0.0mb/s 2.5ms/op 1982us/op-cpu [0ms - 12ms]
deletefile1 1522ops 152ops/s 0.0mb/s 14.7ms/op 11150us/op-cpu [0ms - 83ms]
closefile3 1541ops 154ops/s 0.0mb/s 3.3ms/op 2790us/op-cpu [0ms - 26ms]
readfile1 1541ops 154ops/s 2.9mb/s 22.5ms/op 16496us/op-cpu [0ms - 94ms]
openfile2 1562ops 156ops/s 0.0mb/s 4.7ms/op 3419us/op-cpu [0ms - 29ms]
closefile2 1564ops 156ops/s 0.0mb/s 3.4ms/op 2449us/op-cpu [0ms - 25ms]
appendfilerand1 1567ops 157ops/s 1.2mb/s 8.6ms/op 6331us/op-cpu [0ms - 33ms]
openfile1 1577ops 158ops/s 0.0mb/s 4.7ms/op 3310us/op-cpu [0ms - 29ms]
closefile1 1581ops 158ops/s 0.0mb/s 3.4ms/op 2751us/op-cpu [0ms - 25ms]
wrtfile1 1582ops 158ops/s 2.5mb/s 28.7ms/op 21018us/op-cpu [0ms - 139ms]
createfile1 1601ops 160ops/s 0.0mb/s 30.8ms/op 23167us/op-cpu [1ms - 110ms]
18464: 81.404: IO Summary: 17162 ops, 1716.054 ops/s, (154/315 r/w), 6.8mb/s, 1147us cpu/op, 42.6ms latency
18464: 81.404: Shutting down processes
webserver
mukuls@MUKUL:~/thesis/fsbench/filebench-1.4.9.1$ filebench
Filebench Version 1.4.9.1
IMPORTANT: Virtual address space randomization is enabled on this machine!
It is highly recommended to disable randomization to provide stable Filebench runs.
Echo 0 to /proc/sys/kernel/randomize_va_space file to disable the randomization.
WARNING: Could not open /proc/sys/kernel/shmmax file!
It means that you probably ran Filebench not as a root. Filebench will not increase shared
region limits in this case, which can lead to the failures on certain workloads.
13832: 0.000: Allocated 170MB of shared memory
filebench> load webserver
13832: 11.930: Web-server Version 3.0 personality successfully loaded
13832: 11.930: Usage: set $dir=<dir>
13832: 11.930: set $meanfilesize=<size> defaults to 16384
13832: 11.930: set $nfiles=<value> defaults to 1000
13832: 11.931: set $meandirwidth=<value> defaults to 20
13832: 11.931: set $nthreads=<value> defaults to 100
13832: 11.931: set $iosize=<size> defaults to 1048576
13832: 11.931: run runtime (e.g. run 60)
filebench> set $dir=/tmp/smr/sfs
filebench> run 10
13832: 24.453: Creating/pre-allocating files and filesets
13832: 24.453: Fileset logfiles: 1 files, 0 leafdirs, avg dir width = 20, avg dir depth = 0.0, 0.002MB
13832: 24.455: Removed any existing fileset logfiles in 1 seconds
13832: 24.456: making tree for filset /tmp/smr/sfs/logfiles
13832: 24.456: Creating fileset logfiles...
13832: 24.457: Preallocated 1 of 1 of fileset logfiles in 1 seconds
13832: 24.460: Fileset bigfileset: 1000 files, 0 leafdirs, avg dir width = 20, avg dir depth = 2.3, 14.760MB
13832: 24.462: Removed any existing fileset bigfileset in 1 seconds
13832: 24.462: making tree for filset /tmp/smr/sfs/bigfileset
13832: 24.479: Creating fileset bigfileset...
13832: 24.982: Preallocated 1000 of 1000 of fileset bigfileset in 1 seconds
13832: 24.982: waiting for fileset pre-allocation to finish
13839: 24.982: Starting 1 filereader instances
13840: 24.982: Starting 100 filereaderthread threads
13832: 26.295: Running...
13832: 36.303: Run took 10 seconds...
13832: 36.446: Per-Operation Breakdown
appendlog 2107ops 211ops/s 1.5mb/s 56.8ms/op 41580us/op-cpu [0ms - 349ms]
closefile10 2009ops 201ops/s 0.0mb/s 2.6ms/op 3783us/op-cpu [0ms - 20ms]
readfile10 2009ops 201ops/s 2.9mb/s 18.6ms/op 14017us/op-cpu [0ms - 123ms]
openfile10 2014ops 201ops/s 0.0mb/s 6.6ms/op 5636us/op-cpu [0ms - 44ms]
closefile9 2016ops 202ops/s 0.0mb/s 2.6ms/op 3819us/op-cpu [0ms - 33ms]
readfile9 2017ops 202ops/s 3.0mb/s 19.3ms/op 13922us/op-cpu [0ms - 110ms]
openfile9 2021ops 202ops/s 0.0mb/s 6.5ms/op 5552us/op-cpu [0ms - 44ms]
closefile8 2021ops 202ops/s 0.0mb/s 2.6ms/op 3835us/op-cpu [0ms - 27ms]
readfile8 2021ops 202ops/s 2.9mb/s 18.9ms/op 14334us/op-cpu [0ms - 104ms]
openfile8 2023ops 202ops/s 0.0mb/s 6.4ms/op 5709us/op-cpu [0ms - 42ms]
closefile7 2024ops 202ops/s 0.0mb/s 2.7ms/op 3918us/op-cpu [0ms - 33ms]
readfile7 2024ops 202ops/s 3.0mb/s 20.1ms/op 14699us/op-cpu [0ms - 108ms]
openfile7 2025ops 202ops/s 0.0mb/s 6.3ms/op 5269us/op-cpu [0ms - 40ms]
closefile6 2025ops 202ops/s 0.0mb/s 2.5ms/op 3620us/op-cpu [0ms - 20ms]
readfile6 2029ops 203ops/s 3.0mb/s 19.8ms/op 14672us/op-cpu [0ms - 152ms]
openfile6 2049ops 205ops/s 0.0mb/s 6.3ms/op 5476us/op-cpu [0ms - 39ms]
closefile5 2053ops 205ops/s 0.0mb/s 2.6ms/op 3707us/op-cpu [0ms - 37ms]
readfile5 2056ops 206ops/s 3.0mb/s 20.2ms/op 14966us/op-cpu [0ms - 132ms]
openfile5 2069ops 207ops/s 0.0mb/s 6.4ms/op 5510us/op-cpu [0ms - 39ms]
closefile4 2073ops 207ops/s 0.0mb/s 2.6ms/op 3816us/op-cpu [0ms - 33ms]
readfile4 2077ops 208ops/s 2.9mb/s 19.2ms/op 14035us/op-cpu [0ms - 155ms]
openfile4 2081ops 208ops/s 0.0mb/s 6.3ms/op 5593us/op-cpu [0ms - 41ms]
closefile3 2082ops 208ops/s 0.0mb/s 2.7ms/op 4121us/op-cpu [0ms - 33ms]
readfile3 2084ops 208ops/s 3.0mb/s 18.3ms/op 13455us/op-cpu [0ms - 100ms]
openfile3 2092ops 209ops/s 0.0mb/s 6.4ms/op 5507us/op-cpu [0ms - 36ms]
closefile2 2093ops 209ops/s 0.0mb/s 2.5ms/op 4009us/op-cpu [0ms - 21ms]
readfile2 2093ops 209ops/s 3.1mb/s 18.4ms/op 13774us/op-cpu [0ms - 134ms]
openfile2 2095ops 209ops/s 0.0mb/s 5.9ms/op 4964us/op-cpu [0ms - 39ms]
closefile1 2097ops 210ops/s 0.0mb/s 2.7ms/op 3534us/op-cpu [0ms - 33ms]
readfile1 2102ops 210ops/s 3.0mb/s 17.0ms/op 12721us/op-cpu [0ms - 103ms]
openfile1 2103ops 210ops/s 0.0mb/s 5.1ms/op 6372us/op-cpu [0ms - 39ms]
13832: 36.449: IO Summary: 63684 ops, 6365.331 ops/s, (2050/211 r/w), 31.8mb/s, 1364us cpu/op, 30.5ms latency
13832: 36.449: Shutting down processes
IOR
mukuls@MUKUL:~/thesis/fsbench/IOR/src/C$ ./IOR -b 4k -t 4k -o /tmp/smr/sfs/aa
IOR-2.10.3: MPI Coordinated Test of Parallel I/O
Run began: Fri Oct 25 00:21:54 2013
Command line used: ./IOR -b 4k -t 4k -o /tmp/smr/sfs/aa
Machine: Linux MUKUL
Summary:
api = POSIX
test filename = /tmp/smr/sfs/aa
access = single-shared-file
ordering in a file = sequential offsets
ordering inter file= no tasks offsets
clients = 1 (1 per node)
repetitions = 1
xfersize = 4096 bytes
blocksize = 4096 bytes
aggregate filesize = 4096 bytes
Operation Max (
MiB) Min (
MiB) Mean (
MiB) Std Dev Max (OPs) Min (OPs) Mean (OPs) Std Dev Mean (s)
--------- --------- --------- ---------- ------- --------- --------- ---------- ------- --------
write 3.04 3.04 3.04 0.00 777.73 777.73 777.73 0.00 0.00129 EXCEL
read 8.61 8.61 8.61 0.00 2204.05 2204.05 2204.05 0.00 0.00045 EXCEL
Max Write: 3.04
MiB /sec (3.19 MB/sec)
Max Read: 8.61
MiB /sec (9.03 MB/sec)
Run finished: Fri Oct 25 00:21:54 2013
Bonnie++
mukuls@MUKUL:~/thesis/fsbench/IOR/src/C$ bonnie++ -d /tmp/smr/sfs/ -r 1 -s 2 -f -b -n 12
Writing intelligently...done
Rewriting...done
Reading intelligently...done
start 'em...done...done...done...done...done...
Create files in sequential order...done.
Stat files in sequential order...done.
Delete files in sequential order...done.
Create files in random order...done.
Stat files in random order...done.
Delete files in random order...done.
Version 1.97 ------Sequential Output------ --Sequential Input- --Random-
Concurrency 1 -Per Chr- --Block-- -Rewrite- -Per Chr- --Block-- --Seeks--
Machine Size K/sec %CP K/sec %CP K/sec %CP K/sec %CP K/sec %CP /sec %CP
MUKUL 2M +++++ +++ +++++ +++ +++++ +++ 1467 8
Latency 1399us 745us 131us 2812ms
Version 1.97 ------Sequential Create------ --------Random Create--------
MUKUL -Create-- --Read--- -Delete-- -Create-- --Read--- -Delete--
files /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP /sec %CP
12 675 1 +++++ +++ 5652 5 1829 5 23211 19 5609 5
Latency 1776ms 2759us 12678us 347ms 1534us 60181us
1.97,1.97,MUKUL,1,1382668839,2M,,,,+++++,+++,+++++,+++,,,+++++,+++,1467,8,12,,,,,675,1,+++++,+++,5652,5,1829,5,23211,19,5609,5,,1399us,745us,,131us,2812ms,1776ms,2759us,12678us,347ms,1534us,60181us
mukuls@MUKUL:~/thesis/fsbench/IOR/src/C$
tiobench
mukuls@MUKUL:~/thesis/fsbench/tiobench-0.3.3$ ./tiobench.pl --dir /tmp/smr/sfs --numruns 2 --block 4096 --size 12
Run #2: ./tiotest -t 8 -f 1 -r 500 -b 4096 -d /tmp/smr/sfs -TTT
Unit information
================
File size = megabytes
Blk Size = bytes
Rate = megabytes per second
CPU% = percentage of CPU used during the test
Latency = milliseconds
Lat% = percent of requests that took longer than X seconds
CPU Eff = Rate divided by CPU% - throughput per cpu load
Sequential Reads
3.11.0-12-generic 12 4096 1 685.32 77.58% 0.010 0.34 0.00000 0.00000 883
3.11.0-12-generic 12 4096 2 ###### 153.6% 0.009 0.48 0.00000 0.00000 1004
3.11.0-12-generic 12 4096 4 ###### 269.8% 0.012 0.74 0.00000 0.00000 644
3.11.0-12-generic 12 4096 8 ###### 246.8% 0.028 5.37 0.00000 0.00000 418
Random Reads
3.11.0-12-generic 12 4096 1 166.32 30.56% 0.045 1.00 0.00000 0.00000 544
3.11.0-12-generic 12 4096 2 388.68 74.04% 0.038 0.92 0.00000 0.00000 525
3.11.0-12-generic 12 4096 4 549.69 151.2% 0.050 3.41 0.00000 0.00000 363
3.11.0-12-generic 12 4096 8 716.25 313.9% 0.072 11.43 0.00000 0.00000 228
Sequential Writes
3.11.0-12-generic 12 4096 1 43.38 18.74% 0.166 0.94 0.00000 0.00000 231
3.11.0-12-generic 12 4096 2 28.91 12.46% 0.318 589.96 0.00000 0.00000 232
3.11.0-12-generic 12 4096 4 94.66 87.68% 0.289 14.01 0.00000 0.00000 108
3.11.0-12-generic 12 4096 8 79.36 173.9% 0.678 9.66 0.00000 0.00000 46
Random Writes
3.11.0-12-generic 12 4096 1 35.04 20.10% 0.212 2.13 0.00000 0.00000 174
3.11.0-12-generic 12 4096 2 116.00 39.73% 0.120 16.85 0.00000 0.00000 292
3.11.0-12-generic 12 4096 4 100.20 88.68% 0.271 13.75 0.00000 0.00000 113
3.11.0-12-generic 12 4096 8 79.94 188.6% 0.697 15.70 0.00000 0.00000 42
FSMark
mukuls@MUKUL:/tmp/smr/sfs$ ~/thesis/fsbench/fs_mark-3.3/fs_mark -D 10 -d . -N 10 -L 1 -s 4096 -t 10
# /home/mukuls/thesis/fsbench/fs_mark-3.3/fs_mark -D 10 -d . -N 10 -L 1 -s 4096 -t 10
# Version 3.3, 10 thread(s) starting at Thu Oct 24 18:39:42 2013
# Sync method: INBAND FSYNC: fsync() per file in write loop.
# Directories: Round Robin between directories across 10 subdirectories with 10 files per subdirectory.
# File names: 40 bytes long, (16 initial bytes of time stamp with 24 random bytes at end of name)
# Files info: size 4096 bytes, written with an IO size of 16384 bytes per write
# App overhead is time in microseconds spent in the test not doing file writing related system calls.
FSUse% Count Size Files/sec App Overhead
32 10000 4096 844.6 393174
Flexible File System Benchmark
This multi-threaded tool can be used to generate various workloads using customizable profiles. This can be used to generate small and large file workloads.
http://sourceforge.net/projects/ffsb/
Usage:
ffsb examples/myprofile
Where profile can be defined as
directio = 0
time = 10
[filesystem]
location = /home/mukuls/thesis/fsbench/test
num_dirs = 100
size_weight 4k 33
size_weight 8k 21
size_weight 16k 13
size_weight 32k 10
size_weight 64k 8
size_weight 128k 5
size_weight 256k 4
size_weight 512k 3
size_weight 8m 2
size_weight 32m 1
# size_weight 1g 1
# min_filesize = 4k
# max_filesize = 10m
# num_files = 0
init_size = 100m
# init_size = 6GB
# init_size = 1gb
# init_util = 0.002
agefs = 0
[threadgroup]
num_threads = 10
write_size = 400
write_blocksize = 1024
create_weight = 10
append_weight = 10
delete_weight = 1
[end]
desired_util = 0.005
[end]
#[filesystem]
# location = /mnt/test1
# clone = /mnt/test2
#[end]
[threadgroup]
num_threads = 4
FFSB version 6.0-RC2 started
benchmark time = 10
ThreadGroup 0
================
num_threads = 2
read_random = off
read_size = 40960 (40KB)
read_blocksize = 4096 (4KB)
read_skip = off
read_skipsize = 0 (0B)
write_random = off
write_size = 40960 (40KB)
fsync_file = 0
write_blocksize = 4096 (4KB)
wait time = 1
op weights
read = 0 (0.00%)
readall = 1 (10.00%)
write = 0 (0.00%)
create = 1 (10.00%)
append = 1 (10.00%)
delete = 1 (10.00%)
metaop = 0 (0.00%)
createdir = 0 (0.00%)
stat = 1 (10.00%)
writeall = 1 (10.00%)
writeall_fsync = 1 (10.00%)
open_close = 1 (10.00%)
write_fsync = 0 (0.00%)
create_fsync = 1 (10.00%)
append_fsync = 1 (10.00%)
FileSystem /tmp/smr/sfs
==========
num_dirs = 5
starting files = 0
Fileset weight:
4096 ( 4KB) -> 33 (100.00%)
directio = off
alignedio = off
bufferedio = off
aging is on
current utilization = 25.07%
desired utilization = 0.50%
Aging
ThreadGroup for fs /tmp/smr/sfs
================
num_threads = 2
read_random = off
read_size = 0 (0B)
read_blocksize = 0 (0B)
read_skip = off
read_skipsize = 0 (0B)
write_random = off
write_size = 400 (400B)
fsync_file = 0
write_blocksize = 1024 (1KB)
wait time = 0
op weights
read = 0 (0.00%)
readall = 0 (0.00%)
write = 0 (0.00%)
create = 10 (47.62%)
append = 10 (47.62%)
delete = 1 (4.76%)
metaop = 0 (0.00%)
createdir = 0 (0.00%)
stat = 0 (0.00%)
writeall = 0 (0.00%)
writeall_fsync = 0 (0.00%)
open_close = 0 (0.00%)
write_fsync = 0 (0.00%)
create_fsync = 0 (0.00%)
append_fsync = 0 (0.00%)
creating new fileset /tmp/smr/sfs
aging fs /tmp/smr/sfs from 0.25 to 0.01
fs setup took 8 secs
Syncing()...2 sec
Starting Actual Benchmark At: Thu Oct 24 11:14:51 2013
Syncing()...29 sec
FFSB benchmark finished at: Thu Oct 24 11:15:31 2013
Results:
Benchmark took 40.33 sec
Total Results
===============
Op Name Transactions Trans/sec % Trans % Op Weight Throughput
=== ======== ===== === ======= ==========
readall : 2038 50.54 5.876% 10.030% 202KB/sec
create : 1187 29.44 3.423% 10.370% 118KB/sec
append : 11280 279.73 32.525% 9.855% 1.09MB/sec
delete : 1185 29.39 3.417% 10.353% NA
stat : 1143 28.34 3.296% 9.986% NA
writeall : 2006 49.75 5.784% 9.663% 199KB/sec
writeall_fsync : 2190 54.31 6.315% 9.872% 217KB/sec
open_close : 1143 28.34 3.296% 9.986% NA
create_fsync : 1139 28.25 3.284% 9.951% 113KB/sec
append_fsync : 11370 281.96 32.785% 9.934% 1.1MB/sec
-
860.03 Transactions per Second
Throughput Results
===================
Read Throughput: 202KB/sec
Write Throughput: 2.83MB/sec
System Call Latency statistics in millisecs
=====
Min Avg Max Total Calls
==== ==== ==== ============
[ open] 0.059000 0.493468 374.946991 9118
-
[ read] 0.000000 0.032523 0.438000 2038
-
[ write] 0.018000 0.088117 329.177002 29172
-
[ unlink] 0.087000 0.238140 0.923000 1185
-
[ close] 0.008000 0.039420 0.513000 9118
-
[ stat] 0.004000 0.062983 0.401000 1143
-
0.5% User Time
3.2% System Time
3.7% CPU Utilization
FFSB version 6.0-RC2 started
benchmark time = 10
ThreadGroup 0
================
num_threads = 2
read_random = off
read_size = 40960 (40KB)
read_blocksize = 4096 (4KB)
read_skip = off
read_skipsize = 0 (0B)
write_random = off
write_size = 40960 (40KB)
fsync_file = 0
write_blocksize = 4096 (4KB)
wait time = 1
op weights
read = 0 (0.00%)
readall = 1 (10.00%)
write = 0 (0.00%)
create = 1 (10.00%)
append = 1 (10.00%)
delete = 1 (10.00%)
metaop = 0 (0.00%)
createdir = 0 (0.00%)
stat = 1 (10.00%)
writeall = 1 (10.00%)
writeall_fsync = 1 (10.00%)
open_close = 1 (10.00%)
write_fsync = 0 (0.00%)
create_fsync = 1 (10.00%)
append_fsync = 1 (10.00%)
FileSystem /tmp/smr/sfs
==========
num_dirs = 5
starting files = 0
Fileset weight:
4096 ( 4KB) -> 33 (50.00%)
2048 ( 2KB) -> 33 (50.00%)
directio = off
alignedio = off
bufferedio = off
aging is on
current utilization = 25.07%
desired utilization = 0.50%
Aging
ThreadGroup for fs /tmp/smr/sfs
================
num_threads = 2
read_random = off
read_size = 0 (0B)
read_blocksize = 0 (0B)
read_skip = off
read_skipsize = 0 (0B)
write_random = off
write_size = 4096 (4KB)
fsync_file = 0
write_blocksize = 1024 (1KB)
wait time = 0
op weights
read = 0 (0.00%)
readall = 0 (0.00%)
write = 0 (0.00%)
create = 10 (43.48%)
append = 10 (43.48%)
delete = 3 (13.04%)
metaop = 0 (0.00%)
createdir = 0 (0.00%)
stat = 0 (0.00%)
writeall = 0 (0.00%)
writeall_fsync = 0 (0.00%)
open_close = 0 (0.00%)
write_fsync = 0 (0.00%)
create_fsync = 0 (0.00%)
append_fsync = 0 (0.00%)
creating new fileset /tmp/smr/sfs
aging fs /tmp/smr/sfs from 0.25 to 0.01
fs setup took 8 secs
Syncing()...3 sec
Starting Actual Benchmark At: Thu Oct 24 11:20:42 2013
Syncing()...4 sec
FFSB benchmark finished at: Thu Oct 24 11:20:57 2013
Results:
Benchmark took 14.64 sec
Total Results
===============
Op Name Transactions Trans/sec % Trans % Op Weight Throughput
=== ======== ===== === ======= ==========
readall : 1505 102.82 4.566% 10.210% 484KB/sec
create : 1107 75.63 3.358% 9.888% 230KB/sec
append : 11370 776.75 34.492% 10.156% 3.03MB/sec
delete : 1180 80.61 3.580% 10.540% NA
stat : 1205 82.32 3.656% 10.764% NA
writeall : 1926 131.58 5.843% 10.147% 454KB/sec
writeall_fsync : 1969 134.51 5.973% 9.728% 465KB/sec
open_close : 1082 73.92 3.282% 9.665% NA
create_fsync : 1060 72.41 3.216% 9.469% 218KB/sec
append_fsync : 10560 721.41 32.035% 9.433% 2.82MB/sec
-
2251.96 Transactions per Second
Throughput Results
===================
Read Throughput: 484KB/sec
Write Throughput: 7.19MB/sec
System Call Latency statistics in millisecs
=====
Min Avg Max Total Calls
==== ==== ==== ============
[ open] 0.059000 0.479224 341.735992 8810
-
[ read] 0.000000 0.198503 340.315002 2041
-
[ write] 0.017000 0.077186 2.086000 27992
-
[ unlink] 0.145000 0.246849 6.195000 1180
-
[ close] 0.009000 0.039045 1.807000 8810
-
[ stat] 0.005000 0.067283 0.403000 1205
-
1.3% User Time
8.8% System Time
10.1% CPU Utilization
time = 10
directio = 0
[filesystem0]
location = /tmp/smr/sfs
num_dirs = 5
size_weight 4k 33
# size_weight 8k 21
# size_weight 16k 13
# size_weight 32k 10
# size_weight 64k 8
# size_weight 128k 5
# size_weight 256k 4
# size_weight 512k 3
# size_weight 8m 2
# size_weight 32m 1
# size_weight 1g 1
min_filesize = 1k
max_filesize = 4k
# num_files = 0
init_size = 100m
# init_size = 6GB
# init_size = 1gb
# init_util = 0.002
agefs = 1
[threadgroup]
num_threads = 2
write_size = 400
write_blocksize = 1024
create_weight = 10
append_weight = 10
delete_weight = 1
[end]
desired_util = 0.005
[end0]
#[filesystem]
# location = /mnt/test1
# clone = /mnt/test2
#[end]
[threadgroup0]
num_threads = 2
# bindfs = /mnt/test1
append_weight = 1
append_fsync_weight = 1
stat_weight = 1
# write_weight = 1
# write_fsync_weight = 1
# read_weight = 1
create_weight = 1
create_fsync_weight = 1
delete_weight = 1
readall_weight = 1
writeall_weight = 1
writeall_fsync_weight = 1
open_close_weight = 1
read_random = 0
write_random = 0
write_size = 40k
write_blocksize = 4k
read_size = 40k
read_blocksize = 4k
op_delay = 1
[stats]
enable_stats = 1
enable_range = 0
# ignore = close
# ignore = open
# ignore = lseek
# ignore = write
# ignore = read
msec_range 0.00 0.01
msec_range 0.01 0.02
msec_range 0.02 0.03
msec_range 0.03 0.04
msec_range 0.04 0.05
msec_range 0.05 0.1
msec_range 0.1 0.2
msec_range 0.2 0.5
msec_range 0.5 1.0
msec_range 1.0 2.0
msec_range 2.0 3.0
msec_range 3.0 4.0
msec_range 4.0 5.0
msec_range 5.0 10.0
msec_range 10.0 10000.0
[end]
[end0]
time = 10
directio = 0
[filesystem0]
location = /tmp/smr/sfs
num_dirs = 5
size_weight 2k 33
size_weight 4k 33
min_filesize = 1k
max_filesize = 4k
# num_files = 0
init_size = 100m
# init_size = 6GB
# init_size = 1gb
# init_util = 0.002
agefs = 1
[threadgroup]
num_threads = 2
write_size = 4096
write_blocksize = 1024
create_weight = 10
append_weight = 10
delete_weight = 3
[end]
desired_util = 0.005
[end0]
[threadgroup0]
num_threads = 2
append_weight = 1
append_fsync_weight = 1
stat_weight = 1
# write_weight = 1
# write_fsync_weight = 1
# read_weight = 1
create_weight = 1
create_fsync_weight = 1
delete_weight = 1
readall_weight = 1
writeall_weight = 1
writeall_fsync_weight = 1
open_close_weight = 1
read_random = 0
write_random = 0
write_size = 40k
write_blocksize = 4k
read_size = 40k
read_blocksize = 4k
op_delay = 1
[stats]
enable_stats = 1
enable_range = 0
# ignore = close
# ignore = open
# ignore = lseek
# ignore = write
# ignore = read
msec_range 0.00 0.01
msec_range 0.01 0.02
msec_range 0.02 0.03
msec_range 0.03 0.04
msec_range 0.04 0.05
msec_range 0.05 0.1
msec_range 0.1 0.2
msec_range 0.2 0.5
msec_range 0.5 1.0
msec_range 1.0 2.0
msec_range 2.0 3.0
msec_range 3.0 4.0
msec_range 4.0 5.0
msec_range 5.0 10.0
msec_range 10.0 10000.0
[end]
[end0]
vdbench
It is a multi-threaded benchmark tool. It helps in creating both raw and file system IOs and also provides methods to replay the traffic using a plug-in.
I was able to find another link for vdbench tool on sourcefourge, this is an open source project.
http://vdbench.sourceforge.net/
Usage :
./vdbench -f /tmp/parmfile
where parmfile can be configured as
fsd=fsd1,anchor=/home/mukuls/thesis/fsbench/test,depth=2,width=10,files=100,size=128k
fwd=default,xfersize=4k,fileio=random,fileselect=random,threads=8,
stopafter=100
fwd=fwd1,fsd=fsd1,operation=read
fwd=fwd2,fsd=fsd1,operation=write
rd=rd1,fwd=fwd*,fwdrate=100,format=yes,elapsed=5,interval=1
mukuls@MUKUL:~/thesis/fsbench/vdbench$ ./vdbench -f ../parmfile
Vdbench distribution: vdbench502
For documentation, see 'vdbench.pdf'.
15:21:11.438 input argument scanned: '-f../parmfile'
15:21:11.616 anchor=/tmp/smr/sfs: there will be 110 directories and a maximum of 10000 files under this anchor.
15:21:11.617 Estimated maximum size for this anchor: 39.062m
15:21:11.617
15:21:11.780 Starting slave: /home/mukuls/thesis/fsbench/vdbench/vdbench
SlaveJvm -m localhost -n localhost-10-131024-15.21.11.374 -l localhost-0 -p 5570
15:21:12.418 All slaves are now connected
15:21:15.002 Starting RD=format_for_rd1
15:21:15.099 localhost-0: anchor=/tmp/smr/sfs mkdir complete.
Oct 24, 2013 .Interval. .ReqstdOps.. ...cpu%... ....read.... ...write.... ..mb/sec... mb/sec .xfer. ...mkdir... ...rmdir... ..create... ...open.... ...close... ..delete... ..getattr.. ..se$
attr..
rate resp total sys rate resp rate resp read write total size rate resp rate resp rate resp rate resp rate resp rate resp rate resp rat$
resp
15:21:16.280 1 0.0 0.00 51.7 2.95 0.0 0.00 0.0 0.00 0.00 0.00 0.00 0 110.0 1.58 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.$
0.00
15:21:17.189 2 533.0 1.27 78.8 24.4 0.0 0.00 533.0 1.27 0.00 2.09 2.09 4103 0.0 0.00 0.0 0.00 533.0 13.93 535.0 10.19 533.0 0.86 0.0 0.00 0.0 0.00 0.$
0.00
15:21:18.239 3 576.0 1.16 74.9 29.1 0.0 0.00 576.0 1.16 0.00 2.25 2.25 4096 0.0 0.00 0.0 0.00 575.0 13.75 574.0 10.78 576.0 0.85 0.0 0.00 0.0 0.00 0.$
0.00
15:21:19.153 4 540.0 1.29 76.4 29.2 0.0 0.00 540.0 1.29 0.00 2.11 2.11 4096 0.0 0.00 0.0 0.00 542.0 14.82 540.0 11.48 541.0 0.92 0.0 0.00 0.0 0.00 0.$
0.00
15:21:20.170 5 566.0 1.24 69.2 31.5 0.0 0.00 566.0 1.24 0.00 2.21 2.21 4096 0.0 0.00 0.0 0.00 565.0 13.97 566.0 10.83 565.0 0.88 0.0 0.00 0.0 0.00 0.$
0.00
15:21:21.162 6 446.0 1.67 87.2 25.5 0.0 0.00 446.0 1.67 0.00 1.74 1.74 4096 0.0 0.00 0.0 0.00 445.0 17.88 446.0 13.72 446.0 0.97 0.0 0.00 0.0 0.00 0.$
0.00
15:21:22.257 7 565.0 1.24 78.4 29.7 0.0 0.00 565.0 1.24 0.00 2.21 2.21 4096 0.0 0.00 0.0 0.00 566.0 14.06 566.0 10.82 566.0 0.93 0.0 0.00 0.0 0.00 0.$
0.00
15:21:23.141 8 372.0 2.32 91.6 22.5 0.0 0.00 372.0 2.32 0.00 1.45 1.45 4096 0.0 0.00 0.0 0.00 371.0 21.32 373.0 15.87 372.0 1.07 0.0 0.00 0.0 0.00 0.$
0.00
15:21:24.099 9 585.0 1.24 75.3 30.0 0.0 0.00 585.0 1.24 0.00 2.29 2.29 4096 0.0 0.00 0.0 0.00 585.0 13.65 584.0 10.42 585.0 0.95 0.0 0.00 0.0 0.00 0.$
0.00
15:21:25.088 10 637.0 1.05 70.7 32.1 0.0 0.00 637.0 1.05 0.00 2.49 2.49 4096 0.0 0.00 0.0 0.00 638.0 12.42 637.0 9.66 638.0 0.81 0.0 0.00 0.0 0.00 0.$
0.00
15:21:26.087 11 519.0 1.40 81.1 27.4 0.0 0.00 519.0 1.40 0.00 2.03 2.03 4096 0.0 0.00 0.0 0.00 519.0 15.36 519.0 11.95 518.0 0.93 0.0 0.00 0.0 0.00 0.$
0.00
15:21:27.131 12 555.0 1.04 61.2 29.2 0.0 0.00 555.0 1.04 0.00 2.17 2.17 4096 0.0 0.00 0.0 0.00 555.0 14.32 555.0 11.21 555.0 1.02 0.0 0.00 0.0 0.00 0.$
0.00
15:21:28.107 13 627.0 1.08 71.4 31.4 0.0 0.00 627.0 1.08 0.00 2.45 2.45 4096 0.0 0.00 0.0 0.00 628.0 12.63 628.0 9.88 628.0 0.82 0.0 0.00 0.0 0.00 0.$
0.00
15:21:29.084 14 616.0 1.12 68.8 29.6 0.0 0.00 616.0 1.12 0.00 2.41 2.41 4096 0.0 0.00 0.0 0.00 615.0 12.93 616.0 10.09 616.0 0.86 0.0 0.00 0.0 0.00 0.$
0.00
15:21:30.094 15 644.0 1.00 69.6 32.6 0.0 0.00 644.0 1.00 0.00 2.52 2.52 4096 0.0 0.00 0.0 0.00 645.0 12.37 644.0 9.77 644.0 0.80 0.0 0.00 0.0 0.[3/1989]
0.00
15:21:31.090 16 612.0 1.10 70.2 32.0 0.0 0.00 612.0 1.10 0.00 2.39 2.39 4096 0.0 0.00 0.0 0.00 612.0 12.93 613.0 10.19 612.0 0.79 0.0 0.00 0.0 0.00 0.0
0.00
15:21:32.086 17 545.0 1.34 61.9 29.5 0.0 0.00 545.0 1.34 0.00 2.13 2.13 4096 0.0 0.00 0.0 0.00 545.0 14.57 544.0 11.08 545.0 0.84 0.0 0.00 0.0 0.00 0.0
0.00
15:21:33.142 18 616.0 1.09 71.0 32.4 0.0 0.00 616.0 1.09 0.00 2.41 2.41 4096 0.0 0.00 0.0 0.00 614.0 12.92 616.0 10.14 616.0 0.85 0.0 0.00 0.0 0.00 0.0
0.00
15:21:33.940 localhost-0: anchor=/tmp/smr/sfs create complete.
15:21:34.091 19 444.0 1.44 68.8 25.3 0.0 0.00 444.0 1.44 0.00 1.73 1.73 4096 0.0 0.00 0.0 0.00 446.0 15.75 443.0 12.33 444.0 0.83 0.0 0.00 0.0 0.00 0.0
0.00
15:21:35.066 20 0.0 0.00 29.6 0.75 0.0 0.00 0.0 0.00 0.00 0.00 0.00 0 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0
0.00
15:21:35.089 avg_2-20 526.2 1.25 71.4 27.6 0.0 0.00 526.2 1.25 0.00 2.06 2.06 4096 0.0 0.00 0.0 0.00 526.3 14.16 526.3 10.95 526.3 0.88 0.0 0.00 0.0 0.00 0.0
0.00
15:21:35.347
15:21:35.348 Miscellaneous statistics:
15:21:35.348 (These statistics do not include activity between the last reported interval and shutdown.)
15:21:35.348 FILE_CREATES Files created: 10000 500/sec
15:21:35.349 DIRECTORY_CREATES Directories created: 110 5/sec
15:21:35.349 WRITE_OPENS Files opened for write activity: 10000 500/sec
15:21:35.350 DIR_BUSY_MKDIR Directory busy (mkdir): 4 0/sec
15:21:35.350 DIR_EXISTS Directory may not exist (yet): 57 2/sec
15:21:35.351 FILE_CLOSES Close requests: 10000 500/sec
15:21:35.351
15:21:38.001 Starting RD=rd1; elapsed=5; fwdrate=100; For loops: None
Oct 24, 2013 .Interval. .ReqstdOps.. ...cpu%... ....read.... ...write.... ..mb/sec... mb/sec .xfer. ...mkdir... ...rmdir... ..create... ...open.... ...close... ..delete... ..getattr.. ..set
attr..
rate resp total sys rate resp rate resp read write total size rate resp rate resp rate resp rate resp rate resp rate resp rate resp rate
resp
15:21:39.036 1 95.0 0.28 27.0 1.40 47.0 0.05 48.0 0.51 0.18 0.19 0.37 4096 0.0 0.00 0.0 0.00 0.0 0.00 8.0 2.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0
0.00
15:21:40.073 2 95.0 0.29 6.3 1.95 48.0 0.02 47.0 0.56 0.19 0.18 0.37 4096 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0
0.00
15:21:41.073 3 97.0 0.31 6.3 2.27 47.0 0.02 50.0 0.59 0.18 0.20 0.38 4096 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0
0.00
15:21:42.081 4 101.0 0.29 6.3 1.77 51.0 0.02 50.0 0.57 0.20 0.20 0.39 4096 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0
0.00
15:21:43.069 5 105.0 0.26 8.1 2.28 57.0 0.02 48.0 0.54 0.22 0.19 0.41 4096 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0
0.00
15:21:43.088 avg_2-5 99.5 0.29 6.8 2.07 50.8 0.02 48.8 0.57 0.20 0.19 0.39 4096 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0 0.00 0.0
0.00
15:21:43.509
15:21:43.509 Miscellaneous statistics:
15:21:43.510 (These statistics do not include activity between the last reported interval and shutdown.)
15:21:43.510 READ_OPENS Files opened for read activity: 4 0/sec
15:21:43.510 WRITE_OPENS Files opened for write activity: 4 0/sec
15:21:43.511
15:21:43.744 Vdbench execution completed successfully. Output directory: /home/mukuls/thesis/fsbench/vdbench/output
mukuls@MUKUL:~/thesis/fsbench/vdbench$
IOmeter
Iometer is an I/O subsystem measurement and characterization tool for single and clustered systems, developed by Intel and relased to open source community. Its distributed under Intel Open Source License.
Tested IOmeter with basic workload (512B 100%Read 0% Random) and (4k 100%Read 0% Random) with modifications for max files in buffercache. Further work include populating results in a useful format and trying different workloads.
References
http://www.iometer.org/
http://www.linuxintro.org/wiki/Iometer
http://www.itechstorm.com/iometer-tutorial-introduction
Installation and Setup
http://greg.porter.name/wiki/HowTo:iometer
LFStest
This test is basically derived from Rosenblum's LFS paper where it performs following operations
- Create a large file with sequential writes
- Read back the file with sequential reads.
- Perform random writes to this existing file
- Perform random reads to this existing file
- Calculates the bandwidth of all above 4 operations.
It's very easy to install and run the benchmark. (Basically a C implementation)
Source
http://fsbench.filesystems.org/bench/sprite-lfs.tar.gz
Usage
The test also provides option of specifying IO size for reads and write at application level e.g. 256KB etc.
largefile [-f file_size] [-i IO_size] [-s seed] dirname (dirname would be /tmp/smr/sfs)
This creates a test file in indicated directory. file_size indicates file size in MBs. IO_size specifies I/O transfers in KBs.
This test was useful to check for operations on files larger than bandsize and a bug has been reported to provide multiple extent support for file on SMRfs
PUMA: Purdue MapReduce Benchmarks Suite
There are a total of 13 benchmarks, out of which Tera-Sort, Word-Count, and Grep are from Hadoop distribution, and others. The three benchmarks from Hadoop distribution are also slightly modified to take number of reduce tasks as input from the user and generate final time completion statistics of jobs.
PUMA jar file:
http://web.ics.purdue.edu/~fahmad/benchmarks/hadoop-0.20.3-dev-examples.jar
- Word-Count
- Inverted-Index
- Term-Vector
- Self-Join
- Adjacency-List
- K-Means
- Classification
- Histogram-Movies
- Histogram-Ratings
- Sequence-Count
- Ranked-Inverted-Index
- Tera-Sort
- Grep
Ref:
http://web.ics.purdue.edu/~fahmad/benchmarks.htm
[hduser@iceberg: ~] $ hadoop jar /home/tejas/Downloads/hadoop-0.20.3-dev-examples.jar wordcount /user/hduser/input /user/hduser/output
Warning: $HADOOP_HOME is deprecated.
13/10/24 22:13:53 WARN mapred.JobClient: Use GenericOptionsParser for parsing the arguments. Applications should implement Tool for the same.
13/10/24 22:13:53 INFO input.FileInputFormat: Total input paths to process : 4
13/10/24 22:13:53 INFO util.NativeCodeLoader: Loaded the native-hadoop library
13/10/24 22:13:53 WARN snappy.LoadSnappy: Snappy native library not loaded
13/10/24 22:13:53 INFO mapred.JobClient: Running job: job_201310242209_0005
13/10/24 22:13:54 INFO mapred.JobClient: map 0% reduce 0%
13/10/24 22:14:10 INFO mapred.JobClient: map 25% reduce 0%
13/10/24 22:14:11 INFO mapred.JobClient: map 36% reduce 0%
13/10/24 22:14:14 INFO mapred.JobClient: map 38% reduce 0%
13/10/24 22:14:17 INFO mapred.JobClient: map 42% reduce 0%
13/10/24 22:14:19 INFO mapred.JobClient: map 67% reduce 0%
13/10/24 22:14:20 INFO mapred.JobClient: map 74% reduce 0%
13/10/24 22:14:23 INFO mapred.JobClient: map 75% reduce 0%
13/10/24 22:14:24 INFO mapred.JobClient: map 100% reduce 0%
13/10/24 22:14:28 INFO mapred.JobClient: map 100% reduce 33%
13/10/24 22:14:32 INFO mapred.JobClient: map 100% reduce 70%
13/10/24 22:14:35 INFO mapred.JobClient: map 100% reduce 100%
13/10/24 22:14:36 INFO mapred.JobClient: Job complete: job_201310242209_0005
13/10/24 22:14:36 INFO mapred.JobClient: Counters: 29
13/10/24 22:14:36 INFO mapred.JobClient: Job Counters
13/10/24 22:14:36 INFO mapred.JobClient: Launched reduce tasks=1
13/10/24 22:14:36 INFO mapred.JobClient: SLOTS_MILLIS_MAPS=51806
13/10/24 22:14:36 INFO mapred.JobClient: Total time spent by all reduces waiting after reserving slots (ms)=0
13/10/24 22:14:36 INFO mapred.JobClient: Total time spent by all maps waiting after reserving slots (ms)=0
13/10/24 22:14:36 INFO mapred.JobClient: Launched map tasks=4
13/10/24 22:14:36 INFO mapred.JobClient: Data-local map tasks=4
13/10/24 22:14:36 INFO mapred.JobClient: SLOTS_MILLIS_REDUCES=23575
13/10/24 22:14:36 INFO mapred.JobClient: File Output Format Counters
13/10/24 22:14:36 INFO mapred.JobClient: Bytes Written=75160700
13/10/24 22:14:36 INFO mapred.JobClient: FileSystemCounters
13/10/24 22:14:36 INFO mapred.JobClient: FILE_BYTES_READ=161139412
13/10/24 22:14:36 INFO mapred.JobClient: HDFS_BYTES_READ=78336286
13/10/24 22:14:36 INFO mapred.JobClient: FILE_BYTES_WRITTEN=241821207
13/10/24 22:14:36 INFO mapred.JobClient: HDFS_BYTES_WRITTEN=75160700
13/10/24 22:14:36 INFO mapred.JobClient: File Input Format Counters
13/10/24 22:14:36 INFO mapred.JobClient: Bytes Read=78335822
13/10/24 22:14:36 INFO mapred.JobClient: Map-Reduce Framework
13/10/24 22:14:36 INFO mapred.JobClient: Map output materialized bytes=80373134
13/10/24 22:14:36 INFO mapred.JobClient: Map input records=1327552
13/10/24 22:14:36 INFO mapred.JobClient: Reduce shuffle bytes=80373134
13/10/24 22:14:36 INFO mapred.JobClient: Spilled Records=3827483
13/10/24 22:14:36 INFO mapred.JobClient: Map output bytes=85126210
13/10/24 22:14:36 INFO mapred.JobClient: Total committed heap usage (bytes)=988610560
13/10/24 22:14:36 INFO mapred.JobClient: CPU time spent (ms)=34790
13/10/24 22:14:36 INFO mapred.JobClient: Combine input records=3349872
13/10/24 22:14:36 INFO mapred.JobClient: SPLIT_RAW_BYTES=464
13/10/24 22:14:36 INFO mapred.JobClient: Reduce input records=1267843
13/10/24 22:14:36 INFO mapred.JobClient: Reduce input groups=1252264
13/10/24 22:14:36 INFO mapred.JobClient: Combine output records=2559640
13/10/24 22:14:36 INFO mapred.JobClient: Physical memory (bytes) snapshot=958275584
13/10/24 22:14:36 INFO mapred.JobClient: Reduce output records=1252264
13/10/24 22:14:36 INFO mapred.JobClient: Virtual memory (bytes) snapshot=4764823552
13/10/24 22:14:36 INFO mapred.JobClient: Map output records=2058075
The iteration took 43 seconds.
aio-stress
Test configuration: Tests were run locally using a 4GB image file formatted with Btrfs (The test crashes when run on a raw partition).
Tests were run for file-sizes: 10MB, 50MB, 100MB and 400MB.
Commandline: aio-stress -s <filesize> a1 a2 -u
File size: 10 M
===========
file size 10MB, record size 64KB, depth 64, ios per iteration 8
max io_submit 16, buffer alignment set to 4KB
threads 1 files 2 contexts 1 context offset 2MB verification off
write on a1 (209.63 MB/s) 10.00 MB in 0.05s
write on a2 (209.38 MB/s) 10.00 MB in 0.05s
thread 0 write totals (260.73 MB/s) 20.00 MB in 0.08s
read on a1 (2409.64 MB/s) 10.00 MB in 0.00s
read on a2 (2396.93 MB/s) 10.00 MB in 0.00s
thread 0 read totals (4507.55 MB/s) 20.00 MB in 0.00s
random write on a1 (380.53 MB/s) 10.00 MB in 0.03s
random write on a2 (379.95 MB/s) 10.00 MB in 0.03s
thread 0 random write totals (412.84 MB/s) 20.00 MB in 0.05s
random read on a1 (1583.28 MB/s) 10.00 MB in 0.01s
random read on a2 (1578.78 MB/s) 10.00 MB in 0.01s
thread 0 random read totals (3095.02 MB/s) 20.00 MB in 0.01s
Running single thread version
File size: 50M
===========
file size 50MB, record size 64KB, depth 64, ios per iteration 8
max io_submit 16, buffer alignment set to 4KB
threads 1 files 2 contexts 1 context offset 2MB verification off
write on a1 (223.43 MB/s) 50.00 MB in 0.22s
write on a2 (223.36 MB/s) 50.00 MB in 0.22s
thread 0 write totals (329.13 MB/s) 100.00 MB in 0.30s
read on a1 (1725.51 MB/s) 50.00 MB in 0.03s
read on a2 (1724.20 MB/s) 50.00 MB in 0.03s
thread 0 read totals (3430.06 MB/s) 100.00 MB in 0.03s
random write on a1 (293.10 MB/s) 50.00 MB in 0.17s
random write on a2 (293.03 MB/s) 50.00 MB in 0.17s
thread 0 random write totals (394.91 MB/s) 100.00 MB in 0.25s
random read on a1 (2776.54 MB/s) 50.00 MB in 0.02s
random read on a2 (2773.16 MB/s) 50.00 MB in 0.02s
thread 0 random read totals (5505.09 MB/s) 100.00 MB in 0.02s
Running single thread version
File size: 100MB
=============
file size 100MB, record size 64KB, depth 64, ios per iteration 8
max io_submit 16, buffer alignment set to 4KB
threads 1 files 2 contexts 1 context offset 2MB verification off
write on a1 (246.02 MB/s) 100.00 MB in 0.41s
write on a2 (245.98 MB/s) 100.00 MB in 0.41s
thread 0 write totals (363.83 MB/s) 200.00 MB in 0.55s
read on a1 (2221.83 MB/s) 100.00 MB in 0.05s
read on a2 (2220.64 MB/s) 100.00 MB in 0.05s
thread 0 read totals (4412.28 MB/s) 200.00 MB in 0.05s
random write on a1 (296.46 MB/s) 100.00 MB in 0.34s
random write on a2 (296.45 MB/s) 100.00 MB in 0.34s
thread 0 random write totals (420.21 MB/s) 200.00 MB in 0.48s
random read on a1 (2608.51 MB/s) 100.00 MB in 0.04s
random read on a2 (2607.15 MB/s) 100.00 MB in 0.04s
thread 0 random read totals (5202.64 MB/s) 200.00 MB in 0.04s
Running single thread version
File size: 200MB
=============
file size 200MB, record size 64KB, depth 64, ios per iteration 8
max io_submit 16, buffer alignment set to 4KB
threads 1 files 2 contexts 1 context offset 2MB verification off
write on a1 (272.05 MB/s) 200.00 MB in 0.74s
write on a2 (272.03 MB/s) 200.00 MB in 0.74s
thread 0 write totals (386.49 MB/s) 400.00 MB in 1.03s
read on a1 (2472.68 MB/s) 200.00 MB in 0.08s
read on a2 (2471.94 MB/s) 200.00 MB in 0.08s
thread 0 read totals (4923.99 MB/s) 400.00 MB in 0.08s
random write on a1 (294.27 MB/s) 200.00 MB in 0.68s
random write on a2 (294.25 MB/s) 200.00 MB in 0.68s
thread 0 random write totals (388.89 MB/s) 400.00 MB in 1.03s
random read on a1 (2686.62 MB/s) 200.00 MB in 0.07s
random read on a2 (2685.90 MB/s) 200.00 MB in 0.07s
thread 0 random read totals (5358.05 MB/s) 400.00 MB in 0.07s
Running single thread version
File size: 400MB
=============
file size 400MB, record size 64KB, depth 64, ios per iteration 8
max io_submit 16, buffer alignment set to 4KB
threads 1 files 2 contexts 1 context offset 2MB verification off
write on a1 (283.86 MB/s) 400.00 MB in 1.41s
write on a2 (283.85 MB/s) 400.00 MB in 1.41s
thread 0 write totals (86.62 MB/s) 800.00 MB in 9.24s
read on a1 (1600.06 MB/s) 400.00 MB in 0.25s
read on a2 (1599.90 MB/s) 400.00 MB in 0.25s
thread 0 read totals (3198.81 MB/s) 800.00 MB in 0.25s
random write on a1 (467.55 MB/s) 400.00 MB in 0.86s
random write on a2 (314.20 MB/s) 400.00 MB in 1.27s
thread 0 random write totals (58.13 MB/s) 800.00 MB in 13.76s
random read on a1 (2461.27 MB/s) 400.00 MB in 0.16s
random read on a2 (2460.95 MB/s) 400.00 MB in 0.16s
thread 0 random read totals (4920.75 MB/s) 800.00 MB in 0.16s
Running single thread version
--
MukulSingh - 25 Oct 2013