@@ -9,7 +9,7 @@ Handbook](https://s905060.gitbooks.io/site-reliability-engineer-handbook/content
99# Add the configmap
1010kubectl apply -f https://raw.githubusercontent.com/openinfrastructure/fio-kubernetes/master/configs.yaml
1111# Run the jobs for 60 seconds.
12- kubectl apply -f https://raw.githubusercontent.com/openinfrastructure/fio-kubernetes/master/configs .yaml
12+ kubectl apply -f https://raw.githubusercontent.com/openinfrastructure/fio-kubernetes/master/fio .yaml
1313```
1414
1515Collect the results using ` kubectl get pods ` and ` kubectl logs ` .
@@ -332,6 +332,66 @@ rpool 53.1G 1.03T 0 127 0 3.46M
332332rpool 53.1G 1.03T 0 19 0 285K
333333```
334334
335+ # Perc H310 replacement
336+
337+ Comparison of the random read/write job1 between two configurations on the same node:
338+
339+ 1 . Dell stock firmware for Perc H310.
340+ 2 . LSI initiator target (IT) mode firmware.
341+
342+ The hypothesis is the 512MB of cache on the dell firmware is causing a large
343+ IOPS backlog, which is too much latency for the k8s controller nodes. etcd
344+ leader writes specifically.
345+
346+ ## Dell Firmware
347+
348+ Run on r620c only with the dell firmware:
349+
350+ ```
351+ testjob: (groupid=0, jobs=12): err= 0: pid=10: Fri May 14 16:43:23 2021
352+ read: IOPS=10.1k, BW=79.1MiB/s (82.0MB/s)(4751MiB/60053msec)
353+ slat (usec): min=5, max=382179, avg=463.84, stdev=3584.34
354+ clat (usec): min=735, max=1863.9k, avg=72486.63, stdev=97045.23
355+ lat (usec): min=780, max=1863.9k, avg=72953.53, stdev=97707.59
356+ clat percentiles (msec):
357+ | 1.00th=[ 5], 5.00th=[ 7], 10.00th=[ 9], 20.00th=[ 14],
358+ | 30.00th=[ 21], 40.00th=[ 33], 50.00th=[ 45], 60.00th=[ 59],
359+ | 70.00th=[ 78], 80.00th=[ 105], 90.00th=[ 159], 95.00th=[ 232],
360+ | 99.00th=[ 493], 99.50th=[ 625], 99.90th=[ 1003], 99.95th=[ 1116],
361+ | 99.99th=[ 1351]
362+ bw ( KiB/s): min= 2830, max=504550, per=98.84%, avg=80065.34, stdev=6278.23, samples=1428
363+ iops : min= 352, max=63063, avg=10007.24, stdev=784.71, samples=1428
364+ write: IOPS=10.1k, BW=79.1MiB/s (82.9MB/s)(4750MiB/60053msec); 0 zone resets
365+ slat (usec): min=6, max=349939, avg=470.04, stdev=3621.64
366+ clat (usec): min=1291, max=1864.0k, avg=78277.90, stdev=101175.43
367+ lat (usec): min=1314, max=1910.4k, avg=78751.05, stdev=101814.17
368+ clat percentiles (msec):
369+ | 1.00th=[ 7], 5.00th=[ 9], 10.00th=[ 11], 20.00th=[ 17],
370+ | 30.00th=[ 25], 40.00th=[ 37], 50.00th=[ 50], 60.00th=[ 64],
371+ | 70.00th=[ 84], 80.00th=[ 112], 90.00th=[ 169], 95.00th=[ 247],
372+ | 99.00th=[ 510], 99.50th=[ 651], 99.90th=[ 1036], 99.95th=[ 1183],
373+ | 99.99th=[ 1418]
374+ bw ( KiB/s): min= 2831, max=508061, per=98.83%, avg=80046.54, stdev=6311.22, samples=1428
375+ iops : min= 353, max=63503, avg=10004.93, stdev=788.84, samples=1428
376+ lat (usec) : 750=0.01%, 1000=0.01%
377+ lat (msec) : 2=0.01%, 4=0.39%, 10=11.56%, 20=15.01%, 50=25.51%
378+ lat (msec) : 100=25.07%, 250=17.97%, 500=3.59%, 750=0.70%, 1000=0.20%
379+ lat (msec) : 2000=0.11%
380+ cpu : usr=1.40%, sys=3.56%, ctx=1090869, majf=0, minf=713
381+ IO depths : 1=0.0%, 2=0.0%, 4=0.0%, 8=0.0%, 16=0.0%, 32=0.0%, >=64=100.0%
382+ submit : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.0%
383+ complete : 0=0.0%, 4=100.0%, 8=0.0%, 16=0.0%, 32=0.0%, 64=0.0%, >=64=0.1%
384+ issued rwts: total=607356,607215,0,0 short=0,0,0,0 dropped=0,0,0,0
385+ latency : target=0, window=0, percentile=100.00%, depth=128
386+
387+ Run status group 0 (all jobs):
388+ READ: bw=79.1MiB/s (82.0MB/s), 79.1MiB/s-79.1MiB/s (82.0MB/s-82.0MB/s), io=4751MiB (4982MB), run=60053-60053msec
389+ WRITE: bw=79.1MiB/s (82.9MB/s), 79.1MiB/s-79.1MiB/s (82.9MB/s-82.9MB/s), io=4750MiB (4981MB), run=60053-60053msec
390+
391+ Disk stats (read/write):
392+ sda: ios=660161/660256, merge=172/341, ticks=28084610/30045312, in_queue=61318072, util=100.00%
393+ ```
394+
335395# Reference
336396
337397Forked from [ this post] ( https://medium.com/@joshua_robinson/storage-benchmarking-with-fio-in-kubernetes-14cf29dc5375 )
0 commit comments