[Test Result] I/O bandwidth Control by dm-ioband - partition-based environment
baramsori72 at gmail.com
Fri Aug 28 03:24:56 PDT 2009
I attached new file that includes I/O total bandwidth of evaluation system.
We tested total bandwidth of weight policy by I/O in Dom0 and DomU system
and it is measured through iostat tool and dm-ioband debug patch which I
gave you several months ago.
Of course, the result in prior report was measured by dm-ioband debug patch.
As a result, the big difference in prior report derives from the location
where we measured I/O bandwidth
iostat counts it in application level and dm-ioband debug patch does it in
I think the difference is related with buffer cache.
Have a nice weekend
2009/8/27 Dong-Jae Kang <baramsori72 at gmail.com>
> Hi Ryo
> 2009/8/27 Ryo Tsuruta <ryov at valinux.co.jp>
> Hi Dong-Jae,
>> # I've added dm-devel to Cc:.
>> Dong-Jae Kang <baramsori72 at gmail.com> wrote:
>> > Hi Ryo
>> > I attached new test result file(ioband-partition-based-evaluation.xls)in
>> > this mail.
>> Thanks for your great job.
>> > In this time, it is not virtualization environment.
>> > I evaluated partition-based use cases before I do it in vitualization
>> > environment.
>> > because I think the two cases are smilar each other.
>> > The detailed information about the evaluation can be referred in
>> > file.
>> > If you have any questions or comments after examine it,
>> > please give me your opinion.
>> I would like to know the throughput without dm-ioband in your
>> environment. Because the total throughput of range-bw policy is
>> 8000KB/s, which means the device has a capability to perform over
>> 8000KB/s, but the total throughput of weight policy is lower than
>> the range-bw policy. In my environment, there is no significant
>> difference in average throughput between with and without dm-ioband.
>> I ran fio in the way described in your result file. Here are the
>> results of my environment. The throughputs were calculated from
>> "iostat -k 1" outputs.
>> buffered write test
>> Avg. throughput [KB/s]
>> w/o ioband w/ioband
>> sdb2 14485 5788
>> sdb3 12494 22295
>> total 26979 28030
> OK, good comments.
> I omitted the total bandwidth of the evaluation system.
> I will reply to you about it tomorrow after I check and re-test it again.
>> Ryo Tsuruta
> Thank you for comments.
> Best Regards,
> Dong-Jae Kang
-------------- next part --------------
A non-text attachment was scrubbed...
Name: total bandwidth result.xls
Size: 33792 bytes
Desc: not available
Url : http://lists.linux-foundation.org/pipermail/containers/attachments/20090828/9d864719/attachment-0001.xls
More information about the Containers