[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Qemu-devel] [PATCH RFC] migration: set cpu throttle value by worklo
From: |
Chao Fan |
Subject: |
Re: [Qemu-devel] [PATCH RFC] migration: set cpu throttle value by workload |
Date: |
Thu, 29 Dec 2016 18:38:43 +0800 |
User-agent: |
Mutt/1.7.1 (2016-10-04) |
Hi all,
There is something to explain in this RFC PATCH.
On Thu, Dec 29, 2016 at 05:16:19PM +0800, Chao Fan wrote:
>This RFC PATCH is my demo about the new feature, here is my POC mail:
>https://lists.gnu.org/archive/html/qemu-devel/2016-12/msg00646.html
>
>When migration_bitmap_sync executed, get the time and read bitmap to
>calculate how many dirty pages born between two sync.
>Use inst_dirty_pages / (time_now - time_prev) / ram_size to get
>inst_dirty_pages_rate. Then map from the inst_dirty_pages_rate
>to cpu throttle value. I have no idea how to map it. So I just do
>that in a simple way. The mapping way is just a guess and should
>be improved.
>
>This is just a demo. There are more methods.
>1.In another file, calculate the inst_dirty_pages_rate every second
> or two seconds or another fixed time. Then set the cpu throttle
> value according to the inst_dirty_pages_rate
>2.When inst_dirty_pages_rate gets a threshold, begin cpu throttle
> and set the throttle value.
>
>Any comments will be welcome.
>
>Signed-off-by: Chao Fan <address@hidden>
>---
> include/qemu/bitmap.h | 17 +++++++++++++++++
> migration/ram.c | 49 +++++++++++++++++++++++++++++++++++++++++++++++++
> 2 files changed, 66 insertions(+)
>
>diff --git a/include/qemu/bitmap.h b/include/qemu/bitmap.h
>index 63ea2d0..dc99f9b 100644
>--- a/include/qemu/bitmap.h
>+++ b/include/qemu/bitmap.h
>@@ -235,4 +235,21 @@ static inline unsigned long *bitmap_zero_extend(unsigned
>long *old,
> return new;
> }
>
>+static inline unsigned long bitmap_weight(const unsigned long *src, long
>nbits)
It is a function imported from kernel, to calculate the number of
dirty pages.
>+{
>+ unsigned long i, count = 0, nlong = nbits / BITS_PER_LONG;
>+
>+ if (small_nbits(nbits)) {
>+ return hweight_long(*src & BITMAP_LAST_WORD_MASK(nbits));
>+ }
>+ for (i = 0; i < nlong; i++) {
>+ count += hweight_long(src[i]);
>+ }
>+ if (nbits % BITS_PER_LONG) {
>+ count += hweight_long(src[i] & BITMAP_LAST_WORD_MASK(nbits));
>+ }
>+
>+ return count;
>+}
>+
> #endif /* BITMAP_H */
>diff --git a/migration/ram.c b/migration/ram.c
>index a1c8089..f96e3e3 100644
>--- a/migration/ram.c
>+++ b/migration/ram.c
>@@ -44,6 +44,7 @@
> #include "exec/ram_addr.h"
> #include "qemu/rcu_queue.h"
> #include "migration/colo.h"
>+#include "hw/boards.h"
>
> #ifdef DEBUG_MIGRATION_RAM
> #define DPRINTF(fmt, ...) \
>@@ -599,6 +600,9 @@ static int64_t num_dirty_pages_period;
> static uint64_t xbzrle_cache_miss_prev;
> static uint64_t iterations_prev;
>
>+static int64_t dirty_pages_time_prev;
>+static int64_t dirty_pages_time_now;
>+
> static void migration_bitmap_sync_init(void)
> {
> start_time = 0;
>@@ -606,6 +610,49 @@ static void migration_bitmap_sync_init(void)
> num_dirty_pages_period = 0;
> xbzrle_cache_miss_prev = 0;
> iterations_prev = 0;
>+
>+ dirty_pages_time_prev = 0;
>+ dirty_pages_time_now = 0;
>+}
>+
>+static void migration_inst_rate(void)
>+{
>+ RAMBlock *block;
>+ MigrationState *s = migrate_get_current();
>+ int64_t inst_dirty_pages_rate, inst_dirty_pages = 0;
>+ int64_t i;
>+ unsigned long *num;
>+ unsigned long len = 0;
>+
>+ dirty_pages_time_now = qemu_clock_get_ms(QEMU_CLOCK_REALTIME);
When sync executed, we do this. And maybe every 1 second or another fixed
time to get the pages and time is also OK. But I have no idear which is
better.
>+ if (dirty_pages_time_prev != 0) {
>+ rcu_read_lock();
>+ DirtyMemoryBlocks *blocks = atomic_rcu_read(
>+ &ram_list.dirty_memory[DIRTY_MEMORY_MIGRATION]);
>+ QLIST_FOREACH_RCU(block, &ram_list.blocks, next) {
>+ if (len == 0) {
>+ len = block->offset;
>+ }
>+ len += block->used_length;
>+ }
>+ ram_addr_t idx = (len >> TARGET_PAGE_BITS) / DIRTY_MEMORY_BLOCK_SIZE;
>+ if (((len >> TARGET_PAGE_BITS) % DIRTY_MEMORY_BLOCK_SIZE) != 0) {
>+ idx++;
>+ }
>+ for (i = 0; i < idx; i++) {
>+ num = blocks->blocks[i];
>+ inst_dirty_pages += bitmap_weight(num, DIRTY_MEMORY_BLOCK_SIZE);
>+ }
>+ rcu_read_unlock();
>+
>+ inst_dirty_pages_rate = inst_dirty_pages * TARGET_PAGE_SIZE *
>+ 1024 * 1024 * 1000 /
The time we get is ms, so pages *1000 to make time changed to second.
The two *1024 is just to keep the magnitude, otherwise the
inst_dirty_pages is so small that the rate will be 0.
>+ (dirty_pages_time_now - dirty_pages_time_prev) /
>+ current_machine->ram_size;
>+ s->parameters.cpu_throttle_initial = inst_dirty_pages_rate / 200;
>+ s->parameters.cpu_throttle_increment = inst_dirty_pages_rate / 200;
Here the 200 is just a guess, because I don't know how map from
inst_dirty_pages_rate to throttle value. So just fill in a number.
I think there are better methods to map this. Then there will be a
better way to set the throttle value than the default 20/10.
Thanks,
Chao Fan
>+ }
>+ dirty_pages_time_prev = dirty_pages_time_now;
> }
>
> static void migration_bitmap_sync(void)
>@@ -629,6 +676,8 @@ static void migration_bitmap_sync(void)
> trace_migration_bitmap_sync_start();
> memory_global_dirty_log_sync();
>
>+ migration_inst_rate();
>+
> qemu_mutex_lock(&migration_bitmap_mutex);
> rcu_read_lock();
> QLIST_FOREACH_RCU(block, &ram_list.blocks, next) {
>--
>2.9.3
>