[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [PATCH v3 17/23] multifd: Use normal pages array on the send side
From: |
Dr. David Alan Gilbert |
Subject: |
Re: [PATCH v3 17/23] multifd: Use normal pages array on the send side |
Date: |
Tue, 30 Nov 2021 10:50:22 +0000 |
User-agent: |
Mutt/2.1.3 (2021-09-10) |
* Juan Quintela (quintela@redhat.com) wrote:
> Signed-off-by: Juan Quintela <quintela@redhat.com>
Can you explain a bit more what's going on here?
Dave
> ---
> migration/multifd.h | 8 ++++++--
> migration/multifd-zlib.c | 6 +++---
> migration/multifd-zstd.c | 6 +++---
> migration/multifd.c | 30 +++++++++++++++++++-----------
> migration/trace-events | 4 ++--
> 5 files changed, 33 insertions(+), 21 deletions(-)
>
> diff --git a/migration/multifd.h b/migration/multifd.h
> index 7496f951a7..78e73df3ec 100644
> --- a/migration/multifd.h
> +++ b/migration/multifd.h
> @@ -104,14 +104,18 @@ typedef struct {
> /* thread local variables */
> /* packets sent through this channel */
> uint64_t num_packets;
> - /* pages sent through this channel */
> - uint64_t num_pages;
> + /* non zero pages sent through this channel */
> + uint64_t num_normal_pages;
> /* syncs main thread and channels */
> QemuSemaphore sem_sync;
> /* buffers to send */
> struct iovec *iov;
> /* number of iovs used */
> uint32_t iovs_num;
> + /* Pages that are not zero */
> + ram_addr_t *normal;
> + /* num of non zero pages */
> + uint32_t normal_num;
> /* used for compression methods */
> void *data;
> } MultiFDSendParams;
> diff --git a/migration/multifd-zlib.c b/migration/multifd-zlib.c
> index f65159392a..25ef68a548 100644
> --- a/migration/multifd-zlib.c
> +++ b/migration/multifd-zlib.c
> @@ -106,16 +106,16 @@ static int zlib_send_prepare(MultiFDSendParams *p,
> Error **errp)
> int ret;
> uint32_t i;
>
> - for (i = 0; i < p->pages->num; i++) {
> + for (i = 0; i < p->normal_num; i++) {
> uint32_t available = z->zbuff_len - out_size;
> int flush = Z_NO_FLUSH;
>
> - if (i == p->pages->num - 1) {
> + if (i == p->normal_num - 1) {
> flush = Z_SYNC_FLUSH;
> }
>
> zs->avail_in = page_size;
> - zs->next_in = p->pages->block->host + p->pages->offset[i];
> + zs->next_in = p->pages->block->host + p->normal[i];
>
> zs->avail_out = available;
> zs->next_out = z->zbuff + out_size;
> diff --git a/migration/multifd-zstd.c b/migration/multifd-zstd.c
> index 6933ba622a..61842d713e 100644
> --- a/migration/multifd-zstd.c
> +++ b/migration/multifd-zstd.c
> @@ -121,13 +121,13 @@ static int zstd_send_prepare(MultiFDSendParams *p,
> Error **errp)
> z->out.size = z->zbuff_len;
> z->out.pos = 0;
>
> - for (i = 0; i < p->pages->num; i++) {
> + for (i = 0; i < p->normal_num; i++) {
> ZSTD_EndDirective flush = ZSTD_e_continue;
>
> - if (i == p->pages->num - 1) {
> + if (i == p->normal_num - 1) {
> flush = ZSTD_e_flush;
> }
> - z->in.src = p->pages->block->host + p->pages->offset[i];
> + z->in.src = p->pages->block->host + p->normal[i];
> z->in.size = page_size;
> z->in.pos = 0;
>
> diff --git a/migration/multifd.c b/migration/multifd.c
> index 6983ba3e7c..dbe919b764 100644
> --- a/migration/multifd.c
> +++ b/migration/multifd.c
> @@ -89,13 +89,13 @@ static int nocomp_send_prepare(MultiFDSendParams *p,
> Error **errp)
> MultiFDPages_t *pages = p->pages;
> size_t page_size = qemu_target_page_size();
>
> - for (int i = 0; i < p->pages->num; i++) {
> - p->iov[p->iovs_num].iov_base = pages->block->host + pages->offset[i];
> + for (int i = 0; i < p->normal_num; i++) {
> + p->iov[p->iovs_num].iov_base = pages->block->host + p->normal[i];
> p->iov[p->iovs_num].iov_len = page_size;
> p->iovs_num++;
> }
>
> - p->next_packet_size = p->pages->num * page_size;
> + p->next_packet_size = p->normal_num * page_size;
> p->flags |= MULTIFD_FLAG_NOCOMP;
> return 0;
> }
> @@ -262,7 +262,7 @@ static void multifd_send_fill_packet(MultiFDSendParams *p)
>
> packet->flags = cpu_to_be32(p->flags);
> packet->pages_alloc = cpu_to_be32(p->pages->allocated);
> - packet->pages_used = cpu_to_be32(p->pages->num);
> + packet->pages_used = cpu_to_be32(p->normal_num);
> packet->next_packet_size = cpu_to_be32(p->next_packet_size);
> packet->packet_num = cpu_to_be64(p->packet_num);
>
> @@ -270,9 +270,9 @@ static void multifd_send_fill_packet(MultiFDSendParams *p)
> strncpy(packet->ramblock, p->pages->block->idstr, 256);
> }
>
> - for (i = 0; i < p->pages->num; i++) {
> + for (i = 0; i < p->normal_num; i++) {
> /* there are architectures where ram_addr_t is 32 bit */
> - uint64_t temp = p->pages->offset[i];
> + uint64_t temp = p->normal[i];
>
> packet->offset[i] = cpu_to_be64(temp);
> }
> @@ -556,6 +556,8 @@ void multifd_save_cleanup(void)
> p->packet = NULL;
> g_free(p->iov);
> p->iov = NULL;
> + g_free(p->normal);
> + p->normal = NULL;
> multifd_send_state->ops->send_cleanup(p, &local_err);
> if (local_err) {
> migrate_set_error(migrate_get_current(), local_err);
> @@ -640,12 +642,17 @@ static void *multifd_send_thread(void *opaque)
> qemu_mutex_lock(&p->mutex);
>
> if (p->pending_job) {
> - uint32_t used = p->pages->num;
> uint64_t packet_num = p->packet_num;
> uint32_t flags = p->flags;
> p->iovs_num = 1;
> + p->normal_num = 0;
>
> - if (used) {
> + for (int i = 0; i < p->pages->num; i++) {
> + p->normal[p->normal_num] = p->pages->offset[i];
> + p->normal_num++;
> + }
> +
> + if (p->normal_num) {
> ret = multifd_send_state->ops->send_prepare(p, &local_err);
> if (ret != 0) {
> qemu_mutex_unlock(&p->mutex);
> @@ -655,12 +662,12 @@ static void *multifd_send_thread(void *opaque)
> multifd_send_fill_packet(p);
> p->flags = 0;
> p->num_packets++;
> - p->num_pages += used;
> + p->num_normal_pages += p->normal_num;
> p->pages->num = 0;
> p->pages->block = NULL;
> qemu_mutex_unlock(&p->mutex);
>
> - trace_multifd_send(p->id, packet_num, used, flags,
> + trace_multifd_send(p->id, packet_num, p->normal_num, flags,
> p->next_packet_size);
>
> p->iov[0].iov_len = p->packet_len;
> @@ -710,7 +717,7 @@ out:
> qemu_mutex_unlock(&p->mutex);
>
> rcu_unregister_thread();
> - trace_multifd_send_thread_end(p->id, p->num_packets, p->num_pages);
> + trace_multifd_send_thread_end(p->id, p->num_packets,
> p->num_normal_pages);
>
> return NULL;
> }
> @@ -910,6 +917,7 @@ int multifd_save_setup(Error **errp)
> p->tls_hostname = g_strdup(s->hostname);
> /* We need one extra place for the packet header */
> p->iov = g_new0(struct iovec, page_count + 1);
> + p->normal = g_new0(ram_addr_t, page_count);
> socket_send_channel_create(multifd_new_send_channel_async, p);
> }
>
> diff --git a/migration/trace-events b/migration/trace-events
> index b48d873b8a..af8dee9af0 100644
> --- a/migration/trace-events
> +++ b/migration/trace-events
> @@ -124,13 +124,13 @@ multifd_recv_sync_main_wait(uint8_t id) "channel %d"
> multifd_recv_terminate_threads(bool error) "error %d"
> multifd_recv_thread_end(uint8_t id, uint64_t packets, uint64_t pages)
> "channel %d packets %" PRIu64 " pages %" PRIu64
> multifd_recv_thread_start(uint8_t id) "%d"
> -multifd_send(uint8_t id, uint64_t packet_num, uint32_t used, uint32_t flags,
> uint32_t next_packet_size) "channel %d packet_num %" PRIu64 " pages %d flags
> 0x%x next packet size %d"
> +multifd_send(uint8_t id, uint64_t packet_num, uint32_t normal, uint32_t
> flags, uint32_t next_packet_size) "channel %d packet_num %" PRIu64 " normal
> pages %d flags 0x%x next packet size %d"
> multifd_send_error(uint8_t id) "channel %d"
> multifd_send_sync_main(long packet_num) "packet num %ld"
> multifd_send_sync_main_signal(uint8_t id) "channel %d"
> multifd_send_sync_main_wait(uint8_t id) "channel %d"
> multifd_send_terminate_threads(bool error) "error %d"
> -multifd_send_thread_end(uint8_t id, uint64_t packets, uint64_t pages)
> "channel %d packets %" PRIu64 " pages %" PRIu64
> +multifd_send_thread_end(uint8_t id, uint64_t packets, uint64_t normal_pages)
> "channel %d packets %" PRIu64 " normal pages %" PRIu64
> multifd_send_thread_start(uint8_t id) "%d"
> multifd_tls_outgoing_handshake_start(void *ioc, void *tioc, const char
> *hostname) "ioc=%p tioc=%p hostname=%s"
> multifd_tls_outgoing_handshake_error(void *ioc, const char *err) "ioc=%p
> err=%s"
> --
> 2.33.1
>
--
Dr. David Alan Gilbert / dgilbert@redhat.com / Manchester, UK
- Re: [PATCH v3 12/23] multifd: Make zlib use iov's, (continued)
- [PATCH v3 06/23] multifd: remove used parameter from send_prepare() method, Juan Quintela, 2021/11/24
- [PATCH v3 13/23] multifd: Make zstd use iov's, Juan Quintela, 2021/11/24
- [PATCH v3 15/23] multifd: Use a single writev on the send side, Juan Quintela, 2021/11/24
- [PATCH v3 16/23] multifd: Unfold "used" variable by its value, Juan Quintela, 2021/11/24
- [PATCH v3 17/23] multifd: Use normal pages array on the send side, Juan Quintela, 2021/11/24
- Re: [PATCH v3 17/23] multifd: Use normal pages array on the send side,
Dr. David Alan Gilbert <=
- [PATCH v3 14/23] multifd: Remove send_write() method, Juan Quintela, 2021/11/24
- [PATCH v3 18/23] multifd: Use normal pages array on the recv side, Juan Quintela, 2021/11/24
- [PATCH v3 19/23] multifd: recv side only needs the RAMBlock host address, Juan Quintela, 2021/11/24
- [PATCH v3 20/23] multifd: Rename pages_used to normal_pages, Juan Quintela, 2021/11/24
- [PATCH v3 21/23] multifd: Support for zero pages transmission, Juan Quintela, 2021/11/24
- [PATCH v3 22/23] multifd: Zero pages transmission, Juan Quintela, 2021/11/24
- [PATCH v3 23/23] migration: Use multifd before we check for the zero page, Juan Quintela, 2021/11/24
- Re: [PATCH v3 00/23] Migration: Transmit and detect zero pages in the multifd threads, Peter Xu, 2021/11/24