[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Qemu-devel] [PULL 2/5] migration: move bdrv_invalidate_cache_all of
Denis V. Lunev
Re: [Qemu-devel] [PULL 2/5] migration: move bdrv_invalidate_cache_all of of coroutine context
Tue, 8 Mar 2016 13:54:33 +0300
Mozilla/5.0 (X11; Linux x86_64; rv:38.0) Gecko/20100101 Thunderbird/38.5.1
On 03/08/2016 01:45 PM, Dr. David Alan Gilbert wrote:
* Denis V. Lunev (address@hidden) wrote:
On 03/07/2016 03:49 PM, Dr. David Alan Gilbert wrote:
* Amit Shah (address@hidden) wrote:
From: "Denis V. Lunev" <address@hidden>
There is a possibility to hit an assert in qcow2_get_specific_info that
s->qcow_version is undefined. This happens when VM in starting from
suspended state, i.e. it processes incoming migration, and in the same
time 'info block' is called.
The problem is that qcow2_invalidate_cache() closes the image and
memset()s BDRVQcowState in the middle.
The patch moves processing of bdrv_invalidate_cache_all out of
coroutine context for postcopy migration to avoid that. This function
is called with the following stack:
Signed-off-by: Denis V. Lunev <address@hidden>
Tested-by: Dr. David Alan Gilbert <address@hidden>
hmm; actually - this segs in a variety of different ways;
there are two problems:
a) + bh = qemu_bh_new(loadvm_postcopy_handle_run_bh, NULL);
That's the easy one; that NULL should be 'mis', because
the bh is expecting to use it as a MigrationIncomingState
so it segs fairly reliably in the qemu_bh_delete(mis->bh)
b) The harder problem is that there's a race where qemu_bh_delete
segs, and I'm not 100% sure why yet - it only does it sometime
(i.e. run virt-test and leave it and it occasionally does it).
From the core it looks like qemu->bh is corrupt (0x10101010...)
so maybe mis has been freed at that point?
I'm suspecting this is the postcopy_ram_listen_thread freeing
mis at the end of it, but I don't know yet.
Yes. this is exactly use-after-free. I have looked into the code
and this seems correct.
Could you try this simple patch?
Hmm no, that's not right.
The order for postcopy is that we are running the listen thread and then
receive the 'run', and the listening thread is still running - so you
can't destroy the incoming state during the run.
It can't get destroyed until both the main thread has finished loading
the migration AND the listen thread has finished.
Hmm - that does give me an idea about the other seg I saw; I need to check it;
but I think the problem is probably the case of a very short postcopy
where the listen thread exits before the handle_run_bh is triggered;
(and since I've only seen it in my virt-test setup, and I know it can do
very short postcopies)
I think the fix here is to pass loadvm_postcopy_handle_run_bh a pointer to it's
own bh structure rather than store it in mis->bh; that way it doesn't use mis
diff --git a/migration/savevm.c b/migration/savevm.c
index 96e7db5..9a020ef 100644
@@ -1446,15 +1446,6 @@ static void *postcopy_ram_listen_thread(void *opaque)
- * If everything has worked fine, then the main thread has waited
- * for us to start, and we're the last use of the mis.
- * (If something broke then qemu will have to exit anyway since it's
- * got a bad migration state).
@@ -1533,6 +1524,14 @@ static void loadvm_postcopy_handle_run_bh(void *opaque)
+ * If everything has worked fine, then the main thread has waited
+ * for us to start, and we're the last use of the mis.
+ * (If something broke then qemu will have to exit anyway since it's
+ * got a bad migration state).
/* After all discards we can start running and asking for pages */
Dr. David Alan Gilbert / address@hidden / Manchester, UK
This will help for sure. The idea to reuse migration state seems wrong.