[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]
Re: [Qemu-devel] [RFC] migration/block:limit the time used for block mig
From: |
858585 jemmy |
Subject: |
Re: [Qemu-devel] [RFC] migration/block:limit the time used for block migration |
Date: |
Tue, 28 Mar 2017 17:32:07 +0800 |
when migrate the vm with quick speed, i find vnc response slowly.
the bug can be reproduce by this command:
virsh migrate-setspeed 165cf436-312f-47e7-90f2-f8aa63f34893 900
virsh migrate --live 165cf436-312f-47e7-90f2-f8aa63f34893
--copy-storage-inc qemu+ssh://10.59.163.38/system
and --copy-storage-all have no problem.
virsh migrate --live 165cf436-312f-47e7-90f2-f8aa63f34893
--copy-storage-all qemu+ssh://10.59.163.38/system
mig_save_device_bulk invoke bdrv_is_allocated, but bdrv_is_allocated maybe
wait for a long time.
so cause the main thread wait for a long time.
this patch limit the time wait for bdrv_is_allocated.
i do not find a better way to solve this bug, Any suggestion?
Thanks.
On Tue, Mar 28, 2017 at 5:23 PM, Lidong Chen <address@hidden> wrote:
> when migration with quick speed, mig_save_device_bulk invoke
> bdrv_is_allocated too frequently, and cause vnc reponse slowly.
> this patch limit the time used for bdrv_is_allocated.
>
> Signed-off-by: Lidong Chen <address@hidden>
> ---
> migration/block.c | 39 +++++++++++++++++++++++++++++++--------
> 1 file changed, 31 insertions(+), 8 deletions(-)
>
> diff --git a/migration/block.c b/migration/block.c
> index 7734ff7..d3e81ca 100644
> --- a/migration/block.c
> +++ b/migration/block.c
> @@ -110,6 +110,7 @@ typedef struct BlkMigState {
> int transferred;
> int prev_progress;
> int bulk_completed;
> + int time_ns_used;
>
> /* Lock must be taken _inside_ the iothread lock and any
> AioContexts. */
> QemuMutex lock;
> @@ -263,6 +264,7 @@ static void blk_mig_read_cb(void *opaque, int ret)
> blk_mig_unlock();
> }
>
> +#define BILLION 1000000000L
> /* Called with no lock taken. */
>
> static int mig_save_device_bulk(QEMUFile *f, BlkMigDevState *bmds)
> @@ -272,16 +274,33 @@ static int mig_save_device_bulk(QEMUFile *f,
> BlkMigDevState *bmds)
> BlockBackend *bb = bmds->blk;
> BlkMigBlock *blk;
> int nr_sectors;
> + struct timespec ts1, ts2;
> + int ret = 0;
> + int timeout_flag = 0;
>
> if (bmds->shared_base) {
> qemu_mutex_lock_iothread();
> aio_context_acquire(blk_get_aio_context(bb));
> /* Skip unallocated sectors; intentionally treats failure as
> * an allocated sector */
> - while (cur_sector < total_sectors &&
> - !bdrv_is_allocated(blk_bs(bb), cur_sector,
> - MAX_IS_ALLOCATED_SEARCH, &nr_sectors)) {
> - cur_sector += nr_sectors;
> + while (cur_sector < total_sectors) {
> + clock_gettime(CLOCK_MONOTONIC_RAW, &ts1);
> + ret = bdrv_is_allocated(blk_bs(bb), cur_sector,
> + MAX_IS_ALLOCATED_SEARCH, &nr_sectors);
> + clock_gettime(CLOCK_MONOTONIC_RAW, &ts2);
> +
> + block_mig_state.time_ns_used += (ts2.tv_sec - ts1.tv_sec) *
> BILLION
> + + (ts2.tv_nsec - ts1.tv_nsec);
> +
> + if (!ret) {
> + cur_sector += nr_sectors;
> + if (block_mig_state.time_ns_used > 100000) {
> + timeout_flag = 1;
> + break;
> + }
> + } else {
> + break;
> + }
> }
> aio_context_release(blk_get_aio_context(bb));
> qemu_mutex_unlock_iothread();
> @@ -292,6 +311,11 @@ static int mig_save_device_bulk(QEMUFile *f,
> BlkMigDevState *bmds)
> return 1;
> }
>
> + if (timeout_flag == 1) {
> + bmds->cur_sector = bmds->completed_sectors = cur_sector;
> + return 0;
> + }
> +
> bmds->completed_sectors = cur_sector;
>
> cur_sector &= ~((int64_t)BDRV_SECTORS_PER_DIRTY_CHUNK - 1);
> @@ -576,9 +600,6 @@ static int mig_save_device_dirty(QEMUFile *f,
> BlkMigDevState *bmds,
> }
>
> bdrv_reset_dirty_bitmap(bmds->dirty_bitmap, sector,
> nr_sectors);
> - sector += nr_sectors;
> - bmds->cur_dirty = sector;
> -
> break;
> }
> sector += BDRV_SECTORS_PER_DIRTY_CHUNK;
> @@ -756,6 +777,7 @@ static int block_save_iterate(QEMUFile *f, void
> *opaque)
> }
>
> blk_mig_reset_dirty_cursor();
> + block_mig_state.time_ns_used = 0;
>
> /* control the rate of transfer */
> blk_mig_lock();
> @@ -764,7 +786,8 @@ static int block_save_iterate(QEMUFile *f, void
> *opaque)
> qemu_file_get_rate_limit(f) &&
> (block_mig_state.submitted +
> block_mig_state.read_done) <
> - MAX_INFLIGHT_IO) {
> + MAX_INFLIGHT_IO &&
> + block_mig_state.time_ns_used <= 100000) {
> blk_mig_unlock();
> if (block_mig_state.bulk_completed == 0) {
> /* first finish the bulk phase */
> --
> 1.8.3.1
>
>