qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Qemu-devel] [PATCHv2] migration: move ram migration support


From: Michael S. Tsirkin
Subject: Re: [Qemu-devel] [PATCHv2] migration: move ram migration support
Date: Thu, 7 Mar 2013 18:38:48 +0200

On Thu, Mar 07, 2013 at 04:20:48PM +0100, Paolo Bonzini wrote:
> Il 07/03/2013 14:32, Michael S. Tsirkin ha scritto:
> > +#ifdef DEBUG_ARCH_INIT
> > +#define DPRINTF(fmt, ...) \
> > +    do { fprintf(stdout, "arch_init: " fmt, ## __VA_ARGS__); } while (0)
> 
> These need to be adjusted, but it can be a follow-up.
> 
> Paolo

Thanks, will do a follow-up.

> > +#else
> > +#define DPRINTF(fmt, ...) \
> > +    do { } while (0)
> > +#endif
> > +
> > +/***********************************************************/
> > +/* ram save/restore */
> > +
> > +#define RAM_SAVE_FLAG_FULL     0x01 /* Obsolete, not used anymore */
> > +#define RAM_SAVE_FLAG_COMPRESS 0x02
> > +#define RAM_SAVE_FLAG_MEM_SIZE 0x04
> > +#define RAM_SAVE_FLAG_PAGE     0x08
> > +#define RAM_SAVE_FLAG_EOS      0x10
> > +#define RAM_SAVE_FLAG_CONTINUE 0x20
> > +#define RAM_SAVE_FLAG_XBZRLE   0x40
> > +
> > +#ifdef __ALTIVEC__
> > +#include <altivec.h>
> > +#define VECTYPE        vector unsigned char
> > +#define SPLAT(p)       vec_splat(vec_ld(0, p), 0)
> > +#define ALL_EQ(v1, v2) vec_all_eq(v1, v2)
> > +/* altivec.h may redefine the bool macro as vector type.
> > + * Reset it to POSIX semantics. */
> > +#undef bool
> > +#define bool _Bool
> > +#elif defined __SSE2__
> > +#include <emmintrin.h>
> > +#define VECTYPE        __m128i
> > +#define SPLAT(p)       _mm_set1_epi8(*(p))
> > +#define ALL_EQ(v1, v2) (_mm_movemask_epi8(_mm_cmpeq_epi8(v1, v2)) == 
> > 0xFFFF)
> > +#else
> > +#define VECTYPE        unsigned long
> > +#define SPLAT(p)       (*(p) * (~0UL / 255))
> > +#define ALL_EQ(v1, v2) ((v1) == (v2))
> > +#endif
> > +
> > +static int is_dup_page(uint8_t *page)
> > +{
> > +    VECTYPE *p = (VECTYPE *)page;
> > +    VECTYPE val = SPLAT(page);
> > +    int i;
> > +
> > +    for (i = 0; i < TARGET_PAGE_SIZE / sizeof(VECTYPE); i++) {
> > +        if (!ALL_EQ(val, p[i])) {
> > +            return 0;
> > +        }
> > +    }
> > +
> > +    return 1;
> > +}
> > +
> > +/* struct contains XBZRLE cache and a static page
> > +   used by the compression */
> > +static struct {
> > +    /* buffer used for XBZRLE encoding */
> > +    uint8_t *encoded_buf;
> > +    /* buffer for storing page content */
> > +    uint8_t *current_buf;
> > +    /* buffer used for XBZRLE decoding */
> > +    uint8_t *decoded_buf;
> > +    /* Cache for XBZRLE */
> > +    PageCache *cache;
> > +} XBZRLE = {
> > +    .encoded_buf = NULL,
> > +    .current_buf = NULL,
> > +    .decoded_buf = NULL,
> > +    .cache = NULL,
> > +};
> > +
> > +
> > +int64_t xbzrle_cache_resize(int64_t new_size)
> > +{
> > +    if (XBZRLE.cache != NULL) {
> > +        return cache_resize(XBZRLE.cache, new_size / TARGET_PAGE_SIZE) *
> > +            TARGET_PAGE_SIZE;
> > +    }
> > +    return pow2floor(new_size);
> > +}
> > +
> > +/* accounting for migration statistics */
> > +typedef struct AccountingInfo {
> > +    uint64_t dup_pages;
> > +    uint64_t norm_pages;
> > +    uint64_t iterations;
> > +    uint64_t xbzrle_bytes;
> > +    uint64_t xbzrle_pages;
> > +    uint64_t xbzrle_cache_miss;
> > +    uint64_t xbzrle_overflows;
> > +} AccountingInfo;
> > +
> > +static AccountingInfo acct_info;
> > +
> > +static void acct_clear(void)
> > +{
> > +    memset(&acct_info, 0, sizeof(acct_info));
> > +}
> > +
> > +uint64_t dup_mig_bytes_transferred(void)
> > +{
> > +    return acct_info.dup_pages * TARGET_PAGE_SIZE;
> > +}
> > +
> > +uint64_t dup_mig_pages_transferred(void)
> > +{
> > +    return acct_info.dup_pages;
> > +}
> > +
> > +uint64_t norm_mig_bytes_transferred(void)
> > +{
> > +    return acct_info.norm_pages * TARGET_PAGE_SIZE;
> > +}
> > +
> > +uint64_t norm_mig_pages_transferred(void)
> > +{
> > +    return acct_info.norm_pages;
> > +}
> > +
> > +uint64_t xbzrle_mig_bytes_transferred(void)
> > +{
> > +    return acct_info.xbzrle_bytes;
> > +}
> > +
> > +uint64_t xbzrle_mig_pages_transferred(void)
> > +{
> > +    return acct_info.xbzrle_pages;
> > +}
> > +
> > +uint64_t xbzrle_mig_pages_cache_miss(void)
> > +{
> > +    return acct_info.xbzrle_cache_miss;
> > +}
> > +
> > +uint64_t xbzrle_mig_pages_overflow(void)
> > +{
> > +    return acct_info.xbzrle_overflows;
> > +}
> > +
> > +static size_t save_block_hdr(QEMUFile *f, RAMBlock *block, ram_addr_t 
> > offset,
> > +                             int cont, int flag)
> > +{
> > +    size_t size;
> > +
> > +    qemu_put_be64(f, offset | cont | flag);
> > +    size = 8;
> > +
> > +    if (!cont) {
> > +        qemu_put_byte(f, strlen(block->idstr));
> > +        qemu_put_buffer(f, (uint8_t *)block->idstr,
> > +                        strlen(block->idstr));
> > +        size += 1 + strlen(block->idstr);
> > +    }
> > +    return size;
> > +}
> > +
> > +#define ENCODING_FLAG_XBZRLE 0x1
> > +
> > +static int save_xbzrle_page(QEMUFile *f, uint8_t *current_data,
> > +                            ram_addr_t current_addr, RAMBlock *block,
> > +                            ram_addr_t offset, int cont, bool last_stage)
> > +{
> > +    int encoded_len = 0, bytes_sent = -1;
> > +    uint8_t *prev_cached_page;
> > +
> > +    if (!cache_is_cached(XBZRLE.cache, current_addr)) {
> > +        if (!last_stage) {
> > +            cache_insert(XBZRLE.cache, current_addr, current_data);
> > +        }
> > +        acct_info.xbzrle_cache_miss++;
> > +        return -1;
> > +    }
> > +
> > +    prev_cached_page = get_cached_data(XBZRLE.cache, current_addr);
> > +
> > +    /* save current buffer into memory */
> > +    memcpy(XBZRLE.current_buf, current_data, TARGET_PAGE_SIZE);
> > +
> > +    /* XBZRLE encoding (if there is no overflow) */
> > +    encoded_len = xbzrle_encode_buffer(prev_cached_page, 
> > XBZRLE.current_buf,
> > +                                       TARGET_PAGE_SIZE, 
> > XBZRLE.encoded_buf,
> > +                                       TARGET_PAGE_SIZE);
> > +    if (encoded_len == 0) {
> > +        DPRINTF("Skipping unmodified page\n");
> > +        return 0;
> > +    } else if (encoded_len == -1) {
> > +        DPRINTF("Overflow\n");
> > +        acct_info.xbzrle_overflows++;
> > +        /* update data in the cache */
> > +        memcpy(prev_cached_page, current_data, TARGET_PAGE_SIZE);
> > +        return -1;
> > +    }
> > +
> > +    /* we need to update the data in the cache, in order to get the same 
> > data */
> > +    if (!last_stage) {
> > +        memcpy(prev_cached_page, XBZRLE.current_buf, TARGET_PAGE_SIZE);
> > +    }
> > +
> > +    /* Send XBZRLE based compressed page */
> > +    bytes_sent = save_block_hdr(f, block, offset, cont, 
> > RAM_SAVE_FLAG_XBZRLE);
> > +    qemu_put_byte(f, ENCODING_FLAG_XBZRLE);
> > +    qemu_put_be16(f, encoded_len);
> > +    qemu_put_buffer(f, XBZRLE.encoded_buf, encoded_len);
> > +    bytes_sent += encoded_len + 1 + 2;
> > +    acct_info.xbzrle_pages++;
> > +    acct_info.xbzrle_bytes += bytes_sent;
> > +
> > +    return bytes_sent;
> > +}
> > +
> > +
> > +/* This is the last block that we have visited serching for dirty pages
> > + */
> > +static RAMBlock *last_seen_block;
> > +/* This is the last block from where we have sent data */
> > +static RAMBlock *last_sent_block;
> > +static ram_addr_t last_offset;
> > +static unsigned long *migration_bitmap;
> > +static uint64_t migration_dirty_pages;
> > +static uint32_t last_version;
> > +
> > +static inline
> > +ram_addr_t migration_bitmap_find_and_reset_dirty(MemoryRegion *mr,
> > +                                                 ram_addr_t start)
> > +{
> > +    unsigned long base = mr->ram_addr >> TARGET_PAGE_BITS;
> > +    unsigned long nr = base + (start >> TARGET_PAGE_BITS);
> > +    unsigned long size = base + (int128_get64(mr->size) >> 
> > TARGET_PAGE_BITS);
> > +
> > +    unsigned long next = find_next_bit(migration_bitmap, size, nr);
> > +
> > +    if (next < size) {
> > +        clear_bit(next, migration_bitmap);
> > +        migration_dirty_pages--;
> > +    }
> > +    return (next - base) << TARGET_PAGE_BITS;
> > +}
> > +
> > +static inline bool migration_bitmap_set_dirty(MemoryRegion *mr,
> > +                                              ram_addr_t offset)
> > +{
> > +    bool ret;
> > +    int nr = (mr->ram_addr + offset) >> TARGET_PAGE_BITS;
> > +
> > +    ret = test_and_set_bit(nr, migration_bitmap);
> > +
> > +    if (!ret) {
> > +        migration_dirty_pages++;
> > +    }
> > +    return ret;
> > +}
> > +
> > +/* Needs iothread lock! */
> > +
> > +static void migration_bitmap_sync(void)
> > +{
> > +    RAMBlock *block;
> > +    ram_addr_t addr;
> > +    uint64_t num_dirty_pages_init = migration_dirty_pages;
> > +    MigrationState *s = migrate_get_current();
> > +    static int64_t start_time;
> > +    static int64_t num_dirty_pages_period;
> > +    int64_t end_time;
> > +
> > +    if (!start_time) {
> > +        start_time = qemu_get_clock_ms(rt_clock);
> > +    }
> > +
> > +    trace_migration_bitmap_sync_start();
> > +    memory_global_sync_dirty_bitmap(get_system_memory());
> > +
> > +    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
> > +        for (addr = 0; addr < block->length; addr += TARGET_PAGE_SIZE) {
> > +            if (memory_region_test_and_clear_dirty(block->mr,
> > +                                                   addr, TARGET_PAGE_SIZE,
> > +                                                   
> > DIRTY_MEMORY_MIGRATION)) {
> > +                migration_bitmap_set_dirty(block->mr, addr);
> > +            }
> > +        }
> > +    }
> > +    trace_migration_bitmap_sync_end(migration_dirty_pages
> > +                                    - num_dirty_pages_init);
> > +    num_dirty_pages_period += migration_dirty_pages - num_dirty_pages_init;
> > +    end_time = qemu_get_clock_ms(rt_clock);
> > +
> > +    /* more than 1 second = 1000 millisecons */
> > +    if (end_time > start_time + 1000) {
> > +        s->dirty_pages_rate = num_dirty_pages_period * 1000
> > +            / (end_time - start_time);
> > +        s->dirty_bytes_rate = s->dirty_pages_rate * TARGET_PAGE_SIZE;
> > +        start_time = end_time;
> > +        num_dirty_pages_period = 0;
> > +    }
> > +}
> > +
> > +/*
> > + * ram_save_block: Writes a page of memory to the stream f
> > + *
> > + * Returns:  The number of bytes written.
> > + *           0 means no dirty pages
> > + */
> > +
> > +static int ram_save_block(QEMUFile *f, bool last_stage)
> > +{
> > +    RAMBlock *block = last_seen_block;
> > +    ram_addr_t offset = last_offset;
> > +    bool complete_round = false;
> > +    int bytes_sent = 0;
> > +    MemoryRegion *mr;
> > +    ram_addr_t current_addr;
> > +
> > +    if (!block)
> > +        block = QTAILQ_FIRST(&ram_list.blocks);
> > +
> > +    while (true) {
> > +        mr = block->mr;
> > +        offset = migration_bitmap_find_and_reset_dirty(mr, offset);
> > +        if (complete_round && block == last_seen_block &&
> > +            offset >= last_offset) {
> > +            break;
> > +        }
> > +        if (offset >= block->length) {
> > +            offset = 0;
> > +            block = QTAILQ_NEXT(block, next);
> > +            if (!block) {
> > +                block = QTAILQ_FIRST(&ram_list.blocks);
> > +                complete_round = true;
> > +            }
> > +        } else {
> > +            uint8_t *p;
> > +            int cont = (block == last_sent_block) ?
> > +                RAM_SAVE_FLAG_CONTINUE : 0;
> > +
> > +            p = memory_region_get_ram_ptr(mr) + offset;
> > +
> > +            /* In doubt sent page as normal */
> > +            bytes_sent = -1;
> > +            if (is_dup_page(p)) {
> > +                acct_info.dup_pages++;
> > +                bytes_sent = save_block_hdr(f, block, offset, cont,
> > +                                            RAM_SAVE_FLAG_COMPRESS);
> > +                qemu_put_byte(f, *p);
> > +                bytes_sent += 1;
> > +            } else if (migrate_use_xbzrle()) {
> > +                current_addr = block->offset + offset;
> > +                bytes_sent = save_xbzrle_page(f, p, current_addr, block,
> > +                                              offset, cont, last_stage);
> > +                if (!last_stage) {
> > +                    p = get_cached_data(XBZRLE.cache, current_addr);
> > +                }
> > +            }
> > +
> > +            /* XBZRLE overflow or normal page */
> > +            if (bytes_sent == -1) {
> > +                bytes_sent = save_block_hdr(f, block, offset, cont, 
> > RAM_SAVE_FLAG_PAGE);
> > +                qemu_put_buffer(f, p, TARGET_PAGE_SIZE);
> > +                bytes_sent += TARGET_PAGE_SIZE;
> > +                acct_info.norm_pages++;
> > +            }
> > +
> > +            /* if page is unmodified, continue to the next */
> > +            if (bytes_sent > 0) {
> > +                last_sent_block = block;
> > +                break;
> > +            }
> > +        }
> > +    }
> > +    last_seen_block = block;
> > +    last_offset = offset;
> > +
> > +    return bytes_sent;
> > +}
> > +
> > +static uint64_t bytes_transferred;
> > +
> > +static ram_addr_t ram_save_remaining(void)
> > +{
> > +    return migration_dirty_pages;
> > +}
> > +
> > +uint64_t ram_bytes_remaining(void)
> > +{
> > +    return ram_save_remaining() * TARGET_PAGE_SIZE;
> > +}
> > +
> > +uint64_t ram_bytes_transferred(void)
> > +{
> > +    return bytes_transferred;
> > +}
> > +
> > +uint64_t ram_bytes_total(void)
> > +{
> > +    RAMBlock *block;
> > +    uint64_t total = 0;
> > +
> > +    QTAILQ_FOREACH(block, &ram_list.blocks, next)
> > +        total += block->length;
> > +
> > +    return total;
> > +}
> > +
> > +static void migration_end(void)
> > +{
> > +    if (migration_bitmap) {
> > +        memory_global_dirty_log_stop();
> > +        g_free(migration_bitmap);
> > +        migration_bitmap = NULL;
> > +    }
> > +
> > +    if (XBZRLE.cache) {
> > +        cache_fini(XBZRLE.cache);
> > +        g_free(XBZRLE.cache);
> > +        g_free(XBZRLE.encoded_buf);
> > +        g_free(XBZRLE.current_buf);
> > +        g_free(XBZRLE.decoded_buf);
> > +        XBZRLE.cache = NULL;
> > +    }
> > +}
> > +
> > +static void ram_migration_cancel(void *opaque)
> > +{
> > +    migration_end();
> > +}
> > +
> > +static void reset_ram_globals(void)
> > +{
> > +    last_seen_block = NULL;
> > +    last_sent_block = NULL;
> > +    last_offset = 0;
> > +    last_version = ram_list.version;
> > +}
> > +
> > +#define MAX_WAIT 50 /* ms, half buffered_file limit */
> > +
> > +static int ram_save_setup(QEMUFile *f, void *opaque)
> > +{
> > +    RAMBlock *block;
> > +    int64_t ram_pages = last_ram_offset() >> TARGET_PAGE_BITS;
> > +
> > +    migration_bitmap = bitmap_new(ram_pages);
> > +    bitmap_set(migration_bitmap, 0, ram_pages);
> > +    migration_dirty_pages = ram_pages;
> > +
> > +    if (migrate_use_xbzrle()) {
> > +        XBZRLE.cache = cache_init(migrate_xbzrle_cache_size() /
> > +                                  TARGET_PAGE_SIZE,
> > +                                  TARGET_PAGE_SIZE);
> > +        if (!XBZRLE.cache) {
> > +            DPRINTF("Error creating cache\n");
> > +            return -1;
> > +        }
> > +        XBZRLE.encoded_buf = g_malloc0(TARGET_PAGE_SIZE);
> > +        XBZRLE.current_buf = g_malloc(TARGET_PAGE_SIZE);
> > +        acct_clear();
> > +    }
> > +
> > +    qemu_mutex_lock_iothread();
> > +    qemu_mutex_lock_ramlist();
> > +    bytes_transferred = 0;
> > +    reset_ram_globals();
> > +
> > +    memory_global_dirty_log_start();
> > +    migration_bitmap_sync();
> > +    qemu_mutex_unlock_iothread();
> > +
> > +    qemu_put_be64(f, ram_bytes_total() | RAM_SAVE_FLAG_MEM_SIZE);
> > +
> > +    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
> > +        qemu_put_byte(f, strlen(block->idstr));
> > +        qemu_put_buffer(f, (uint8_t *)block->idstr, strlen(block->idstr));
> > +        qemu_put_be64(f, block->length);
> > +    }
> > +
> > +    qemu_mutex_unlock_ramlist();
> > +    qemu_put_be64(f, RAM_SAVE_FLAG_EOS);
> > +
> > +    return 0;
> > +}
> > +
> > +static int ram_save_iterate(QEMUFile *f, void *opaque)
> > +{
> > +    int ret;
> > +    int i;
> > +    int64_t t0;
> > +    int total_sent = 0;
> > +
> > +    qemu_mutex_lock_ramlist();
> > +
> > +    if (ram_list.version != last_version) {
> > +        reset_ram_globals();
> > +    }
> > +
> > +    t0 = qemu_get_clock_ns(rt_clock);
> > +    i = 0;
> > +    while ((ret = qemu_file_rate_limit(f)) == 0) {
> > +        int bytes_sent;
> > +
> > +        bytes_sent = ram_save_block(f, false);
> > +        /* no more blocks to sent */
> > +        if (bytes_sent == 0) {
> > +            break;
> > +        }
> > +        total_sent += bytes_sent;
> > +        acct_info.iterations++;
> > +        /* we want to check in the 1st loop, just in case it was the 1st 
> > time
> > +           and we had to sync the dirty bitmap.
> > +           qemu_get_clock_ns() is a bit expensive, so we only check each 
> > some
> > +           iterations
> > +        */
> > +        if ((i & 63) == 0) {
> > +            uint64_t t1 = (qemu_get_clock_ns(rt_clock) - t0) / 1000000;
> > +            if (t1 > MAX_WAIT) {
> > +                DPRINTF("big wait: %" PRIu64 " milliseconds, %d 
> > iterations\n",
> > +                        t1, i);
> > +                break;
> > +            }
> > +        }
> > +        i++;
> > +    }
> > +
> > +    qemu_mutex_unlock_ramlist();
> > +
> > +    if (ret < 0) {
> > +        bytes_transferred += total_sent;
> > +        return ret;
> > +    }
> > +
> > +    qemu_put_be64(f, RAM_SAVE_FLAG_EOS);
> > +    total_sent += 8;
> > +    bytes_transferred += total_sent;
> > +
> > +    return total_sent;
> > +}
> > +
> > +static int ram_save_complete(QEMUFile *f, void *opaque)
> > +{
> > +    qemu_mutex_lock_ramlist();
> > +    migration_bitmap_sync();
> > +
> > +    /* try transferring iterative blocks of memory */
> > +
> > +    /* flush all remaining blocks regardless of rate limiting */
> > +    while (true) {
> > +        int bytes_sent;
> > +
> > +        bytes_sent = ram_save_block(f, true);
> > +        /* no more blocks to sent */
> > +        if (bytes_sent == 0) {
> > +            break;
> > +        }
> > +        bytes_transferred += bytes_sent;
> > +    }
> > +    migration_end();
> > +
> > +    qemu_mutex_unlock_ramlist();
> > +    qemu_put_be64(f, RAM_SAVE_FLAG_EOS);
> > +
> > +    return 0;
> > +}
> > +
> > +static uint64_t ram_save_pending(QEMUFile *f, void *opaque, uint64_t 
> > max_size)
> > +{
> > +    uint64_t remaining_size;
> > +
> > +    remaining_size = ram_save_remaining() * TARGET_PAGE_SIZE;
> > +
> > +    if (remaining_size < max_size) {
> > +        qemu_mutex_lock_iothread();
> > +        migration_bitmap_sync();
> > +        qemu_mutex_unlock_iothread();
> > +        remaining_size = ram_save_remaining() * TARGET_PAGE_SIZE;
> > +    }
> > +    return remaining_size;
> > +}
> > +
> > +static int load_xbzrle(QEMUFile *f, ram_addr_t addr, void *host)
> > +{
> > +    int ret, rc = 0;
> > +    unsigned int xh_len;
> > +    int xh_flags;
> > +
> > +    if (!XBZRLE.decoded_buf) {
> > +        XBZRLE.decoded_buf = g_malloc(TARGET_PAGE_SIZE);
> > +    }
> > +
> > +    /* extract RLE header */
> > +    xh_flags = qemu_get_byte(f);
> > +    xh_len = qemu_get_be16(f);
> > +
> > +    if (xh_flags != ENCODING_FLAG_XBZRLE) {
> > +        fprintf(stderr, "Failed to load XBZRLE page - wrong 
> > compression!\n");
> > +        return -1;
> > +    }
> > +
> > +    if (xh_len > TARGET_PAGE_SIZE) {
> > +        fprintf(stderr, "Failed to load XBZRLE page - len overflow!\n");
> > +        return -1;
> > +    }
> > +    /* load data and decode */
> > +    qemu_get_buffer(f, XBZRLE.decoded_buf, xh_len);
> > +
> > +    /* decode RLE */
> > +    ret = xbzrle_decode_buffer(XBZRLE.decoded_buf, xh_len, host,
> > +                               TARGET_PAGE_SIZE);
> > +    if (ret == -1) {
> > +        fprintf(stderr, "Failed to load XBZRLE page - decode error!\n");
> > +        rc = -1;
> > +    } else  if (ret > TARGET_PAGE_SIZE) {
> > +        fprintf(stderr, "Failed to load XBZRLE page - size %d exceeds 
> > %d!\n",
> > +                ret, TARGET_PAGE_SIZE);
> > +        abort();
> > +    }
> > +
> > +    return rc;
> > +}
> > +
> > +static inline void *host_from_stream_offset(QEMUFile *f,
> > +                                            ram_addr_t offset,
> > +                                            int flags)
> > +{
> > +    static RAMBlock *block = NULL;
> > +    char id[256];
> > +    uint8_t len;
> > +
> > +    if (flags & RAM_SAVE_FLAG_CONTINUE) {
> > +        if (!block) {
> > +            fprintf(stderr, "Ack, bad migration stream!\n");
> > +            return NULL;
> > +        }
> > +
> > +        return memory_region_get_ram_ptr(block->mr) + offset;
> > +    }
> > +
> > +    len = qemu_get_byte(f);
> > +    qemu_get_buffer(f, (uint8_t *)id, len);
> > +    id[len] = 0;
> > +
> > +    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
> > +        if (!strncmp(id, block->idstr, sizeof(id)))
> > +            return memory_region_get_ram_ptr(block->mr) + offset;
> > +    }
> > +
> > +    fprintf(stderr, "Can't find block %s!\n", id);
> > +    return NULL;
> > +}
> > +
> > +static int ram_load(QEMUFile *f, void *opaque, int version_id)
> > +{
> > +    ram_addr_t addr;
> > +    int flags, ret = 0;
> > +    int error;
> > +    static uint64_t seq_iter;
> > +
> > +    seq_iter++;
> > +
> > +    if (version_id < 4 || version_id > 4) {
> > +        return -EINVAL;
> > +    }
> > +
> > +    do {
> > +        addr = qemu_get_be64(f);
> > +
> > +        flags = addr & ~TARGET_PAGE_MASK;
> > +        addr &= TARGET_PAGE_MASK;
> > +
> > +        if (flags & RAM_SAVE_FLAG_MEM_SIZE) {
> > +            if (version_id == 4) {
> > +                /* Synchronize RAM block list */
> > +                char id[256];
> > +                ram_addr_t length;
> > +                ram_addr_t total_ram_bytes = addr;
> > +
> > +                while (total_ram_bytes) {
> > +                    RAMBlock *block;
> > +                    uint8_t len;
> > +
> > +                    len = qemu_get_byte(f);
> > +                    qemu_get_buffer(f, (uint8_t *)id, len);
> > +                    id[len] = 0;
> > +                    length = qemu_get_be64(f);
> > +
> > +                    QTAILQ_FOREACH(block, &ram_list.blocks, next) {
> > +                        if (!strncmp(id, block->idstr, sizeof(id))) {
> > +                            if (block->length != length) {
> > +                                ret =  -EINVAL;
> > +                                goto done;
> > +                            }
> > +                            break;
> > +                        }
> > +                    }
> > +
> > +                    if (!block) {
> > +                        fprintf(stderr, "Unknown ramblock \"%s\", cannot "
> > +                                "accept migration\n", id);
> > +                        ret = -EINVAL;
> > +                        goto done;
> > +                    }
> > +
> > +                    total_ram_bytes -= length;
> > +                }
> > +            }
> > +        }
> > +
> > +        if (flags & RAM_SAVE_FLAG_COMPRESS) {
> > +            void *host;
> > +            uint8_t ch;
> > +
> > +            host = host_from_stream_offset(f, addr, flags);
> > +            if (!host) {
> > +                return -EINVAL;
> > +            }
> > +
> > +            ch = qemu_get_byte(f);
> > +            memset(host, ch, TARGET_PAGE_SIZE);
> > +#ifndef _WIN32
> > +            if (ch == 0 &&
> > +                (!kvm_enabled() || kvm_has_sync_mmu()) &&
> > +                getpagesize() <= TARGET_PAGE_SIZE) {
> > +                qemu_madvise(host, TARGET_PAGE_SIZE, QEMU_MADV_DONTNEED);
> > +            }
> > +#endif
> > +        } else if (flags & RAM_SAVE_FLAG_PAGE) {
> > +            void *host;
> > +
> > +            host = host_from_stream_offset(f, addr, flags);
> > +            if (!host) {
> > +                return -EINVAL;
> > +            }
> > +
> > +            qemu_get_buffer(f, host, TARGET_PAGE_SIZE);
> > +        } else if (flags & RAM_SAVE_FLAG_XBZRLE) {
> > +            void *host = host_from_stream_offset(f, addr, flags);
> > +            if (!host) {
> > +                return -EINVAL;
> > +            }
> > +
> > +            if (load_xbzrle(f, addr, host) < 0) {
> > +                ret = -EINVAL;
> > +                goto done;
> > +            }
> > +        }
> > +        error = qemu_file_get_error(f);
> > +        if (error) {
> > +            ret = error;
> > +            goto done;
> > +        }
> > +    } while (!(flags & RAM_SAVE_FLAG_EOS));
> > +
> > +done:
> > +    DPRINTF("Completed load of VM with exit code %d seq iteration "
> > +            "%" PRIu64 "\n", ret, seq_iter);
> > +    return ret;
> > +}
> > +
> > +SaveVMHandlers savevm_ram_handlers = {
> > +    .save_live_setup = ram_save_setup,
> > +    .save_live_iterate = ram_save_iterate,
> > +    .save_live_complete = ram_save_complete,
> > +    .save_live_pending = ram_save_pending,
> > +    .load_state = ram_load,
> > +    .cancel = ram_migration_cancel,
> > +};
> > 



reply via email to

[Prev in Thread] Current Thread [Next in Thread]