qemu-devel
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Qemu-devel] [PATCH v6 2/2] block/vxhs.c: Add qemu-iotests for new b


From: Ketan Nilangekar
Subject: Re: [Qemu-devel] [PATCH v6 2/2] block/vxhs.c: Add qemu-iotests for new block device type "vxhs"
Date: Tue, 14 Feb 2017 03:02:00 +0000
User-agent: Microsoft-MacOutlook/f.1c.1.161117


On 2/13/17, 3:23 PM, "Jeff Cody" <address@hidden> wrote:

    On Mon, Feb 13, 2017 at 10:36:53PM +0000, Ketan Nilangekar wrote:
    > 
    > 
    > On 2/13/17, 8:32 AM, "Jeff Cody" <address@hidden> wrote:
    > 
    >     On Mon, Feb 13, 2017 at 01:37:25PM +0000, Stefan Hajnoczi wrote:
    >     > On Tue, Feb 07, 2017 at 03:12:36PM -0800, ashish mittal wrote:
    >     > > On Tue, Nov 8, 2016 at 12:44 PM, Jeff Cody <address@hidden> wrote:
    >     > > > On Mon, Nov 07, 2016 at 04:59:45PM -0800, Ashish Mittal wrote:
    >     > > >> These changes use a vxhs test server that is a part of the 
following
    >     > > >> repository:
    >     > > >> https://github.com/MittalAshish/libqnio.git
    >     > > >>
    >     > > >> Signed-off-by: Ashish Mittal <address@hidden>
    >     > > >> ---
    >     > > >> v6 changelog:
    >     > > >> (1) Added iotests for VxHS block device.
    >     > > >>
    >     > > >>  tests/qemu-iotests/common        |  6 ++++++
    >     > > >>  tests/qemu-iotests/common.config | 13 +++++++++++++
    >     > > >>  tests/qemu-iotests/common.filter |  1 +
    >     > > >>  tests/qemu-iotests/common.rc     | 19 +++++++++++++++++++
    >     > > >>  4 files changed, 39 insertions(+)
    >     > > >>
    >     > > >> diff --git a/tests/qemu-iotests/common 
b/tests/qemu-iotests/common
    >     > > >> index d60ea2c..41430d8 100644
    >     > > >> --- a/tests/qemu-iotests/common
    >     > > >> +++ b/tests/qemu-iotests/common
    >     > > >
    >     > > > When using raw format, I was able to run the test successfully 
for all
    >     > > > supported test cases (26 of them).
    >     > > >
    >     > > > With qcow2, they fail - but not the fault of this patch, I 
think; but
    >     > > > rather, the fault of the test server.  Can qnio_server be 
modified so that
    >     > > > it does not work on just raw files?
    >     > > >
    >     > > >
    >     > > 
    >     > > VxHS supports and uses only the raw format.
    >     > 
    >     > That's like saying only ext4 guest file systems are supported on 
VxHS
    >     > and not ZFS.  The VxHS driver should not care what file system is 
used,
    >     > it just does block I/O without interpreting the data.
    >     > 
    >     > It must be possible to use any format on top of the VxHS protocol.
    >     > After all, the image format drivers just do block I/O.  If there is 
a
    >     > case where qcow2 on VxHS fails then it needs to be investigated.
    >     > 
    >     > The VxHS driver can't be merged until we at least understand the 
cause
    >     > of the qcow2 test failures.
    >     >
    >     
    >     A quick run with the test server and a QEMU process showed an abort() 
in the
    >     test server, so I just sent a pull req to libqnio to fix that.  
    >     
    >     But playing with it in gdb right now with a test qcow2 file, I see 
that we
    >     are waiting in aio_poll() forever for the test server to respond to a 
read
    >     request, when using qcow2 format.  
    >     
    >     As Stefan said, this doesn't really make any sense - why would VXHS 
behave
    >     differently based on the file contents?
    > 
    > [Ketan] To read/write a qcow2 backed device VxHS server implementation
    > will need to understand the qcow2 format. This is not just block IO but
    > actually does involve interpreting the qcow2 header and cluster formats.
    > Clearly the test server implementation does not handle it as it was never
    > intended to. VxHS backend won't handle it either because VxHS virtual
    > disks are written as non-sparse files.  There are space savings with the
    > qcow2 format but performance penalties as well because of the metadata
    > overhead. As a block storage provider, VxHS does not support sparse file
    > formats like qcow2 primarily because of performance reasons.  Implementing
    > a qcow2 backend in the test server would be a non-trivial and non-useful
    > exercise since the VxHS server won't support it.
    >
    
    What?  Why would the backend need to know anything about qcow2 formats; are
    you manipulating the guest image data directly yourself?  But regardless,
    since the test server is naive and just reads and writes data, the fact that
    the test server breaks on qcow2 image formats means that the test server is
    broken for raw images, as well [1].
    
    The backend should not need to know anything about the image file format in
    order to serve data, as the backend is essentially serving bytes. (I guess
    the only thing the backend would need to be aware of is how to invoke
    qemu-img to create a qcow2 image file initially).
    
    
    QEMU is what interprets the qcow2 format (or any image format).  Every read
    from QEMU will look like a raw file read from the perspective of libqnio /
    vxhs.  I can't see any reason why vxhs would need to know anything about the
    contents of the image file itself.
    
    The only thing the test server needs to do, in order to be able to serve
    qcow2 files correctly, is to be able to serve raw files correctly.
    
    Looking at the libqnio implementation from the test server with gdb, I think
    that the reason why qcow2 format does not work is that the current
    libqnio implementation does not handle short reads correctly.  For instance,
    if the file is not a even multiple of the read size, and we try to read 512
    bytes at the end of the file, it appears libqnio hangs on the short read.  I
    am not sure if this is a bug exclusive to the test server, or in the libqnio
    implementation.
    
    [1]
    This bug isn't limited to qcow2, since as mentioned above, QEMU is just
    relying on libqnio to read raw bytes.  So, you can reproduce this bug with a
    raw image file, as well:
    
    # dd if=/dev/zero of=test.raw bs=1 count=196616
    
    # qemu-io -c "read 196608 512" vxhs://localhost:9999/test.raw
    
    The above qemu-io process will hang in aio_poll, because libqnio never
    processes the callback. This is because process_incoming_messages() thinks
    it needs a full 512 bytes before it will call the registered callback, even
    though there will only ever be 8 bytes read.
    
    The above scenario is also exactly what is happening when I try to use a
    qcow2 format; QEMU issues a read for 8 bytes at offset 196608, and the file
    is 196616 bytes long.
    
    (The sparse image thing seems shouldn't be an issue - first, the test server
    works just fine with sparse raw images... as matter of fact, the
    "create_vdisk.sh" script in the libqnio repository does exactly that.  Also,
    if you choose to not use sparse images for whatever reason on the backend,
    qcow2 does not require it to be sparse, as image files can be created with
    the "-o preallocation=full" option).
    
[Ketan] You are correct. We have identified the problem with test server and 
have a patch for it which returns zero filled reads to align the read buffers 
to requested block size. We could also fix this in libvxhs by not 
expecting/waiting for requested block size in case the reads are partial.
Ashish will upload the patch for the test server fix. Please let us know how 
the qcow tests go.

    -Jeff
    


reply via email to

[Prev in Thread] Current Thread [Next in Thread]