qemu-ppc
[Top][All Lists]
Advanced

[Date Prev][Date Next][Thread Prev][Thread Next][Date Index][Thread Index]

Re: [Qemu-ppc] [Qemu-devel] [PULL 075/118] macio: handle non-block ATAPI


From: Kevin Wolf
Subject: Re: [Qemu-ppc] [Qemu-devel] [PULL 075/118] macio: handle non-block ATAPI DMA transfers
Date: Tue, 24 Jun 2014 14:07:03 +0200
User-agent: Mutt/1.5.21 (2010-09-15)

Am 24.06.2014 um 13:27 hat Alexander Graf geschrieben:
> 
> On 24.06.14 13:22, Kevin Wolf wrote:
> >Am 24.06.2014 um 13:02 hat Alexander Graf geschrieben:
> >>The way DBDMA works is that you put in something similar to a
> >>scatter-gather list: A list of chunks to read / write and where in
> >>memory those chunks live. DBDMA then goes over its list and does the
> >>pokes. So for example if the list is
> >>
> >>   [ memaddr = 0x12000 | len = 500 ]
> >>   [ memaddr = 0x13098 | len = 12 ]
> >>
> >>then it reads 500 bytes from IDE, writes them at memory offset
> >>0x12000 and after that reads another 12 bytes from IDE and puts them
> >>at memory offset 0x13098.
> >>
> >>The reason we have such complicated code for real DMA is that we
> >>can't model this easily with our direct block-to-memory API. That
> >>one can only work on a 512 byte granularity. So when we see
> >>unaligned accesses like above, we have to split them out and handle
> >>them lazily.
> >Wait... What kind of granularity are you talking about?
> >
> >We do need disk accesses with a 512 byte granularity, because the API
> >takes a sector number. This is also what real IDE disks do, they don't
> >provide byte access.
> >
> >However, for the memory, I can't see why you couldn't pass a s/g list
> >like what you wrote above to the DMA functions. This is not unusual at
> >all and is the same as ide/pci.c does. There is no 512-byte alignment
> >needed for the individual s/g list entries, only the total size should
> >obviously be a multiple of 512 in the general case (otherwise the list
> >would be too short or too long for the request).
> >
> >If this is really what we're talking about, then I think your problem is
> >just that you try to handle the 500 byte and the 12 byte as individual
> >requests instead of building up the s/g list and then sending a single
> >request.
> 
> The 500 and 12 byte requests can come in as separate requests that
> require previous requests to have finished. What Mac OS X does for
> example is
> 
>   [ memaddr = 0x2000 | len = 1024 ]
>   [ memaddr = 0x1000 | len = 510 ]
> 
> <wait for ack>
> 
>   [ memaddr = 0x10fe | len = 2 ]
>   [ memaddr = 0x3000 | len = 2048 ]
> 
> If it was as simple as creating a working sglist, I would've
> certainly done so long ago :).

Thanks, that's the explanation that was missing for me (I'm sure you
explained it more than once to me in the past few years, but I keep
forgetting).

This means, however, that exposing the byte access in the block layer is
probably not what you want. Otherwise you would read the same sector
twice from the image (assuming cache=none, so the backend must have
512-byte alignment). If you do the handling in the device emulation you
can read the full request once and then only do the DMA part with a byte
granularity. I suppose this is the complicated code that you have today?

Kevin



reply via email to

[Prev in Thread] Current Thread [Next in Thread]