|
From: | Peter Lieven |
Subject: | Re: [Qemu-devel] [RFC PATCH] qemu-io: add drain/undrain cmd |
Date: | Mon, 15 May 2017 16:02:05 +0200 |
User-agent: | Mozilla/5.0 (X11; Linux x86_64; rv:45.0) Gecko/20100101 Thunderbird/45.5.1 |
Am 15.05.2017 um 15:35 schrieb Fam Zheng:
On Mon, 05/15 15:01, Peter Lieven wrote:Am 15.05.2017 um 14:52 schrieb Fam Zheng:On Mon, 05/15 14:32, Peter Lieven wrote:Am 15.05.2017 um 14:28 schrieb Fam Zheng:On Mon, 05/15 13:58, Peter Lieven wrote:Am 15.05.2017 um 13:53 schrieb Fam Zheng:On Mon, 05/15 13:26, Peter Lieven wrote:Am 15.05.2017 um 12:50 schrieb Fam Zheng:On Mon, 05/15 12:02, Peter Lieven wrote:Hi Block developers, I would like to add a feature to Qemu to drain all traffic from a block so that I can take external snaphosts without the risk to that in the middle of a write operation. Its meant for cases where where QGA freeze/thaw is not available. For me its enough to have this through qemu-io, but Kevin asked me to check if its not worth to have a stable API for it and present it via QMP/HMP. What are your thoughts?For debugging purpose or a "hacky" usage where you know what you are doing, it may be fine to have this. The only issue is it should be a separate flag, like BlockJob.user_paused.How can I add, remove such a flag?Like bs->user_drained. Set it in "drain" command, then increment bs->quiesce_counter if toggled; vise versa.Ah okay. You wouldn't use bdrv_drained_begin/end? Because in these functions the counter is incremented already.You're right, calling bdrv_drained_begin() is better.What happens from guest perspective? In the case of virtio, the request queue is not handled and -ETIMEDOUT may happen. With IDE, I/O commands are still handled, the command is not effective (or rather the implementation is not complete).That it only works with virtio is fine. However, the thing it does not work correctly apply then also to all other users of the drained_begin/end functions, right? As for the timeout I only plan to drain the device for about 1 second.It didn't matter because for IDE, the invariant (staying quiesced as long as necessary) is already ensured by BQL. Virtio is different because it supports ioeventfd and data plane.Okay understood. So my use of bdrv_drained_begin/end is more an abuse of these functions?Sort of. But it's not unreasonable to "extend" bdrv_drained_begin/end to cover IDE, I just haven't thought about "how".Do you have another idea how to achieve what I want? I was thinking of throttle the I/O to zero. It would be enough to do this for writes, reading doesn't hurt in my case.Maybe add a block filter on top of the drained node, drain it when doing so, then queue all further requests with a CoQueue until "undrain". (It is then not quite to "drain" but to "halt" or "pause", though.)To get the drain for free was why I was looking at this approach. If I read correctly if I keep using bdrv_drained_begin/end its too hacky to implement it in QMP?I think so.If yes, would support adding it to qemu-io?I'm under the impression that you are looking to a real use case, I don't think I like the idea. Also, accessing the image from other processes while QEMU is using it is strongly discouraged, and there is the coming image locking mechanism to prevent this from happening. Why is the blockdev-snapshot command not enough?blockdev-snapshot is enough, but I still fear the case there is suddenly too much I/O for the live-commit. And that the whole snapshot / commit code is more senstive than just stopping I/O for a second or two.In this case, the image fleecing approach may be what you need. It creates a temporary point in time snapshot which is lightweight and disposable. Something like: https://lists.gnu.org/archive/html/qemu-devel/2014-03/msg01359.html (Ccing John who may have more up-to-date pointers)
We discussed about it a few months ago. But maybe he has an update.
do you have a pointer to the image locking mechanism?It hit qemu.git master just a moment ago. See raw_check_perm.
which master are you looking at? $ git fetch upstream $ git log upstream/master --oneline 3a87606 Merge tag 'tracing-pull-request' into staging b54933e Merge tag 'block-pull-request' into staging 3753e25 Merge remote-tracking branch 'kwolf/tags/for-upstream' into staging 5651743 trace: add sanity check 321d1db aio: add missing aio_notify() to aio_enable_external() ee29d6a block: Simplify BDRV_BLOCK_RAW recursion 33c53c5 coroutine: remove GThread implementation ecc1f5a maintainers: Add myself as linux-user reviewer d541e20 Merge remote-tracking branch 'mreitz/tags/pull-block-2017-05-11' into queue-block 8dd30c8 MAINTAINERS: Add qemu-progress to the block layer d2cb36a qcow2: Discard/zero clusters by byte count f10ee13 qcow2: Assert that cluster operations are aligned fbaa6bb qcow2: Optimize write zero of unaligned tail cluster e249d51 iotests: Add test 179 to cover write zeroes with unmap d9ca221 iotests: Improve _filter_qemu_img_map 06cc5e2 qcow2: Optimize zero_single_l2() to minimize L2 churn fdfab37 qcow2: Make distinction between zero cluster types obvious 3ef9521 qcow2: Name typedef for cluster type 4341df8 qcow2: Correctly report status of preallocated zero clusters 4c41cb4 block: Update comments on BDRV_BLOCK_* meanings Thanks, Peter
[Prev in Thread] | Current Thread | [Next in Thread] |