mirror of https://github.com/xemu-project/xemu.git
jobs: use job locks also in the unit tests
Add missing job synchronization in the unit tests, with explicit locks. We are deliberately using _locked functions wrapped by a guard instead of a normal call because the normal call will be removed in future, as the only usage is limited to the tests. In other words, if a function like job_pause() is/will be only used in tests to avoid: WITH_JOB_LOCK_GUARD(){ job_pause_locked(); } then it is not worth keeping job_pause(), and just use the guard. Note: at this stage, job_{lock/unlock} and job lock guard macros are *nop*. Signed-off-by: Emanuele Giuseppe Esposito <eesposit@redhat.com> Reviewed-by: Vladimir Sementsov-Ogievskiy <vsementsov@yandex-team.ru> Reviewed-by: Kevin Wolf <kwolf@redhat.com> Message-Id: <20220926093214.506243-10-eesposit@redhat.com> Signed-off-by: Kevin Wolf <kwolf@redhat.com>
This commit is contained in:
parent
9624112441
commit
191e7af394
|
@ -943,13 +943,16 @@ static void test_blockjob_common_drain_node(enum drain_type drain_type,
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
g_assert_cmpint(job->job.pause_count, ==, 0);
|
g_assert_cmpint(job->job.pause_count, ==, 0);
|
||||||
g_assert_false(job->job.paused);
|
g_assert_false(job->job.paused);
|
||||||
g_assert_true(tjob->running);
|
g_assert_true(tjob->running);
|
||||||
g_assert_true(job->job.busy); /* We're in qemu_co_sleep_ns() */
|
g_assert_true(job->job.busy); /* We're in qemu_co_sleep_ns() */
|
||||||
|
}
|
||||||
|
|
||||||
do_drain_begin_unlocked(drain_type, drain_bs);
|
do_drain_begin_unlocked(drain_type, drain_bs);
|
||||||
|
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
if (drain_type == BDRV_DRAIN_ALL) {
|
if (drain_type == BDRV_DRAIN_ALL) {
|
||||||
/* bdrv_drain_all() drains both src and target */
|
/* bdrv_drain_all() drains both src and target */
|
||||||
g_assert_cmpint(job->job.pause_count, ==, 2);
|
g_assert_cmpint(job->job.pause_count, ==, 2);
|
||||||
|
@ -958,22 +961,31 @@ static void test_blockjob_common_drain_node(enum drain_type drain_type,
|
||||||
}
|
}
|
||||||
g_assert_true(job->job.paused);
|
g_assert_true(job->job.paused);
|
||||||
g_assert_false(job->job.busy); /* The job is paused */
|
g_assert_false(job->job.busy); /* The job is paused */
|
||||||
|
}
|
||||||
|
|
||||||
do_drain_end_unlocked(drain_type, drain_bs);
|
do_drain_end_unlocked(drain_type, drain_bs);
|
||||||
|
|
||||||
if (use_iothread) {
|
if (use_iothread) {
|
||||||
/* paused is reset in the I/O thread, wait for it */
|
/*
|
||||||
|
* Here we are waiting for the paused status to change,
|
||||||
|
* so don't bother protecting the read every time.
|
||||||
|
*
|
||||||
|
* paused is reset in the I/O thread, wait for it
|
||||||
|
*/
|
||||||
while (job->job.paused) {
|
while (job->job.paused) {
|
||||||
aio_poll(qemu_get_aio_context(), false);
|
aio_poll(qemu_get_aio_context(), false);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
g_assert_cmpint(job->job.pause_count, ==, 0);
|
g_assert_cmpint(job->job.pause_count, ==, 0);
|
||||||
g_assert_false(job->job.paused);
|
g_assert_false(job->job.paused);
|
||||||
g_assert_true(job->job.busy); /* We're in qemu_co_sleep_ns() */
|
g_assert_true(job->job.busy); /* We're in qemu_co_sleep_ns() */
|
||||||
|
}
|
||||||
|
|
||||||
do_drain_begin_unlocked(drain_type, target);
|
do_drain_begin_unlocked(drain_type, target);
|
||||||
|
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
if (drain_type == BDRV_DRAIN_ALL) {
|
if (drain_type == BDRV_DRAIN_ALL) {
|
||||||
/* bdrv_drain_all() drains both src and target */
|
/* bdrv_drain_all() drains both src and target */
|
||||||
g_assert_cmpint(job->job.pause_count, ==, 2);
|
g_assert_cmpint(job->job.pause_count, ==, 2);
|
||||||
|
@ -982,22 +994,32 @@ static void test_blockjob_common_drain_node(enum drain_type drain_type,
|
||||||
}
|
}
|
||||||
g_assert_true(job->job.paused);
|
g_assert_true(job->job.paused);
|
||||||
g_assert_false(job->job.busy); /* The job is paused */
|
g_assert_false(job->job.busy); /* The job is paused */
|
||||||
|
}
|
||||||
|
|
||||||
do_drain_end_unlocked(drain_type, target);
|
do_drain_end_unlocked(drain_type, target);
|
||||||
|
|
||||||
if (use_iothread) {
|
if (use_iothread) {
|
||||||
/* paused is reset in the I/O thread, wait for it */
|
/*
|
||||||
|
* Here we are waiting for the paused status to change,
|
||||||
|
* so don't bother protecting the read every time.
|
||||||
|
*
|
||||||
|
* paused is reset in the I/O thread, wait for it
|
||||||
|
*/
|
||||||
while (job->job.paused) {
|
while (job->job.paused) {
|
||||||
aio_poll(qemu_get_aio_context(), false);
|
aio_poll(qemu_get_aio_context(), false);
|
||||||
}
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
g_assert_cmpint(job->job.pause_count, ==, 0);
|
g_assert_cmpint(job->job.pause_count, ==, 0);
|
||||||
g_assert_false(job->job.paused);
|
g_assert_false(job->job.paused);
|
||||||
g_assert_true(job->job.busy); /* We're in qemu_co_sleep_ns() */
|
g_assert_true(job->job.busy); /* We're in qemu_co_sleep_ns() */
|
||||||
|
}
|
||||||
|
|
||||||
aio_context_acquire(ctx);
|
aio_context_acquire(ctx);
|
||||||
ret = job_complete_sync(&job->job, &error_abort);
|
WITH_JOB_LOCK_GUARD() {
|
||||||
|
ret = job_complete_sync_locked(&job->job, &error_abort);
|
||||||
|
}
|
||||||
g_assert_cmpint(ret, ==, (result == TEST_JOB_SUCCESS ? 0 : -EIO));
|
g_assert_cmpint(ret, ==, (result == TEST_JOB_SUCCESS ? 0 : -EIO));
|
||||||
|
|
||||||
if (use_iothread) {
|
if (use_iothread) {
|
||||||
|
|
|
@ -583,7 +583,9 @@ static void test_attach_blockjob(void)
|
||||||
}
|
}
|
||||||
|
|
||||||
aio_context_acquire(ctx);
|
aio_context_acquire(ctx);
|
||||||
job_complete_sync(&tjob->common.job, &error_abort);
|
WITH_JOB_LOCK_GUARD() {
|
||||||
|
job_complete_sync_locked(&tjob->common.job, &error_abort);
|
||||||
|
}
|
||||||
blk_set_aio_context(blk, qemu_get_aio_context(), &error_abort);
|
blk_set_aio_context(blk, qemu_get_aio_context(), &error_abort);
|
||||||
aio_context_release(ctx);
|
aio_context_release(ctx);
|
||||||
|
|
||||||
|
@ -757,7 +759,9 @@ static void test_propagate_mirror(void)
|
||||||
BLOCKDEV_ON_ERROR_REPORT, BLOCKDEV_ON_ERROR_REPORT,
|
BLOCKDEV_ON_ERROR_REPORT, BLOCKDEV_ON_ERROR_REPORT,
|
||||||
false, "filter_node", MIRROR_COPY_MODE_BACKGROUND,
|
false, "filter_node", MIRROR_COPY_MODE_BACKGROUND,
|
||||||
&error_abort);
|
&error_abort);
|
||||||
job = job_get("job0");
|
WITH_JOB_LOCK_GUARD() {
|
||||||
|
job = job_get_locked("job0");
|
||||||
|
}
|
||||||
filter = bdrv_find_node("filter_node");
|
filter = bdrv_find_node("filter_node");
|
||||||
|
|
||||||
/* Change the AioContext of src */
|
/* Change the AioContext of src */
|
||||||
|
|
|
@ -116,8 +116,10 @@ static void test_single_job(int expected)
|
||||||
job = test_block_job_start(1, true, expected, &result, txn);
|
job = test_block_job_start(1, true, expected, &result, txn);
|
||||||
job_start(&job->job);
|
job_start(&job->job);
|
||||||
|
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
if (expected == -ECANCELED) {
|
if (expected == -ECANCELED) {
|
||||||
job_cancel(&job->job, false);
|
job_cancel_locked(&job->job, false);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
while (result == -EINPROGRESS) {
|
while (result == -EINPROGRESS) {
|
||||||
|
@ -160,13 +162,15 @@ static void test_pair_jobs(int expected1, int expected2)
|
||||||
/* Release our reference now to trigger as many nice
|
/* Release our reference now to trigger as many nice
|
||||||
* use-after-free bugs as possible.
|
* use-after-free bugs as possible.
|
||||||
*/
|
*/
|
||||||
job_txn_unref(txn);
|
WITH_JOB_LOCK_GUARD() {
|
||||||
|
job_txn_unref_locked(txn);
|
||||||
|
|
||||||
if (expected1 == -ECANCELED) {
|
if (expected1 == -ECANCELED) {
|
||||||
job_cancel(&job1->job, false);
|
job_cancel_locked(&job1->job, false);
|
||||||
}
|
}
|
||||||
if (expected2 == -ECANCELED) {
|
if (expected2 == -ECANCELED) {
|
||||||
job_cancel(&job2->job, false);
|
job_cancel_locked(&job2->job, false);
|
||||||
|
}
|
||||||
}
|
}
|
||||||
|
|
||||||
while (result1 == -EINPROGRESS || result2 == -EINPROGRESS) {
|
while (result1 == -EINPROGRESS || result2 == -EINPROGRESS) {
|
||||||
|
@ -219,7 +223,9 @@ static void test_pair_jobs_fail_cancel_race(void)
|
||||||
job_start(&job1->job);
|
job_start(&job1->job);
|
||||||
job_start(&job2->job);
|
job_start(&job2->job);
|
||||||
|
|
||||||
job_cancel(&job1->job, false);
|
WITH_JOB_LOCK_GUARD() {
|
||||||
|
job_cancel_locked(&job1->job, false);
|
||||||
|
}
|
||||||
|
|
||||||
/* Now make job2 finish before the main loop kicks jobs. This simulates
|
/* Now make job2 finish before the main loop kicks jobs. This simulates
|
||||||
* the race between a pending kick and another job completing.
|
* the race between a pending kick and another job completing.
|
||||||
|
|
|
@ -211,8 +211,11 @@ static CancelJob *create_common(Job **pjob)
|
||||||
bjob = mk_job(blk, "Steve", &test_cancel_driver, true,
|
bjob = mk_job(blk, "Steve", &test_cancel_driver, true,
|
||||||
JOB_MANUAL_FINALIZE | JOB_MANUAL_DISMISS);
|
JOB_MANUAL_FINALIZE | JOB_MANUAL_DISMISS);
|
||||||
job = &bjob->job;
|
job = &bjob->job;
|
||||||
job_ref(job);
|
WITH_JOB_LOCK_GUARD() {
|
||||||
|
job_ref_locked(job);
|
||||||
assert(job->status == JOB_STATUS_CREATED);
|
assert(job->status == JOB_STATUS_CREATED);
|
||||||
|
}
|
||||||
|
|
||||||
s = container_of(bjob, CancelJob, common);
|
s = container_of(bjob, CancelJob, common);
|
||||||
s->blk = blk;
|
s->blk = blk;
|
||||||
|
|
||||||
|
@ -231,12 +234,14 @@ static void cancel_common(CancelJob *s)
|
||||||
aio_context_acquire(ctx);
|
aio_context_acquire(ctx);
|
||||||
|
|
||||||
job_cancel_sync(&job->job, true);
|
job_cancel_sync(&job->job, true);
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
if (sts != JOB_STATUS_CREATED && sts != JOB_STATUS_CONCLUDED) {
|
if (sts != JOB_STATUS_CREATED && sts != JOB_STATUS_CONCLUDED) {
|
||||||
Job *dummy = &job->job;
|
Job *dummy = &job->job;
|
||||||
job_dismiss(&dummy, &error_abort);
|
job_dismiss_locked(&dummy, &error_abort);
|
||||||
}
|
}
|
||||||
assert(job->job.status == JOB_STATUS_NULL);
|
assert(job->job.status == JOB_STATUS_NULL);
|
||||||
job_unref(&job->job);
|
job_unref_locked(&job->job);
|
||||||
|
}
|
||||||
destroy_blk(blk);
|
destroy_blk(blk);
|
||||||
|
|
||||||
aio_context_release(ctx);
|
aio_context_release(ctx);
|
||||||
|
@ -251,6 +256,13 @@ static void test_cancel_created(void)
|
||||||
cancel_common(s);
|
cancel_common(s);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void assert_job_status_is(Job *job, int status)
|
||||||
|
{
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
|
assert(job->status == status);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
|
||||||
static void test_cancel_running(void)
|
static void test_cancel_running(void)
|
||||||
{
|
{
|
||||||
Job *job;
|
Job *job;
|
||||||
|
@ -259,7 +271,7 @@ static void test_cancel_running(void)
|
||||||
s = create_common(&job);
|
s = create_common(&job);
|
||||||
|
|
||||||
job_start(job);
|
job_start(job);
|
||||||
assert(job->status == JOB_STATUS_RUNNING);
|
assert_job_status_is(job, JOB_STATUS_RUNNING);
|
||||||
|
|
||||||
cancel_common(s);
|
cancel_common(s);
|
||||||
}
|
}
|
||||||
|
@ -272,11 +284,12 @@ static void test_cancel_paused(void)
|
||||||
s = create_common(&job);
|
s = create_common(&job);
|
||||||
|
|
||||||
job_start(job);
|
job_start(job);
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
assert(job->status == JOB_STATUS_RUNNING);
|
assert(job->status == JOB_STATUS_RUNNING);
|
||||||
|
job_user_pause_locked(job, &error_abort);
|
||||||
job_user_pause(job, &error_abort);
|
}
|
||||||
job_enter(job);
|
job_enter(job);
|
||||||
assert(job->status == JOB_STATUS_PAUSED);
|
assert_job_status_is(job, JOB_STATUS_PAUSED);
|
||||||
|
|
||||||
cancel_common(s);
|
cancel_common(s);
|
||||||
}
|
}
|
||||||
|
@ -289,11 +302,11 @@ static void test_cancel_ready(void)
|
||||||
s = create_common(&job);
|
s = create_common(&job);
|
||||||
|
|
||||||
job_start(job);
|
job_start(job);
|
||||||
assert(job->status == JOB_STATUS_RUNNING);
|
assert_job_status_is(job, JOB_STATUS_RUNNING);
|
||||||
|
|
||||||
s->should_converge = true;
|
s->should_converge = true;
|
||||||
job_enter(job);
|
job_enter(job);
|
||||||
assert(job->status == JOB_STATUS_READY);
|
assert_job_status_is(job, JOB_STATUS_READY);
|
||||||
|
|
||||||
cancel_common(s);
|
cancel_common(s);
|
||||||
}
|
}
|
||||||
|
@ -306,15 +319,16 @@ static void test_cancel_standby(void)
|
||||||
s = create_common(&job);
|
s = create_common(&job);
|
||||||
|
|
||||||
job_start(job);
|
job_start(job);
|
||||||
assert(job->status == JOB_STATUS_RUNNING);
|
assert_job_status_is(job, JOB_STATUS_RUNNING);
|
||||||
|
|
||||||
s->should_converge = true;
|
s->should_converge = true;
|
||||||
job_enter(job);
|
job_enter(job);
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
assert(job->status == JOB_STATUS_READY);
|
assert(job->status == JOB_STATUS_READY);
|
||||||
|
job_user_pause_locked(job, &error_abort);
|
||||||
job_user_pause(job, &error_abort);
|
}
|
||||||
job_enter(job);
|
job_enter(job);
|
||||||
assert(job->status == JOB_STATUS_STANDBY);
|
assert_job_status_is(job, JOB_STATUS_STANDBY);
|
||||||
|
|
||||||
cancel_common(s);
|
cancel_common(s);
|
||||||
}
|
}
|
||||||
|
@ -327,20 +341,21 @@ static void test_cancel_pending(void)
|
||||||
s = create_common(&job);
|
s = create_common(&job);
|
||||||
|
|
||||||
job_start(job);
|
job_start(job);
|
||||||
assert(job->status == JOB_STATUS_RUNNING);
|
assert_job_status_is(job, JOB_STATUS_RUNNING);
|
||||||
|
|
||||||
s->should_converge = true;
|
s->should_converge = true;
|
||||||
job_enter(job);
|
job_enter(job);
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
assert(job->status == JOB_STATUS_READY);
|
assert(job->status == JOB_STATUS_READY);
|
||||||
|
job_complete_locked(job, &error_abort);
|
||||||
job_complete(job, &error_abort);
|
}
|
||||||
job_enter(job);
|
job_enter(job);
|
||||||
while (!job->deferred_to_main_loop) {
|
while (!job->deferred_to_main_loop) {
|
||||||
aio_poll(qemu_get_aio_context(), true);
|
aio_poll(qemu_get_aio_context(), true);
|
||||||
}
|
}
|
||||||
assert(job->status == JOB_STATUS_READY);
|
assert_job_status_is(job, JOB_STATUS_READY);
|
||||||
aio_poll(qemu_get_aio_context(), true);
|
aio_poll(qemu_get_aio_context(), true);
|
||||||
assert(job->status == JOB_STATUS_PENDING);
|
assert_job_status_is(job, JOB_STATUS_PENDING);
|
||||||
|
|
||||||
cancel_common(s);
|
cancel_common(s);
|
||||||
}
|
}
|
||||||
|
@ -353,25 +368,28 @@ static void test_cancel_concluded(void)
|
||||||
s = create_common(&job);
|
s = create_common(&job);
|
||||||
|
|
||||||
job_start(job);
|
job_start(job);
|
||||||
assert(job->status == JOB_STATUS_RUNNING);
|
assert_job_status_is(job, JOB_STATUS_RUNNING);
|
||||||
|
|
||||||
s->should_converge = true;
|
s->should_converge = true;
|
||||||
job_enter(job);
|
job_enter(job);
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
assert(job->status == JOB_STATUS_READY);
|
assert(job->status == JOB_STATUS_READY);
|
||||||
|
job_complete_locked(job, &error_abort);
|
||||||
job_complete(job, &error_abort);
|
}
|
||||||
job_enter(job);
|
job_enter(job);
|
||||||
while (!job->deferred_to_main_loop) {
|
while (!job->deferred_to_main_loop) {
|
||||||
aio_poll(qemu_get_aio_context(), true);
|
aio_poll(qemu_get_aio_context(), true);
|
||||||
}
|
}
|
||||||
assert(job->status == JOB_STATUS_READY);
|
assert_job_status_is(job, JOB_STATUS_READY);
|
||||||
aio_poll(qemu_get_aio_context(), true);
|
aio_poll(qemu_get_aio_context(), true);
|
||||||
assert(job->status == JOB_STATUS_PENDING);
|
assert_job_status_is(job, JOB_STATUS_PENDING);
|
||||||
|
|
||||||
aio_context_acquire(job->aio_context);
|
aio_context_acquire(job->aio_context);
|
||||||
job_finalize(job, &error_abort);
|
WITH_JOB_LOCK_GUARD() {
|
||||||
|
job_finalize_locked(job, &error_abort);
|
||||||
|
}
|
||||||
aio_context_release(job->aio_context);
|
aio_context_release(job->aio_context);
|
||||||
assert(job->status == JOB_STATUS_CONCLUDED);
|
assert_job_status_is(job, JOB_STATUS_CONCLUDED);
|
||||||
|
|
||||||
cancel_common(s);
|
cancel_common(s);
|
||||||
}
|
}
|
||||||
|
@ -459,36 +477,43 @@ static void test_complete_in_standby(void)
|
||||||
bjob = mk_job(blk, "job", &test_yielding_driver, true,
|
bjob = mk_job(blk, "job", &test_yielding_driver, true,
|
||||||
JOB_MANUAL_FINALIZE | JOB_MANUAL_DISMISS);
|
JOB_MANUAL_FINALIZE | JOB_MANUAL_DISMISS);
|
||||||
job = &bjob->job;
|
job = &bjob->job;
|
||||||
assert(job->status == JOB_STATUS_CREATED);
|
assert_job_status_is(job, JOB_STATUS_CREATED);
|
||||||
|
|
||||||
/* Wait for the job to become READY */
|
/* Wait for the job to become READY */
|
||||||
job_start(job);
|
job_start(job);
|
||||||
aio_context_acquire(ctx);
|
aio_context_acquire(ctx);
|
||||||
|
/*
|
||||||
|
* Here we are waiting for the status to change, so don't bother
|
||||||
|
* protecting the read every time.
|
||||||
|
*/
|
||||||
AIO_WAIT_WHILE(ctx, job->status != JOB_STATUS_READY);
|
AIO_WAIT_WHILE(ctx, job->status != JOB_STATUS_READY);
|
||||||
aio_context_release(ctx);
|
aio_context_release(ctx);
|
||||||
|
|
||||||
/* Begin the drained section, pausing the job */
|
/* Begin the drained section, pausing the job */
|
||||||
bdrv_drain_all_begin();
|
bdrv_drain_all_begin();
|
||||||
assert(job->status == JOB_STATUS_STANDBY);
|
assert_job_status_is(job, JOB_STATUS_STANDBY);
|
||||||
|
|
||||||
/* Lock the IO thread to prevent the job from being run */
|
/* Lock the IO thread to prevent the job from being run */
|
||||||
aio_context_acquire(ctx);
|
aio_context_acquire(ctx);
|
||||||
/* This will schedule the job to resume it */
|
/* This will schedule the job to resume it */
|
||||||
bdrv_drain_all_end();
|
bdrv_drain_all_end();
|
||||||
|
|
||||||
|
WITH_JOB_LOCK_GUARD() {
|
||||||
/* But the job cannot run, so it will remain on standby */
|
/* But the job cannot run, so it will remain on standby */
|
||||||
assert(job->status == JOB_STATUS_STANDBY);
|
assert(job->status == JOB_STATUS_STANDBY);
|
||||||
|
|
||||||
/* Even though the job is on standby, this should work */
|
/* Even though the job is on standby, this should work */
|
||||||
job_complete(job, &error_abort);
|
job_complete_locked(job, &error_abort);
|
||||||
|
|
||||||
/* The test is done now, clean up. */
|
/* The test is done now, clean up. */
|
||||||
job_finish_sync(job, NULL, &error_abort);
|
job_finish_sync_locked(job, NULL, &error_abort);
|
||||||
assert(job->status == JOB_STATUS_PENDING);
|
assert(job->status == JOB_STATUS_PENDING);
|
||||||
|
|
||||||
job_finalize(job, &error_abort);
|
job_finalize_locked(job, &error_abort);
|
||||||
assert(job->status == JOB_STATUS_CONCLUDED);
|
assert(job->status == JOB_STATUS_CONCLUDED);
|
||||||
|
|
||||||
job_dismiss(&job, &error_abort);
|
job_dismiss_locked(&job, &error_abort);
|
||||||
|
}
|
||||||
|
|
||||||
destroy_blk(blk);
|
destroy_blk(blk);
|
||||||
aio_context_release(ctx);
|
aio_context_release(ctx);
|
||||||
|
|
Loading…
Reference in New Issue