aboutsummaryrefslogtreecommitdiffstats
path: root/tests/unit/test-blockjob.c
diff options
context:
space:
mode:
authorTimos Ampelikiotis <t.ampelikiotis@virtualopensystems.com>2023-10-10 11:40:56 +0000
committerTimos Ampelikiotis <t.ampelikiotis@virtualopensystems.com>2023-10-10 11:40:56 +0000
commite02cda008591317b1625707ff8e115a4841aa889 (patch)
treeaee302e3cf8b59ec2d32ec481be3d1afddfc8968 /tests/unit/test-blockjob.c
parentcc668e6b7e0ffd8c9d130513d12053cf5eda1d3b (diff)
Introduce Virtio-loopback epsilon release:
Epsilon release introduces a new compatibility layer which make virtio-loopback design to work with QEMU and rust-vmm vhost-user backend without require any changes. Signed-off-by: Timos Ampelikiotis <t.ampelikiotis@virtualopensystems.com> Change-Id: I52e57563e08a7d0bdc002f8e928ee61ba0c53dd9
Diffstat (limited to 'tests/unit/test-blockjob.c')
-rw-r--r--tests/unit/test-blockjob.c514
1 files changed, 514 insertions, 0 deletions
diff --git a/tests/unit/test-blockjob.c b/tests/unit/test-blockjob.c
new file mode 100644
index 000000000..4c9e1bf1e
--- /dev/null
+++ b/tests/unit/test-blockjob.c
@@ -0,0 +1,514 @@
+/*
+ * Blockjob tests
+ *
+ * Copyright Igalia, S.L. 2016
+ *
+ * Authors:
+ * Alberto Garcia <berto@igalia.com>
+ *
+ * This work is licensed under the terms of the GNU LGPL, version 2 or later.
+ * See the COPYING.LIB file in the top-level directory.
+ */
+
+#include "qemu/osdep.h"
+#include "qapi/error.h"
+#include "qemu/main-loop.h"
+#include "block/blockjob_int.h"
+#include "sysemu/block-backend.h"
+#include "qapi/qmp/qdict.h"
+#include "iothread.h"
+
+static const BlockJobDriver test_block_job_driver = {
+ .job_driver = {
+ .instance_size = sizeof(BlockJob),
+ .free = block_job_free,
+ .user_resume = block_job_user_resume,
+ },
+};
+
+static void block_job_cb(void *opaque, int ret)
+{
+}
+
+static BlockJob *mk_job(BlockBackend *blk, const char *id,
+ const BlockJobDriver *drv, bool should_succeed,
+ int flags)
+{
+ BlockJob *job;
+ Error *err = NULL;
+
+ job = block_job_create(id, drv, NULL, blk_bs(blk),
+ 0, BLK_PERM_ALL, 0, flags, block_job_cb,
+ NULL, &err);
+ if (should_succeed) {
+ g_assert_null(err);
+ g_assert_nonnull(job);
+ if (id) {
+ g_assert_cmpstr(job->job.id, ==, id);
+ } else {
+ g_assert_cmpstr(job->job.id, ==, blk_name(blk));
+ }
+ } else {
+ error_free_or_abort(&err);
+ g_assert_null(job);
+ }
+
+ return job;
+}
+
+static BlockJob *do_test_id(BlockBackend *blk, const char *id,
+ bool should_succeed)
+{
+ return mk_job(blk, id, &test_block_job_driver,
+ should_succeed, JOB_DEFAULT);
+}
+
+/* This creates a BlockBackend (optionally with a name) with a
+ * BlockDriverState inserted. */
+static BlockBackend *create_blk(const char *name)
+{
+ /* No I/O is performed on this device */
+ BlockBackend *blk = blk_new(qemu_get_aio_context(), 0, BLK_PERM_ALL);
+ BlockDriverState *bs;
+
+ QDict *opt = qdict_new();
+ qdict_put_str(opt, "file.read-zeroes", "on");
+ bs = bdrv_open("null-co://", NULL, opt, 0, &error_abort);
+ g_assert_nonnull(bs);
+
+ blk_insert_bs(blk, bs, &error_abort);
+ bdrv_unref(bs);
+
+ if (name) {
+ Error *err = NULL;
+ monitor_add_blk(blk, name, &err);
+ g_assert_null(err);
+ }
+
+ return blk;
+}
+
+/* This destroys the backend */
+static void destroy_blk(BlockBackend *blk)
+{
+ if (blk_name(blk)[0] != '\0') {
+ monitor_remove_blk(blk);
+ }
+
+ blk_remove_bs(blk);
+ blk_unref(blk);
+}
+
+static void test_job_ids(void)
+{
+ BlockBackend *blk[3];
+ BlockJob *job[3];
+
+ blk[0] = create_blk(NULL);
+ blk[1] = create_blk("drive1");
+ blk[2] = create_blk("drive2");
+
+ /* No job ID provided and the block backend has no name */
+ job[0] = do_test_id(blk[0], NULL, false);
+
+ /* These are all invalid job IDs */
+ job[0] = do_test_id(blk[0], "0id", false);
+ job[0] = do_test_id(blk[0], "", false);
+ job[0] = do_test_id(blk[0], " ", false);
+ job[0] = do_test_id(blk[0], "123", false);
+ job[0] = do_test_id(blk[0], "_id", false);
+ job[0] = do_test_id(blk[0], "-id", false);
+ job[0] = do_test_id(blk[0], ".id", false);
+ job[0] = do_test_id(blk[0], "#id", false);
+
+ /* This one is valid */
+ job[0] = do_test_id(blk[0], "id0", true);
+
+ /* We can have two jobs in the same BDS */
+ job[1] = do_test_id(blk[0], "id1", true);
+ job_early_fail(&job[1]->job);
+
+ /* Duplicate job IDs are not allowed */
+ job[1] = do_test_id(blk[1], "id0", false);
+
+ /* But once job[0] finishes we can reuse its ID */
+ job_early_fail(&job[0]->job);
+ job[1] = do_test_id(blk[1], "id0", true);
+
+ /* No job ID specified, defaults to the backend name ('drive1') */
+ job_early_fail(&job[1]->job);
+ job[1] = do_test_id(blk[1], NULL, true);
+
+ /* Duplicate job ID */
+ job[2] = do_test_id(blk[2], "drive1", false);
+
+ /* The ID of job[2] would default to 'drive2' but it is already in use */
+ job[0] = do_test_id(blk[0], "drive2", true);
+ job[2] = do_test_id(blk[2], NULL, false);
+
+ /* This one is valid */
+ job[2] = do_test_id(blk[2], "id_2", true);
+
+ job_early_fail(&job[0]->job);
+ job_early_fail(&job[1]->job);
+ job_early_fail(&job[2]->job);
+
+ destroy_blk(blk[0]);
+ destroy_blk(blk[1]);
+ destroy_blk(blk[2]);
+}
+
+typedef struct CancelJob {
+ BlockJob common;
+ BlockBackend *blk;
+ bool should_converge;
+ bool should_complete;
+} CancelJob;
+
+static void cancel_job_complete(Job *job, Error **errp)
+{
+ CancelJob *s = container_of(job, CancelJob, common.job);
+ s->should_complete = true;
+}
+
+static int coroutine_fn cancel_job_run(Job *job, Error **errp)
+{
+ CancelJob *s = container_of(job, CancelJob, common.job);
+
+ while (!s->should_complete) {
+ if (job_is_cancelled(&s->common.job)) {
+ return 0;
+ }
+
+ if (!job_is_ready(&s->common.job) && s->should_converge) {
+ job_transition_to_ready(&s->common.job);
+ }
+
+ job_sleep_ns(&s->common.job, 100000);
+ }
+
+ return 0;
+}
+
+static const BlockJobDriver test_cancel_driver = {
+ .job_driver = {
+ .instance_size = sizeof(CancelJob),
+ .free = block_job_free,
+ .user_resume = block_job_user_resume,
+ .run = cancel_job_run,
+ .complete = cancel_job_complete,
+ },
+};
+
+static CancelJob *create_common(Job **pjob)
+{
+ BlockBackend *blk;
+ Job *job;
+ BlockJob *bjob;
+ CancelJob *s;
+
+ blk = create_blk(NULL);
+ bjob = mk_job(blk, "Steve", &test_cancel_driver, true,
+ JOB_MANUAL_FINALIZE | JOB_MANUAL_DISMISS);
+ job = &bjob->job;
+ job_ref(job);
+ assert(job->status == JOB_STATUS_CREATED);
+ s = container_of(bjob, CancelJob, common);
+ s->blk = blk;
+
+ *pjob = job;
+ return s;
+}
+
+static void cancel_common(CancelJob *s)
+{
+ BlockJob *job = &s->common;
+ BlockBackend *blk = s->blk;
+ JobStatus sts = job->job.status;
+ AioContext *ctx;
+
+ ctx = job->job.aio_context;
+ aio_context_acquire(ctx);
+
+ job_cancel_sync(&job->job, true);
+ if (sts != JOB_STATUS_CREATED && sts != JOB_STATUS_CONCLUDED) {
+ Job *dummy = &job->job;
+ job_dismiss(&dummy, &error_abort);
+ }
+ assert(job->job.status == JOB_STATUS_NULL);
+ job_unref(&job->job);
+ destroy_blk(blk);
+
+ aio_context_release(ctx);
+}
+
+static void test_cancel_created(void)
+{
+ Job *job;
+ CancelJob *s;
+
+ s = create_common(&job);
+ cancel_common(s);
+}
+
+static void test_cancel_running(void)
+{
+ Job *job;
+ CancelJob *s;
+
+ s = create_common(&job);
+
+ job_start(job);
+ assert(job->status == JOB_STATUS_RUNNING);
+
+ cancel_common(s);
+}
+
+static void test_cancel_paused(void)
+{
+ Job *job;
+ CancelJob *s;
+
+ s = create_common(&job);
+
+ job_start(job);
+ assert(job->status == JOB_STATUS_RUNNING);
+
+ job_user_pause(job, &error_abort);
+ job_enter(job);
+ assert(job->status == JOB_STATUS_PAUSED);
+
+ cancel_common(s);
+}
+
+static void test_cancel_ready(void)
+{
+ Job *job;
+ CancelJob *s;
+
+ s = create_common(&job);
+
+ job_start(job);
+ assert(job->status == JOB_STATUS_RUNNING);
+
+ s->should_converge = true;
+ job_enter(job);
+ assert(job->status == JOB_STATUS_READY);
+
+ cancel_common(s);
+}
+
+static void test_cancel_standby(void)
+{
+ Job *job;
+ CancelJob *s;
+
+ s = create_common(&job);
+
+ job_start(job);
+ assert(job->status == JOB_STATUS_RUNNING);
+
+ s->should_converge = true;
+ job_enter(job);
+ assert(job->status == JOB_STATUS_READY);
+
+ job_user_pause(job, &error_abort);
+ job_enter(job);
+ assert(job->status == JOB_STATUS_STANDBY);
+
+ cancel_common(s);
+}
+
+static void test_cancel_pending(void)
+{
+ Job *job;
+ CancelJob *s;
+
+ s = create_common(&job);
+
+ job_start(job);
+ assert(job->status == JOB_STATUS_RUNNING);
+
+ s->should_converge = true;
+ job_enter(job);
+ assert(job->status == JOB_STATUS_READY);
+
+ job_complete(job, &error_abort);
+ job_enter(job);
+ while (!job->deferred_to_main_loop) {
+ aio_poll(qemu_get_aio_context(), true);
+ }
+ assert(job->status == JOB_STATUS_READY);
+ aio_poll(qemu_get_aio_context(), true);
+ assert(job->status == JOB_STATUS_PENDING);
+
+ cancel_common(s);
+}
+
+static void test_cancel_concluded(void)
+{
+ Job *job;
+ CancelJob *s;
+
+ s = create_common(&job);
+
+ job_start(job);
+ assert(job->status == JOB_STATUS_RUNNING);
+
+ s->should_converge = true;
+ job_enter(job);
+ assert(job->status == JOB_STATUS_READY);
+
+ job_complete(job, &error_abort);
+ job_enter(job);
+ while (!job->deferred_to_main_loop) {
+ aio_poll(qemu_get_aio_context(), true);
+ }
+ assert(job->status == JOB_STATUS_READY);
+ aio_poll(qemu_get_aio_context(), true);
+ assert(job->status == JOB_STATUS_PENDING);
+
+ aio_context_acquire(job->aio_context);
+ job_finalize(job, &error_abort);
+ aio_context_release(job->aio_context);
+ assert(job->status == JOB_STATUS_CONCLUDED);
+
+ cancel_common(s);
+}
+
+/* (See test_yielding_driver for the job description) */
+typedef struct YieldingJob {
+ BlockJob common;
+ bool should_complete;
+} YieldingJob;
+
+static void yielding_job_complete(Job *job, Error **errp)
+{
+ YieldingJob *s = container_of(job, YieldingJob, common.job);
+ s->should_complete = true;
+ job_enter(job);
+}
+
+static int coroutine_fn yielding_job_run(Job *job, Error **errp)
+{
+ YieldingJob *s = container_of(job, YieldingJob, common.job);
+
+ job_transition_to_ready(job);
+
+ while (!s->should_complete) {
+ job_yield(job);
+ }
+
+ return 0;
+}
+
+/*
+ * This job transitions immediately to the READY state, and then
+ * yields until it is to complete.
+ */
+static const BlockJobDriver test_yielding_driver = {
+ .job_driver = {
+ .instance_size = sizeof(YieldingJob),
+ .free = block_job_free,
+ .user_resume = block_job_user_resume,
+ .run = yielding_job_run,
+ .complete = yielding_job_complete,
+ },
+};
+
+/*
+ * Test that job_complete() works even on jobs that are in a paused
+ * state (i.e., STANDBY).
+ *
+ * To do this, run YieldingJob in an IO thread, get it into the READY
+ * state, then have a drained section. Before ending the section,
+ * acquire the context so the job will not be entered and will thus
+ * remain on STANDBY.
+ *
+ * job_complete() should still work without error.
+ *
+ * Note that on the QMP interface, it is impossible to lock an IO
+ * thread before a drained section ends. In practice, the
+ * bdrv_drain_all_end() and the aio_context_acquire() will be
+ * reversed. However, that makes for worse reproducibility here:
+ * Sometimes, the job would no longer be in STANDBY then but already
+ * be started. We cannot prevent that, because the IO thread runs
+ * concurrently. We can only prevent it by taking the lock before
+ * ending the drained section, so we do that.
+ *
+ * (You can reverse the order of operations and most of the time the
+ * test will pass, but sometimes the assert(status == STANDBY) will
+ * fail.)
+ */
+static void test_complete_in_standby(void)
+{
+ BlockBackend *blk;
+ IOThread *iothread;
+ AioContext *ctx;
+ Job *job;
+ BlockJob *bjob;
+
+ /* Create a test drive, move it to an IO thread */
+ blk = create_blk(NULL);
+ iothread = iothread_new();
+
+ ctx = iothread_get_aio_context(iothread);
+ blk_set_aio_context(blk, ctx, &error_abort);
+
+ /* Create our test job */
+ bjob = mk_job(blk, "job", &test_yielding_driver, true,
+ JOB_MANUAL_FINALIZE | JOB_MANUAL_DISMISS);
+ job = &bjob->job;
+ assert(job->status == JOB_STATUS_CREATED);
+
+ /* Wait for the job to become READY */
+ job_start(job);
+ aio_context_acquire(ctx);
+ AIO_WAIT_WHILE(ctx, job->status != JOB_STATUS_READY);
+ aio_context_release(ctx);
+
+ /* Begin the drained section, pausing the job */
+ bdrv_drain_all_begin();
+ assert(job->status == JOB_STATUS_STANDBY);
+ /* Lock the IO thread to prevent the job from being run */
+ aio_context_acquire(ctx);
+ /* This will schedule the job to resume it */
+ bdrv_drain_all_end();
+
+ /* But the job cannot run, so it will remain on standby */
+ assert(job->status == JOB_STATUS_STANDBY);
+
+ /* Even though the job is on standby, this should work */
+ job_complete(job, &error_abort);
+
+ /* The test is done now, clean up. */
+ job_finish_sync(job, NULL, &error_abort);
+ assert(job->status == JOB_STATUS_PENDING);
+
+ job_finalize(job, &error_abort);
+ assert(job->status == JOB_STATUS_CONCLUDED);
+
+ job_dismiss(&job, &error_abort);
+
+ destroy_blk(blk);
+ aio_context_release(ctx);
+ iothread_join(iothread);
+}
+
+int main(int argc, char **argv)
+{
+ qemu_init_main_loop(&error_abort);
+ bdrv_init();
+
+ g_test_init(&argc, &argv, NULL);
+ g_test_add_func("/blockjob/ids", test_job_ids);
+ g_test_add_func("/blockjob/cancel/created", test_cancel_created);
+ g_test_add_func("/blockjob/cancel/running", test_cancel_running);
+ g_test_add_func("/blockjob/cancel/paused", test_cancel_paused);
+ g_test_add_func("/blockjob/cancel/ready", test_cancel_ready);
+ g_test_add_func("/blockjob/cancel/standby", test_cancel_standby);
+ g_test_add_func("/blockjob/cancel/pending", test_cancel_pending);
+ g_test_add_func("/blockjob/cancel/concluded", test_cancel_concluded);
+ g_test_add_func("/blockjob/complete_in_standby", test_complete_in_standby);
+ return g_test_run();
+}