mirror of
https://gitlab.com/qemu-project/qemu
synced 2024-11-05 20:35:44 +00:00
blockjobs: model single jobs as transactions
model all independent jobs as single job transactions. It's one less case we have to worry about when we add more states to the transition machine. This way, we can just treat all job lifetimes exactly the same. This helps tighten assertions of the STM graph and removes some conditionals that would have been needed in the coming commits adding a more explicit job lifetime management API. Signed-off-by: John Snow <jsnow@redhat.com> Reviewed-by: Eric Blake <eblake@redhat.com> Reviewed-by: Kevin Wolf <kwolf@redhat.com> Signed-off-by: Kevin Wolf <kwolf@redhat.com>
This commit is contained in:
parent
d4fce18844
commit
75859b9420
10 changed files with 32 additions and 31 deletions
|
@ -621,7 +621,7 @@ BlockJob *backup_job_create(const char *job_id, BlockDriverState *bs,
|
|||
}
|
||||
|
||||
/* job->common.len is fixed, so we can't allow resize */
|
||||
job = block_job_create(job_id, &backup_job_driver, bs,
|
||||
job = block_job_create(job_id, &backup_job_driver, txn, bs,
|
||||
BLK_PERM_CONSISTENT_READ,
|
||||
BLK_PERM_CONSISTENT_READ | BLK_PERM_WRITE |
|
||||
BLK_PERM_WRITE_UNCHANGED | BLK_PERM_GRAPH_MOD,
|
||||
|
@ -677,7 +677,6 @@ BlockJob *backup_job_create(const char *job_id, BlockDriverState *bs,
|
|||
block_job_add_bdrv(&job->common, "target", target, 0, BLK_PERM_ALL,
|
||||
&error_abort);
|
||||
job->common.len = len;
|
||||
block_job_txn_add_job(txn, &job->common);
|
||||
|
||||
return &job->common;
|
||||
|
||||
|
|
|
@ -289,7 +289,7 @@ void commit_start(const char *job_id, BlockDriverState *bs,
|
|||
return;
|
||||
}
|
||||
|
||||
s = block_job_create(job_id, &commit_job_driver, bs, 0, BLK_PERM_ALL,
|
||||
s = block_job_create(job_id, &commit_job_driver, NULL, bs, 0, BLK_PERM_ALL,
|
||||
speed, BLOCK_JOB_DEFAULT, NULL, NULL, errp);
|
||||
if (!s) {
|
||||
return;
|
||||
|
|
|
@ -1166,7 +1166,7 @@ static void mirror_start_job(const char *job_id, BlockDriverState *bs,
|
|||
}
|
||||
|
||||
/* Make sure that the source is not resized while the job is running */
|
||||
s = block_job_create(job_id, driver, mirror_top_bs,
|
||||
s = block_job_create(job_id, driver, NULL, mirror_top_bs,
|
||||
BLK_PERM_CONSISTENT_READ,
|
||||
BLK_PERM_CONSISTENT_READ | BLK_PERM_WRITE_UNCHANGED |
|
||||
BLK_PERM_WRITE | BLK_PERM_GRAPH_MOD, speed,
|
||||
|
|
|
@ -244,7 +244,7 @@ void stream_start(const char *job_id, BlockDriverState *bs,
|
|||
/* Prevent concurrent jobs trying to modify the graph structure here, we
|
||||
* already have our own plans. Also don't allow resize as the image size is
|
||||
* queried only at the job start and then cached. */
|
||||
s = block_job_create(job_id, &stream_job_driver, bs,
|
||||
s = block_job_create(job_id, &stream_job_driver, NULL, bs,
|
||||
BLK_PERM_CONSISTENT_READ | BLK_PERM_WRITE_UNCHANGED |
|
||||
BLK_PERM_GRAPH_MOD,
|
||||
BLK_PERM_CONSISTENT_READ | BLK_PERM_WRITE_UNCHANGED |
|
||||
|
|
25
blockjob.c
25
blockjob.c
|
@ -357,10 +357,8 @@ static void block_job_completed_single(BlockJob *job)
|
|||
}
|
||||
}
|
||||
|
||||
if (job->txn) {
|
||||
QLIST_REMOVE(job, txn_list);
|
||||
block_job_txn_unref(job->txn);
|
||||
}
|
||||
QLIST_REMOVE(job, txn_list);
|
||||
block_job_txn_unref(job->txn);
|
||||
block_job_unref(job);
|
||||
}
|
||||
|
||||
|
@ -647,7 +645,7 @@ static void block_job_event_completed(BlockJob *job, const char *msg)
|
|||
*/
|
||||
|
||||
void *block_job_create(const char *job_id, const BlockJobDriver *driver,
|
||||
BlockDriverState *bs, uint64_t perm,
|
||||
BlockJobTxn *txn, BlockDriverState *bs, uint64_t perm,
|
||||
uint64_t shared_perm, int64_t speed, int flags,
|
||||
BlockCompletionFunc *cb, void *opaque, Error **errp)
|
||||
{
|
||||
|
@ -729,6 +727,17 @@ void *block_job_create(const char *job_id, const BlockJobDriver *driver,
|
|||
return NULL;
|
||||
}
|
||||
}
|
||||
|
||||
/* Single jobs are modeled as single-job transactions for sake of
|
||||
* consolidating the job management logic */
|
||||
if (!txn) {
|
||||
txn = block_job_txn_new();
|
||||
block_job_txn_add_job(txn, job);
|
||||
block_job_txn_unref(txn);
|
||||
} else {
|
||||
block_job_txn_add_job(txn, job);
|
||||
}
|
||||
|
||||
return job;
|
||||
}
|
||||
|
||||
|
@ -752,13 +761,11 @@ void block_job_early_fail(BlockJob *job)
|
|||
|
||||
void block_job_completed(BlockJob *job, int ret)
|
||||
{
|
||||
assert(job && job->txn && !job->completed);
|
||||
assert(blk_bs(job->blk)->job == job);
|
||||
assert(!job->completed);
|
||||
job->completed = true;
|
||||
job->ret = ret;
|
||||
if (!job->txn) {
|
||||
block_job_completed_single(job);
|
||||
} else if (ret < 0 || block_job_is_cancelled(job)) {
|
||||
if (ret < 0 || block_job_is_cancelled(job)) {
|
||||
block_job_completed_txn_abort(job);
|
||||
} else {
|
||||
block_job_completed_txn_success(job);
|
||||
|
|
|
@ -141,7 +141,6 @@ typedef struct BlockJob {
|
|||
*/
|
||||
QEMUTimer sleep_timer;
|
||||
|
||||
/** Non-NULL if this job is part of a transaction */
|
||||
BlockJobTxn *txn;
|
||||
QLIST_ENTRY(BlockJob) txn_list;
|
||||
} BlockJob;
|
||||
|
|
|
@ -115,6 +115,7 @@ struct BlockJobDriver {
|
|||
* @job_id: The id of the newly-created job, or %NULL to have one
|
||||
* generated automatically.
|
||||
* @job_type: The class object for the newly-created job.
|
||||
* @txn: The transaction this job belongs to, if any. %NULL otherwise.
|
||||
* @bs: The block
|
||||
* @perm, @shared_perm: Permissions to request for @bs
|
||||
* @speed: The maximum speed, in bytes per second, or 0 for unlimited.
|
||||
|
@ -132,7 +133,7 @@ struct BlockJobDriver {
|
|||
* called from a wrapper that is specific to the job type.
|
||||
*/
|
||||
void *block_job_create(const char *job_id, const BlockJobDriver *driver,
|
||||
BlockDriverState *bs, uint64_t perm,
|
||||
BlockJobTxn *txn, BlockDriverState *bs, uint64_t perm,
|
||||
uint64_t shared_perm, int64_t speed, int flags,
|
||||
BlockCompletionFunc *cb, void *opaque, Error **errp);
|
||||
|
||||
|
|
|
@ -541,8 +541,8 @@ static void test_blockjob_common(enum drain_type drain_type)
|
|||
blk_target = blk_new(BLK_PERM_ALL, BLK_PERM_ALL);
|
||||
blk_insert_bs(blk_target, target, &error_abort);
|
||||
|
||||
job = block_job_create("job0", &test_job_driver, src, 0, BLK_PERM_ALL, 0,
|
||||
0, NULL, NULL, &error_abort);
|
||||
job = block_job_create("job0", &test_job_driver, NULL, src, 0, BLK_PERM_ALL,
|
||||
0, 0, NULL, NULL, &error_abort);
|
||||
block_job_add_bdrv(job, "target", target, 0, BLK_PERM_ALL, &error_abort);
|
||||
block_job_start(job);
|
||||
|
||||
|
|
|
@ -87,7 +87,7 @@ static const BlockJobDriver test_block_job_driver = {
|
|||
*/
|
||||
static BlockJob *test_block_job_start(unsigned int iterations,
|
||||
bool use_timer,
|
||||
int rc, int *result)
|
||||
int rc, int *result, BlockJobTxn *txn)
|
||||
{
|
||||
BlockDriverState *bs;
|
||||
TestBlockJob *s;
|
||||
|
@ -101,7 +101,7 @@ static BlockJob *test_block_job_start(unsigned int iterations,
|
|||
g_assert_nonnull(bs);
|
||||
|
||||
snprintf(job_id, sizeof(job_id), "job%u", counter++);
|
||||
s = block_job_create(job_id, &test_block_job_driver, bs,
|
||||
s = block_job_create(job_id, &test_block_job_driver, txn, bs,
|
||||
0, BLK_PERM_ALL, 0, BLOCK_JOB_DEFAULT,
|
||||
test_block_job_cb, data, &error_abort);
|
||||
s->iterations = iterations;
|
||||
|
@ -120,8 +120,7 @@ static void test_single_job(int expected)
|
|||
int result = -EINPROGRESS;
|
||||
|
||||
txn = block_job_txn_new();
|
||||
job = test_block_job_start(1, true, expected, &result);
|
||||
block_job_txn_add_job(txn, job);
|
||||
job = test_block_job_start(1, true, expected, &result, txn);
|
||||
block_job_start(job);
|
||||
|
||||
if (expected == -ECANCELED) {
|
||||
|
@ -160,10 +159,8 @@ static void test_pair_jobs(int expected1, int expected2)
|
|||
int result2 = -EINPROGRESS;
|
||||
|
||||
txn = block_job_txn_new();
|
||||
job1 = test_block_job_start(1, true, expected1, &result1);
|
||||
block_job_txn_add_job(txn, job1);
|
||||
job2 = test_block_job_start(2, true, expected2, &result2);
|
||||
block_job_txn_add_job(txn, job2);
|
||||
job1 = test_block_job_start(1, true, expected1, &result1, txn);
|
||||
job2 = test_block_job_start(2, true, expected2, &result2, txn);
|
||||
block_job_start(job1);
|
||||
block_job_start(job2);
|
||||
|
||||
|
@ -224,10 +221,8 @@ static void test_pair_jobs_fail_cancel_race(void)
|
|||
int result2 = -EINPROGRESS;
|
||||
|
||||
txn = block_job_txn_new();
|
||||
job1 = test_block_job_start(1, true, -ECANCELED, &result1);
|
||||
block_job_txn_add_job(txn, job1);
|
||||
job2 = test_block_job_start(2, false, 0, &result2);
|
||||
block_job_txn_add_job(txn, job2);
|
||||
job1 = test_block_job_start(1, true, -ECANCELED, &result1, txn);
|
||||
job2 = test_block_job_start(2, false, 0, &result2, txn);
|
||||
block_job_start(job1);
|
||||
block_job_start(job2);
|
||||
|
||||
|
|
|
@ -30,7 +30,7 @@ static BlockJob *do_test_id(BlockBackend *blk, const char *id,
|
|||
BlockJob *job;
|
||||
Error *errp = NULL;
|
||||
|
||||
job = block_job_create(id, &test_block_job_driver, blk_bs(blk),
|
||||
job = block_job_create(id, &test_block_job_driver, NULL, blk_bs(blk),
|
||||
0, BLK_PERM_ALL, 0, BLOCK_JOB_DEFAULT, block_job_cb,
|
||||
NULL, &errp);
|
||||
if (should_succeed) {
|
||||
|
|
Loading…
Reference in a new issue