mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/stable/linux.git
synced 2024-09-30 06:10:56 +00:00
dm: rename map_info flush_request to target_request_nr
'target_request_nr' is a more generic name that reflects the fact that it will be used for both flush and discard support. Signed-off-by: Mike Snitzer <snitzer@redhat.com> Signed-off-by: Alasdair G Kergon <agk@redhat.com>
This commit is contained in:
parent
26803b9f06
commit
57cba5d365
4 changed files with 16 additions and 14 deletions
|
@ -1692,7 +1692,7 @@ static int snapshot_merge_map(struct dm_target *ti, struct bio *bio,
|
||||||
chunk_t chunk;
|
chunk_t chunk;
|
||||||
|
|
||||||
if (unlikely(bio_empty_barrier(bio))) {
|
if (unlikely(bio_empty_barrier(bio))) {
|
||||||
if (!map_context->flush_request)
|
if (!map_context->target_request_nr)
|
||||||
bio->bi_bdev = s->origin->bdev;
|
bio->bi_bdev = s->origin->bdev;
|
||||||
else
|
else
|
||||||
bio->bi_bdev = s->cow->bdev;
|
bio->bi_bdev = s->cow->bdev;
|
||||||
|
|
|
@ -213,10 +213,12 @@ static int stripe_map(struct dm_target *ti, struct bio *bio,
|
||||||
struct stripe_c *sc = (struct stripe_c *) ti->private;
|
struct stripe_c *sc = (struct stripe_c *) ti->private;
|
||||||
sector_t offset, chunk;
|
sector_t offset, chunk;
|
||||||
uint32_t stripe;
|
uint32_t stripe;
|
||||||
|
unsigned target_request_nr;
|
||||||
|
|
||||||
if (unlikely(bio_empty_barrier(bio))) {
|
if (unlikely(bio_empty_barrier(bio))) {
|
||||||
BUG_ON(map_context->flush_request >= sc->stripes);
|
target_request_nr = map_context->target_request_nr;
|
||||||
bio->bi_bdev = sc->stripe[map_context->flush_request].dev->bdev;
|
BUG_ON(target_request_nr >= sc->stripes);
|
||||||
|
bio->bi_bdev = sc->stripe[target_request_nr].dev->bdev;
|
||||||
return DM_MAPIO_REMAPPED;
|
return DM_MAPIO_REMAPPED;
|
||||||
}
|
}
|
||||||
|
|
||||||
|
|
|
@ -1183,12 +1183,12 @@ static struct dm_target_io *alloc_tio(struct clone_info *ci,
|
||||||
}
|
}
|
||||||
|
|
||||||
static void __flush_target(struct clone_info *ci, struct dm_target *ti,
|
static void __flush_target(struct clone_info *ci, struct dm_target *ti,
|
||||||
unsigned flush_nr)
|
unsigned request_nr)
|
||||||
{
|
{
|
||||||
struct dm_target_io *tio = alloc_tio(ci, ti);
|
struct dm_target_io *tio = alloc_tio(ci, ti);
|
||||||
struct bio *clone;
|
struct bio *clone;
|
||||||
|
|
||||||
tio->info.flush_request = flush_nr;
|
tio->info.target_request_nr = request_nr;
|
||||||
|
|
||||||
clone = bio_alloc_bioset(GFP_NOIO, 0, ci->md->bs);
|
clone = bio_alloc_bioset(GFP_NOIO, 0, ci->md->bs);
|
||||||
__bio_clone(clone, ci->bio);
|
__bio_clone(clone, ci->bio);
|
||||||
|
@ -1199,13 +1199,13 @@ static void __flush_target(struct clone_info *ci, struct dm_target *ti,
|
||||||
|
|
||||||
static int __clone_and_map_empty_barrier(struct clone_info *ci)
|
static int __clone_and_map_empty_barrier(struct clone_info *ci)
|
||||||
{
|
{
|
||||||
unsigned target_nr = 0, flush_nr;
|
unsigned target_nr = 0, request_nr;
|
||||||
struct dm_target *ti;
|
struct dm_target *ti;
|
||||||
|
|
||||||
while ((ti = dm_table_get_target(ci->map, target_nr++)))
|
while ((ti = dm_table_get_target(ci->map, target_nr++)))
|
||||||
for (flush_nr = 0; flush_nr < ti->num_flush_requests;
|
for (request_nr = 0; request_nr < ti->num_flush_requests;
|
||||||
flush_nr++)
|
request_nr++)
|
||||||
__flush_target(ci, ti, flush_nr);
|
__flush_target(ci, ti, request_nr);
|
||||||
|
|
||||||
ci->sector_count = 0;
|
ci->sector_count = 0;
|
||||||
|
|
||||||
|
@ -2424,11 +2424,11 @@ static void dm_queue_flush(struct mapped_device *md)
|
||||||
queue_work(md->wq, &md->work);
|
queue_work(md->wq, &md->work);
|
||||||
}
|
}
|
||||||
|
|
||||||
static void dm_rq_set_flush_nr(struct request *clone, unsigned flush_nr)
|
static void dm_rq_set_target_request_nr(struct request *clone, unsigned request_nr)
|
||||||
{
|
{
|
||||||
struct dm_rq_target_io *tio = clone->end_io_data;
|
struct dm_rq_target_io *tio = clone->end_io_data;
|
||||||
|
|
||||||
tio->info.flush_request = flush_nr;
|
tio->info.target_request_nr = request_nr;
|
||||||
}
|
}
|
||||||
|
|
||||||
/* Issue barrier requests to targets and wait for their completion. */
|
/* Issue barrier requests to targets and wait for their completion. */
|
||||||
|
@ -2446,7 +2446,7 @@ static int dm_rq_barrier(struct mapped_device *md)
|
||||||
ti = dm_table_get_target(map, i);
|
ti = dm_table_get_target(map, i);
|
||||||
for (j = 0; j < ti->num_flush_requests; j++) {
|
for (j = 0; j < ti->num_flush_requests; j++) {
|
||||||
clone = clone_rq(md->flush_request, md, GFP_NOIO);
|
clone = clone_rq(md->flush_request, md, GFP_NOIO);
|
||||||
dm_rq_set_flush_nr(clone, j);
|
dm_rq_set_target_request_nr(clone, j);
|
||||||
atomic_inc(&md->pending[rq_data_dir(clone)]);
|
atomic_inc(&md->pending[rq_data_dir(clone)]);
|
||||||
map_request(ti, clone, md);
|
map_request(ti, clone, md);
|
||||||
}
|
}
|
||||||
|
|
|
@ -22,7 +22,7 @@ typedef enum { STATUSTYPE_INFO, STATUSTYPE_TABLE } status_type_t;
|
||||||
union map_info {
|
union map_info {
|
||||||
void *ptr;
|
void *ptr;
|
||||||
unsigned long long ll;
|
unsigned long long ll;
|
||||||
unsigned flush_request;
|
unsigned target_request_nr;
|
||||||
};
|
};
|
||||||
|
|
||||||
/*
|
/*
|
||||||
|
@ -174,7 +174,7 @@ struct dm_target {
|
||||||
* A number of zero-length barrier requests that will be submitted
|
* A number of zero-length barrier requests that will be submitted
|
||||||
* to the target for the purpose of flushing cache.
|
* to the target for the purpose of flushing cache.
|
||||||
*
|
*
|
||||||
* The request number will be placed in union map_info->flush_request.
|
* The request number will be placed in union map_info->target_request_nr.
|
||||||
* It is a responsibility of the target driver to remap these requests
|
* It is a responsibility of the target driver to remap these requests
|
||||||
* to the real underlying devices.
|
* to the real underlying devices.
|
||||||
*/
|
*/
|
||||||
|
|
Loading…
Reference in a new issue