| .. | .. |
|---|
| 63 | 63 | void dm_table_presuspend_undo_targets(struct dm_table *t); |
|---|
| 64 | 64 | void dm_table_postsuspend_targets(struct dm_table *t); |
|---|
| 65 | 65 | int dm_table_resume_targets(struct dm_table *t); |
|---|
| 66 | | -int dm_table_any_congested(struct dm_table *t, int bdi_bits); |
|---|
| 67 | 66 | enum dm_queue_mode dm_table_get_type(struct dm_table *t); |
|---|
| 68 | 67 | struct target_type *dm_table_get_immutable_target_type(struct dm_table *t); |
|---|
| 69 | 68 | struct dm_target *dm_table_get_immutable_target(struct dm_table *t); |
|---|
| 70 | 69 | struct dm_target *dm_table_get_wildcard_target(struct dm_table *t); |
|---|
| 71 | 70 | bool dm_table_bio_based(struct dm_table *t); |
|---|
| 72 | 71 | bool dm_table_request_based(struct dm_table *t); |
|---|
| 73 | | -bool dm_table_all_blk_mq_devices(struct dm_table *t); |
|---|
| 74 | 72 | void dm_table_free_md_mempools(struct dm_table *t); |
|---|
| 75 | 73 | struct dm_md_mempools *dm_table_get_md_mempools(struct dm_table *t); |
|---|
| 74 | +bool dm_table_supports_dax(struct dm_table *t, iterate_devices_callout_fn fn, |
|---|
| 75 | + int *blocksize); |
|---|
| 76 | +int device_not_dax_capable(struct dm_target *ti, struct dm_dev *dev, |
|---|
| 77 | + sector_t start, sector_t len, void *data); |
|---|
| 76 | 78 | |
|---|
| 77 | 79 | void dm_lock_md_type(struct mapped_device *md); |
|---|
| 78 | 80 | void dm_unlock_md_type(struct mapped_device *md); |
|---|
| .. | .. |
|---|
| 81 | 83 | struct target_type *dm_get_immutable_target_type(struct mapped_device *md); |
|---|
| 82 | 84 | |
|---|
| 83 | 85 | int dm_setup_md_queue(struct mapped_device *md, struct dm_table *t); |
|---|
| 84 | | - |
|---|
| 85 | | -/* |
|---|
| 86 | | - * To check the return value from dm_table_find_target(). |
|---|
| 87 | | - */ |
|---|
| 88 | | -#define dm_target_is_valid(t) ((t)->table) |
|---|
| 89 | 86 | |
|---|
| 90 | 87 | /* |
|---|
| 91 | 88 | * To check whether the target type is bio-based or not (request-based). |
|---|
| .. | .. |
|---|
| 182 | 179 | int dm_lock_for_deletion(struct mapped_device *md, bool mark_deferred, bool only_deferred); |
|---|
| 183 | 180 | int dm_cancel_deferred_remove(struct mapped_device *md); |
|---|
| 184 | 181 | int dm_request_based(struct mapped_device *md); |
|---|
| 185 | | -sector_t dm_get_size(struct mapped_device *md); |
|---|
| 186 | | -struct request_queue *dm_get_md_queue(struct mapped_device *md); |
|---|
| 187 | 182 | int dm_get_table_device(struct mapped_device *md, dev_t dev, fmode_t mode, |
|---|
| 188 | 183 | struct dm_dev **result); |
|---|
| 189 | 184 | void dm_put_table_device(struct mapped_device *md, struct dm_dev *d); |
|---|
| 190 | | -struct dm_stats *dm_get_stats(struct mapped_device *md); |
|---|
| 191 | 185 | |
|---|
| 192 | 186 | int dm_kobject_uevent(struct mapped_device *md, enum kobject_action action, |
|---|
| 193 | 187 | unsigned cookie); |
|---|