.. | .. |
---|
| 1 | +// SPDX-License-Identifier: GPL-2.0-only |
---|
1 | 2 | /* |
---|
2 | 3 | * Add configfs and memory store: Kyungchan Koh <kkc6196@fb.com> and |
---|
3 | 4 | * Shaohua Li <shli@fb.com> |
---|
.. | .. |
---|
22 | 23 | #ifdef CONFIG_BLK_DEV_NULL_BLK_FAULT_INJECTION |
---|
23 | 24 | static DECLARE_FAULT_ATTR(null_timeout_attr); |
---|
24 | 25 | static DECLARE_FAULT_ATTR(null_requeue_attr); |
---|
| 26 | +static DECLARE_FAULT_ATTR(null_init_hctx_attr); |
---|
25 | 27 | #endif |
---|
26 | 28 | |
---|
27 | 29 | static inline u64 mb_per_tick(int mbps) |
---|
.. | .. |
---|
95 | 97 | MODULE_PARM_DESC(home_node, "Home node for the device"); |
---|
96 | 98 | |
---|
97 | 99 | #ifdef CONFIG_BLK_DEV_NULL_BLK_FAULT_INJECTION |
---|
| 100 | +/* |
---|
| 101 | + * For more details about fault injection, please refer to |
---|
| 102 | + * Documentation/fault-injection/fault-injection.rst. |
---|
| 103 | + */ |
---|
98 | 104 | static char g_timeout_str[80]; |
---|
99 | 105 | module_param_string(timeout, g_timeout_str, sizeof(g_timeout_str), 0444); |
---|
| 106 | +MODULE_PARM_DESC(timeout, "Fault injection. timeout=<interval>,<probability>,<space>,<times>"); |
---|
100 | 107 | |
---|
101 | 108 | static char g_requeue_str[80]; |
---|
102 | 109 | module_param_string(requeue, g_requeue_str, sizeof(g_requeue_str), 0444); |
---|
| 110 | +MODULE_PARM_DESC(requeue, "Fault injection. requeue=<interval>,<probability>,<space>,<times>"); |
---|
| 111 | + |
---|
| 112 | +static char g_init_hctx_str[80]; |
---|
| 113 | +module_param_string(init_hctx, g_init_hctx_str, sizeof(g_init_hctx_str), 0444); |
---|
| 114 | +MODULE_PARM_DESC(init_hctx, "Fault injection to fail hctx init. init_hctx=<interval>,<probability>,<space>,<times>"); |
---|
103 | 115 | #endif |
---|
104 | 116 | |
---|
105 | 117 | static int g_queue_mode = NULL_Q_MQ; |
---|
.. | .. |
---|
140 | 152 | module_param_named(bs, g_bs, int, 0444); |
---|
141 | 153 | MODULE_PARM_DESC(bs, "Block size (in bytes)"); |
---|
142 | 154 | |
---|
143 | | -static int nr_devices = 1; |
---|
144 | | -module_param(nr_devices, int, 0444); |
---|
| 155 | +static unsigned int nr_devices = 1; |
---|
| 156 | +module_param(nr_devices, uint, 0444); |
---|
145 | 157 | MODULE_PARM_DESC(nr_devices, "Number of devices to register"); |
---|
146 | 158 | |
---|
147 | 159 | static bool g_blocking; |
---|
.. | .. |
---|
151 | 163 | static bool shared_tags; |
---|
152 | 164 | module_param(shared_tags, bool, 0444); |
---|
153 | 165 | MODULE_PARM_DESC(shared_tags, "Share tag set between devices for blk-mq"); |
---|
| 166 | + |
---|
| 167 | +static bool g_shared_tag_bitmap; |
---|
| 168 | +module_param_named(shared_tag_bitmap, g_shared_tag_bitmap, bool, 0444); |
---|
| 169 | +MODULE_PARM_DESC(shared_tag_bitmap, "Use shared tag bitmap for all submission queues for blk-mq"); |
---|
154 | 170 | |
---|
155 | 171 | static int g_irqmode = NULL_IRQ_SOFTIRQ; |
---|
156 | 172 | |
---|
.. | .. |
---|
188 | 204 | module_param_named(zone_size, g_zone_size, ulong, S_IRUGO); |
---|
189 | 205 | MODULE_PARM_DESC(zone_size, "Zone size in MB when block device is zoned. Must be power-of-two: Default: 256"); |
---|
190 | 206 | |
---|
| 207 | +static unsigned long g_zone_capacity; |
---|
| 208 | +module_param_named(zone_capacity, g_zone_capacity, ulong, 0444); |
---|
| 209 | +MODULE_PARM_DESC(zone_capacity, "Zone capacity in MB when block device is zoned. Can be less than or equal to zone size. Default: Zone size"); |
---|
| 210 | + |
---|
| 211 | +static unsigned int g_zone_nr_conv; |
---|
| 212 | +module_param_named(zone_nr_conv, g_zone_nr_conv, uint, 0444); |
---|
| 213 | +MODULE_PARM_DESC(zone_nr_conv, "Number of conventional zones when block device is zoned. Default: 0"); |
---|
| 214 | + |
---|
| 215 | +static unsigned int g_zone_max_open; |
---|
| 216 | +module_param_named(zone_max_open, g_zone_max_open, uint, 0444); |
---|
| 217 | +MODULE_PARM_DESC(zone_max_open, "Maximum number of open zones when block device is zoned. Default: 0 (no limit)"); |
---|
| 218 | + |
---|
| 219 | +static unsigned int g_zone_max_active; |
---|
| 220 | +module_param_named(zone_max_active, g_zone_max_active, uint, 0444); |
---|
| 221 | +MODULE_PARM_DESC(zone_max_active, "Maximum number of active zones when block device is zoned. Default: 0 (no limit)"); |
---|
| 222 | + |
---|
191 | 223 | static struct nullb_device *null_alloc_dev(void); |
---|
192 | 224 | static void null_free_dev(struct nullb_device *dev); |
---|
193 | 225 | static void null_del_dev(struct nullb *nullb); |
---|
.. | .. |
---|
222 | 254 | int result; |
---|
223 | 255 | |
---|
224 | 256 | result = kstrtouint(page, 0, &tmp); |
---|
225 | | - if (result) |
---|
| 257 | + if (result < 0) |
---|
226 | 258 | return result; |
---|
227 | 259 | |
---|
228 | 260 | *val = tmp; |
---|
.. | .. |
---|
236 | 268 | unsigned long tmp; |
---|
237 | 269 | |
---|
238 | 270 | result = kstrtoul(page, 0, &tmp); |
---|
239 | | - if (result) |
---|
| 271 | + if (result < 0) |
---|
240 | 272 | return result; |
---|
241 | 273 | |
---|
242 | 274 | *val = tmp; |
---|
.. | .. |
---|
250 | 282 | int result; |
---|
251 | 283 | |
---|
252 | 284 | result = kstrtobool(page, &tmp); |
---|
253 | | - if (result) |
---|
| 285 | + if (result < 0) |
---|
254 | 286 | return result; |
---|
255 | 287 | |
---|
256 | 288 | *val = tmp; |
---|
.. | .. |
---|
258 | 290 | } |
---|
259 | 291 | |
---|
260 | 292 | /* The following macro should only be used with TYPE = {uint, ulong, bool}. */ |
---|
261 | | -#define NULLB_DEVICE_ATTR(NAME, TYPE) \ |
---|
262 | | -static ssize_t \ |
---|
263 | | -nullb_device_##NAME##_show(struct config_item *item, char *page) \ |
---|
264 | | -{ \ |
---|
265 | | - return nullb_device_##TYPE##_attr_show( \ |
---|
266 | | - to_nullb_device(item)->NAME, page); \ |
---|
267 | | -} \ |
---|
268 | | -static ssize_t \ |
---|
269 | | -nullb_device_##NAME##_store(struct config_item *item, const char *page, \ |
---|
270 | | - size_t count) \ |
---|
271 | | -{ \ |
---|
272 | | - if (test_bit(NULLB_DEV_FL_CONFIGURED, &to_nullb_device(item)->flags)) \ |
---|
273 | | - return -EBUSY; \ |
---|
274 | | - return nullb_device_##TYPE##_attr_store( \ |
---|
275 | | - &to_nullb_device(item)->NAME, page, count); \ |
---|
276 | | -} \ |
---|
| 293 | +#define NULLB_DEVICE_ATTR(NAME, TYPE, APPLY) \ |
---|
| 294 | +static ssize_t \ |
---|
| 295 | +nullb_device_##NAME##_show(struct config_item *item, char *page) \ |
---|
| 296 | +{ \ |
---|
| 297 | + return nullb_device_##TYPE##_attr_show( \ |
---|
| 298 | + to_nullb_device(item)->NAME, page); \ |
---|
| 299 | +} \ |
---|
| 300 | +static ssize_t \ |
---|
| 301 | +nullb_device_##NAME##_store(struct config_item *item, const char *page, \ |
---|
| 302 | + size_t count) \ |
---|
| 303 | +{ \ |
---|
| 304 | + int (*apply_fn)(struct nullb_device *dev, TYPE new_value) = APPLY;\ |
---|
| 305 | + struct nullb_device *dev = to_nullb_device(item); \ |
---|
| 306 | + TYPE new_value = 0; \ |
---|
| 307 | + int ret; \ |
---|
| 308 | + \ |
---|
| 309 | + ret = nullb_device_##TYPE##_attr_store(&new_value, page, count);\ |
---|
| 310 | + if (ret < 0) \ |
---|
| 311 | + return ret; \ |
---|
| 312 | + if (apply_fn) \ |
---|
| 313 | + ret = apply_fn(dev, new_value); \ |
---|
| 314 | + else if (test_bit(NULLB_DEV_FL_CONFIGURED, &dev->flags)) \ |
---|
| 315 | + ret = -EBUSY; \ |
---|
| 316 | + if (ret < 0) \ |
---|
| 317 | + return ret; \ |
---|
| 318 | + dev->NAME = new_value; \ |
---|
| 319 | + return count; \ |
---|
| 320 | +} \ |
---|
277 | 321 | CONFIGFS_ATTR(nullb_device_, NAME); |
---|
278 | 322 | |
---|
279 | | -NULLB_DEVICE_ATTR(size, ulong); |
---|
280 | | -NULLB_DEVICE_ATTR(completion_nsec, ulong); |
---|
281 | | -NULLB_DEVICE_ATTR(submit_queues, uint); |
---|
282 | | -NULLB_DEVICE_ATTR(home_node, uint); |
---|
283 | | -NULLB_DEVICE_ATTR(queue_mode, uint); |
---|
284 | | -NULLB_DEVICE_ATTR(blocksize, uint); |
---|
285 | | -NULLB_DEVICE_ATTR(irqmode, uint); |
---|
286 | | -NULLB_DEVICE_ATTR(hw_queue_depth, uint); |
---|
287 | | -NULLB_DEVICE_ATTR(index, uint); |
---|
288 | | -NULLB_DEVICE_ATTR(blocking, bool); |
---|
289 | | -NULLB_DEVICE_ATTR(use_per_node_hctx, bool); |
---|
290 | | -NULLB_DEVICE_ATTR(memory_backed, bool); |
---|
291 | | -NULLB_DEVICE_ATTR(discard, bool); |
---|
292 | | -NULLB_DEVICE_ATTR(mbps, uint); |
---|
293 | | -NULLB_DEVICE_ATTR(cache_size, ulong); |
---|
294 | | -NULLB_DEVICE_ATTR(zoned, bool); |
---|
295 | | -NULLB_DEVICE_ATTR(zone_size, ulong); |
---|
| 323 | +static int nullb_apply_submit_queues(struct nullb_device *dev, |
---|
| 324 | + unsigned int submit_queues) |
---|
| 325 | +{ |
---|
| 326 | + struct nullb *nullb = dev->nullb; |
---|
| 327 | + struct blk_mq_tag_set *set; |
---|
| 328 | + |
---|
| 329 | + if (!nullb) |
---|
| 330 | + return 0; |
---|
| 331 | + |
---|
| 332 | + /* |
---|
| 333 | + * Make sure that null_init_hctx() does not access nullb->queues[] past |
---|
| 334 | + * the end of that array. |
---|
| 335 | + */ |
---|
| 336 | + if (submit_queues > nr_cpu_ids) |
---|
| 337 | + return -EINVAL; |
---|
| 338 | + set = nullb->tag_set; |
---|
| 339 | + blk_mq_update_nr_hw_queues(set, submit_queues); |
---|
| 340 | + return set->nr_hw_queues == submit_queues ? 0 : -ENOMEM; |
---|
| 341 | +} |
---|
| 342 | + |
---|
| 343 | +NULLB_DEVICE_ATTR(size, ulong, NULL); |
---|
| 344 | +NULLB_DEVICE_ATTR(completion_nsec, ulong, NULL); |
---|
| 345 | +NULLB_DEVICE_ATTR(submit_queues, uint, nullb_apply_submit_queues); |
---|
| 346 | +NULLB_DEVICE_ATTR(home_node, uint, NULL); |
---|
| 347 | +NULLB_DEVICE_ATTR(queue_mode, uint, NULL); |
---|
| 348 | +NULLB_DEVICE_ATTR(blocksize, uint, NULL); |
---|
| 349 | +NULLB_DEVICE_ATTR(irqmode, uint, NULL); |
---|
| 350 | +NULLB_DEVICE_ATTR(hw_queue_depth, uint, NULL); |
---|
| 351 | +NULLB_DEVICE_ATTR(index, uint, NULL); |
---|
| 352 | +NULLB_DEVICE_ATTR(blocking, bool, NULL); |
---|
| 353 | +NULLB_DEVICE_ATTR(use_per_node_hctx, bool, NULL); |
---|
| 354 | +NULLB_DEVICE_ATTR(memory_backed, bool, NULL); |
---|
| 355 | +NULLB_DEVICE_ATTR(discard, bool, NULL); |
---|
| 356 | +NULLB_DEVICE_ATTR(mbps, uint, NULL); |
---|
| 357 | +NULLB_DEVICE_ATTR(cache_size, ulong, NULL); |
---|
| 358 | +NULLB_DEVICE_ATTR(zoned, bool, NULL); |
---|
| 359 | +NULLB_DEVICE_ATTR(zone_size, ulong, NULL); |
---|
| 360 | +NULLB_DEVICE_ATTR(zone_capacity, ulong, NULL); |
---|
| 361 | +NULLB_DEVICE_ATTR(zone_nr_conv, uint, NULL); |
---|
| 362 | +NULLB_DEVICE_ATTR(zone_max_open, uint, NULL); |
---|
| 363 | +NULLB_DEVICE_ATTR(zone_max_active, uint, NULL); |
---|
296 | 364 | |
---|
297 | 365 | static ssize_t nullb_device_power_show(struct config_item *item, char *page) |
---|
298 | 366 | { |
---|
.. | .. |
---|
408 | 476 | &nullb_device_attr_badblocks, |
---|
409 | 477 | &nullb_device_attr_zoned, |
---|
410 | 478 | &nullb_device_attr_zone_size, |
---|
| 479 | + &nullb_device_attr_zone_capacity, |
---|
| 480 | + &nullb_device_attr_zone_nr_conv, |
---|
| 481 | + &nullb_device_attr_zone_max_open, |
---|
| 482 | + &nullb_device_attr_zone_max_active, |
---|
411 | 483 | NULL, |
---|
412 | 484 | }; |
---|
413 | 485 | |
---|
.. | .. |
---|
460 | 532 | |
---|
461 | 533 | static ssize_t memb_group_features_show(struct config_item *item, char *page) |
---|
462 | 534 | { |
---|
463 | | - return snprintf(page, PAGE_SIZE, "memory_backed,discard,bandwidth,cache,badblocks,zoned,zone_size\n"); |
---|
| 535 | + return snprintf(page, PAGE_SIZE, |
---|
| 536 | + "memory_backed,discard,bandwidth,cache,badblocks,zoned,zone_size,zone_capacity,zone_nr_conv,zone_max_open,zone_max_active\n"); |
---|
464 | 537 | } |
---|
465 | 538 | |
---|
466 | 539 | CONFIGFS_ATTR_RO(memb_group_, features); |
---|
.. | .. |
---|
521 | 594 | dev->use_per_node_hctx = g_use_per_node_hctx; |
---|
522 | 595 | dev->zoned = g_zoned; |
---|
523 | 596 | dev->zone_size = g_zone_size; |
---|
| 597 | + dev->zone_capacity = g_zone_capacity; |
---|
| 598 | + dev->zone_nr_conv = g_zone_nr_conv; |
---|
| 599 | + dev->zone_max_open = g_zone_max_open; |
---|
| 600 | + dev->zone_max_active = g_zone_max_active; |
---|
524 | 601 | return dev; |
---|
525 | 602 | } |
---|
526 | 603 | |
---|
.. | .. |
---|
529 | 606 | if (!dev) |
---|
530 | 607 | return; |
---|
531 | 608 | |
---|
532 | | - null_zone_exit(dev); |
---|
| 609 | + null_free_zoned_dev(dev); |
---|
533 | 610 | badblocks_exit(&dev->badblocks); |
---|
534 | 611 | kfree(dev); |
---|
535 | 612 | } |
---|
.. | .. |
---|
608 | 685 | |
---|
609 | 686 | static void end_cmd(struct nullb_cmd *cmd) |
---|
610 | 687 | { |
---|
611 | | - struct request_queue *q = NULL; |
---|
612 | 688 | int queue_mode = cmd->nq->dev->queue_mode; |
---|
613 | | - |
---|
614 | | - if (cmd->rq) |
---|
615 | | - q = cmd->rq->q; |
---|
616 | 689 | |
---|
617 | 690 | switch (queue_mode) { |
---|
618 | 691 | case NULL_Q_MQ: |
---|
619 | 692 | blk_mq_end_request(cmd->rq, cmd->error); |
---|
620 | 693 | return; |
---|
621 | | - case NULL_Q_RQ: |
---|
622 | | - INIT_LIST_HEAD(&cmd->rq->queuelist); |
---|
623 | | - blk_end_request_all(cmd->rq, cmd->error); |
---|
624 | | - break; |
---|
625 | 694 | case NULL_Q_BIO: |
---|
626 | 695 | cmd->bio->bi_status = cmd->error; |
---|
627 | 696 | bio_endio(cmd->bio); |
---|
.. | .. |
---|
629 | 698 | } |
---|
630 | 699 | |
---|
631 | 700 | free_cmd(cmd); |
---|
632 | | - |
---|
633 | | - /* Restart queue if needed, as we are freeing a tag */ |
---|
634 | | - if (queue_mode == NULL_Q_RQ && blk_queue_stopped(q)) { |
---|
635 | | - unsigned long flags; |
---|
636 | | - |
---|
637 | | - spin_lock_irqsave(q->queue_lock, flags); |
---|
638 | | - blk_start_queue_async(q); |
---|
639 | | - spin_unlock_irqrestore(q->queue_lock, flags); |
---|
640 | | - } |
---|
641 | 701 | } |
---|
642 | 702 | |
---|
643 | 703 | static enum hrtimer_restart null_cmd_timer_expired(struct hrtimer *timer) |
---|
.. | .. |
---|
654 | 714 | hrtimer_start(&cmd->timer, kt, HRTIMER_MODE_REL); |
---|
655 | 715 | } |
---|
656 | 716 | |
---|
657 | | -static void null_softirq_done_fn(struct request *rq) |
---|
| 717 | +static void null_complete_rq(struct request *rq) |
---|
658 | 718 | { |
---|
659 | | - struct nullb *nullb = rq->q->queuedata; |
---|
660 | | - |
---|
661 | | - if (nullb->dev->queue_mode == NULL_Q_MQ) |
---|
662 | | - end_cmd(blk_mq_rq_to_pdu(rq)); |
---|
663 | | - else |
---|
664 | | - end_cmd(rq->special); |
---|
| 719 | + end_cmd(blk_mq_rq_to_pdu(rq)); |
---|
665 | 720 | } |
---|
666 | 721 | |
---|
667 | 722 | static struct nullb_page *null_alloc_page(gfp_t gfp_flags) |
---|
.. | .. |
---|
1011 | 1066 | return 0; |
---|
1012 | 1067 | } |
---|
1013 | 1068 | |
---|
| 1069 | +static void nullb_fill_pattern(struct nullb *nullb, struct page *page, |
---|
| 1070 | + unsigned int len, unsigned int off) |
---|
| 1071 | +{ |
---|
| 1072 | + void *dst; |
---|
| 1073 | + |
---|
| 1074 | + dst = kmap_atomic(page); |
---|
| 1075 | + memset(dst + off, 0xFF, len); |
---|
| 1076 | + kunmap_atomic(dst); |
---|
| 1077 | +} |
---|
| 1078 | + |
---|
1014 | 1079 | static void null_handle_discard(struct nullb *nullb, sector_t sector, size_t n) |
---|
1015 | 1080 | { |
---|
1016 | 1081 | size_t temp; |
---|
.. | .. |
---|
1051 | 1116 | unsigned int len, unsigned int off, bool is_write, sector_t sector, |
---|
1052 | 1117 | bool is_fua) |
---|
1053 | 1118 | { |
---|
| 1119 | + struct nullb_device *dev = nullb->dev; |
---|
| 1120 | + unsigned int valid_len = len; |
---|
1054 | 1121 | int err = 0; |
---|
1055 | 1122 | |
---|
1056 | 1123 | if (!is_write) { |
---|
1057 | | - err = copy_from_nullb(nullb, page, off, sector, len); |
---|
| 1124 | + if (dev->zoned) |
---|
| 1125 | + valid_len = null_zone_valid_read_len(nullb, |
---|
| 1126 | + sector, len); |
---|
| 1127 | + |
---|
| 1128 | + if (valid_len) { |
---|
| 1129 | + err = copy_from_nullb(nullb, page, off, |
---|
| 1130 | + sector, valid_len); |
---|
| 1131 | + off += valid_len; |
---|
| 1132 | + len -= valid_len; |
---|
| 1133 | + } |
---|
| 1134 | + |
---|
| 1135 | + if (len) |
---|
| 1136 | + nullb_fill_pattern(nullb, page, len, off); |
---|
1058 | 1137 | flush_dcache_page(page); |
---|
1059 | 1138 | } else { |
---|
1060 | 1139 | flush_dcache_page(page); |
---|
.. | .. |
---|
1121 | 1200 | len = bvec.bv_len; |
---|
1122 | 1201 | err = null_transfer(nullb, bvec.bv_page, len, bvec.bv_offset, |
---|
1123 | 1202 | op_is_write(bio_op(bio)), sector, |
---|
1124 | | - bio_op(bio) & REQ_FUA); |
---|
| 1203 | + bio->bi_opf & REQ_FUA); |
---|
1125 | 1204 | if (err) { |
---|
1126 | 1205 | spin_unlock_irq(&nullb->lock); |
---|
1127 | 1206 | return err; |
---|
.. | .. |
---|
1138 | 1217 | |
---|
1139 | 1218 | if (nullb->dev->queue_mode == NULL_Q_MQ) |
---|
1140 | 1219 | blk_mq_stop_hw_queues(q); |
---|
1141 | | - else { |
---|
1142 | | - spin_lock_irq(q->queue_lock); |
---|
1143 | | - blk_stop_queue(q); |
---|
1144 | | - spin_unlock_irq(q->queue_lock); |
---|
1145 | | - } |
---|
1146 | 1220 | } |
---|
1147 | 1221 | |
---|
1148 | 1222 | static void null_restart_queue_async(struct nullb *nullb) |
---|
1149 | 1223 | { |
---|
1150 | 1224 | struct request_queue *q = nullb->q; |
---|
1151 | | - unsigned long flags; |
---|
1152 | 1225 | |
---|
1153 | 1226 | if (nullb->dev->queue_mode == NULL_Q_MQ) |
---|
1154 | 1227 | blk_mq_start_stopped_hw_queues(q, true); |
---|
1155 | | - else { |
---|
1156 | | - spin_lock_irqsave(q->queue_lock, flags); |
---|
1157 | | - blk_start_queue_async(q); |
---|
1158 | | - spin_unlock_irqrestore(q->queue_lock, flags); |
---|
1159 | | - } |
---|
1160 | 1228 | } |
---|
1161 | 1229 | |
---|
1162 | | -static bool cmd_report_zone(struct nullb *nullb, struct nullb_cmd *cmd) |
---|
1163 | | -{ |
---|
1164 | | - struct nullb_device *dev = cmd->nq->dev; |
---|
1165 | | - |
---|
1166 | | - if (dev->queue_mode == NULL_Q_BIO) { |
---|
1167 | | - if (bio_op(cmd->bio) == REQ_OP_ZONE_REPORT) { |
---|
1168 | | - cmd->error = null_zone_report(nullb, cmd->bio); |
---|
1169 | | - return true; |
---|
1170 | | - } |
---|
1171 | | - } else { |
---|
1172 | | - if (req_op(cmd->rq) == REQ_OP_ZONE_REPORT) { |
---|
1173 | | - cmd->error = null_zone_report(nullb, cmd->rq->bio); |
---|
1174 | | - return true; |
---|
1175 | | - } |
---|
1176 | | - } |
---|
1177 | | - |
---|
1178 | | - return false; |
---|
1179 | | -} |
---|
1180 | | - |
---|
1181 | | -static blk_status_t null_handle_cmd(struct nullb_cmd *cmd) |
---|
| 1230 | +static inline blk_status_t null_handle_throttled(struct nullb_cmd *cmd) |
---|
1182 | 1231 | { |
---|
1183 | 1232 | struct nullb_device *dev = cmd->nq->dev; |
---|
1184 | 1233 | struct nullb *nullb = dev->nullb; |
---|
1185 | | - int err = 0; |
---|
| 1234 | + blk_status_t sts = BLK_STS_OK; |
---|
| 1235 | + struct request *rq = cmd->rq; |
---|
1186 | 1236 | |
---|
1187 | | - if (cmd_report_zone(nullb, cmd)) |
---|
1188 | | - goto out; |
---|
| 1237 | + if (!hrtimer_active(&nullb->bw_timer)) |
---|
| 1238 | + hrtimer_restart(&nullb->bw_timer); |
---|
1189 | 1239 | |
---|
1190 | | - if (test_bit(NULLB_DEV_FL_THROTTLED, &dev->flags)) { |
---|
1191 | | - struct request *rq = cmd->rq; |
---|
1192 | | - |
---|
1193 | | - if (!hrtimer_active(&nullb->bw_timer)) |
---|
1194 | | - hrtimer_restart(&nullb->bw_timer); |
---|
1195 | | - |
---|
1196 | | - if (atomic_long_sub_return(blk_rq_bytes(rq), |
---|
1197 | | - &nullb->cur_bytes) < 0) { |
---|
1198 | | - null_stop_queue(nullb); |
---|
1199 | | - /* race with timer */ |
---|
1200 | | - if (atomic_long_read(&nullb->cur_bytes) > 0) |
---|
1201 | | - null_restart_queue_async(nullb); |
---|
1202 | | - if (dev->queue_mode == NULL_Q_RQ) { |
---|
1203 | | - struct request_queue *q = nullb->q; |
---|
1204 | | - |
---|
1205 | | - spin_lock_irq(q->queue_lock); |
---|
1206 | | - rq->rq_flags |= RQF_DONTPREP; |
---|
1207 | | - blk_requeue_request(q, rq); |
---|
1208 | | - spin_unlock_irq(q->queue_lock); |
---|
1209 | | - return BLK_STS_OK; |
---|
1210 | | - } else |
---|
1211 | | - /* requeue request */ |
---|
1212 | | - return BLK_STS_DEV_RESOURCE; |
---|
1213 | | - } |
---|
| 1240 | + if (atomic_long_sub_return(blk_rq_bytes(rq), &nullb->cur_bytes) < 0) { |
---|
| 1241 | + null_stop_queue(nullb); |
---|
| 1242 | + /* race with timer */ |
---|
| 1243 | + if (atomic_long_read(&nullb->cur_bytes) > 0) |
---|
| 1244 | + null_restart_queue_async(nullb); |
---|
| 1245 | + /* requeue request */ |
---|
| 1246 | + sts = BLK_STS_DEV_RESOURCE; |
---|
1214 | 1247 | } |
---|
| 1248 | + return sts; |
---|
| 1249 | +} |
---|
1215 | 1250 | |
---|
1216 | | - if (nullb->dev->badblocks.shift != -1) { |
---|
1217 | | - int bad_sectors; |
---|
1218 | | - sector_t sector, size, first_bad; |
---|
1219 | | - bool is_flush = true; |
---|
| 1251 | +static inline blk_status_t null_handle_badblocks(struct nullb_cmd *cmd, |
---|
| 1252 | + sector_t sector, |
---|
| 1253 | + sector_t nr_sectors) |
---|
| 1254 | +{ |
---|
| 1255 | + struct badblocks *bb = &cmd->nq->dev->badblocks; |
---|
| 1256 | + sector_t first_bad; |
---|
| 1257 | + int bad_sectors; |
---|
1220 | 1258 | |
---|
1221 | | - if (dev->queue_mode == NULL_Q_BIO && |
---|
1222 | | - bio_op(cmd->bio) != REQ_OP_FLUSH) { |
---|
1223 | | - is_flush = false; |
---|
1224 | | - sector = cmd->bio->bi_iter.bi_sector; |
---|
1225 | | - size = bio_sectors(cmd->bio); |
---|
1226 | | - } |
---|
1227 | | - if (dev->queue_mode != NULL_Q_BIO && |
---|
1228 | | - req_op(cmd->rq) != REQ_OP_FLUSH) { |
---|
1229 | | - is_flush = false; |
---|
1230 | | - sector = blk_rq_pos(cmd->rq); |
---|
1231 | | - size = blk_rq_sectors(cmd->rq); |
---|
1232 | | - } |
---|
1233 | | - if (!is_flush && badblocks_check(&nullb->dev->badblocks, sector, |
---|
1234 | | - size, &first_bad, &bad_sectors)) { |
---|
1235 | | - cmd->error = BLK_STS_IOERR; |
---|
1236 | | - goto out; |
---|
1237 | | - } |
---|
| 1259 | + if (badblocks_check(bb, sector, nr_sectors, &first_bad, &bad_sectors)) |
---|
| 1260 | + return BLK_STS_IOERR; |
---|
| 1261 | + |
---|
| 1262 | + return BLK_STS_OK; |
---|
| 1263 | +} |
---|
| 1264 | + |
---|
| 1265 | +static inline blk_status_t null_handle_memory_backed(struct nullb_cmd *cmd, |
---|
| 1266 | + enum req_opf op) |
---|
| 1267 | +{ |
---|
| 1268 | + struct nullb_device *dev = cmd->nq->dev; |
---|
| 1269 | + int err; |
---|
| 1270 | + |
---|
| 1271 | + if (dev->queue_mode == NULL_Q_BIO) |
---|
| 1272 | + err = null_handle_bio(cmd); |
---|
| 1273 | + else |
---|
| 1274 | + err = null_handle_rq(cmd); |
---|
| 1275 | + |
---|
| 1276 | + return errno_to_blk_status(err); |
---|
| 1277 | +} |
---|
| 1278 | + |
---|
| 1279 | +static void nullb_zero_read_cmd_buffer(struct nullb_cmd *cmd) |
---|
| 1280 | +{ |
---|
| 1281 | + struct nullb_device *dev = cmd->nq->dev; |
---|
| 1282 | + struct bio *bio; |
---|
| 1283 | + |
---|
| 1284 | + if (dev->memory_backed) |
---|
| 1285 | + return; |
---|
| 1286 | + |
---|
| 1287 | + if (dev->queue_mode == NULL_Q_BIO && bio_op(cmd->bio) == REQ_OP_READ) { |
---|
| 1288 | + zero_fill_bio(cmd->bio); |
---|
| 1289 | + } else if (req_op(cmd->rq) == REQ_OP_READ) { |
---|
| 1290 | + __rq_for_each_bio(bio, cmd->rq) |
---|
| 1291 | + zero_fill_bio(bio); |
---|
1238 | 1292 | } |
---|
| 1293 | +} |
---|
1239 | 1294 | |
---|
1240 | | - if (dev->memory_backed) { |
---|
1241 | | - if (dev->queue_mode == NULL_Q_BIO) { |
---|
1242 | | - if (bio_op(cmd->bio) == REQ_OP_FLUSH) |
---|
1243 | | - err = null_handle_flush(nullb); |
---|
1244 | | - else |
---|
1245 | | - err = null_handle_bio(cmd); |
---|
1246 | | - } else { |
---|
1247 | | - if (req_op(cmd->rq) == REQ_OP_FLUSH) |
---|
1248 | | - err = null_handle_flush(nullb); |
---|
1249 | | - else |
---|
1250 | | - err = null_handle_rq(cmd); |
---|
1251 | | - } |
---|
1252 | | - } |
---|
1253 | | - cmd->error = errno_to_blk_status(err); |
---|
| 1295 | +static inline void nullb_complete_cmd(struct nullb_cmd *cmd) |
---|
| 1296 | +{ |
---|
| 1297 | + /* |
---|
| 1298 | + * Since root privileges are required to configure the null_blk |
---|
| 1299 | + * driver, it is fine that this driver does not initialize the |
---|
| 1300 | + * data buffers of read commands. Zero-initialize these buffers |
---|
| 1301 | + * anyway if KMSAN is enabled to prevent that KMSAN complains |
---|
| 1302 | + * about null_blk not initializing read data buffers. |
---|
| 1303 | + */ |
---|
| 1304 | + if (IS_ENABLED(CONFIG_KMSAN)) |
---|
| 1305 | + nullb_zero_read_cmd_buffer(cmd); |
---|
1254 | 1306 | |
---|
1255 | | - if (!cmd->error && dev->zoned) { |
---|
1256 | | - sector_t sector; |
---|
1257 | | - unsigned int nr_sectors; |
---|
1258 | | - int op; |
---|
1259 | | - |
---|
1260 | | - if (dev->queue_mode == NULL_Q_BIO) { |
---|
1261 | | - op = bio_op(cmd->bio); |
---|
1262 | | - sector = cmd->bio->bi_iter.bi_sector; |
---|
1263 | | - nr_sectors = cmd->bio->bi_iter.bi_size >> 9; |
---|
1264 | | - } else { |
---|
1265 | | - op = req_op(cmd->rq); |
---|
1266 | | - sector = blk_rq_pos(cmd->rq); |
---|
1267 | | - nr_sectors = blk_rq_sectors(cmd->rq); |
---|
1268 | | - } |
---|
1269 | | - |
---|
1270 | | - if (op == REQ_OP_WRITE) |
---|
1271 | | - null_zone_write(cmd, sector, nr_sectors); |
---|
1272 | | - else if (op == REQ_OP_ZONE_RESET) |
---|
1273 | | - null_zone_reset(cmd, sector); |
---|
1274 | | - } |
---|
1275 | | -out: |
---|
1276 | 1307 | /* Complete IO by inline, softirq or timer */ |
---|
1277 | | - switch (dev->irqmode) { |
---|
| 1308 | + switch (cmd->nq->dev->irqmode) { |
---|
1278 | 1309 | case NULL_IRQ_SOFTIRQ: |
---|
1279 | | - switch (dev->queue_mode) { |
---|
| 1310 | + switch (cmd->nq->dev->queue_mode) { |
---|
1280 | 1311 | case NULL_Q_MQ: |
---|
1281 | | - blk_mq_complete_request(cmd->rq); |
---|
1282 | | - break; |
---|
1283 | | - case NULL_Q_RQ: |
---|
1284 | | - blk_complete_request(cmd->rq); |
---|
| 1312 | + if (likely(!blk_should_fake_timeout(cmd->rq->q))) |
---|
| 1313 | + blk_mq_complete_request(cmd->rq); |
---|
1285 | 1314 | break; |
---|
1286 | 1315 | case NULL_Q_BIO: |
---|
1287 | 1316 | /* |
---|
.. | .. |
---|
1298 | 1327 | null_cmd_end_timer(cmd); |
---|
1299 | 1328 | break; |
---|
1300 | 1329 | } |
---|
| 1330 | +} |
---|
| 1331 | + |
---|
| 1332 | +blk_status_t null_process_cmd(struct nullb_cmd *cmd, |
---|
| 1333 | + enum req_opf op, sector_t sector, |
---|
| 1334 | + unsigned int nr_sectors) |
---|
| 1335 | +{ |
---|
| 1336 | + struct nullb_device *dev = cmd->nq->dev; |
---|
| 1337 | + blk_status_t ret; |
---|
| 1338 | + |
---|
| 1339 | + if (dev->badblocks.shift != -1) { |
---|
| 1340 | + ret = null_handle_badblocks(cmd, sector, nr_sectors); |
---|
| 1341 | + if (ret != BLK_STS_OK) |
---|
| 1342 | + return ret; |
---|
| 1343 | + } |
---|
| 1344 | + |
---|
| 1345 | + if (dev->memory_backed) |
---|
| 1346 | + return null_handle_memory_backed(cmd, op); |
---|
| 1347 | + |
---|
| 1348 | + return BLK_STS_OK; |
---|
| 1349 | +} |
---|
| 1350 | + |
---|
| 1351 | +static blk_status_t null_handle_cmd(struct nullb_cmd *cmd, sector_t sector, |
---|
| 1352 | + sector_t nr_sectors, enum req_opf op) |
---|
| 1353 | +{ |
---|
| 1354 | + struct nullb_device *dev = cmd->nq->dev; |
---|
| 1355 | + struct nullb *nullb = dev->nullb; |
---|
| 1356 | + blk_status_t sts; |
---|
| 1357 | + |
---|
| 1358 | + if (test_bit(NULLB_DEV_FL_THROTTLED, &dev->flags)) { |
---|
| 1359 | + sts = null_handle_throttled(cmd); |
---|
| 1360 | + if (sts != BLK_STS_OK) |
---|
| 1361 | + return sts; |
---|
| 1362 | + } |
---|
| 1363 | + |
---|
| 1364 | + if (op == REQ_OP_FLUSH) { |
---|
| 1365 | + cmd->error = errno_to_blk_status(null_handle_flush(nullb)); |
---|
| 1366 | + goto out; |
---|
| 1367 | + } |
---|
| 1368 | + |
---|
| 1369 | + if (dev->zoned) |
---|
| 1370 | + sts = null_process_zoned_cmd(cmd, op, sector, nr_sectors); |
---|
| 1371 | + else |
---|
| 1372 | + sts = null_process_cmd(cmd, op, sector, nr_sectors); |
---|
| 1373 | + |
---|
| 1374 | + /* Do not overwrite errors (e.g. timeout errors) */ |
---|
| 1375 | + if (cmd->error == BLK_STS_OK) |
---|
| 1376 | + cmd->error = sts; |
---|
| 1377 | + |
---|
| 1378 | +out: |
---|
| 1379 | + nullb_complete_cmd(cmd); |
---|
1301 | 1380 | return BLK_STS_OK; |
---|
1302 | 1381 | } |
---|
1303 | 1382 | |
---|
.. | .. |
---|
1338 | 1417 | return &nullb->queues[index]; |
---|
1339 | 1418 | } |
---|
1340 | 1419 | |
---|
1341 | | -static blk_qc_t null_queue_bio(struct request_queue *q, struct bio *bio) |
---|
| 1420 | +static blk_qc_t null_submit_bio(struct bio *bio) |
---|
1342 | 1421 | { |
---|
1343 | | - struct nullb *nullb = q->queuedata; |
---|
| 1422 | + sector_t sector = bio->bi_iter.bi_sector; |
---|
| 1423 | + sector_t nr_sectors = bio_sectors(bio); |
---|
| 1424 | + struct nullb *nullb = bio->bi_disk->private_data; |
---|
1344 | 1425 | struct nullb_queue *nq = nullb_to_queue(nullb); |
---|
1345 | 1426 | struct nullb_cmd *cmd; |
---|
1346 | 1427 | |
---|
1347 | 1428 | cmd = alloc_cmd(nq, 1); |
---|
1348 | 1429 | cmd->bio = bio; |
---|
1349 | 1430 | |
---|
1350 | | - null_handle_cmd(cmd); |
---|
| 1431 | + null_handle_cmd(cmd, sector, nr_sectors, bio_op(bio)); |
---|
1351 | 1432 | return BLK_QC_T_NONE; |
---|
1352 | | -} |
---|
1353 | | - |
---|
1354 | | -static enum blk_eh_timer_return null_rq_timed_out_fn(struct request *rq) |
---|
1355 | | -{ |
---|
1356 | | - pr_info("null: rq %p timed out\n", rq); |
---|
1357 | | - __blk_complete_request(rq); |
---|
1358 | | - return BLK_EH_DONE; |
---|
1359 | | -} |
---|
1360 | | - |
---|
1361 | | -static int null_rq_prep_fn(struct request_queue *q, struct request *req) |
---|
1362 | | -{ |
---|
1363 | | - struct nullb *nullb = q->queuedata; |
---|
1364 | | - struct nullb_queue *nq = nullb_to_queue(nullb); |
---|
1365 | | - struct nullb_cmd *cmd; |
---|
1366 | | - |
---|
1367 | | - cmd = alloc_cmd(nq, 0); |
---|
1368 | | - if (cmd) { |
---|
1369 | | - cmd->rq = req; |
---|
1370 | | - req->special = cmd; |
---|
1371 | | - return BLKPREP_OK; |
---|
1372 | | - } |
---|
1373 | | - blk_stop_queue(q); |
---|
1374 | | - |
---|
1375 | | - return BLKPREP_DEFER; |
---|
1376 | 1433 | } |
---|
1377 | 1434 | |
---|
1378 | 1435 | static bool should_timeout_request(struct request *rq) |
---|
.. | .. |
---|
1393 | 1450 | return false; |
---|
1394 | 1451 | } |
---|
1395 | 1452 | |
---|
1396 | | -static void null_request_fn(struct request_queue *q) |
---|
1397 | | -{ |
---|
1398 | | - struct request *rq; |
---|
1399 | | - |
---|
1400 | | - while ((rq = blk_fetch_request(q)) != NULL) { |
---|
1401 | | - struct nullb_cmd *cmd = rq->special; |
---|
1402 | | - |
---|
1403 | | - /* just ignore the request */ |
---|
1404 | | - if (should_timeout_request(rq)) |
---|
1405 | | - continue; |
---|
1406 | | - if (should_requeue_request(rq)) { |
---|
1407 | | - blk_requeue_request(q, rq); |
---|
1408 | | - continue; |
---|
1409 | | - } |
---|
1410 | | - |
---|
1411 | | - spin_unlock_irq(q->queue_lock); |
---|
1412 | | - null_handle_cmd(cmd); |
---|
1413 | | - spin_lock_irq(q->queue_lock); |
---|
1414 | | - } |
---|
1415 | | -} |
---|
1416 | | - |
---|
1417 | 1453 | static enum blk_eh_timer_return null_timeout_rq(struct request *rq, bool res) |
---|
1418 | 1454 | { |
---|
1419 | | - pr_info("null: rq %p timed out\n", rq); |
---|
1420 | | - blk_mq_complete_request(rq); |
---|
| 1455 | + struct nullb_cmd *cmd = blk_mq_rq_to_pdu(rq); |
---|
| 1456 | + |
---|
| 1457 | + pr_info("rq %p timed out\n", rq); |
---|
| 1458 | + |
---|
| 1459 | + /* |
---|
| 1460 | + * If the device is marked as blocking (i.e. memory backed or zoned |
---|
| 1461 | + * device), the submission path may be blocked waiting for resources |
---|
| 1462 | + * and cause real timeouts. For these real timeouts, the submission |
---|
| 1463 | + * path will complete the request using blk_mq_complete_request(). |
---|
| 1464 | + * Only fake timeouts need to execute blk_mq_complete_request() here. |
---|
| 1465 | + */ |
---|
| 1466 | + cmd->error = BLK_STS_TIMEOUT; |
---|
| 1467 | + if (cmd->fake_timeout) |
---|
| 1468 | + blk_mq_complete_request(rq); |
---|
1421 | 1469 | return BLK_EH_DONE; |
---|
1422 | 1470 | } |
---|
1423 | 1471 | |
---|
.. | .. |
---|
1426 | 1474 | { |
---|
1427 | 1475 | struct nullb_cmd *cmd = blk_mq_rq_to_pdu(bd->rq); |
---|
1428 | 1476 | struct nullb_queue *nq = hctx->driver_data; |
---|
| 1477 | + sector_t nr_sectors = blk_rq_sectors(bd->rq); |
---|
| 1478 | + sector_t sector = blk_rq_pos(bd->rq); |
---|
1429 | 1479 | |
---|
1430 | 1480 | might_sleep_if(hctx->flags & BLK_MQ_F_BLOCKING); |
---|
1431 | 1481 | |
---|
.. | .. |
---|
1436 | 1486 | cmd->rq = bd->rq; |
---|
1437 | 1487 | cmd->error = BLK_STS_OK; |
---|
1438 | 1488 | cmd->nq = nq; |
---|
| 1489 | + cmd->fake_timeout = should_timeout_request(bd->rq); |
---|
1439 | 1490 | |
---|
1440 | 1491 | blk_mq_start_request(bd->rq); |
---|
1441 | 1492 | |
---|
.. | .. |
---|
1452 | 1503 | return BLK_STS_OK; |
---|
1453 | 1504 | } |
---|
1454 | 1505 | } |
---|
1455 | | - if (should_timeout_request(bd->rq)) |
---|
| 1506 | + if (cmd->fake_timeout) |
---|
1456 | 1507 | return BLK_STS_OK; |
---|
1457 | 1508 | |
---|
1458 | | - return null_handle_cmd(cmd); |
---|
| 1509 | + return null_handle_cmd(cmd, sector, nr_sectors, req_op(bd->rq)); |
---|
1459 | 1510 | } |
---|
1460 | | - |
---|
1461 | | -static const struct blk_mq_ops null_mq_ops = { |
---|
1462 | | - .queue_rq = null_queue_rq, |
---|
1463 | | - .complete = null_softirq_done_fn, |
---|
1464 | | - .timeout = null_timeout_rq, |
---|
1465 | | -}; |
---|
1466 | 1511 | |
---|
1467 | 1512 | static void cleanup_queue(struct nullb_queue *nq) |
---|
1468 | 1513 | { |
---|
.. | .. |
---|
1479 | 1524 | |
---|
1480 | 1525 | kfree(nullb->queues); |
---|
1481 | 1526 | } |
---|
| 1527 | + |
---|
| 1528 | +static void null_exit_hctx(struct blk_mq_hw_ctx *hctx, unsigned int hctx_idx) |
---|
| 1529 | +{ |
---|
| 1530 | + struct nullb_queue *nq = hctx->driver_data; |
---|
| 1531 | + struct nullb *nullb = nq->dev->nullb; |
---|
| 1532 | + |
---|
| 1533 | + nullb->nr_queues--; |
---|
| 1534 | +} |
---|
| 1535 | + |
---|
| 1536 | +static void null_init_queue(struct nullb *nullb, struct nullb_queue *nq) |
---|
| 1537 | +{ |
---|
| 1538 | + init_waitqueue_head(&nq->wait); |
---|
| 1539 | + nq->queue_depth = nullb->queue_depth; |
---|
| 1540 | + nq->dev = nullb->dev; |
---|
| 1541 | +} |
---|
| 1542 | + |
---|
| 1543 | +static int null_init_hctx(struct blk_mq_hw_ctx *hctx, void *driver_data, |
---|
| 1544 | + unsigned int hctx_idx) |
---|
| 1545 | +{ |
---|
| 1546 | + struct nullb *nullb = hctx->queue->queuedata; |
---|
| 1547 | + struct nullb_queue *nq; |
---|
| 1548 | + |
---|
| 1549 | +#ifdef CONFIG_BLK_DEV_NULL_BLK_FAULT_INJECTION |
---|
| 1550 | + if (g_init_hctx_str[0] && should_fail(&null_init_hctx_attr, 1)) |
---|
| 1551 | + return -EFAULT; |
---|
| 1552 | +#endif |
---|
| 1553 | + |
---|
| 1554 | + nq = &nullb->queues[hctx_idx]; |
---|
| 1555 | + hctx->driver_data = nq; |
---|
| 1556 | + null_init_queue(nullb, nq); |
---|
| 1557 | + nullb->nr_queues++; |
---|
| 1558 | + |
---|
| 1559 | + return 0; |
---|
| 1560 | +} |
---|
| 1561 | + |
---|
| 1562 | +static const struct blk_mq_ops null_mq_ops = { |
---|
| 1563 | + .queue_rq = null_queue_rq, |
---|
| 1564 | + .complete = null_complete_rq, |
---|
| 1565 | + .timeout = null_timeout_rq, |
---|
| 1566 | + .init_hctx = null_init_hctx, |
---|
| 1567 | + .exit_hctx = null_exit_hctx, |
---|
| 1568 | +}; |
---|
1482 | 1569 | |
---|
1483 | 1570 | static void null_del_dev(struct nullb *nullb) |
---|
1484 | 1571 | { |
---|
.. | .. |
---|
1517 | 1604 | { |
---|
1518 | 1605 | if (nullb->dev->discard == false) |
---|
1519 | 1606 | return; |
---|
| 1607 | + |
---|
| 1608 | + if (nullb->dev->zoned) { |
---|
| 1609 | + nullb->dev->discard = false; |
---|
| 1610 | + pr_info("discard option is ignored in zoned mode\n"); |
---|
| 1611 | + return; |
---|
| 1612 | + } |
---|
| 1613 | + |
---|
1520 | 1614 | nullb->q->limits.discard_granularity = nullb->dev->blocksize; |
---|
1521 | 1615 | nullb->q->limits.discard_alignment = nullb->dev->blocksize; |
---|
1522 | 1616 | blk_queue_max_discard_sectors(nullb->q, UINT_MAX >> 9); |
---|
1523 | 1617 | blk_queue_flag_set(QUEUE_FLAG_DISCARD, nullb->q); |
---|
1524 | 1618 | } |
---|
1525 | 1619 | |
---|
1526 | | -static int null_open(struct block_device *bdev, fmode_t mode) |
---|
1527 | | -{ |
---|
1528 | | - return 0; |
---|
1529 | | -} |
---|
1530 | | - |
---|
1531 | | -static void null_release(struct gendisk *disk, fmode_t mode) |
---|
1532 | | -{ |
---|
1533 | | -} |
---|
1534 | | - |
---|
1535 | | -static const struct block_device_operations null_fops = { |
---|
1536 | | - .owner = THIS_MODULE, |
---|
1537 | | - .open = null_open, |
---|
1538 | | - .release = null_release, |
---|
| 1620 | +static const struct block_device_operations null_bio_ops = { |
---|
| 1621 | + .owner = THIS_MODULE, |
---|
| 1622 | + .submit_bio = null_submit_bio, |
---|
| 1623 | + .report_zones = null_report_zones, |
---|
1539 | 1624 | }; |
---|
1540 | 1625 | |
---|
1541 | | -static void null_init_queue(struct nullb *nullb, struct nullb_queue *nq) |
---|
1542 | | -{ |
---|
1543 | | - BUG_ON(!nullb); |
---|
1544 | | - BUG_ON(!nq); |
---|
1545 | | - |
---|
1546 | | - init_waitqueue_head(&nq->wait); |
---|
1547 | | - nq->queue_depth = nullb->queue_depth; |
---|
1548 | | - nq->dev = nullb->dev; |
---|
1549 | | -} |
---|
1550 | | - |
---|
1551 | | -static void null_init_queues(struct nullb *nullb) |
---|
1552 | | -{ |
---|
1553 | | - struct request_queue *q = nullb->q; |
---|
1554 | | - struct blk_mq_hw_ctx *hctx; |
---|
1555 | | - struct nullb_queue *nq; |
---|
1556 | | - int i; |
---|
1557 | | - |
---|
1558 | | - queue_for_each_hw_ctx(q, hctx, i) { |
---|
1559 | | - if (!hctx->nr_ctx || !hctx->tags) |
---|
1560 | | - continue; |
---|
1561 | | - nq = &nullb->queues[i]; |
---|
1562 | | - hctx->driver_data = nq; |
---|
1563 | | - null_init_queue(nullb, nq); |
---|
1564 | | - nullb->nr_queues++; |
---|
1565 | | - } |
---|
1566 | | -} |
---|
| 1626 | +static const struct block_device_operations null_rq_ops = { |
---|
| 1627 | + .owner = THIS_MODULE, |
---|
| 1628 | + .report_zones = null_report_zones, |
---|
| 1629 | +}; |
---|
1567 | 1630 | |
---|
1568 | 1631 | static int setup_commands(struct nullb_queue *nq) |
---|
1569 | 1632 | { |
---|
.. | .. |
---|
1583 | 1646 | |
---|
1584 | 1647 | for (i = 0; i < nq->queue_depth; i++) { |
---|
1585 | 1648 | cmd = &nq->cmds[i]; |
---|
1586 | | - INIT_LIST_HEAD(&cmd->list); |
---|
1587 | | - cmd->ll_list.next = NULL; |
---|
1588 | 1649 | cmd->tag = -1U; |
---|
1589 | 1650 | } |
---|
1590 | 1651 | |
---|
.. | .. |
---|
1593 | 1654 | |
---|
1594 | 1655 | static int setup_queues(struct nullb *nullb) |
---|
1595 | 1656 | { |
---|
1596 | | - nullb->queues = kcalloc(nullb->dev->submit_queues, |
---|
1597 | | - sizeof(struct nullb_queue), |
---|
| 1657 | + nullb->queues = kcalloc(nr_cpu_ids, sizeof(struct nullb_queue), |
---|
1598 | 1658 | GFP_KERNEL); |
---|
1599 | 1659 | if (!nullb->queues) |
---|
1600 | 1660 | return -ENOMEM; |
---|
1601 | 1661 | |
---|
1602 | | - nullb->nr_queues = 0; |
---|
1603 | 1662 | nullb->queue_depth = nullb->dev->hw_queue_depth; |
---|
1604 | 1663 | |
---|
1605 | 1664 | return 0; |
---|
.. | .. |
---|
1625 | 1684 | |
---|
1626 | 1685 | static int null_gendisk_register(struct nullb *nullb) |
---|
1627 | 1686 | { |
---|
| 1687 | + sector_t size = ((sector_t)nullb->dev->size * SZ_1M) >> SECTOR_SHIFT; |
---|
1628 | 1688 | struct gendisk *disk; |
---|
1629 | | - sector_t size; |
---|
1630 | 1689 | |
---|
1631 | 1690 | disk = nullb->disk = alloc_disk_node(1, nullb->dev->home_node); |
---|
1632 | 1691 | if (!disk) |
---|
1633 | 1692 | return -ENOMEM; |
---|
1634 | | - size = (sector_t)nullb->dev->size * 1024 * 1024ULL; |
---|
1635 | | - set_capacity(disk, size >> 9); |
---|
| 1693 | + set_capacity(disk, size); |
---|
1636 | 1694 | |
---|
1637 | 1695 | disk->flags |= GENHD_FL_EXT_DEVT | GENHD_FL_SUPPRESS_PARTITION_INFO; |
---|
1638 | 1696 | disk->major = null_major; |
---|
1639 | 1697 | disk->first_minor = nullb->index; |
---|
1640 | | - disk->fops = &null_fops; |
---|
| 1698 | + if (queue_is_mq(nullb->q)) |
---|
| 1699 | + disk->fops = &null_rq_ops; |
---|
| 1700 | + else |
---|
| 1701 | + disk->fops = &null_bio_ops; |
---|
1641 | 1702 | disk->private_data = nullb; |
---|
1642 | 1703 | disk->queue = nullb->q; |
---|
1643 | 1704 | strncpy(disk->disk_name, nullb->disk_name, DISK_NAME_LEN); |
---|
| 1705 | + |
---|
| 1706 | + if (nullb->dev->zoned) { |
---|
| 1707 | + int ret = null_register_zoned_dev(nullb); |
---|
| 1708 | + |
---|
| 1709 | + if (ret) |
---|
| 1710 | + return ret; |
---|
| 1711 | + } |
---|
1644 | 1712 | |
---|
1645 | 1713 | add_disk(disk); |
---|
1646 | 1714 | return 0; |
---|
.. | .. |
---|
1658 | 1726 | set->flags = BLK_MQ_F_SHOULD_MERGE; |
---|
1659 | 1727 | if (g_no_sched) |
---|
1660 | 1728 | set->flags |= BLK_MQ_F_NO_SCHED; |
---|
| 1729 | + if (g_shared_tag_bitmap) |
---|
| 1730 | + set->flags |= BLK_MQ_F_TAG_HCTX_SHARED; |
---|
1661 | 1731 | set->driver_data = NULL; |
---|
1662 | 1732 | |
---|
1663 | 1733 | if ((nullb && nullb->dev->blocking) || g_blocking) |
---|
.. | .. |
---|
1666 | 1736 | return blk_mq_alloc_tag_set(set); |
---|
1667 | 1737 | } |
---|
1668 | 1738 | |
---|
1669 | | -static void null_validate_conf(struct nullb_device *dev) |
---|
| 1739 | +static int null_validate_conf(struct nullb_device *dev) |
---|
1670 | 1740 | { |
---|
1671 | 1741 | dev->blocksize = round_down(dev->blocksize, 512); |
---|
1672 | 1742 | dev->blocksize = clamp_t(unsigned int, dev->blocksize, 512, 4096); |
---|
.. | .. |
---|
1693 | 1763 | /* can not stop a queue */ |
---|
1694 | 1764 | if (dev->queue_mode == NULL_Q_BIO) |
---|
1695 | 1765 | dev->mbps = 0; |
---|
| 1766 | + |
---|
| 1767 | + if (dev->zoned && |
---|
| 1768 | + (!dev->zone_size || !is_power_of_2(dev->zone_size))) { |
---|
| 1769 | + pr_err("zone_size must be power-of-two\n"); |
---|
| 1770 | + return -EINVAL; |
---|
| 1771 | + } |
---|
| 1772 | + |
---|
| 1773 | + return 0; |
---|
1696 | 1774 | } |
---|
1697 | 1775 | |
---|
1698 | 1776 | #ifdef CONFIG_BLK_DEV_NULL_BLK_FAULT_INJECTION |
---|
.. | .. |
---|
1716 | 1794 | return false; |
---|
1717 | 1795 | if (!__null_setup_fault(&null_requeue_attr, g_requeue_str)) |
---|
1718 | 1796 | return false; |
---|
| 1797 | + if (!__null_setup_fault(&null_init_hctx_attr, g_init_hctx_str)) |
---|
| 1798 | + return false; |
---|
1719 | 1799 | #endif |
---|
1720 | 1800 | return true; |
---|
1721 | 1801 | } |
---|
.. | .. |
---|
1725 | 1805 | struct nullb *nullb; |
---|
1726 | 1806 | int rv; |
---|
1727 | 1807 | |
---|
1728 | | - null_validate_conf(dev); |
---|
| 1808 | + rv = null_validate_conf(dev); |
---|
| 1809 | + if (rv) |
---|
| 1810 | + return rv; |
---|
1729 | 1811 | |
---|
1730 | 1812 | nullb = kzalloc_node(sizeof(*nullb), GFP_KERNEL, dev->home_node); |
---|
1731 | 1813 | if (!nullb) { |
---|
.. | .. |
---|
1757 | 1839 | goto out_cleanup_queues; |
---|
1758 | 1840 | |
---|
1759 | 1841 | nullb->tag_set->timeout = 5 * HZ; |
---|
1760 | | - nullb->q = blk_mq_init_queue(nullb->tag_set); |
---|
| 1842 | + nullb->q = blk_mq_init_queue_data(nullb->tag_set, nullb); |
---|
1761 | 1843 | if (IS_ERR(nullb->q)) { |
---|
1762 | 1844 | rv = -ENOMEM; |
---|
1763 | 1845 | goto out_cleanup_tags; |
---|
1764 | 1846 | } |
---|
1765 | | - null_init_queues(nullb); |
---|
1766 | 1847 | } else if (dev->queue_mode == NULL_Q_BIO) { |
---|
1767 | | - nullb->q = blk_alloc_queue_node(GFP_KERNEL, dev->home_node, |
---|
1768 | | - NULL); |
---|
| 1848 | + nullb->q = blk_alloc_queue(dev->home_node); |
---|
1769 | 1849 | if (!nullb->q) { |
---|
1770 | 1850 | rv = -ENOMEM; |
---|
1771 | 1851 | goto out_cleanup_queues; |
---|
1772 | 1852 | } |
---|
1773 | | - blk_queue_make_request(nullb->q, null_queue_bio); |
---|
1774 | | - rv = init_driver_queues(nullb); |
---|
1775 | | - if (rv) |
---|
1776 | | - goto out_cleanup_blk_queue; |
---|
1777 | | - } else { |
---|
1778 | | - nullb->q = blk_init_queue_node(null_request_fn, &nullb->lock, |
---|
1779 | | - dev->home_node); |
---|
1780 | | - if (!nullb->q) { |
---|
1781 | | - rv = -ENOMEM; |
---|
1782 | | - goto out_cleanup_queues; |
---|
1783 | | - } |
---|
1784 | | - |
---|
1785 | | - if (!null_setup_fault()) |
---|
1786 | | - goto out_cleanup_blk_queue; |
---|
1787 | | - |
---|
1788 | | - blk_queue_prep_rq(nullb->q, null_rq_prep_fn); |
---|
1789 | | - blk_queue_softirq_done(nullb->q, null_softirq_done_fn); |
---|
1790 | | - blk_queue_rq_timed_out(nullb->q, null_rq_timed_out_fn); |
---|
1791 | | - nullb->q->rq_timeout = 5 * HZ; |
---|
1792 | 1853 | rv = init_driver_queues(nullb); |
---|
1793 | 1854 | if (rv) |
---|
1794 | 1855 | goto out_cleanup_blk_queue; |
---|
.. | .. |
---|
1802 | 1863 | if (dev->cache_size > 0) { |
---|
1803 | 1864 | set_bit(NULLB_DEV_FL_CACHE, &nullb->dev->flags); |
---|
1804 | 1865 | blk_queue_write_cache(nullb->q, true, true); |
---|
1805 | | - blk_queue_flush_queueable(nullb->q, true); |
---|
1806 | 1866 | } |
---|
1807 | 1867 | |
---|
1808 | 1868 | if (dev->zoned) { |
---|
1809 | | - rv = null_zone_init(dev); |
---|
| 1869 | + rv = null_init_zoned_dev(dev, nullb->q); |
---|
1810 | 1870 | if (rv) |
---|
1811 | 1871 | goto out_cleanup_blk_queue; |
---|
1812 | | - |
---|
1813 | | - blk_queue_chunk_sectors(nullb->q, dev->zone_size_sects); |
---|
1814 | | - nullb->q->limits.zoned = BLK_ZONED_HM; |
---|
1815 | 1872 | } |
---|
1816 | 1873 | |
---|
1817 | 1874 | nullb->q->queuedata = nullb; |
---|
.. | .. |
---|
1819 | 1876 | blk_queue_flag_clear(QUEUE_FLAG_ADD_RANDOM, nullb->q); |
---|
1820 | 1877 | |
---|
1821 | 1878 | mutex_lock(&lock); |
---|
1822 | | - nullb->index = ida_simple_get(&nullb_indexes, 0, 0, GFP_KERNEL); |
---|
1823 | | - dev->index = nullb->index; |
---|
| 1879 | + rv = ida_simple_get(&nullb_indexes, 0, 0, GFP_KERNEL); |
---|
| 1880 | + if (rv < 0) { |
---|
| 1881 | + mutex_unlock(&lock); |
---|
| 1882 | + goto out_cleanup_zone; |
---|
| 1883 | + } |
---|
| 1884 | + nullb->index = rv; |
---|
| 1885 | + dev->index = rv; |
---|
1824 | 1886 | mutex_unlock(&lock); |
---|
1825 | 1887 | |
---|
1826 | 1888 | blk_queue_logical_block_size(nullb->q, dev->blocksize); |
---|
.. | .. |
---|
1832 | 1894 | |
---|
1833 | 1895 | rv = null_gendisk_register(nullb); |
---|
1834 | 1896 | if (rv) |
---|
1835 | | - goto out_cleanup_zone; |
---|
| 1897 | + goto out_ida_free; |
---|
1836 | 1898 | |
---|
1837 | 1899 | mutex_lock(&lock); |
---|
1838 | 1900 | list_add_tail(&nullb->list, &nullb_list); |
---|
1839 | 1901 | mutex_unlock(&lock); |
---|
1840 | 1902 | |
---|
1841 | 1903 | return 0; |
---|
| 1904 | + |
---|
| 1905 | +out_ida_free: |
---|
| 1906 | + ida_free(&nullb_indexes, nullb->index); |
---|
1842 | 1907 | out_cleanup_zone: |
---|
1843 | | - if (dev->zoned) |
---|
1844 | | - null_zone_exit(dev); |
---|
| 1908 | + null_free_zoned_dev(dev); |
---|
1845 | 1909 | out_cleanup_blk_queue: |
---|
1846 | 1910 | blk_cleanup_queue(nullb->q); |
---|
1847 | 1911 | out_cleanup_tags: |
---|
.. | .. |
---|
1864 | 1928 | struct nullb_device *dev; |
---|
1865 | 1929 | |
---|
1866 | 1930 | if (g_bs > PAGE_SIZE) { |
---|
1867 | | - pr_warn("null_blk: invalid block size\n"); |
---|
1868 | | - pr_warn("null_blk: defaults block size to %lu\n", PAGE_SIZE); |
---|
| 1931 | + pr_warn("invalid block size\n"); |
---|
| 1932 | + pr_warn("defaults block size to %lu\n", PAGE_SIZE); |
---|
1869 | 1933 | g_bs = PAGE_SIZE; |
---|
1870 | 1934 | } |
---|
1871 | 1935 | |
---|
1872 | | - if (!is_power_of_2(g_zone_size)) { |
---|
1873 | | - pr_err("null_blk: zone_size must be power-of-two\n"); |
---|
1874 | | - return -EINVAL; |
---|
| 1936 | + if (g_home_node != NUMA_NO_NODE && g_home_node >= nr_online_nodes) { |
---|
| 1937 | + pr_err("invalid home_node value\n"); |
---|
| 1938 | + g_home_node = NUMA_NO_NODE; |
---|
1875 | 1939 | } |
---|
1876 | 1940 | |
---|
| 1941 | + if (g_queue_mode == NULL_Q_RQ) { |
---|
| 1942 | + pr_err("legacy IO path no longer available\n"); |
---|
| 1943 | + return -EINVAL; |
---|
| 1944 | + } |
---|
1877 | 1945 | if (g_queue_mode == NULL_Q_MQ && g_use_per_node_hctx) { |
---|
1878 | 1946 | if (g_submit_queues != nr_online_nodes) { |
---|
1879 | | - pr_warn("null_blk: submit_queues param is set to %u.\n", |
---|
| 1947 | + pr_warn("submit_queues param is set to %u.\n", |
---|
1880 | 1948 | nr_online_nodes); |
---|
1881 | 1949 | g_submit_queues = nr_online_nodes; |
---|
1882 | 1950 | } |
---|
.. | .. |
---|
1919 | 1987 | } |
---|
1920 | 1988 | } |
---|
1921 | 1989 | |
---|
1922 | | - pr_info("null: module loaded\n"); |
---|
| 1990 | + pr_info("module loaded\n"); |
---|
1923 | 1991 | return 0; |
---|
1924 | 1992 | |
---|
1925 | 1993 | err_dev: |
---|