.. | .. |
---|
| 1 | +// SPDX-License-Identifier: GPL-2.0 |
---|
1 | 2 | /* |
---|
2 | 3 | * Copyright (C) 2016 CNEX Labs |
---|
3 | 4 | * Initial release: Javier Gonzalez <javier@cnexlabs.com> |
---|
.. | .. |
---|
16 | 17 | * |
---|
17 | 18 | */ |
---|
18 | 19 | |
---|
| 20 | +#define CREATE_TRACE_POINTS |
---|
| 21 | + |
---|
19 | 22 | #include "pblk.h" |
---|
| 23 | +#include "pblk-trace.h" |
---|
20 | 24 | |
---|
21 | 25 | static void pblk_line_mark_bb(struct work_struct *work) |
---|
22 | 26 | { |
---|
.. | .. |
---|
27 | 31 | struct ppa_addr *ppa = line_ws->priv; |
---|
28 | 32 | int ret; |
---|
29 | 33 | |
---|
30 | | - ret = nvm_set_tgt_bb_tbl(dev, ppa, 1, NVM_BLK_T_GRWN_BAD); |
---|
| 34 | + ret = nvm_set_chunk_meta(dev, ppa, 1, NVM_BLK_T_GRWN_BAD); |
---|
31 | 35 | if (ret) { |
---|
32 | 36 | struct pblk_line *line; |
---|
33 | 37 | int pos; |
---|
34 | 38 | |
---|
35 | | - line = &pblk->lines[pblk_ppa_to_line(*ppa)]; |
---|
| 39 | + line = pblk_ppa_to_line(pblk, *ppa); |
---|
36 | 40 | pos = pblk_ppa_to_pos(&dev->geo, *ppa); |
---|
37 | 41 | |
---|
38 | 42 | pblk_err(pblk, "failed to mark bb, line:%d, pos:%d\n", |
---|
.. | .. |
---|
80 | 84 | struct pblk_line *line; |
---|
81 | 85 | int pos; |
---|
82 | 86 | |
---|
83 | | - line = &pblk->lines[pblk_ppa_to_line(rqd->ppa_addr)]; |
---|
| 87 | + line = pblk_ppa_to_line(pblk, rqd->ppa_addr); |
---|
84 | 88 | pos = pblk_ppa_to_pos(geo, rqd->ppa_addr); |
---|
85 | 89 | chunk = &line->chks[pos]; |
---|
86 | 90 | |
---|
87 | 91 | atomic_dec(&line->left_seblks); |
---|
88 | 92 | |
---|
89 | 93 | if (rqd->error) { |
---|
| 94 | + trace_pblk_chunk_reset(pblk_disk_name(pblk), |
---|
| 95 | + &rqd->ppa_addr, PBLK_CHUNK_RESET_FAILED); |
---|
| 96 | + |
---|
90 | 97 | chunk->state = NVM_CHK_ST_OFFLINE; |
---|
91 | 98 | pblk_mark_bb(pblk, line, rqd->ppa_addr); |
---|
92 | 99 | } else { |
---|
| 100 | + trace_pblk_chunk_reset(pblk_disk_name(pblk), |
---|
| 101 | + &rqd->ppa_addr, PBLK_CHUNK_RESET_DONE); |
---|
| 102 | + |
---|
93 | 103 | chunk->state = NVM_CHK_ST_FREE; |
---|
94 | 104 | } |
---|
| 105 | + |
---|
| 106 | + trace_pblk_chunk_state(pblk_disk_name(pblk), &rqd->ppa_addr, |
---|
| 107 | + chunk->state); |
---|
95 | 108 | |
---|
96 | 109 | atomic_dec(&pblk->inflight_io); |
---|
97 | 110 | } |
---|
.. | .. |
---|
108 | 121 | /* |
---|
109 | 122 | * Get information for all chunks from the device. |
---|
110 | 123 | * |
---|
111 | | - * The caller is responsible for freeing the returned structure |
---|
| 124 | + * The caller is responsible for freeing (vmalloc) the returned structure |
---|
112 | 125 | */ |
---|
113 | | -struct nvm_chk_meta *pblk_chunk_get_info(struct pblk *pblk) |
---|
| 126 | +struct nvm_chk_meta *pblk_get_chunk_meta(struct pblk *pblk) |
---|
114 | 127 | { |
---|
115 | 128 | struct nvm_tgt_dev *dev = pblk->dev; |
---|
116 | 129 | struct nvm_geo *geo = &dev->geo; |
---|
.. | .. |
---|
122 | 135 | ppa.ppa = 0; |
---|
123 | 136 | |
---|
124 | 137 | len = geo->all_chunks * sizeof(*meta); |
---|
125 | | - meta = kzalloc(len, GFP_KERNEL); |
---|
| 138 | + meta = vzalloc(len); |
---|
126 | 139 | if (!meta) |
---|
127 | 140 | return ERR_PTR(-ENOMEM); |
---|
128 | 141 | |
---|
129 | | - ret = nvm_get_chunk_meta(dev, meta, ppa, geo->all_chunks); |
---|
| 142 | + ret = nvm_get_chunk_meta(dev, ppa, geo->all_chunks, meta); |
---|
130 | 143 | if (ret) { |
---|
131 | | - kfree(meta); |
---|
| 144 | + vfree(meta); |
---|
132 | 145 | return ERR_PTR(-EIO); |
---|
133 | 146 | } |
---|
134 | 147 | |
---|
.. | .. |
---|
192 | 205 | { |
---|
193 | 206 | struct pblk_line *line; |
---|
194 | 207 | u64 paddr; |
---|
195 | | - int line_id; |
---|
196 | 208 | |
---|
197 | 209 | #ifdef CONFIG_NVM_PBLK_DEBUG |
---|
198 | 210 | /* Callers must ensure that the ppa points to a device address */ |
---|
.. | .. |
---|
200 | 212 | BUG_ON(pblk_ppa_empty(ppa)); |
---|
201 | 213 | #endif |
---|
202 | 214 | |
---|
203 | | - line_id = pblk_ppa_to_line(ppa); |
---|
204 | | - line = &pblk->lines[line_id]; |
---|
| 215 | + line = pblk_ppa_to_line(pblk, ppa); |
---|
205 | 216 | paddr = pblk_dev_ppa_to_line_addr(pblk, ppa); |
---|
206 | 217 | |
---|
207 | 218 | __pblk_map_invalidate(pblk, line, paddr); |
---|
.. | .. |
---|
225 | 236 | pblk_trans_map_set(pblk, lba, ppa); |
---|
226 | 237 | } |
---|
227 | 238 | spin_unlock(&pblk->trans_lock); |
---|
| 239 | +} |
---|
| 240 | + |
---|
| 241 | +int pblk_alloc_rqd_meta(struct pblk *pblk, struct nvm_rq *rqd) |
---|
| 242 | +{ |
---|
| 243 | + struct nvm_tgt_dev *dev = pblk->dev; |
---|
| 244 | + |
---|
| 245 | + rqd->meta_list = nvm_dev_dma_alloc(dev->parent, GFP_KERNEL, |
---|
| 246 | + &rqd->dma_meta_list); |
---|
| 247 | + if (!rqd->meta_list) |
---|
| 248 | + return -ENOMEM; |
---|
| 249 | + |
---|
| 250 | + if (rqd->nr_ppas == 1) |
---|
| 251 | + return 0; |
---|
| 252 | + |
---|
| 253 | + rqd->ppa_list = rqd->meta_list + pblk_dma_meta_size(pblk); |
---|
| 254 | + rqd->dma_ppa_list = rqd->dma_meta_list + pblk_dma_meta_size(pblk); |
---|
| 255 | + |
---|
| 256 | + return 0; |
---|
| 257 | +} |
---|
| 258 | + |
---|
| 259 | +void pblk_free_rqd_meta(struct pblk *pblk, struct nvm_rq *rqd) |
---|
| 260 | +{ |
---|
| 261 | + struct nvm_tgt_dev *dev = pblk->dev; |
---|
| 262 | + |
---|
| 263 | + if (rqd->meta_list) |
---|
| 264 | + nvm_dev_dma_free(dev->parent, rqd->meta_list, |
---|
| 265 | + rqd->dma_meta_list); |
---|
228 | 266 | } |
---|
229 | 267 | |
---|
230 | 268 | /* Caller must guarantee that the request is a valid type */ |
---|
.. | .. |
---|
258 | 296 | /* Typically used on completion path. Cannot guarantee request consistency */ |
---|
259 | 297 | void pblk_free_rqd(struct pblk *pblk, struct nvm_rq *rqd, int type) |
---|
260 | 298 | { |
---|
261 | | - struct nvm_tgt_dev *dev = pblk->dev; |
---|
262 | 299 | mempool_t *pool; |
---|
263 | 300 | |
---|
264 | 301 | switch (type) { |
---|
265 | 302 | case PBLK_WRITE: |
---|
266 | 303 | kfree(((struct pblk_c_ctx *)nvm_rq_to_pdu(rqd))->lun_bitmap); |
---|
267 | | - /* fall through */ |
---|
| 304 | + fallthrough; |
---|
268 | 305 | case PBLK_WRITE_INT: |
---|
269 | 306 | pool = &pblk->w_rq_pool; |
---|
270 | 307 | break; |
---|
.. | .. |
---|
279 | 316 | return; |
---|
280 | 317 | } |
---|
281 | 318 | |
---|
282 | | - if (rqd->meta_list) |
---|
283 | | - nvm_dev_dma_free(dev->parent, rqd->meta_list, |
---|
284 | | - rqd->dma_meta_list); |
---|
| 319 | + pblk_free_rqd_meta(pblk, rqd); |
---|
285 | 320 | mempool_free(rqd, pool); |
---|
286 | 321 | } |
---|
287 | 322 | |
---|
.. | .. |
---|
343 | 378 | { |
---|
344 | 379 | unsigned int secs_avail = pblk_rb_read_count(&pblk->rwb); |
---|
345 | 380 | |
---|
346 | | - if (secs_avail >= pblk->min_write_pgs) |
---|
| 381 | + if (secs_avail >= pblk->min_write_pgs_data) |
---|
347 | 382 | pblk_write_kick(pblk); |
---|
348 | 383 | } |
---|
349 | 384 | |
---|
.. | .. |
---|
374 | 409 | struct pblk_line_meta *lm = &pblk->lm; |
---|
375 | 410 | struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
---|
376 | 411 | struct list_head *move_list = NULL; |
---|
377 | | - int vsc = le32_to_cpu(*line->vsc); |
---|
| 412 | + int packed_meta = (le32_to_cpu(*line->vsc) / pblk->min_write_pgs_data) |
---|
| 413 | + * (pblk->min_write_pgs - pblk->min_write_pgs_data); |
---|
| 414 | + int vsc = le32_to_cpu(*line->vsc) + packed_meta; |
---|
378 | 415 | |
---|
379 | 416 | lockdep_assert_held(&line->lock); |
---|
380 | 417 | |
---|
.. | .. |
---|
411 | 448 | } |
---|
412 | 449 | } else { |
---|
413 | 450 | line->state = PBLK_LINESTATE_CORRUPT; |
---|
| 451 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
---|
| 452 | + line->state); |
---|
| 453 | + |
---|
414 | 454 | line->gc_group = PBLK_LINEGC_NONE; |
---|
415 | 455 | move_list = &l_mg->corrupt_list; |
---|
416 | 456 | pblk_err(pblk, "corrupted vsc for line %d, vsc:%d (%d/%d/%d)\n", |
---|
.. | .. |
---|
467 | 507 | pblk->sec_per_write = sec_per_write; |
---|
468 | 508 | } |
---|
469 | 509 | |
---|
470 | | -int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd) |
---|
| 510 | +int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd, void *buf) |
---|
471 | 511 | { |
---|
472 | 512 | struct nvm_tgt_dev *dev = pblk->dev; |
---|
473 | 513 | |
---|
.. | .. |
---|
478 | 518 | return NVM_IO_ERR; |
---|
479 | 519 | #endif |
---|
480 | 520 | |
---|
481 | | - return nvm_submit_io(dev, rqd); |
---|
| 521 | + return nvm_submit_io(dev, rqd, buf); |
---|
482 | 522 | } |
---|
483 | 523 | |
---|
484 | | -int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd) |
---|
| 524 | +void pblk_check_chunk_state_update(struct pblk *pblk, struct nvm_rq *rqd) |
---|
485 | 525 | { |
---|
486 | | - struct nvm_tgt_dev *dev = pblk->dev; |
---|
| 526 | + struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); |
---|
487 | 527 | |
---|
488 | | - atomic_inc(&pblk->inflight_io); |
---|
| 528 | + int i; |
---|
489 | 529 | |
---|
490 | | -#ifdef CONFIG_NVM_PBLK_DEBUG |
---|
491 | | - if (pblk_check_io(pblk, rqd)) |
---|
492 | | - return NVM_IO_ERR; |
---|
493 | | -#endif |
---|
| 530 | + for (i = 0; i < rqd->nr_ppas; i++) { |
---|
| 531 | + struct ppa_addr *ppa = &ppa_list[i]; |
---|
| 532 | + struct nvm_chk_meta *chunk = pblk_dev_ppa_to_chunk(pblk, *ppa); |
---|
| 533 | + u64 caddr = pblk_dev_ppa_to_chunk_addr(pblk, *ppa); |
---|
494 | 534 | |
---|
495 | | - return nvm_submit_io_sync(dev, rqd); |
---|
496 | | -} |
---|
497 | | - |
---|
498 | | -static void pblk_bio_map_addr_endio(struct bio *bio) |
---|
499 | | -{ |
---|
500 | | - bio_put(bio); |
---|
501 | | -} |
---|
502 | | - |
---|
503 | | -struct bio *pblk_bio_map_addr(struct pblk *pblk, void *data, |
---|
504 | | - unsigned int nr_secs, unsigned int len, |
---|
505 | | - int alloc_type, gfp_t gfp_mask) |
---|
506 | | -{ |
---|
507 | | - struct nvm_tgt_dev *dev = pblk->dev; |
---|
508 | | - void *kaddr = data; |
---|
509 | | - struct page *page; |
---|
510 | | - struct bio *bio; |
---|
511 | | - int i, ret; |
---|
512 | | - |
---|
513 | | - if (alloc_type == PBLK_KMALLOC_META) |
---|
514 | | - return bio_map_kern(dev->q, kaddr, len, gfp_mask); |
---|
515 | | - |
---|
516 | | - bio = bio_kmalloc(gfp_mask, nr_secs); |
---|
517 | | - if (!bio) |
---|
518 | | - return ERR_PTR(-ENOMEM); |
---|
519 | | - |
---|
520 | | - for (i = 0; i < nr_secs; i++) { |
---|
521 | | - page = vmalloc_to_page(kaddr); |
---|
522 | | - if (!page) { |
---|
523 | | - pblk_err(pblk, "could not map vmalloc bio\n"); |
---|
524 | | - bio_put(bio); |
---|
525 | | - bio = ERR_PTR(-ENOMEM); |
---|
526 | | - goto out; |
---|
527 | | - } |
---|
528 | | - |
---|
529 | | - ret = bio_add_pc_page(dev->q, bio, page, PAGE_SIZE, 0); |
---|
530 | | - if (ret != PAGE_SIZE) { |
---|
531 | | - pblk_err(pblk, "could not add page to bio\n"); |
---|
532 | | - bio_put(bio); |
---|
533 | | - bio = ERR_PTR(-ENOMEM); |
---|
534 | | - goto out; |
---|
535 | | - } |
---|
536 | | - |
---|
537 | | - kaddr += PAGE_SIZE; |
---|
| 535 | + if (caddr == 0) |
---|
| 536 | + trace_pblk_chunk_state(pblk_disk_name(pblk), |
---|
| 537 | + ppa, NVM_CHK_ST_OPEN); |
---|
| 538 | + else if (caddr == (chunk->cnlb - 1)) |
---|
| 539 | + trace_pblk_chunk_state(pblk_disk_name(pblk), |
---|
| 540 | + ppa, NVM_CHK_ST_CLOSED); |
---|
538 | 541 | } |
---|
| 542 | +} |
---|
539 | 543 | |
---|
540 | | - bio->bi_end_io = pblk_bio_map_addr_endio; |
---|
541 | | -out: |
---|
542 | | - return bio; |
---|
| 544 | +int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd, void *buf) |
---|
| 545 | +{ |
---|
| 546 | + struct nvm_tgt_dev *dev = pblk->dev; |
---|
| 547 | + int ret; |
---|
| 548 | + |
---|
| 549 | + atomic_inc(&pblk->inflight_io); |
---|
| 550 | + |
---|
| 551 | +#ifdef CONFIG_NVM_PBLK_DEBUG |
---|
| 552 | + if (pblk_check_io(pblk, rqd)) |
---|
| 553 | + return NVM_IO_ERR; |
---|
| 554 | +#endif |
---|
| 555 | + |
---|
| 556 | + ret = nvm_submit_io_sync(dev, rqd, buf); |
---|
| 557 | + |
---|
| 558 | + if (trace_pblk_chunk_state_enabled() && !ret && |
---|
| 559 | + rqd->opcode == NVM_OP_PWRITE) |
---|
| 560 | + pblk_check_chunk_state_update(pblk, rqd); |
---|
| 561 | + |
---|
| 562 | + return ret; |
---|
| 563 | +} |
---|
| 564 | + |
---|
| 565 | +static int pblk_submit_io_sync_sem(struct pblk *pblk, struct nvm_rq *rqd, |
---|
| 566 | + void *buf) |
---|
| 567 | +{ |
---|
| 568 | + struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); |
---|
| 569 | + int ret; |
---|
| 570 | + |
---|
| 571 | + pblk_down_chunk(pblk, ppa_list[0]); |
---|
| 572 | + ret = pblk_submit_io_sync(pblk, rqd, buf); |
---|
| 573 | + pblk_up_chunk(pblk, ppa_list[0]); |
---|
| 574 | + |
---|
| 575 | + return ret; |
---|
543 | 576 | } |
---|
544 | 577 | |
---|
545 | 578 | int pblk_calc_secs(struct pblk *pblk, unsigned long secs_avail, |
---|
546 | | - unsigned long secs_to_flush) |
---|
| 579 | + unsigned long secs_to_flush, bool skip_meta) |
---|
547 | 580 | { |
---|
548 | 581 | int max = pblk->sec_per_write; |
---|
549 | 582 | int min = pblk->min_write_pgs; |
---|
550 | 583 | int secs_to_sync = 0; |
---|
| 584 | + |
---|
| 585 | + if (skip_meta && pblk->min_write_pgs_data != pblk->min_write_pgs) |
---|
| 586 | + min = max = pblk->min_write_pgs_data; |
---|
551 | 587 | |
---|
552 | 588 | if (secs_avail >= max) |
---|
553 | 589 | secs_to_sync = max; |
---|
.. | .. |
---|
623 | 659 | return paddr; |
---|
624 | 660 | } |
---|
625 | 661 | |
---|
626 | | -/* |
---|
627 | | - * Submit emeta to one LUN in the raid line at the time to avoid a deadlock when |
---|
628 | | - * taking the per LUN semaphore. |
---|
629 | | - */ |
---|
630 | | -static int pblk_line_submit_emeta_io(struct pblk *pblk, struct pblk_line *line, |
---|
631 | | - void *emeta_buf, u64 paddr, int dir) |
---|
632 | | -{ |
---|
633 | | - struct nvm_tgt_dev *dev = pblk->dev; |
---|
634 | | - struct nvm_geo *geo = &dev->geo; |
---|
635 | | - struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
---|
636 | | - struct pblk_line_meta *lm = &pblk->lm; |
---|
637 | | - void *ppa_list, *meta_list; |
---|
638 | | - struct bio *bio; |
---|
639 | | - struct nvm_rq rqd; |
---|
640 | | - dma_addr_t dma_ppa_list, dma_meta_list; |
---|
641 | | - int min = pblk->min_write_pgs; |
---|
642 | | - int left_ppas = lm->emeta_sec[0]; |
---|
643 | | - int id = line->id; |
---|
644 | | - int rq_ppas, rq_len; |
---|
645 | | - int cmd_op, bio_op; |
---|
646 | | - int i, j; |
---|
647 | | - int ret; |
---|
648 | | - |
---|
649 | | - if (dir == PBLK_WRITE) { |
---|
650 | | - bio_op = REQ_OP_WRITE; |
---|
651 | | - cmd_op = NVM_OP_PWRITE; |
---|
652 | | - } else if (dir == PBLK_READ) { |
---|
653 | | - bio_op = REQ_OP_READ; |
---|
654 | | - cmd_op = NVM_OP_PREAD; |
---|
655 | | - } else |
---|
656 | | - return -EINVAL; |
---|
657 | | - |
---|
658 | | - meta_list = nvm_dev_dma_alloc(dev->parent, GFP_KERNEL, |
---|
659 | | - &dma_meta_list); |
---|
660 | | - if (!meta_list) |
---|
661 | | - return -ENOMEM; |
---|
662 | | - |
---|
663 | | - ppa_list = meta_list + pblk_dma_meta_size; |
---|
664 | | - dma_ppa_list = dma_meta_list + pblk_dma_meta_size; |
---|
665 | | - |
---|
666 | | -next_rq: |
---|
667 | | - memset(&rqd, 0, sizeof(struct nvm_rq)); |
---|
668 | | - |
---|
669 | | - rq_ppas = pblk_calc_secs(pblk, left_ppas, 0); |
---|
670 | | - rq_len = rq_ppas * geo->csecs; |
---|
671 | | - |
---|
672 | | - bio = pblk_bio_map_addr(pblk, emeta_buf, rq_ppas, rq_len, |
---|
673 | | - l_mg->emeta_alloc_type, GFP_KERNEL); |
---|
674 | | - if (IS_ERR(bio)) { |
---|
675 | | - ret = PTR_ERR(bio); |
---|
676 | | - goto free_rqd_dma; |
---|
677 | | - } |
---|
678 | | - |
---|
679 | | - bio->bi_iter.bi_sector = 0; /* internal bio */ |
---|
680 | | - bio_set_op_attrs(bio, bio_op, 0); |
---|
681 | | - |
---|
682 | | - rqd.bio = bio; |
---|
683 | | - rqd.meta_list = meta_list; |
---|
684 | | - rqd.ppa_list = ppa_list; |
---|
685 | | - rqd.dma_meta_list = dma_meta_list; |
---|
686 | | - rqd.dma_ppa_list = dma_ppa_list; |
---|
687 | | - rqd.opcode = cmd_op; |
---|
688 | | - rqd.nr_ppas = rq_ppas; |
---|
689 | | - |
---|
690 | | - if (dir == PBLK_WRITE) { |
---|
691 | | - struct pblk_sec_meta *meta_list = rqd.meta_list; |
---|
692 | | - |
---|
693 | | - rqd.flags = pblk_set_progr_mode(pblk, PBLK_WRITE); |
---|
694 | | - for (i = 0; i < rqd.nr_ppas; ) { |
---|
695 | | - spin_lock(&line->lock); |
---|
696 | | - paddr = __pblk_alloc_page(pblk, line, min); |
---|
697 | | - spin_unlock(&line->lock); |
---|
698 | | - for (j = 0; j < min; j++, i++, paddr++) { |
---|
699 | | - meta_list[i].lba = cpu_to_le64(ADDR_EMPTY); |
---|
700 | | - rqd.ppa_list[i] = |
---|
701 | | - addr_to_gen_ppa(pblk, paddr, id); |
---|
702 | | - } |
---|
703 | | - } |
---|
704 | | - } else { |
---|
705 | | - for (i = 0; i < rqd.nr_ppas; ) { |
---|
706 | | - struct ppa_addr ppa = addr_to_gen_ppa(pblk, paddr, id); |
---|
707 | | - int pos = pblk_ppa_to_pos(geo, ppa); |
---|
708 | | - int read_type = PBLK_READ_RANDOM; |
---|
709 | | - |
---|
710 | | - if (pblk_io_aligned(pblk, rq_ppas)) |
---|
711 | | - read_type = PBLK_READ_SEQUENTIAL; |
---|
712 | | - rqd.flags = pblk_set_read_mode(pblk, read_type); |
---|
713 | | - |
---|
714 | | - while (test_bit(pos, line->blk_bitmap)) { |
---|
715 | | - paddr += min; |
---|
716 | | - if (pblk_boundary_paddr_checks(pblk, paddr)) { |
---|
717 | | - pblk_err(pblk, "corrupt emeta line:%d\n", |
---|
718 | | - line->id); |
---|
719 | | - bio_put(bio); |
---|
720 | | - ret = -EINTR; |
---|
721 | | - goto free_rqd_dma; |
---|
722 | | - } |
---|
723 | | - |
---|
724 | | - ppa = addr_to_gen_ppa(pblk, paddr, id); |
---|
725 | | - pos = pblk_ppa_to_pos(geo, ppa); |
---|
726 | | - } |
---|
727 | | - |
---|
728 | | - if (pblk_boundary_paddr_checks(pblk, paddr + min)) { |
---|
729 | | - pblk_err(pblk, "corrupt emeta line:%d\n", |
---|
730 | | - line->id); |
---|
731 | | - bio_put(bio); |
---|
732 | | - ret = -EINTR; |
---|
733 | | - goto free_rqd_dma; |
---|
734 | | - } |
---|
735 | | - |
---|
736 | | - for (j = 0; j < min; j++, i++, paddr++) |
---|
737 | | - rqd.ppa_list[i] = |
---|
738 | | - addr_to_gen_ppa(pblk, paddr, line->id); |
---|
739 | | - } |
---|
740 | | - } |
---|
741 | | - |
---|
742 | | - ret = pblk_submit_io_sync(pblk, &rqd); |
---|
743 | | - if (ret) { |
---|
744 | | - pblk_err(pblk, "emeta I/O submission failed: %d\n", ret); |
---|
745 | | - bio_put(bio); |
---|
746 | | - goto free_rqd_dma; |
---|
747 | | - } |
---|
748 | | - |
---|
749 | | - atomic_dec(&pblk->inflight_io); |
---|
750 | | - |
---|
751 | | - if (rqd.error) { |
---|
752 | | - if (dir == PBLK_WRITE) |
---|
753 | | - pblk_log_write_err(pblk, &rqd); |
---|
754 | | - else |
---|
755 | | - pblk_log_read_err(pblk, &rqd); |
---|
756 | | - } |
---|
757 | | - |
---|
758 | | - emeta_buf += rq_len; |
---|
759 | | - left_ppas -= rq_ppas; |
---|
760 | | - if (left_ppas) |
---|
761 | | - goto next_rq; |
---|
762 | | -free_rqd_dma: |
---|
763 | | - nvm_dev_dma_free(dev->parent, rqd.meta_list, rqd.dma_meta_list); |
---|
764 | | - return ret; |
---|
765 | | -} |
---|
766 | | - |
---|
767 | 662 | u64 pblk_line_smeta_start(struct pblk *pblk, struct pblk_line *line) |
---|
768 | 663 | { |
---|
769 | 664 | struct nvm_tgt_dev *dev = pblk->dev; |
---|
.. | .. |
---|
779 | 674 | return bit * geo->ws_opt; |
---|
780 | 675 | } |
---|
781 | 676 | |
---|
782 | | -static int pblk_line_submit_smeta_io(struct pblk *pblk, struct pblk_line *line, |
---|
783 | | - u64 paddr, int dir) |
---|
| 677 | +int pblk_line_smeta_read(struct pblk *pblk, struct pblk_line *line) |
---|
784 | 678 | { |
---|
785 | | - struct nvm_tgt_dev *dev = pblk->dev; |
---|
786 | 679 | struct pblk_line_meta *lm = &pblk->lm; |
---|
787 | | - struct bio *bio; |
---|
| 680 | + struct ppa_addr *ppa_list; |
---|
788 | 681 | struct nvm_rq rqd; |
---|
789 | | - __le64 *lba_list = NULL; |
---|
| 682 | + u64 paddr = pblk_line_smeta_start(pblk, line); |
---|
790 | 683 | int i, ret; |
---|
791 | | - int cmd_op, bio_op; |
---|
792 | | - int flags; |
---|
793 | | - |
---|
794 | | - if (dir == PBLK_WRITE) { |
---|
795 | | - bio_op = REQ_OP_WRITE; |
---|
796 | | - cmd_op = NVM_OP_PWRITE; |
---|
797 | | - flags = pblk_set_progr_mode(pblk, PBLK_WRITE); |
---|
798 | | - lba_list = emeta_to_lbas(pblk, line->emeta->buf); |
---|
799 | | - } else if (dir == PBLK_READ_RECOV || dir == PBLK_READ) { |
---|
800 | | - bio_op = REQ_OP_READ; |
---|
801 | | - cmd_op = NVM_OP_PREAD; |
---|
802 | | - flags = pblk_set_read_mode(pblk, PBLK_READ_SEQUENTIAL); |
---|
803 | | - } else |
---|
804 | | - return -EINVAL; |
---|
805 | 684 | |
---|
806 | 685 | memset(&rqd, 0, sizeof(struct nvm_rq)); |
---|
807 | 686 | |
---|
808 | | - rqd.meta_list = nvm_dev_dma_alloc(dev->parent, GFP_KERNEL, |
---|
809 | | - &rqd.dma_meta_list); |
---|
810 | | - if (!rqd.meta_list) |
---|
811 | | - return -ENOMEM; |
---|
| 687 | + ret = pblk_alloc_rqd_meta(pblk, &rqd); |
---|
| 688 | + if (ret) |
---|
| 689 | + return ret; |
---|
812 | 690 | |
---|
813 | | - rqd.ppa_list = rqd.meta_list + pblk_dma_meta_size; |
---|
814 | | - rqd.dma_ppa_list = rqd.dma_meta_list + pblk_dma_meta_size; |
---|
815 | | - |
---|
816 | | - bio = bio_map_kern(dev->q, line->smeta, lm->smeta_len, GFP_KERNEL); |
---|
817 | | - if (IS_ERR(bio)) { |
---|
818 | | - ret = PTR_ERR(bio); |
---|
819 | | - goto free_ppa_list; |
---|
820 | | - } |
---|
821 | | - |
---|
822 | | - bio->bi_iter.bi_sector = 0; /* internal bio */ |
---|
823 | | - bio_set_op_attrs(bio, bio_op, 0); |
---|
824 | | - |
---|
825 | | - rqd.bio = bio; |
---|
826 | | - rqd.opcode = cmd_op; |
---|
827 | | - rqd.flags = flags; |
---|
| 691 | + rqd.opcode = NVM_OP_PREAD; |
---|
828 | 692 | rqd.nr_ppas = lm->smeta_sec; |
---|
| 693 | + rqd.is_seq = 1; |
---|
| 694 | + ppa_list = nvm_rq_to_ppa_list(&rqd); |
---|
829 | 695 | |
---|
830 | | - for (i = 0; i < lm->smeta_sec; i++, paddr++) { |
---|
831 | | - struct pblk_sec_meta *meta_list = rqd.meta_list; |
---|
| 696 | + for (i = 0; i < lm->smeta_sec; i++, paddr++) |
---|
| 697 | + ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line->id); |
---|
832 | 698 | |
---|
833 | | - rqd.ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line->id); |
---|
834 | | - |
---|
835 | | - if (dir == PBLK_WRITE) { |
---|
836 | | - __le64 addr_empty = cpu_to_le64(ADDR_EMPTY); |
---|
837 | | - |
---|
838 | | - meta_list[i].lba = lba_list[paddr] = addr_empty; |
---|
839 | | - } |
---|
840 | | - } |
---|
841 | | - |
---|
842 | | - /* |
---|
843 | | - * This I/O is sent by the write thread when a line is replace. Since |
---|
844 | | - * the write thread is the only one sending write and erase commands, |
---|
845 | | - * there is no need to take the LUN semaphore. |
---|
846 | | - */ |
---|
847 | | - ret = pblk_submit_io_sync(pblk, &rqd); |
---|
| 699 | + ret = pblk_submit_io_sync(pblk, &rqd, line->smeta); |
---|
848 | 700 | if (ret) { |
---|
849 | 701 | pblk_err(pblk, "smeta I/O submission failed: %d\n", ret); |
---|
850 | | - bio_put(bio); |
---|
851 | | - goto free_ppa_list; |
---|
| 702 | + goto clear_rqd; |
---|
| 703 | + } |
---|
| 704 | + |
---|
| 705 | + atomic_dec(&pblk->inflight_io); |
---|
| 706 | + |
---|
| 707 | + if (rqd.error && rqd.error != NVM_RSP_WARN_HIGHECC) { |
---|
| 708 | + pblk_log_read_err(pblk, &rqd); |
---|
| 709 | + ret = -EIO; |
---|
| 710 | + } |
---|
| 711 | + |
---|
| 712 | +clear_rqd: |
---|
| 713 | + pblk_free_rqd_meta(pblk, &rqd); |
---|
| 714 | + return ret; |
---|
| 715 | +} |
---|
| 716 | + |
---|
| 717 | +static int pblk_line_smeta_write(struct pblk *pblk, struct pblk_line *line, |
---|
| 718 | + u64 paddr) |
---|
| 719 | +{ |
---|
| 720 | + struct pblk_line_meta *lm = &pblk->lm; |
---|
| 721 | + struct ppa_addr *ppa_list; |
---|
| 722 | + struct nvm_rq rqd; |
---|
| 723 | + __le64 *lba_list = emeta_to_lbas(pblk, line->emeta->buf); |
---|
| 724 | + __le64 addr_empty = cpu_to_le64(ADDR_EMPTY); |
---|
| 725 | + int i, ret; |
---|
| 726 | + |
---|
| 727 | + memset(&rqd, 0, sizeof(struct nvm_rq)); |
---|
| 728 | + |
---|
| 729 | + ret = pblk_alloc_rqd_meta(pblk, &rqd); |
---|
| 730 | + if (ret) |
---|
| 731 | + return ret; |
---|
| 732 | + |
---|
| 733 | + rqd.opcode = NVM_OP_PWRITE; |
---|
| 734 | + rqd.nr_ppas = lm->smeta_sec; |
---|
| 735 | + rqd.is_seq = 1; |
---|
| 736 | + ppa_list = nvm_rq_to_ppa_list(&rqd); |
---|
| 737 | + |
---|
| 738 | + for (i = 0; i < lm->smeta_sec; i++, paddr++) { |
---|
| 739 | + struct pblk_sec_meta *meta = pblk_get_meta(pblk, |
---|
| 740 | + rqd.meta_list, i); |
---|
| 741 | + |
---|
| 742 | + ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line->id); |
---|
| 743 | + meta->lba = lba_list[paddr] = addr_empty; |
---|
| 744 | + } |
---|
| 745 | + |
---|
| 746 | + ret = pblk_submit_io_sync_sem(pblk, &rqd, line->smeta); |
---|
| 747 | + if (ret) { |
---|
| 748 | + pblk_err(pblk, "smeta I/O submission failed: %d\n", ret); |
---|
| 749 | + goto clear_rqd; |
---|
852 | 750 | } |
---|
853 | 751 | |
---|
854 | 752 | atomic_dec(&pblk->inflight_io); |
---|
855 | 753 | |
---|
856 | 754 | if (rqd.error) { |
---|
857 | | - if (dir == PBLK_WRITE) { |
---|
858 | | - pblk_log_write_err(pblk, &rqd); |
---|
859 | | - ret = 1; |
---|
860 | | - } else if (dir == PBLK_READ) |
---|
861 | | - pblk_log_read_err(pblk, &rqd); |
---|
| 755 | + pblk_log_write_err(pblk, &rqd); |
---|
| 756 | + ret = -EIO; |
---|
862 | 757 | } |
---|
863 | 758 | |
---|
864 | | -free_ppa_list: |
---|
865 | | - nvm_dev_dma_free(dev->parent, rqd.meta_list, rqd.dma_meta_list); |
---|
866 | | - |
---|
| 759 | +clear_rqd: |
---|
| 760 | + pblk_free_rqd_meta(pblk, &rqd); |
---|
867 | 761 | return ret; |
---|
868 | 762 | } |
---|
869 | 763 | |
---|
870 | | -int pblk_line_read_smeta(struct pblk *pblk, struct pblk_line *line) |
---|
871 | | -{ |
---|
872 | | - u64 bpaddr = pblk_line_smeta_start(pblk, line); |
---|
873 | | - |
---|
874 | | - return pblk_line_submit_smeta_io(pblk, line, bpaddr, PBLK_READ_RECOV); |
---|
875 | | -} |
---|
876 | | - |
---|
877 | | -int pblk_line_read_emeta(struct pblk *pblk, struct pblk_line *line, |
---|
| 764 | +int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line, |
---|
878 | 765 | void *emeta_buf) |
---|
879 | 766 | { |
---|
880 | | - return pblk_line_submit_emeta_io(pblk, line, emeta_buf, |
---|
881 | | - line->emeta_ssec, PBLK_READ); |
---|
| 767 | + struct nvm_tgt_dev *dev = pblk->dev; |
---|
| 768 | + struct nvm_geo *geo = &dev->geo; |
---|
| 769 | + struct pblk_line_meta *lm = &pblk->lm; |
---|
| 770 | + void *ppa_list_buf, *meta_list; |
---|
| 771 | + struct ppa_addr *ppa_list; |
---|
| 772 | + struct nvm_rq rqd; |
---|
| 773 | + u64 paddr = line->emeta_ssec; |
---|
| 774 | + dma_addr_t dma_ppa_list, dma_meta_list; |
---|
| 775 | + int min = pblk->min_write_pgs; |
---|
| 776 | + int left_ppas = lm->emeta_sec[0]; |
---|
| 777 | + int line_id = line->id; |
---|
| 778 | + int rq_ppas, rq_len; |
---|
| 779 | + int i, j; |
---|
| 780 | + int ret; |
---|
| 781 | + |
---|
| 782 | + meta_list = nvm_dev_dma_alloc(dev->parent, GFP_KERNEL, |
---|
| 783 | + &dma_meta_list); |
---|
| 784 | + if (!meta_list) |
---|
| 785 | + return -ENOMEM; |
---|
| 786 | + |
---|
| 787 | + ppa_list_buf = meta_list + pblk_dma_meta_size(pblk); |
---|
| 788 | + dma_ppa_list = dma_meta_list + pblk_dma_meta_size(pblk); |
---|
| 789 | + |
---|
| 790 | +next_rq: |
---|
| 791 | + memset(&rqd, 0, sizeof(struct nvm_rq)); |
---|
| 792 | + |
---|
| 793 | + rq_ppas = pblk_calc_secs(pblk, left_ppas, 0, false); |
---|
| 794 | + rq_len = rq_ppas * geo->csecs; |
---|
| 795 | + |
---|
| 796 | + rqd.meta_list = meta_list; |
---|
| 797 | + rqd.ppa_list = ppa_list_buf; |
---|
| 798 | + rqd.dma_meta_list = dma_meta_list; |
---|
| 799 | + rqd.dma_ppa_list = dma_ppa_list; |
---|
| 800 | + rqd.opcode = NVM_OP_PREAD; |
---|
| 801 | + rqd.nr_ppas = rq_ppas; |
---|
| 802 | + ppa_list = nvm_rq_to_ppa_list(&rqd); |
---|
| 803 | + |
---|
| 804 | + for (i = 0; i < rqd.nr_ppas; ) { |
---|
| 805 | + struct ppa_addr ppa = addr_to_gen_ppa(pblk, paddr, line_id); |
---|
| 806 | + int pos = pblk_ppa_to_pos(geo, ppa); |
---|
| 807 | + |
---|
| 808 | + if (pblk_io_aligned(pblk, rq_ppas)) |
---|
| 809 | + rqd.is_seq = 1; |
---|
| 810 | + |
---|
| 811 | + while (test_bit(pos, line->blk_bitmap)) { |
---|
| 812 | + paddr += min; |
---|
| 813 | + if (pblk_boundary_paddr_checks(pblk, paddr)) { |
---|
| 814 | + ret = -EINTR; |
---|
| 815 | + goto free_rqd_dma; |
---|
| 816 | + } |
---|
| 817 | + |
---|
| 818 | + ppa = addr_to_gen_ppa(pblk, paddr, line_id); |
---|
| 819 | + pos = pblk_ppa_to_pos(geo, ppa); |
---|
| 820 | + } |
---|
| 821 | + |
---|
| 822 | + if (pblk_boundary_paddr_checks(pblk, paddr + min)) { |
---|
| 823 | + ret = -EINTR; |
---|
| 824 | + goto free_rqd_dma; |
---|
| 825 | + } |
---|
| 826 | + |
---|
| 827 | + for (j = 0; j < min; j++, i++, paddr++) |
---|
| 828 | + ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line_id); |
---|
| 829 | + } |
---|
| 830 | + |
---|
| 831 | + ret = pblk_submit_io_sync(pblk, &rqd, emeta_buf); |
---|
| 832 | + if (ret) { |
---|
| 833 | + pblk_err(pblk, "emeta I/O submission failed: %d\n", ret); |
---|
| 834 | + goto free_rqd_dma; |
---|
| 835 | + } |
---|
| 836 | + |
---|
| 837 | + atomic_dec(&pblk->inflight_io); |
---|
| 838 | + |
---|
| 839 | + if (rqd.error && rqd.error != NVM_RSP_WARN_HIGHECC) { |
---|
| 840 | + pblk_log_read_err(pblk, &rqd); |
---|
| 841 | + ret = -EIO; |
---|
| 842 | + goto free_rqd_dma; |
---|
| 843 | + } |
---|
| 844 | + |
---|
| 845 | + emeta_buf += rq_len; |
---|
| 846 | + left_ppas -= rq_ppas; |
---|
| 847 | + if (left_ppas) |
---|
| 848 | + goto next_rq; |
---|
| 849 | + |
---|
| 850 | +free_rqd_dma: |
---|
| 851 | + nvm_dev_dma_free(dev->parent, rqd.meta_list, rqd.dma_meta_list); |
---|
| 852 | + return ret; |
---|
882 | 853 | } |
---|
883 | 854 | |
---|
884 | 855 | static void pblk_setup_e_rq(struct pblk *pblk, struct nvm_rq *rqd, |
---|
.. | .. |
---|
887 | 858 | rqd->opcode = NVM_OP_ERASE; |
---|
888 | 859 | rqd->ppa_addr = ppa; |
---|
889 | 860 | rqd->nr_ppas = 1; |
---|
890 | | - rqd->flags = pblk_set_progr_mode(pblk, PBLK_ERASE); |
---|
| 861 | + rqd->is_seq = 1; |
---|
891 | 862 | rqd->bio = NULL; |
---|
892 | 863 | } |
---|
893 | 864 | |
---|
.. | .. |
---|
896 | 867 | struct nvm_rq rqd = {NULL}; |
---|
897 | 868 | int ret; |
---|
898 | 869 | |
---|
| 870 | + trace_pblk_chunk_reset(pblk_disk_name(pblk), &ppa, |
---|
| 871 | + PBLK_CHUNK_RESET_START); |
---|
| 872 | + |
---|
899 | 873 | pblk_setup_e_rq(pblk, &rqd, ppa); |
---|
900 | 874 | |
---|
901 | 875 | /* The write thread schedules erases so that it minimizes disturbances |
---|
902 | 876 | * with writes. Thus, there is no need to take the LUN semaphore. |
---|
903 | 877 | */ |
---|
904 | | - ret = pblk_submit_io_sync(pblk, &rqd); |
---|
| 878 | + ret = pblk_submit_io_sync(pblk, &rqd, NULL); |
---|
905 | 879 | rqd.private = pblk; |
---|
906 | 880 | __pblk_end_io_erase(pblk, &rqd); |
---|
907 | 881 | |
---|
.. | .. |
---|
995 | 969 | spin_lock(&l_mg->free_lock); |
---|
996 | 970 | spin_lock(&line->lock); |
---|
997 | 971 | line->state = PBLK_LINESTATE_BAD; |
---|
| 972 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
---|
| 973 | + line->state); |
---|
998 | 974 | spin_unlock(&line->lock); |
---|
999 | 975 | |
---|
1000 | 976 | list_add_tail(&line->list, &l_mg->bad_list); |
---|
.. | .. |
---|
1012 | 988 | bitmap_set(line->lun_bitmap, 0, lm->lun_bitmap_len); |
---|
1013 | 989 | |
---|
1014 | 990 | smeta_buf->header.identifier = cpu_to_le32(PBLK_MAGIC); |
---|
1015 | | - memcpy(smeta_buf->header.uuid, pblk->instance_uuid, 16); |
---|
| 991 | + guid_copy((guid_t *)&smeta_buf->header.uuid, &pblk->instance_uuid); |
---|
1016 | 992 | smeta_buf->header.id = cpu_to_le32(line->id); |
---|
1017 | 993 | smeta_buf->header.type = cpu_to_le16(line->type); |
---|
1018 | 994 | smeta_buf->header.version_major = SMETA_VERSION_MAJOR; |
---|
.. | .. |
---|
1058 | 1034 | static int pblk_line_alloc_bitmaps(struct pblk *pblk, struct pblk_line *line) |
---|
1059 | 1035 | { |
---|
1060 | 1036 | struct pblk_line_meta *lm = &pblk->lm; |
---|
| 1037 | + struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
---|
1061 | 1038 | |
---|
1062 | | - line->map_bitmap = kzalloc(lm->sec_bitmap_len, GFP_KERNEL); |
---|
| 1039 | + line->map_bitmap = mempool_alloc(l_mg->bitmap_pool, GFP_KERNEL); |
---|
1063 | 1040 | if (!line->map_bitmap) |
---|
1064 | 1041 | return -ENOMEM; |
---|
1065 | 1042 | |
---|
| 1043 | + memset(line->map_bitmap, 0, lm->sec_bitmap_len); |
---|
| 1044 | + |
---|
1066 | 1045 | /* will be initialized using bb info from map_bitmap */ |
---|
1067 | | - line->invalid_bitmap = kmalloc(lm->sec_bitmap_len, GFP_KERNEL); |
---|
| 1046 | + line->invalid_bitmap = mempool_alloc(l_mg->bitmap_pool, GFP_KERNEL); |
---|
1068 | 1047 | if (!line->invalid_bitmap) { |
---|
1069 | | - kfree(line->map_bitmap); |
---|
| 1048 | + mempool_free(line->map_bitmap, l_mg->bitmap_pool); |
---|
1070 | 1049 | line->map_bitmap = NULL; |
---|
1071 | 1050 | return -ENOMEM; |
---|
1072 | 1051 | } |
---|
.. | .. |
---|
1106 | 1085 | off = bit * geo->ws_opt; |
---|
1107 | 1086 | bitmap_set(line->map_bitmap, off, lm->smeta_sec); |
---|
1108 | 1087 | line->sec_in_line -= lm->smeta_sec; |
---|
1109 | | - line->smeta_ssec = off; |
---|
1110 | 1088 | line->cur_sec = off + lm->smeta_sec; |
---|
1111 | 1089 | |
---|
1112 | | - if (init && pblk_line_submit_smeta_io(pblk, line, off, PBLK_WRITE)) { |
---|
| 1090 | + if (init && pblk_line_smeta_write(pblk, line, off)) { |
---|
1113 | 1091 | pblk_debug(pblk, "line smeta I/O failed. Retry\n"); |
---|
1114 | 1092 | return 0; |
---|
1115 | 1093 | } |
---|
.. | .. |
---|
1139 | 1117 | bitmap_weight(line->invalid_bitmap, lm->sec_per_line)) { |
---|
1140 | 1118 | spin_lock(&line->lock); |
---|
1141 | 1119 | line->state = PBLK_LINESTATE_BAD; |
---|
| 1120 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
---|
| 1121 | + line->state); |
---|
1142 | 1122 | spin_unlock(&line->lock); |
---|
1143 | 1123 | |
---|
1144 | 1124 | list_add_tail(&line->list, &l_mg->bad_list); |
---|
.. | .. |
---|
1191 | 1171 | if (line->state == PBLK_LINESTATE_NEW) { |
---|
1192 | 1172 | blk_to_erase = pblk_prepare_new_line(pblk, line); |
---|
1193 | 1173 | line->state = PBLK_LINESTATE_FREE; |
---|
| 1174 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
---|
| 1175 | + line->state); |
---|
1194 | 1176 | } else { |
---|
1195 | 1177 | blk_to_erase = blk_in_line; |
---|
1196 | 1178 | } |
---|
.. | .. |
---|
1208 | 1190 | } |
---|
1209 | 1191 | |
---|
1210 | 1192 | line->state = PBLK_LINESTATE_OPEN; |
---|
| 1193 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
---|
| 1194 | + line->state); |
---|
1211 | 1195 | |
---|
1212 | 1196 | atomic_set(&line->left_eblks, blk_to_erase); |
---|
1213 | 1197 | atomic_set(&line->left_seblks, blk_to_erase); |
---|
.. | .. |
---|
1216 | 1200 | spin_unlock(&line->lock); |
---|
1217 | 1201 | |
---|
1218 | 1202 | kref_init(&line->ref); |
---|
| 1203 | + atomic_set(&line->sec_to_update, 0); |
---|
1219 | 1204 | |
---|
1220 | 1205 | return 0; |
---|
1221 | 1206 | } |
---|
1222 | 1207 | |
---|
| 1208 | +/* Line allocations in the recovery path are always single threaded */ |
---|
1223 | 1209 | int pblk_line_recov_alloc(struct pblk *pblk, struct pblk_line *line) |
---|
1224 | 1210 | { |
---|
1225 | 1211 | struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
---|
.. | .. |
---|
1259 | 1245 | |
---|
1260 | 1246 | void pblk_line_recov_close(struct pblk *pblk, struct pblk_line *line) |
---|
1261 | 1247 | { |
---|
1262 | | - kfree(line->map_bitmap); |
---|
| 1248 | + struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
---|
| 1249 | + |
---|
| 1250 | + mempool_free(line->map_bitmap, l_mg->bitmap_pool); |
---|
1263 | 1251 | line->map_bitmap = NULL; |
---|
1264 | 1252 | line->smeta = NULL; |
---|
1265 | 1253 | line->emeta = NULL; |
---|
.. | .. |
---|
1277 | 1265 | |
---|
1278 | 1266 | void pblk_line_free(struct pblk_line *line) |
---|
1279 | 1267 | { |
---|
1280 | | - kfree(line->map_bitmap); |
---|
1281 | | - kfree(line->invalid_bitmap); |
---|
| 1268 | + struct pblk *pblk = line->pblk; |
---|
| 1269 | + struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
---|
| 1270 | + |
---|
| 1271 | + mempool_free(line->map_bitmap, l_mg->bitmap_pool); |
---|
| 1272 | + mempool_free(line->invalid_bitmap, l_mg->bitmap_pool); |
---|
1282 | 1273 | |
---|
1283 | 1274 | pblk_line_reinit(line); |
---|
1284 | 1275 | } |
---|
.. | .. |
---|
1306 | 1297 | if (unlikely(bit >= lm->blk_per_line)) { |
---|
1307 | 1298 | spin_lock(&line->lock); |
---|
1308 | 1299 | line->state = PBLK_LINESTATE_BAD; |
---|
| 1300 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
---|
| 1301 | + line->state); |
---|
1309 | 1302 | spin_unlock(&line->lock); |
---|
1310 | 1303 | |
---|
1311 | 1304 | list_add_tail(&line->list, &l_mg->bad_list); |
---|
.. | .. |
---|
1440 | 1433 | return line; |
---|
1441 | 1434 | } |
---|
1442 | 1435 | |
---|
| 1436 | +void pblk_ppa_to_line_put(struct pblk *pblk, struct ppa_addr ppa) |
---|
| 1437 | +{ |
---|
| 1438 | + struct pblk_line *line; |
---|
| 1439 | + |
---|
| 1440 | + line = pblk_ppa_to_line(pblk, ppa); |
---|
| 1441 | + kref_put(&line->ref, pblk_line_put_wq); |
---|
| 1442 | +} |
---|
| 1443 | + |
---|
| 1444 | +void pblk_rq_to_line_put(struct pblk *pblk, struct nvm_rq *rqd) |
---|
| 1445 | +{ |
---|
| 1446 | + struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); |
---|
| 1447 | + int i; |
---|
| 1448 | + |
---|
| 1449 | + for (i = 0; i < rqd->nr_ppas; i++) |
---|
| 1450 | + pblk_ppa_to_line_put(pblk, ppa_list[i]); |
---|
| 1451 | +} |
---|
| 1452 | + |
---|
1443 | 1453 | static void pblk_stop_writes(struct pblk *pblk, struct pblk_line *line) |
---|
1444 | 1454 | { |
---|
1445 | 1455 | lockdep_assert_held(&pblk->l_mg.free_lock); |
---|
1446 | 1456 | |
---|
1447 | 1457 | pblk_set_space_limit(pblk); |
---|
1448 | 1458 | pblk->state = PBLK_STATE_STOPPING; |
---|
| 1459 | + trace_pblk_state(pblk_disk_name(pblk), pblk->state); |
---|
1449 | 1460 | } |
---|
1450 | 1461 | |
---|
1451 | 1462 | static void pblk_line_close_meta_sync(struct pblk *pblk) |
---|
.. | .. |
---|
1495 | 1506 | return; |
---|
1496 | 1507 | } |
---|
1497 | 1508 | pblk->state = PBLK_STATE_RECOVERING; |
---|
| 1509 | + trace_pblk_state(pblk_disk_name(pblk), pblk->state); |
---|
1498 | 1510 | spin_unlock(&l_mg->free_lock); |
---|
1499 | 1511 | |
---|
1500 | 1512 | pblk_flush_writer(pblk); |
---|
.. | .. |
---|
1516 | 1528 | |
---|
1517 | 1529 | spin_lock(&l_mg->free_lock); |
---|
1518 | 1530 | pblk->state = PBLK_STATE_STOPPED; |
---|
| 1531 | + trace_pblk_state(pblk_disk_name(pblk), pblk->state); |
---|
1519 | 1532 | l_mg->data_line = NULL; |
---|
1520 | 1533 | l_mg->data_next = NULL; |
---|
1521 | 1534 | spin_unlock(&l_mg->free_lock); |
---|
.. | .. |
---|
1606 | 1619 | |
---|
1607 | 1620 | spin_lock(&line->lock); |
---|
1608 | 1621 | WARN_ON(line->state != PBLK_LINESTATE_GC); |
---|
| 1622 | + if (line->w_err_gc->has_gc_err) { |
---|
| 1623 | + spin_unlock(&line->lock); |
---|
| 1624 | + pblk_err(pblk, "line %d had errors during GC\n", line->id); |
---|
| 1625 | + pblk_put_line_back(pblk, line); |
---|
| 1626 | + line->w_err_gc->has_gc_err = 0; |
---|
| 1627 | + return; |
---|
| 1628 | + } |
---|
| 1629 | + |
---|
1609 | 1630 | line->state = PBLK_LINESTATE_FREE; |
---|
| 1631 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
---|
| 1632 | + line->state); |
---|
1610 | 1633 | line->gc_group = PBLK_LINEGC_NONE; |
---|
1611 | 1634 | pblk_line_free(line); |
---|
1612 | 1635 | |
---|
.. | .. |
---|
1675 | 1698 | rqd->end_io = pblk_end_io_erase; |
---|
1676 | 1699 | rqd->private = pblk; |
---|
1677 | 1700 | |
---|
| 1701 | + trace_pblk_chunk_reset(pblk_disk_name(pblk), |
---|
| 1702 | + &ppa, PBLK_CHUNK_RESET_START); |
---|
| 1703 | + |
---|
1678 | 1704 | /* The write thread schedules erases so that it minimizes disturbances |
---|
1679 | 1705 | * with writes. Thus, there is no need to take the LUN semaphore. |
---|
1680 | 1706 | */ |
---|
1681 | | - err = pblk_submit_io(pblk, rqd); |
---|
| 1707 | + err = pblk_submit_io(pblk, rqd, NULL); |
---|
1682 | 1708 | if (err) { |
---|
1683 | 1709 | struct nvm_tgt_dev *dev = pblk->dev; |
---|
1684 | 1710 | struct nvm_geo *geo = &dev->geo; |
---|
1685 | 1711 | |
---|
1686 | 1712 | pblk_err(pblk, "could not async erase line:%d,blk:%d\n", |
---|
1687 | | - pblk_ppa_to_line(ppa), |
---|
| 1713 | + pblk_ppa_to_line_id(ppa), |
---|
1688 | 1714 | pblk_ppa_to_pos(geo, ppa)); |
---|
1689 | 1715 | } |
---|
1690 | 1716 | |
---|
.. | .. |
---|
1736 | 1762 | WARN_ON(line->state != PBLK_LINESTATE_OPEN); |
---|
1737 | 1763 | line->state = PBLK_LINESTATE_CLOSED; |
---|
1738 | 1764 | move_list = pblk_line_gc_list(pblk, line); |
---|
1739 | | - |
---|
1740 | 1765 | list_add_tail(&line->list, move_list); |
---|
1741 | 1766 | |
---|
1742 | | - kfree(line->map_bitmap); |
---|
| 1767 | + mempool_free(line->map_bitmap, l_mg->bitmap_pool); |
---|
1743 | 1768 | line->map_bitmap = NULL; |
---|
1744 | 1769 | line->smeta = NULL; |
---|
1745 | 1770 | line->emeta = NULL; |
---|
.. | .. |
---|
1755 | 1780 | |
---|
1756 | 1781 | spin_unlock(&line->lock); |
---|
1757 | 1782 | spin_unlock(&l_mg->gc_lock); |
---|
| 1783 | + |
---|
| 1784 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
---|
| 1785 | + line->state); |
---|
1758 | 1786 | } |
---|
1759 | 1787 | |
---|
1760 | 1788 | void pblk_line_close_meta(struct pblk *pblk, struct pblk_line *line) |
---|
.. | .. |
---|
1775 | 1803 | |
---|
1776 | 1804 | if (le32_to_cpu(emeta_buf->header.identifier) != PBLK_MAGIC) { |
---|
1777 | 1805 | emeta_buf->header.identifier = cpu_to_le32(PBLK_MAGIC); |
---|
1778 | | - memcpy(emeta_buf->header.uuid, pblk->instance_uuid, 16); |
---|
| 1806 | + guid_copy((guid_t *)&emeta_buf->header.uuid, |
---|
| 1807 | + &pblk->instance_uuid); |
---|
1779 | 1808 | emeta_buf->header.id = cpu_to_le32(line->id); |
---|
1780 | 1809 | emeta_buf->header.type = cpu_to_le16(line->type); |
---|
1781 | 1810 | emeta_buf->header.version_major = EMETA_VERSION_MAJOR; |
---|
.. | .. |
---|
1806 | 1835 | static void pblk_save_lba_list(struct pblk *pblk, struct pblk_line *line) |
---|
1807 | 1836 | { |
---|
1808 | 1837 | struct pblk_line_meta *lm = &pblk->lm; |
---|
1809 | | - struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
---|
1810 | 1838 | unsigned int lba_list_size = lm->emeta_len[2]; |
---|
1811 | 1839 | struct pblk_w_err_gc *w_err_gc = line->w_err_gc; |
---|
1812 | 1840 | struct pblk_emeta *emeta = line->emeta; |
---|
1813 | 1841 | |
---|
1814 | | - w_err_gc->lba_list = pblk_malloc(lba_list_size, |
---|
1815 | | - l_mg->emeta_alloc_type, GFP_KERNEL); |
---|
| 1842 | + w_err_gc->lba_list = kvmalloc(lba_list_size, GFP_KERNEL); |
---|
1816 | 1843 | memcpy(w_err_gc->lba_list, emeta_to_lbas(pblk, emeta->buf), |
---|
1817 | 1844 | lba_list_size); |
---|
1818 | 1845 | } |
---|
.. | .. |
---|
1851 | 1878 | queue_work(wq, &line_ws->ws); |
---|
1852 | 1879 | } |
---|
1853 | 1880 | |
---|
1854 | | -static void __pblk_down_page(struct pblk *pblk, struct ppa_addr *ppa_list, |
---|
1855 | | - int nr_ppas, int pos) |
---|
| 1881 | +static void __pblk_down_chunk(struct pblk *pblk, int pos) |
---|
1856 | 1882 | { |
---|
1857 | 1883 | struct pblk_lun *rlun = &pblk->luns[pos]; |
---|
1858 | 1884 | int ret; |
---|
.. | .. |
---|
1861 | 1887 | * Only send one inflight I/O per LUN. Since we map at a page |
---|
1862 | 1888 | * granurality, all ppas in the I/O will map to the same LUN |
---|
1863 | 1889 | */ |
---|
1864 | | -#ifdef CONFIG_NVM_PBLK_DEBUG |
---|
1865 | | - int i; |
---|
1866 | | - |
---|
1867 | | - for (i = 1; i < nr_ppas; i++) |
---|
1868 | | - WARN_ON(ppa_list[0].a.lun != ppa_list[i].a.lun || |
---|
1869 | | - ppa_list[0].a.ch != ppa_list[i].a.ch); |
---|
1870 | | -#endif |
---|
1871 | 1890 | |
---|
1872 | 1891 | ret = down_timeout(&rlun->wr_sem, msecs_to_jiffies(30000)); |
---|
1873 | 1892 | if (ret == -ETIME || ret == -EINTR) |
---|
.. | .. |
---|
1875 | 1894 | -ret); |
---|
1876 | 1895 | } |
---|
1877 | 1896 | |
---|
1878 | | -void pblk_down_page(struct pblk *pblk, struct ppa_addr *ppa_list, int nr_ppas) |
---|
| 1897 | +void pblk_down_chunk(struct pblk *pblk, struct ppa_addr ppa) |
---|
1879 | 1898 | { |
---|
1880 | 1899 | struct nvm_tgt_dev *dev = pblk->dev; |
---|
1881 | 1900 | struct nvm_geo *geo = &dev->geo; |
---|
1882 | | - int pos = pblk_ppa_to_pos(geo, ppa_list[0]); |
---|
| 1901 | + int pos = pblk_ppa_to_pos(geo, ppa); |
---|
1883 | 1902 | |
---|
1884 | | - __pblk_down_page(pblk, ppa_list, nr_ppas, pos); |
---|
| 1903 | + __pblk_down_chunk(pblk, pos); |
---|
1885 | 1904 | } |
---|
1886 | 1905 | |
---|
1887 | | -void pblk_down_rq(struct pblk *pblk, struct ppa_addr *ppa_list, int nr_ppas, |
---|
| 1906 | +void pblk_down_rq(struct pblk *pblk, struct ppa_addr ppa, |
---|
1888 | 1907 | unsigned long *lun_bitmap) |
---|
1889 | 1908 | { |
---|
1890 | 1909 | struct nvm_tgt_dev *dev = pblk->dev; |
---|
1891 | 1910 | struct nvm_geo *geo = &dev->geo; |
---|
1892 | | - int pos = pblk_ppa_to_pos(geo, ppa_list[0]); |
---|
| 1911 | + int pos = pblk_ppa_to_pos(geo, ppa); |
---|
1893 | 1912 | |
---|
1894 | 1913 | /* If the LUN has been locked for this same request, do no attempt to |
---|
1895 | 1914 | * lock it again |
---|
.. | .. |
---|
1897 | 1916 | if (test_and_set_bit(pos, lun_bitmap)) |
---|
1898 | 1917 | return; |
---|
1899 | 1918 | |
---|
1900 | | - __pblk_down_page(pblk, ppa_list, nr_ppas, pos); |
---|
| 1919 | + __pblk_down_chunk(pblk, pos); |
---|
1901 | 1920 | } |
---|
1902 | 1921 | |
---|
1903 | | -void pblk_up_page(struct pblk *pblk, struct ppa_addr *ppa_list, int nr_ppas) |
---|
| 1922 | +void pblk_up_chunk(struct pblk *pblk, struct ppa_addr ppa) |
---|
1904 | 1923 | { |
---|
1905 | 1924 | struct nvm_tgt_dev *dev = pblk->dev; |
---|
1906 | 1925 | struct nvm_geo *geo = &dev->geo; |
---|
1907 | 1926 | struct pblk_lun *rlun; |
---|
1908 | | - int pos = pblk_ppa_to_pos(geo, ppa_list[0]); |
---|
1909 | | - |
---|
1910 | | -#ifdef CONFIG_NVM_PBLK_DEBUG |
---|
1911 | | - int i; |
---|
1912 | | - |
---|
1913 | | - for (i = 1; i < nr_ppas; i++) |
---|
1914 | | - WARN_ON(ppa_list[0].a.lun != ppa_list[i].a.lun || |
---|
1915 | | - ppa_list[0].a.ch != ppa_list[i].a.ch); |
---|
1916 | | -#endif |
---|
| 1927 | + int pos = pblk_ppa_to_pos(geo, ppa); |
---|
1917 | 1928 | |
---|
1918 | 1929 | rlun = &pblk->luns[pos]; |
---|
1919 | 1930 | up(&rlun->wr_sem); |
---|
1920 | 1931 | } |
---|
1921 | 1932 | |
---|
1922 | | -void pblk_up_rq(struct pblk *pblk, struct ppa_addr *ppa_list, int nr_ppas, |
---|
1923 | | - unsigned long *lun_bitmap) |
---|
| 1933 | +void pblk_up_rq(struct pblk *pblk, unsigned long *lun_bitmap) |
---|
1924 | 1934 | { |
---|
1925 | 1935 | struct nvm_tgt_dev *dev = pblk->dev; |
---|
1926 | 1936 | struct nvm_geo *geo = &dev->geo; |
---|
.. | .. |
---|
1939 | 1949 | struct ppa_addr ppa_l2p; |
---|
1940 | 1950 | |
---|
1941 | 1951 | /* logic error: lba out-of-bounds. Ignore update */ |
---|
1942 | | - if (!(lba < pblk->rl.nr_secs)) { |
---|
| 1952 | + if (!(lba < pblk->capacity)) { |
---|
1943 | 1953 | WARN(1, "pblk: corrupted L2P map request\n"); |
---|
1944 | 1954 | return; |
---|
1945 | 1955 | } |
---|
.. | .. |
---|
1979 | 1989 | #endif |
---|
1980 | 1990 | |
---|
1981 | 1991 | /* logic error: lba out-of-bounds. Ignore update */ |
---|
1982 | | - if (!(lba < pblk->rl.nr_secs)) { |
---|
| 1992 | + if (!(lba < pblk->capacity)) { |
---|
1983 | 1993 | WARN(1, "pblk: corrupted L2P map request\n"); |
---|
1984 | 1994 | return 0; |
---|
1985 | 1995 | } |
---|
.. | .. |
---|
2025 | 2035 | } |
---|
2026 | 2036 | |
---|
2027 | 2037 | /* logic error: lba out-of-bounds. Ignore update */ |
---|
2028 | | - if (!(lba < pblk->rl.nr_secs)) { |
---|
| 2038 | + if (!(lba < pblk->capacity)) { |
---|
2029 | 2039 | WARN(1, "pblk: corrupted L2P map request\n"); |
---|
2030 | 2040 | return; |
---|
2031 | 2041 | } |
---|
.. | .. |
---|
2051 | 2061 | spin_unlock(&pblk->trans_lock); |
---|
2052 | 2062 | } |
---|
2053 | 2063 | |
---|
2054 | | -void pblk_lookup_l2p_seq(struct pblk *pblk, struct ppa_addr *ppas, |
---|
2055 | | - sector_t blba, int nr_secs) |
---|
| 2064 | +int pblk_lookup_l2p_seq(struct pblk *pblk, struct ppa_addr *ppas, |
---|
| 2065 | + sector_t blba, int nr_secs, bool *from_cache) |
---|
2056 | 2066 | { |
---|
2057 | 2067 | int i; |
---|
2058 | 2068 | |
---|
.. | .. |
---|
2064 | 2074 | |
---|
2065 | 2075 | /* If the L2P entry maps to a line, the reference is valid */ |
---|
2066 | 2076 | if (!pblk_ppa_empty(ppa) && !pblk_addr_in_cache(ppa)) { |
---|
2067 | | - int line_id = pblk_ppa_to_line(ppa); |
---|
2068 | | - struct pblk_line *line = &pblk->lines[line_id]; |
---|
| 2077 | + struct pblk_line *line = pblk_ppa_to_line(pblk, ppa); |
---|
| 2078 | + |
---|
| 2079 | + if (i > 0 && *from_cache) |
---|
| 2080 | + break; |
---|
| 2081 | + *from_cache = false; |
---|
2069 | 2082 | |
---|
2070 | 2083 | kref_get(&line->ref); |
---|
| 2084 | + } else { |
---|
| 2085 | + if (i > 0 && !*from_cache) |
---|
| 2086 | + break; |
---|
| 2087 | + *from_cache = true; |
---|
2071 | 2088 | } |
---|
2072 | 2089 | } |
---|
2073 | 2090 | spin_unlock(&pblk->trans_lock); |
---|
| 2091 | + return i; |
---|
2074 | 2092 | } |
---|
2075 | 2093 | |
---|
2076 | 2094 | void pblk_lookup_l2p_rand(struct pblk *pblk, struct ppa_addr *ppas, |
---|
.. | .. |
---|
2084 | 2102 | lba = lba_list[i]; |
---|
2085 | 2103 | if (lba != ADDR_EMPTY) { |
---|
2086 | 2104 | /* logic error: lba out-of-bounds. Ignore update */ |
---|
2087 | | - if (!(lba < pblk->rl.nr_secs)) { |
---|
| 2105 | + if (!(lba < pblk->capacity)) { |
---|
2088 | 2106 | WARN(1, "pblk: corrupted L2P map request\n"); |
---|
2089 | 2107 | continue; |
---|
2090 | 2108 | } |
---|
.. | .. |
---|
2093 | 2111 | } |
---|
2094 | 2112 | spin_unlock(&pblk->trans_lock); |
---|
2095 | 2113 | } |
---|
| 2114 | + |
---|
| 2115 | +void *pblk_get_meta_for_writes(struct pblk *pblk, struct nvm_rq *rqd) |
---|
| 2116 | +{ |
---|
| 2117 | + void *buffer; |
---|
| 2118 | + |
---|
| 2119 | + if (pblk_is_oob_meta_supported(pblk)) { |
---|
| 2120 | + /* Just use OOB metadata buffer as always */ |
---|
| 2121 | + buffer = rqd->meta_list; |
---|
| 2122 | + } else { |
---|
| 2123 | + /* We need to reuse last page of request (packed metadata) |
---|
| 2124 | + * in similar way as traditional oob metadata |
---|
| 2125 | + */ |
---|
| 2126 | + buffer = page_to_virt( |
---|
| 2127 | + rqd->bio->bi_io_vec[rqd->bio->bi_vcnt - 1].bv_page); |
---|
| 2128 | + } |
---|
| 2129 | + |
---|
| 2130 | + return buffer; |
---|
| 2131 | +} |
---|
| 2132 | + |
---|
| 2133 | +void pblk_get_packed_meta(struct pblk *pblk, struct nvm_rq *rqd) |
---|
| 2134 | +{ |
---|
| 2135 | + void *meta_list = rqd->meta_list; |
---|
| 2136 | + void *page; |
---|
| 2137 | + int i = 0; |
---|
| 2138 | + |
---|
| 2139 | + if (pblk_is_oob_meta_supported(pblk)) |
---|
| 2140 | + return; |
---|
| 2141 | + |
---|
| 2142 | + page = page_to_virt(rqd->bio->bi_io_vec[rqd->bio->bi_vcnt - 1].bv_page); |
---|
| 2143 | + /* We need to fill oob meta buffer with data from packed metadata */ |
---|
| 2144 | + for (; i < rqd->nr_ppas; i++) |
---|
| 2145 | + memcpy(pblk_get_meta(pblk, meta_list, i), |
---|
| 2146 | + page + (i * sizeof(struct pblk_sec_meta)), |
---|
| 2147 | + sizeof(struct pblk_sec_meta)); |
---|
| 2148 | +} |
---|