| .. | .. |
|---|
| 1 | +// SPDX-License-Identifier: GPL-2.0 |
|---|
| 1 | 2 | /* |
|---|
| 2 | 3 | * Copyright (C) 2016 CNEX Labs |
|---|
| 3 | 4 | * Initial release: Javier Gonzalez <javier@cnexlabs.com> |
|---|
| .. | .. |
|---|
| 16 | 17 | * |
|---|
| 17 | 18 | */ |
|---|
| 18 | 19 | |
|---|
| 20 | +#define CREATE_TRACE_POINTS |
|---|
| 21 | + |
|---|
| 19 | 22 | #include "pblk.h" |
|---|
| 23 | +#include "pblk-trace.h" |
|---|
| 20 | 24 | |
|---|
| 21 | 25 | static void pblk_line_mark_bb(struct work_struct *work) |
|---|
| 22 | 26 | { |
|---|
| .. | .. |
|---|
| 27 | 31 | struct ppa_addr *ppa = line_ws->priv; |
|---|
| 28 | 32 | int ret; |
|---|
| 29 | 33 | |
|---|
| 30 | | - ret = nvm_set_tgt_bb_tbl(dev, ppa, 1, NVM_BLK_T_GRWN_BAD); |
|---|
| 34 | + ret = nvm_set_chunk_meta(dev, ppa, 1, NVM_BLK_T_GRWN_BAD); |
|---|
| 31 | 35 | if (ret) { |
|---|
| 32 | 36 | struct pblk_line *line; |
|---|
| 33 | 37 | int pos; |
|---|
| 34 | 38 | |
|---|
| 35 | | - line = &pblk->lines[pblk_ppa_to_line(*ppa)]; |
|---|
| 39 | + line = pblk_ppa_to_line(pblk, *ppa); |
|---|
| 36 | 40 | pos = pblk_ppa_to_pos(&dev->geo, *ppa); |
|---|
| 37 | 41 | |
|---|
| 38 | 42 | pblk_err(pblk, "failed to mark bb, line:%d, pos:%d\n", |
|---|
| .. | .. |
|---|
| 80 | 84 | struct pblk_line *line; |
|---|
| 81 | 85 | int pos; |
|---|
| 82 | 86 | |
|---|
| 83 | | - line = &pblk->lines[pblk_ppa_to_line(rqd->ppa_addr)]; |
|---|
| 87 | + line = pblk_ppa_to_line(pblk, rqd->ppa_addr); |
|---|
| 84 | 88 | pos = pblk_ppa_to_pos(geo, rqd->ppa_addr); |
|---|
| 85 | 89 | chunk = &line->chks[pos]; |
|---|
| 86 | 90 | |
|---|
| 87 | 91 | atomic_dec(&line->left_seblks); |
|---|
| 88 | 92 | |
|---|
| 89 | 93 | if (rqd->error) { |
|---|
| 94 | + trace_pblk_chunk_reset(pblk_disk_name(pblk), |
|---|
| 95 | + &rqd->ppa_addr, PBLK_CHUNK_RESET_FAILED); |
|---|
| 96 | + |
|---|
| 90 | 97 | chunk->state = NVM_CHK_ST_OFFLINE; |
|---|
| 91 | 98 | pblk_mark_bb(pblk, line, rqd->ppa_addr); |
|---|
| 92 | 99 | } else { |
|---|
| 100 | + trace_pblk_chunk_reset(pblk_disk_name(pblk), |
|---|
| 101 | + &rqd->ppa_addr, PBLK_CHUNK_RESET_DONE); |
|---|
| 102 | + |
|---|
| 93 | 103 | chunk->state = NVM_CHK_ST_FREE; |
|---|
| 94 | 104 | } |
|---|
| 105 | + |
|---|
| 106 | + trace_pblk_chunk_state(pblk_disk_name(pblk), &rqd->ppa_addr, |
|---|
| 107 | + chunk->state); |
|---|
| 95 | 108 | |
|---|
| 96 | 109 | atomic_dec(&pblk->inflight_io); |
|---|
| 97 | 110 | } |
|---|
| .. | .. |
|---|
| 108 | 121 | /* |
|---|
| 109 | 122 | * Get information for all chunks from the device. |
|---|
| 110 | 123 | * |
|---|
| 111 | | - * The caller is responsible for freeing the returned structure |
|---|
| 124 | + * The caller is responsible for freeing (vmalloc) the returned structure |
|---|
| 112 | 125 | */ |
|---|
| 113 | | -struct nvm_chk_meta *pblk_chunk_get_info(struct pblk *pblk) |
|---|
| 126 | +struct nvm_chk_meta *pblk_get_chunk_meta(struct pblk *pblk) |
|---|
| 114 | 127 | { |
|---|
| 115 | 128 | struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 116 | 129 | struct nvm_geo *geo = &dev->geo; |
|---|
| .. | .. |
|---|
| 122 | 135 | ppa.ppa = 0; |
|---|
| 123 | 136 | |
|---|
| 124 | 137 | len = geo->all_chunks * sizeof(*meta); |
|---|
| 125 | | - meta = kzalloc(len, GFP_KERNEL); |
|---|
| 138 | + meta = vzalloc(len); |
|---|
| 126 | 139 | if (!meta) |
|---|
| 127 | 140 | return ERR_PTR(-ENOMEM); |
|---|
| 128 | 141 | |
|---|
| 129 | | - ret = nvm_get_chunk_meta(dev, meta, ppa, geo->all_chunks); |
|---|
| 142 | + ret = nvm_get_chunk_meta(dev, ppa, geo->all_chunks, meta); |
|---|
| 130 | 143 | if (ret) { |
|---|
| 131 | | - kfree(meta); |
|---|
| 144 | + vfree(meta); |
|---|
| 132 | 145 | return ERR_PTR(-EIO); |
|---|
| 133 | 146 | } |
|---|
| 134 | 147 | |
|---|
| .. | .. |
|---|
| 192 | 205 | { |
|---|
| 193 | 206 | struct pblk_line *line; |
|---|
| 194 | 207 | u64 paddr; |
|---|
| 195 | | - int line_id; |
|---|
| 196 | 208 | |
|---|
| 197 | 209 | #ifdef CONFIG_NVM_PBLK_DEBUG |
|---|
| 198 | 210 | /* Callers must ensure that the ppa points to a device address */ |
|---|
| .. | .. |
|---|
| 200 | 212 | BUG_ON(pblk_ppa_empty(ppa)); |
|---|
| 201 | 213 | #endif |
|---|
| 202 | 214 | |
|---|
| 203 | | - line_id = pblk_ppa_to_line(ppa); |
|---|
| 204 | | - line = &pblk->lines[line_id]; |
|---|
| 215 | + line = pblk_ppa_to_line(pblk, ppa); |
|---|
| 205 | 216 | paddr = pblk_dev_ppa_to_line_addr(pblk, ppa); |
|---|
| 206 | 217 | |
|---|
| 207 | 218 | __pblk_map_invalidate(pblk, line, paddr); |
|---|
| .. | .. |
|---|
| 225 | 236 | pblk_trans_map_set(pblk, lba, ppa); |
|---|
| 226 | 237 | } |
|---|
| 227 | 238 | spin_unlock(&pblk->trans_lock); |
|---|
| 239 | +} |
|---|
| 240 | + |
|---|
| 241 | +int pblk_alloc_rqd_meta(struct pblk *pblk, struct nvm_rq *rqd) |
|---|
| 242 | +{ |
|---|
| 243 | + struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 244 | + |
|---|
| 245 | + rqd->meta_list = nvm_dev_dma_alloc(dev->parent, GFP_KERNEL, |
|---|
| 246 | + &rqd->dma_meta_list); |
|---|
| 247 | + if (!rqd->meta_list) |
|---|
| 248 | + return -ENOMEM; |
|---|
| 249 | + |
|---|
| 250 | + if (rqd->nr_ppas == 1) |
|---|
| 251 | + return 0; |
|---|
| 252 | + |
|---|
| 253 | + rqd->ppa_list = rqd->meta_list + pblk_dma_meta_size(pblk); |
|---|
| 254 | + rqd->dma_ppa_list = rqd->dma_meta_list + pblk_dma_meta_size(pblk); |
|---|
| 255 | + |
|---|
| 256 | + return 0; |
|---|
| 257 | +} |
|---|
| 258 | + |
|---|
| 259 | +void pblk_free_rqd_meta(struct pblk *pblk, struct nvm_rq *rqd) |
|---|
| 260 | +{ |
|---|
| 261 | + struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 262 | + |
|---|
| 263 | + if (rqd->meta_list) |
|---|
| 264 | + nvm_dev_dma_free(dev->parent, rqd->meta_list, |
|---|
| 265 | + rqd->dma_meta_list); |
|---|
| 228 | 266 | } |
|---|
| 229 | 267 | |
|---|
| 230 | 268 | /* Caller must guarantee that the request is a valid type */ |
|---|
| .. | .. |
|---|
| 258 | 296 | /* Typically used on completion path. Cannot guarantee request consistency */ |
|---|
| 259 | 297 | void pblk_free_rqd(struct pblk *pblk, struct nvm_rq *rqd, int type) |
|---|
| 260 | 298 | { |
|---|
| 261 | | - struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 262 | 299 | mempool_t *pool; |
|---|
| 263 | 300 | |
|---|
| 264 | 301 | switch (type) { |
|---|
| 265 | 302 | case PBLK_WRITE: |
|---|
| 266 | 303 | kfree(((struct pblk_c_ctx *)nvm_rq_to_pdu(rqd))->lun_bitmap); |
|---|
| 267 | | - /* fall through */ |
|---|
| 304 | + fallthrough; |
|---|
| 268 | 305 | case PBLK_WRITE_INT: |
|---|
| 269 | 306 | pool = &pblk->w_rq_pool; |
|---|
| 270 | 307 | break; |
|---|
| .. | .. |
|---|
| 279 | 316 | return; |
|---|
| 280 | 317 | } |
|---|
| 281 | 318 | |
|---|
| 282 | | - if (rqd->meta_list) |
|---|
| 283 | | - nvm_dev_dma_free(dev->parent, rqd->meta_list, |
|---|
| 284 | | - rqd->dma_meta_list); |
|---|
| 319 | + pblk_free_rqd_meta(pblk, rqd); |
|---|
| 285 | 320 | mempool_free(rqd, pool); |
|---|
| 286 | 321 | } |
|---|
| 287 | 322 | |
|---|
| .. | .. |
|---|
| 343 | 378 | { |
|---|
| 344 | 379 | unsigned int secs_avail = pblk_rb_read_count(&pblk->rwb); |
|---|
| 345 | 380 | |
|---|
| 346 | | - if (secs_avail >= pblk->min_write_pgs) |
|---|
| 381 | + if (secs_avail >= pblk->min_write_pgs_data) |
|---|
| 347 | 382 | pblk_write_kick(pblk); |
|---|
| 348 | 383 | } |
|---|
| 349 | 384 | |
|---|
| .. | .. |
|---|
| 374 | 409 | struct pblk_line_meta *lm = &pblk->lm; |
|---|
| 375 | 410 | struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
|---|
| 376 | 411 | struct list_head *move_list = NULL; |
|---|
| 377 | | - int vsc = le32_to_cpu(*line->vsc); |
|---|
| 412 | + int packed_meta = (le32_to_cpu(*line->vsc) / pblk->min_write_pgs_data) |
|---|
| 413 | + * (pblk->min_write_pgs - pblk->min_write_pgs_data); |
|---|
| 414 | + int vsc = le32_to_cpu(*line->vsc) + packed_meta; |
|---|
| 378 | 415 | |
|---|
| 379 | 416 | lockdep_assert_held(&line->lock); |
|---|
| 380 | 417 | |
|---|
| .. | .. |
|---|
| 411 | 448 | } |
|---|
| 412 | 449 | } else { |
|---|
| 413 | 450 | line->state = PBLK_LINESTATE_CORRUPT; |
|---|
| 451 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
|---|
| 452 | + line->state); |
|---|
| 453 | + |
|---|
| 414 | 454 | line->gc_group = PBLK_LINEGC_NONE; |
|---|
| 415 | 455 | move_list = &l_mg->corrupt_list; |
|---|
| 416 | 456 | pblk_err(pblk, "corrupted vsc for line %d, vsc:%d (%d/%d/%d)\n", |
|---|
| .. | .. |
|---|
| 467 | 507 | pblk->sec_per_write = sec_per_write; |
|---|
| 468 | 508 | } |
|---|
| 469 | 509 | |
|---|
| 470 | | -int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd) |
|---|
| 510 | +int pblk_submit_io(struct pblk *pblk, struct nvm_rq *rqd, void *buf) |
|---|
| 471 | 511 | { |
|---|
| 472 | 512 | struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 473 | 513 | |
|---|
| .. | .. |
|---|
| 478 | 518 | return NVM_IO_ERR; |
|---|
| 479 | 519 | #endif |
|---|
| 480 | 520 | |
|---|
| 481 | | - return nvm_submit_io(dev, rqd); |
|---|
| 521 | + return nvm_submit_io(dev, rqd, buf); |
|---|
| 482 | 522 | } |
|---|
| 483 | 523 | |
|---|
| 484 | | -int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd) |
|---|
| 524 | +void pblk_check_chunk_state_update(struct pblk *pblk, struct nvm_rq *rqd) |
|---|
| 485 | 525 | { |
|---|
| 486 | | - struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 526 | + struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); |
|---|
| 487 | 527 | |
|---|
| 488 | | - atomic_inc(&pblk->inflight_io); |
|---|
| 528 | + int i; |
|---|
| 489 | 529 | |
|---|
| 490 | | -#ifdef CONFIG_NVM_PBLK_DEBUG |
|---|
| 491 | | - if (pblk_check_io(pblk, rqd)) |
|---|
| 492 | | - return NVM_IO_ERR; |
|---|
| 493 | | -#endif |
|---|
| 530 | + for (i = 0; i < rqd->nr_ppas; i++) { |
|---|
| 531 | + struct ppa_addr *ppa = &ppa_list[i]; |
|---|
| 532 | + struct nvm_chk_meta *chunk = pblk_dev_ppa_to_chunk(pblk, *ppa); |
|---|
| 533 | + u64 caddr = pblk_dev_ppa_to_chunk_addr(pblk, *ppa); |
|---|
| 494 | 534 | |
|---|
| 495 | | - return nvm_submit_io_sync(dev, rqd); |
|---|
| 496 | | -} |
|---|
| 497 | | - |
|---|
| 498 | | -static void pblk_bio_map_addr_endio(struct bio *bio) |
|---|
| 499 | | -{ |
|---|
| 500 | | - bio_put(bio); |
|---|
| 501 | | -} |
|---|
| 502 | | - |
|---|
| 503 | | -struct bio *pblk_bio_map_addr(struct pblk *pblk, void *data, |
|---|
| 504 | | - unsigned int nr_secs, unsigned int len, |
|---|
| 505 | | - int alloc_type, gfp_t gfp_mask) |
|---|
| 506 | | -{ |
|---|
| 507 | | - struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 508 | | - void *kaddr = data; |
|---|
| 509 | | - struct page *page; |
|---|
| 510 | | - struct bio *bio; |
|---|
| 511 | | - int i, ret; |
|---|
| 512 | | - |
|---|
| 513 | | - if (alloc_type == PBLK_KMALLOC_META) |
|---|
| 514 | | - return bio_map_kern(dev->q, kaddr, len, gfp_mask); |
|---|
| 515 | | - |
|---|
| 516 | | - bio = bio_kmalloc(gfp_mask, nr_secs); |
|---|
| 517 | | - if (!bio) |
|---|
| 518 | | - return ERR_PTR(-ENOMEM); |
|---|
| 519 | | - |
|---|
| 520 | | - for (i = 0; i < nr_secs; i++) { |
|---|
| 521 | | - page = vmalloc_to_page(kaddr); |
|---|
| 522 | | - if (!page) { |
|---|
| 523 | | - pblk_err(pblk, "could not map vmalloc bio\n"); |
|---|
| 524 | | - bio_put(bio); |
|---|
| 525 | | - bio = ERR_PTR(-ENOMEM); |
|---|
| 526 | | - goto out; |
|---|
| 527 | | - } |
|---|
| 528 | | - |
|---|
| 529 | | - ret = bio_add_pc_page(dev->q, bio, page, PAGE_SIZE, 0); |
|---|
| 530 | | - if (ret != PAGE_SIZE) { |
|---|
| 531 | | - pblk_err(pblk, "could not add page to bio\n"); |
|---|
| 532 | | - bio_put(bio); |
|---|
| 533 | | - bio = ERR_PTR(-ENOMEM); |
|---|
| 534 | | - goto out; |
|---|
| 535 | | - } |
|---|
| 536 | | - |
|---|
| 537 | | - kaddr += PAGE_SIZE; |
|---|
| 535 | + if (caddr == 0) |
|---|
| 536 | + trace_pblk_chunk_state(pblk_disk_name(pblk), |
|---|
| 537 | + ppa, NVM_CHK_ST_OPEN); |
|---|
| 538 | + else if (caddr == (chunk->cnlb - 1)) |
|---|
| 539 | + trace_pblk_chunk_state(pblk_disk_name(pblk), |
|---|
| 540 | + ppa, NVM_CHK_ST_CLOSED); |
|---|
| 538 | 541 | } |
|---|
| 542 | +} |
|---|
| 539 | 543 | |
|---|
| 540 | | - bio->bi_end_io = pblk_bio_map_addr_endio; |
|---|
| 541 | | -out: |
|---|
| 542 | | - return bio; |
|---|
| 544 | +int pblk_submit_io_sync(struct pblk *pblk, struct nvm_rq *rqd, void *buf) |
|---|
| 545 | +{ |
|---|
| 546 | + struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 547 | + int ret; |
|---|
| 548 | + |
|---|
| 549 | + atomic_inc(&pblk->inflight_io); |
|---|
| 550 | + |
|---|
| 551 | +#ifdef CONFIG_NVM_PBLK_DEBUG |
|---|
| 552 | + if (pblk_check_io(pblk, rqd)) |
|---|
| 553 | + return NVM_IO_ERR; |
|---|
| 554 | +#endif |
|---|
| 555 | + |
|---|
| 556 | + ret = nvm_submit_io_sync(dev, rqd, buf); |
|---|
| 557 | + |
|---|
| 558 | + if (trace_pblk_chunk_state_enabled() && !ret && |
|---|
| 559 | + rqd->opcode == NVM_OP_PWRITE) |
|---|
| 560 | + pblk_check_chunk_state_update(pblk, rqd); |
|---|
| 561 | + |
|---|
| 562 | + return ret; |
|---|
| 563 | +} |
|---|
| 564 | + |
|---|
| 565 | +static int pblk_submit_io_sync_sem(struct pblk *pblk, struct nvm_rq *rqd, |
|---|
| 566 | + void *buf) |
|---|
| 567 | +{ |
|---|
| 568 | + struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); |
|---|
| 569 | + int ret; |
|---|
| 570 | + |
|---|
| 571 | + pblk_down_chunk(pblk, ppa_list[0]); |
|---|
| 572 | + ret = pblk_submit_io_sync(pblk, rqd, buf); |
|---|
| 573 | + pblk_up_chunk(pblk, ppa_list[0]); |
|---|
| 574 | + |
|---|
| 575 | + return ret; |
|---|
| 543 | 576 | } |
|---|
| 544 | 577 | |
|---|
| 545 | 578 | int pblk_calc_secs(struct pblk *pblk, unsigned long secs_avail, |
|---|
| 546 | | - unsigned long secs_to_flush) |
|---|
| 579 | + unsigned long secs_to_flush, bool skip_meta) |
|---|
| 547 | 580 | { |
|---|
| 548 | 581 | int max = pblk->sec_per_write; |
|---|
| 549 | 582 | int min = pblk->min_write_pgs; |
|---|
| 550 | 583 | int secs_to_sync = 0; |
|---|
| 584 | + |
|---|
| 585 | + if (skip_meta && pblk->min_write_pgs_data != pblk->min_write_pgs) |
|---|
| 586 | + min = max = pblk->min_write_pgs_data; |
|---|
| 551 | 587 | |
|---|
| 552 | 588 | if (secs_avail >= max) |
|---|
| 553 | 589 | secs_to_sync = max; |
|---|
| .. | .. |
|---|
| 623 | 659 | return paddr; |
|---|
| 624 | 660 | } |
|---|
| 625 | 661 | |
|---|
| 626 | | -/* |
|---|
| 627 | | - * Submit emeta to one LUN in the raid line at the time to avoid a deadlock when |
|---|
| 628 | | - * taking the per LUN semaphore. |
|---|
| 629 | | - */ |
|---|
| 630 | | -static int pblk_line_submit_emeta_io(struct pblk *pblk, struct pblk_line *line, |
|---|
| 631 | | - void *emeta_buf, u64 paddr, int dir) |
|---|
| 632 | | -{ |
|---|
| 633 | | - struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 634 | | - struct nvm_geo *geo = &dev->geo; |
|---|
| 635 | | - struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
|---|
| 636 | | - struct pblk_line_meta *lm = &pblk->lm; |
|---|
| 637 | | - void *ppa_list, *meta_list; |
|---|
| 638 | | - struct bio *bio; |
|---|
| 639 | | - struct nvm_rq rqd; |
|---|
| 640 | | - dma_addr_t dma_ppa_list, dma_meta_list; |
|---|
| 641 | | - int min = pblk->min_write_pgs; |
|---|
| 642 | | - int left_ppas = lm->emeta_sec[0]; |
|---|
| 643 | | - int id = line->id; |
|---|
| 644 | | - int rq_ppas, rq_len; |
|---|
| 645 | | - int cmd_op, bio_op; |
|---|
| 646 | | - int i, j; |
|---|
| 647 | | - int ret; |
|---|
| 648 | | - |
|---|
| 649 | | - if (dir == PBLK_WRITE) { |
|---|
| 650 | | - bio_op = REQ_OP_WRITE; |
|---|
| 651 | | - cmd_op = NVM_OP_PWRITE; |
|---|
| 652 | | - } else if (dir == PBLK_READ) { |
|---|
| 653 | | - bio_op = REQ_OP_READ; |
|---|
| 654 | | - cmd_op = NVM_OP_PREAD; |
|---|
| 655 | | - } else |
|---|
| 656 | | - return -EINVAL; |
|---|
| 657 | | - |
|---|
| 658 | | - meta_list = nvm_dev_dma_alloc(dev->parent, GFP_KERNEL, |
|---|
| 659 | | - &dma_meta_list); |
|---|
| 660 | | - if (!meta_list) |
|---|
| 661 | | - return -ENOMEM; |
|---|
| 662 | | - |
|---|
| 663 | | - ppa_list = meta_list + pblk_dma_meta_size; |
|---|
| 664 | | - dma_ppa_list = dma_meta_list + pblk_dma_meta_size; |
|---|
| 665 | | - |
|---|
| 666 | | -next_rq: |
|---|
| 667 | | - memset(&rqd, 0, sizeof(struct nvm_rq)); |
|---|
| 668 | | - |
|---|
| 669 | | - rq_ppas = pblk_calc_secs(pblk, left_ppas, 0); |
|---|
| 670 | | - rq_len = rq_ppas * geo->csecs; |
|---|
| 671 | | - |
|---|
| 672 | | - bio = pblk_bio_map_addr(pblk, emeta_buf, rq_ppas, rq_len, |
|---|
| 673 | | - l_mg->emeta_alloc_type, GFP_KERNEL); |
|---|
| 674 | | - if (IS_ERR(bio)) { |
|---|
| 675 | | - ret = PTR_ERR(bio); |
|---|
| 676 | | - goto free_rqd_dma; |
|---|
| 677 | | - } |
|---|
| 678 | | - |
|---|
| 679 | | - bio->bi_iter.bi_sector = 0; /* internal bio */ |
|---|
| 680 | | - bio_set_op_attrs(bio, bio_op, 0); |
|---|
| 681 | | - |
|---|
| 682 | | - rqd.bio = bio; |
|---|
| 683 | | - rqd.meta_list = meta_list; |
|---|
| 684 | | - rqd.ppa_list = ppa_list; |
|---|
| 685 | | - rqd.dma_meta_list = dma_meta_list; |
|---|
| 686 | | - rqd.dma_ppa_list = dma_ppa_list; |
|---|
| 687 | | - rqd.opcode = cmd_op; |
|---|
| 688 | | - rqd.nr_ppas = rq_ppas; |
|---|
| 689 | | - |
|---|
| 690 | | - if (dir == PBLK_WRITE) { |
|---|
| 691 | | - struct pblk_sec_meta *meta_list = rqd.meta_list; |
|---|
| 692 | | - |
|---|
| 693 | | - rqd.flags = pblk_set_progr_mode(pblk, PBLK_WRITE); |
|---|
| 694 | | - for (i = 0; i < rqd.nr_ppas; ) { |
|---|
| 695 | | - spin_lock(&line->lock); |
|---|
| 696 | | - paddr = __pblk_alloc_page(pblk, line, min); |
|---|
| 697 | | - spin_unlock(&line->lock); |
|---|
| 698 | | - for (j = 0; j < min; j++, i++, paddr++) { |
|---|
| 699 | | - meta_list[i].lba = cpu_to_le64(ADDR_EMPTY); |
|---|
| 700 | | - rqd.ppa_list[i] = |
|---|
| 701 | | - addr_to_gen_ppa(pblk, paddr, id); |
|---|
| 702 | | - } |
|---|
| 703 | | - } |
|---|
| 704 | | - } else { |
|---|
| 705 | | - for (i = 0; i < rqd.nr_ppas; ) { |
|---|
| 706 | | - struct ppa_addr ppa = addr_to_gen_ppa(pblk, paddr, id); |
|---|
| 707 | | - int pos = pblk_ppa_to_pos(geo, ppa); |
|---|
| 708 | | - int read_type = PBLK_READ_RANDOM; |
|---|
| 709 | | - |
|---|
| 710 | | - if (pblk_io_aligned(pblk, rq_ppas)) |
|---|
| 711 | | - read_type = PBLK_READ_SEQUENTIAL; |
|---|
| 712 | | - rqd.flags = pblk_set_read_mode(pblk, read_type); |
|---|
| 713 | | - |
|---|
| 714 | | - while (test_bit(pos, line->blk_bitmap)) { |
|---|
| 715 | | - paddr += min; |
|---|
| 716 | | - if (pblk_boundary_paddr_checks(pblk, paddr)) { |
|---|
| 717 | | - pblk_err(pblk, "corrupt emeta line:%d\n", |
|---|
| 718 | | - line->id); |
|---|
| 719 | | - bio_put(bio); |
|---|
| 720 | | - ret = -EINTR; |
|---|
| 721 | | - goto free_rqd_dma; |
|---|
| 722 | | - } |
|---|
| 723 | | - |
|---|
| 724 | | - ppa = addr_to_gen_ppa(pblk, paddr, id); |
|---|
| 725 | | - pos = pblk_ppa_to_pos(geo, ppa); |
|---|
| 726 | | - } |
|---|
| 727 | | - |
|---|
| 728 | | - if (pblk_boundary_paddr_checks(pblk, paddr + min)) { |
|---|
| 729 | | - pblk_err(pblk, "corrupt emeta line:%d\n", |
|---|
| 730 | | - line->id); |
|---|
| 731 | | - bio_put(bio); |
|---|
| 732 | | - ret = -EINTR; |
|---|
| 733 | | - goto free_rqd_dma; |
|---|
| 734 | | - } |
|---|
| 735 | | - |
|---|
| 736 | | - for (j = 0; j < min; j++, i++, paddr++) |
|---|
| 737 | | - rqd.ppa_list[i] = |
|---|
| 738 | | - addr_to_gen_ppa(pblk, paddr, line->id); |
|---|
| 739 | | - } |
|---|
| 740 | | - } |
|---|
| 741 | | - |
|---|
| 742 | | - ret = pblk_submit_io_sync(pblk, &rqd); |
|---|
| 743 | | - if (ret) { |
|---|
| 744 | | - pblk_err(pblk, "emeta I/O submission failed: %d\n", ret); |
|---|
| 745 | | - bio_put(bio); |
|---|
| 746 | | - goto free_rqd_dma; |
|---|
| 747 | | - } |
|---|
| 748 | | - |
|---|
| 749 | | - atomic_dec(&pblk->inflight_io); |
|---|
| 750 | | - |
|---|
| 751 | | - if (rqd.error) { |
|---|
| 752 | | - if (dir == PBLK_WRITE) |
|---|
| 753 | | - pblk_log_write_err(pblk, &rqd); |
|---|
| 754 | | - else |
|---|
| 755 | | - pblk_log_read_err(pblk, &rqd); |
|---|
| 756 | | - } |
|---|
| 757 | | - |
|---|
| 758 | | - emeta_buf += rq_len; |
|---|
| 759 | | - left_ppas -= rq_ppas; |
|---|
| 760 | | - if (left_ppas) |
|---|
| 761 | | - goto next_rq; |
|---|
| 762 | | -free_rqd_dma: |
|---|
| 763 | | - nvm_dev_dma_free(dev->parent, rqd.meta_list, rqd.dma_meta_list); |
|---|
| 764 | | - return ret; |
|---|
| 765 | | -} |
|---|
| 766 | | - |
|---|
| 767 | 662 | u64 pblk_line_smeta_start(struct pblk *pblk, struct pblk_line *line) |
|---|
| 768 | 663 | { |
|---|
| 769 | 664 | struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| .. | .. |
|---|
| 779 | 674 | return bit * geo->ws_opt; |
|---|
| 780 | 675 | } |
|---|
| 781 | 676 | |
|---|
| 782 | | -static int pblk_line_submit_smeta_io(struct pblk *pblk, struct pblk_line *line, |
|---|
| 783 | | - u64 paddr, int dir) |
|---|
| 677 | +int pblk_line_smeta_read(struct pblk *pblk, struct pblk_line *line) |
|---|
| 784 | 678 | { |
|---|
| 785 | | - struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 786 | 679 | struct pblk_line_meta *lm = &pblk->lm; |
|---|
| 787 | | - struct bio *bio; |
|---|
| 680 | + struct ppa_addr *ppa_list; |
|---|
| 788 | 681 | struct nvm_rq rqd; |
|---|
| 789 | | - __le64 *lba_list = NULL; |
|---|
| 682 | + u64 paddr = pblk_line_smeta_start(pblk, line); |
|---|
| 790 | 683 | int i, ret; |
|---|
| 791 | | - int cmd_op, bio_op; |
|---|
| 792 | | - int flags; |
|---|
| 793 | | - |
|---|
| 794 | | - if (dir == PBLK_WRITE) { |
|---|
| 795 | | - bio_op = REQ_OP_WRITE; |
|---|
| 796 | | - cmd_op = NVM_OP_PWRITE; |
|---|
| 797 | | - flags = pblk_set_progr_mode(pblk, PBLK_WRITE); |
|---|
| 798 | | - lba_list = emeta_to_lbas(pblk, line->emeta->buf); |
|---|
| 799 | | - } else if (dir == PBLK_READ_RECOV || dir == PBLK_READ) { |
|---|
| 800 | | - bio_op = REQ_OP_READ; |
|---|
| 801 | | - cmd_op = NVM_OP_PREAD; |
|---|
| 802 | | - flags = pblk_set_read_mode(pblk, PBLK_READ_SEQUENTIAL); |
|---|
| 803 | | - } else |
|---|
| 804 | | - return -EINVAL; |
|---|
| 805 | 684 | |
|---|
| 806 | 685 | memset(&rqd, 0, sizeof(struct nvm_rq)); |
|---|
| 807 | 686 | |
|---|
| 808 | | - rqd.meta_list = nvm_dev_dma_alloc(dev->parent, GFP_KERNEL, |
|---|
| 809 | | - &rqd.dma_meta_list); |
|---|
| 810 | | - if (!rqd.meta_list) |
|---|
| 811 | | - return -ENOMEM; |
|---|
| 687 | + ret = pblk_alloc_rqd_meta(pblk, &rqd); |
|---|
| 688 | + if (ret) |
|---|
| 689 | + return ret; |
|---|
| 812 | 690 | |
|---|
| 813 | | - rqd.ppa_list = rqd.meta_list + pblk_dma_meta_size; |
|---|
| 814 | | - rqd.dma_ppa_list = rqd.dma_meta_list + pblk_dma_meta_size; |
|---|
| 815 | | - |
|---|
| 816 | | - bio = bio_map_kern(dev->q, line->smeta, lm->smeta_len, GFP_KERNEL); |
|---|
| 817 | | - if (IS_ERR(bio)) { |
|---|
| 818 | | - ret = PTR_ERR(bio); |
|---|
| 819 | | - goto free_ppa_list; |
|---|
| 820 | | - } |
|---|
| 821 | | - |
|---|
| 822 | | - bio->bi_iter.bi_sector = 0; /* internal bio */ |
|---|
| 823 | | - bio_set_op_attrs(bio, bio_op, 0); |
|---|
| 824 | | - |
|---|
| 825 | | - rqd.bio = bio; |
|---|
| 826 | | - rqd.opcode = cmd_op; |
|---|
| 827 | | - rqd.flags = flags; |
|---|
| 691 | + rqd.opcode = NVM_OP_PREAD; |
|---|
| 828 | 692 | rqd.nr_ppas = lm->smeta_sec; |
|---|
| 693 | + rqd.is_seq = 1; |
|---|
| 694 | + ppa_list = nvm_rq_to_ppa_list(&rqd); |
|---|
| 829 | 695 | |
|---|
| 830 | | - for (i = 0; i < lm->smeta_sec; i++, paddr++) { |
|---|
| 831 | | - struct pblk_sec_meta *meta_list = rqd.meta_list; |
|---|
| 696 | + for (i = 0; i < lm->smeta_sec; i++, paddr++) |
|---|
| 697 | + ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line->id); |
|---|
| 832 | 698 | |
|---|
| 833 | | - rqd.ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line->id); |
|---|
| 834 | | - |
|---|
| 835 | | - if (dir == PBLK_WRITE) { |
|---|
| 836 | | - __le64 addr_empty = cpu_to_le64(ADDR_EMPTY); |
|---|
| 837 | | - |
|---|
| 838 | | - meta_list[i].lba = lba_list[paddr] = addr_empty; |
|---|
| 839 | | - } |
|---|
| 840 | | - } |
|---|
| 841 | | - |
|---|
| 842 | | - /* |
|---|
| 843 | | - * This I/O is sent by the write thread when a line is replace. Since |
|---|
| 844 | | - * the write thread is the only one sending write and erase commands, |
|---|
| 845 | | - * there is no need to take the LUN semaphore. |
|---|
| 846 | | - */ |
|---|
| 847 | | - ret = pblk_submit_io_sync(pblk, &rqd); |
|---|
| 699 | + ret = pblk_submit_io_sync(pblk, &rqd, line->smeta); |
|---|
| 848 | 700 | if (ret) { |
|---|
| 849 | 701 | pblk_err(pblk, "smeta I/O submission failed: %d\n", ret); |
|---|
| 850 | | - bio_put(bio); |
|---|
| 851 | | - goto free_ppa_list; |
|---|
| 702 | + goto clear_rqd; |
|---|
| 703 | + } |
|---|
| 704 | + |
|---|
| 705 | + atomic_dec(&pblk->inflight_io); |
|---|
| 706 | + |
|---|
| 707 | + if (rqd.error && rqd.error != NVM_RSP_WARN_HIGHECC) { |
|---|
| 708 | + pblk_log_read_err(pblk, &rqd); |
|---|
| 709 | + ret = -EIO; |
|---|
| 710 | + } |
|---|
| 711 | + |
|---|
| 712 | +clear_rqd: |
|---|
| 713 | + pblk_free_rqd_meta(pblk, &rqd); |
|---|
| 714 | + return ret; |
|---|
| 715 | +} |
|---|
| 716 | + |
|---|
| 717 | +static int pblk_line_smeta_write(struct pblk *pblk, struct pblk_line *line, |
|---|
| 718 | + u64 paddr) |
|---|
| 719 | +{ |
|---|
| 720 | + struct pblk_line_meta *lm = &pblk->lm; |
|---|
| 721 | + struct ppa_addr *ppa_list; |
|---|
| 722 | + struct nvm_rq rqd; |
|---|
| 723 | + __le64 *lba_list = emeta_to_lbas(pblk, line->emeta->buf); |
|---|
| 724 | + __le64 addr_empty = cpu_to_le64(ADDR_EMPTY); |
|---|
| 725 | + int i, ret; |
|---|
| 726 | + |
|---|
| 727 | + memset(&rqd, 0, sizeof(struct nvm_rq)); |
|---|
| 728 | + |
|---|
| 729 | + ret = pblk_alloc_rqd_meta(pblk, &rqd); |
|---|
| 730 | + if (ret) |
|---|
| 731 | + return ret; |
|---|
| 732 | + |
|---|
| 733 | + rqd.opcode = NVM_OP_PWRITE; |
|---|
| 734 | + rqd.nr_ppas = lm->smeta_sec; |
|---|
| 735 | + rqd.is_seq = 1; |
|---|
| 736 | + ppa_list = nvm_rq_to_ppa_list(&rqd); |
|---|
| 737 | + |
|---|
| 738 | + for (i = 0; i < lm->smeta_sec; i++, paddr++) { |
|---|
| 739 | + struct pblk_sec_meta *meta = pblk_get_meta(pblk, |
|---|
| 740 | + rqd.meta_list, i); |
|---|
| 741 | + |
|---|
| 742 | + ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line->id); |
|---|
| 743 | + meta->lba = lba_list[paddr] = addr_empty; |
|---|
| 744 | + } |
|---|
| 745 | + |
|---|
| 746 | + ret = pblk_submit_io_sync_sem(pblk, &rqd, line->smeta); |
|---|
| 747 | + if (ret) { |
|---|
| 748 | + pblk_err(pblk, "smeta I/O submission failed: %d\n", ret); |
|---|
| 749 | + goto clear_rqd; |
|---|
| 852 | 750 | } |
|---|
| 853 | 751 | |
|---|
| 854 | 752 | atomic_dec(&pblk->inflight_io); |
|---|
| 855 | 753 | |
|---|
| 856 | 754 | if (rqd.error) { |
|---|
| 857 | | - if (dir == PBLK_WRITE) { |
|---|
| 858 | | - pblk_log_write_err(pblk, &rqd); |
|---|
| 859 | | - ret = 1; |
|---|
| 860 | | - } else if (dir == PBLK_READ) |
|---|
| 861 | | - pblk_log_read_err(pblk, &rqd); |
|---|
| 755 | + pblk_log_write_err(pblk, &rqd); |
|---|
| 756 | + ret = -EIO; |
|---|
| 862 | 757 | } |
|---|
| 863 | 758 | |
|---|
| 864 | | -free_ppa_list: |
|---|
| 865 | | - nvm_dev_dma_free(dev->parent, rqd.meta_list, rqd.dma_meta_list); |
|---|
| 866 | | - |
|---|
| 759 | +clear_rqd: |
|---|
| 760 | + pblk_free_rqd_meta(pblk, &rqd); |
|---|
| 867 | 761 | return ret; |
|---|
| 868 | 762 | } |
|---|
| 869 | 763 | |
|---|
| 870 | | -int pblk_line_read_smeta(struct pblk *pblk, struct pblk_line *line) |
|---|
| 871 | | -{ |
|---|
| 872 | | - u64 bpaddr = pblk_line_smeta_start(pblk, line); |
|---|
| 873 | | - |
|---|
| 874 | | - return pblk_line_submit_smeta_io(pblk, line, bpaddr, PBLK_READ_RECOV); |
|---|
| 875 | | -} |
|---|
| 876 | | - |
|---|
| 877 | | -int pblk_line_read_emeta(struct pblk *pblk, struct pblk_line *line, |
|---|
| 764 | +int pblk_line_emeta_read(struct pblk *pblk, struct pblk_line *line, |
|---|
| 878 | 765 | void *emeta_buf) |
|---|
| 879 | 766 | { |
|---|
| 880 | | - return pblk_line_submit_emeta_io(pblk, line, emeta_buf, |
|---|
| 881 | | - line->emeta_ssec, PBLK_READ); |
|---|
| 767 | + struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 768 | + struct nvm_geo *geo = &dev->geo; |
|---|
| 769 | + struct pblk_line_meta *lm = &pblk->lm; |
|---|
| 770 | + void *ppa_list_buf, *meta_list; |
|---|
| 771 | + struct ppa_addr *ppa_list; |
|---|
| 772 | + struct nvm_rq rqd; |
|---|
| 773 | + u64 paddr = line->emeta_ssec; |
|---|
| 774 | + dma_addr_t dma_ppa_list, dma_meta_list; |
|---|
| 775 | + int min = pblk->min_write_pgs; |
|---|
| 776 | + int left_ppas = lm->emeta_sec[0]; |
|---|
| 777 | + int line_id = line->id; |
|---|
| 778 | + int rq_ppas, rq_len; |
|---|
| 779 | + int i, j; |
|---|
| 780 | + int ret; |
|---|
| 781 | + |
|---|
| 782 | + meta_list = nvm_dev_dma_alloc(dev->parent, GFP_KERNEL, |
|---|
| 783 | + &dma_meta_list); |
|---|
| 784 | + if (!meta_list) |
|---|
| 785 | + return -ENOMEM; |
|---|
| 786 | + |
|---|
| 787 | + ppa_list_buf = meta_list + pblk_dma_meta_size(pblk); |
|---|
| 788 | + dma_ppa_list = dma_meta_list + pblk_dma_meta_size(pblk); |
|---|
| 789 | + |
|---|
| 790 | +next_rq: |
|---|
| 791 | + memset(&rqd, 0, sizeof(struct nvm_rq)); |
|---|
| 792 | + |
|---|
| 793 | + rq_ppas = pblk_calc_secs(pblk, left_ppas, 0, false); |
|---|
| 794 | + rq_len = rq_ppas * geo->csecs; |
|---|
| 795 | + |
|---|
| 796 | + rqd.meta_list = meta_list; |
|---|
| 797 | + rqd.ppa_list = ppa_list_buf; |
|---|
| 798 | + rqd.dma_meta_list = dma_meta_list; |
|---|
| 799 | + rqd.dma_ppa_list = dma_ppa_list; |
|---|
| 800 | + rqd.opcode = NVM_OP_PREAD; |
|---|
| 801 | + rqd.nr_ppas = rq_ppas; |
|---|
| 802 | + ppa_list = nvm_rq_to_ppa_list(&rqd); |
|---|
| 803 | + |
|---|
| 804 | + for (i = 0; i < rqd.nr_ppas; ) { |
|---|
| 805 | + struct ppa_addr ppa = addr_to_gen_ppa(pblk, paddr, line_id); |
|---|
| 806 | + int pos = pblk_ppa_to_pos(geo, ppa); |
|---|
| 807 | + |
|---|
| 808 | + if (pblk_io_aligned(pblk, rq_ppas)) |
|---|
| 809 | + rqd.is_seq = 1; |
|---|
| 810 | + |
|---|
| 811 | + while (test_bit(pos, line->blk_bitmap)) { |
|---|
| 812 | + paddr += min; |
|---|
| 813 | + if (pblk_boundary_paddr_checks(pblk, paddr)) { |
|---|
| 814 | + ret = -EINTR; |
|---|
| 815 | + goto free_rqd_dma; |
|---|
| 816 | + } |
|---|
| 817 | + |
|---|
| 818 | + ppa = addr_to_gen_ppa(pblk, paddr, line_id); |
|---|
| 819 | + pos = pblk_ppa_to_pos(geo, ppa); |
|---|
| 820 | + } |
|---|
| 821 | + |
|---|
| 822 | + if (pblk_boundary_paddr_checks(pblk, paddr + min)) { |
|---|
| 823 | + ret = -EINTR; |
|---|
| 824 | + goto free_rqd_dma; |
|---|
| 825 | + } |
|---|
| 826 | + |
|---|
| 827 | + for (j = 0; j < min; j++, i++, paddr++) |
|---|
| 828 | + ppa_list[i] = addr_to_gen_ppa(pblk, paddr, line_id); |
|---|
| 829 | + } |
|---|
| 830 | + |
|---|
| 831 | + ret = pblk_submit_io_sync(pblk, &rqd, emeta_buf); |
|---|
| 832 | + if (ret) { |
|---|
| 833 | + pblk_err(pblk, "emeta I/O submission failed: %d\n", ret); |
|---|
| 834 | + goto free_rqd_dma; |
|---|
| 835 | + } |
|---|
| 836 | + |
|---|
| 837 | + atomic_dec(&pblk->inflight_io); |
|---|
| 838 | + |
|---|
| 839 | + if (rqd.error && rqd.error != NVM_RSP_WARN_HIGHECC) { |
|---|
| 840 | + pblk_log_read_err(pblk, &rqd); |
|---|
| 841 | + ret = -EIO; |
|---|
| 842 | + goto free_rqd_dma; |
|---|
| 843 | + } |
|---|
| 844 | + |
|---|
| 845 | + emeta_buf += rq_len; |
|---|
| 846 | + left_ppas -= rq_ppas; |
|---|
| 847 | + if (left_ppas) |
|---|
| 848 | + goto next_rq; |
|---|
| 849 | + |
|---|
| 850 | +free_rqd_dma: |
|---|
| 851 | + nvm_dev_dma_free(dev->parent, rqd.meta_list, rqd.dma_meta_list); |
|---|
| 852 | + return ret; |
|---|
| 882 | 853 | } |
|---|
| 883 | 854 | |
|---|
| 884 | 855 | static void pblk_setup_e_rq(struct pblk *pblk, struct nvm_rq *rqd, |
|---|
| .. | .. |
|---|
| 887 | 858 | rqd->opcode = NVM_OP_ERASE; |
|---|
| 888 | 859 | rqd->ppa_addr = ppa; |
|---|
| 889 | 860 | rqd->nr_ppas = 1; |
|---|
| 890 | | - rqd->flags = pblk_set_progr_mode(pblk, PBLK_ERASE); |
|---|
| 861 | + rqd->is_seq = 1; |
|---|
| 891 | 862 | rqd->bio = NULL; |
|---|
| 892 | 863 | } |
|---|
| 893 | 864 | |
|---|
| .. | .. |
|---|
| 896 | 867 | struct nvm_rq rqd = {NULL}; |
|---|
| 897 | 868 | int ret; |
|---|
| 898 | 869 | |
|---|
| 870 | + trace_pblk_chunk_reset(pblk_disk_name(pblk), &ppa, |
|---|
| 871 | + PBLK_CHUNK_RESET_START); |
|---|
| 872 | + |
|---|
| 899 | 873 | pblk_setup_e_rq(pblk, &rqd, ppa); |
|---|
| 900 | 874 | |
|---|
| 901 | 875 | /* The write thread schedules erases so that it minimizes disturbances |
|---|
| 902 | 876 | * with writes. Thus, there is no need to take the LUN semaphore. |
|---|
| 903 | 877 | */ |
|---|
| 904 | | - ret = pblk_submit_io_sync(pblk, &rqd); |
|---|
| 878 | + ret = pblk_submit_io_sync(pblk, &rqd, NULL); |
|---|
| 905 | 879 | rqd.private = pblk; |
|---|
| 906 | 880 | __pblk_end_io_erase(pblk, &rqd); |
|---|
| 907 | 881 | |
|---|
| .. | .. |
|---|
| 995 | 969 | spin_lock(&l_mg->free_lock); |
|---|
| 996 | 970 | spin_lock(&line->lock); |
|---|
| 997 | 971 | line->state = PBLK_LINESTATE_BAD; |
|---|
| 972 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
|---|
| 973 | + line->state); |
|---|
| 998 | 974 | spin_unlock(&line->lock); |
|---|
| 999 | 975 | |
|---|
| 1000 | 976 | list_add_tail(&line->list, &l_mg->bad_list); |
|---|
| .. | .. |
|---|
| 1012 | 988 | bitmap_set(line->lun_bitmap, 0, lm->lun_bitmap_len); |
|---|
| 1013 | 989 | |
|---|
| 1014 | 990 | smeta_buf->header.identifier = cpu_to_le32(PBLK_MAGIC); |
|---|
| 1015 | | - memcpy(smeta_buf->header.uuid, pblk->instance_uuid, 16); |
|---|
| 991 | + guid_copy((guid_t *)&smeta_buf->header.uuid, &pblk->instance_uuid); |
|---|
| 1016 | 992 | smeta_buf->header.id = cpu_to_le32(line->id); |
|---|
| 1017 | 993 | smeta_buf->header.type = cpu_to_le16(line->type); |
|---|
| 1018 | 994 | smeta_buf->header.version_major = SMETA_VERSION_MAJOR; |
|---|
| .. | .. |
|---|
| 1058 | 1034 | static int pblk_line_alloc_bitmaps(struct pblk *pblk, struct pblk_line *line) |
|---|
| 1059 | 1035 | { |
|---|
| 1060 | 1036 | struct pblk_line_meta *lm = &pblk->lm; |
|---|
| 1037 | + struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
|---|
| 1061 | 1038 | |
|---|
| 1062 | | - line->map_bitmap = kzalloc(lm->sec_bitmap_len, GFP_KERNEL); |
|---|
| 1039 | + line->map_bitmap = mempool_alloc(l_mg->bitmap_pool, GFP_KERNEL); |
|---|
| 1063 | 1040 | if (!line->map_bitmap) |
|---|
| 1064 | 1041 | return -ENOMEM; |
|---|
| 1065 | 1042 | |
|---|
| 1043 | + memset(line->map_bitmap, 0, lm->sec_bitmap_len); |
|---|
| 1044 | + |
|---|
| 1066 | 1045 | /* will be initialized using bb info from map_bitmap */ |
|---|
| 1067 | | - line->invalid_bitmap = kmalloc(lm->sec_bitmap_len, GFP_KERNEL); |
|---|
| 1046 | + line->invalid_bitmap = mempool_alloc(l_mg->bitmap_pool, GFP_KERNEL); |
|---|
| 1068 | 1047 | if (!line->invalid_bitmap) { |
|---|
| 1069 | | - kfree(line->map_bitmap); |
|---|
| 1048 | + mempool_free(line->map_bitmap, l_mg->bitmap_pool); |
|---|
| 1070 | 1049 | line->map_bitmap = NULL; |
|---|
| 1071 | 1050 | return -ENOMEM; |
|---|
| 1072 | 1051 | } |
|---|
| .. | .. |
|---|
| 1106 | 1085 | off = bit * geo->ws_opt; |
|---|
| 1107 | 1086 | bitmap_set(line->map_bitmap, off, lm->smeta_sec); |
|---|
| 1108 | 1087 | line->sec_in_line -= lm->smeta_sec; |
|---|
| 1109 | | - line->smeta_ssec = off; |
|---|
| 1110 | 1088 | line->cur_sec = off + lm->smeta_sec; |
|---|
| 1111 | 1089 | |
|---|
| 1112 | | - if (init && pblk_line_submit_smeta_io(pblk, line, off, PBLK_WRITE)) { |
|---|
| 1090 | + if (init && pblk_line_smeta_write(pblk, line, off)) { |
|---|
| 1113 | 1091 | pblk_debug(pblk, "line smeta I/O failed. Retry\n"); |
|---|
| 1114 | 1092 | return 0; |
|---|
| 1115 | 1093 | } |
|---|
| .. | .. |
|---|
| 1139 | 1117 | bitmap_weight(line->invalid_bitmap, lm->sec_per_line)) { |
|---|
| 1140 | 1118 | spin_lock(&line->lock); |
|---|
| 1141 | 1119 | line->state = PBLK_LINESTATE_BAD; |
|---|
| 1120 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
|---|
| 1121 | + line->state); |
|---|
| 1142 | 1122 | spin_unlock(&line->lock); |
|---|
| 1143 | 1123 | |
|---|
| 1144 | 1124 | list_add_tail(&line->list, &l_mg->bad_list); |
|---|
| .. | .. |
|---|
| 1191 | 1171 | if (line->state == PBLK_LINESTATE_NEW) { |
|---|
| 1192 | 1172 | blk_to_erase = pblk_prepare_new_line(pblk, line); |
|---|
| 1193 | 1173 | line->state = PBLK_LINESTATE_FREE; |
|---|
| 1174 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
|---|
| 1175 | + line->state); |
|---|
| 1194 | 1176 | } else { |
|---|
| 1195 | 1177 | blk_to_erase = blk_in_line; |
|---|
| 1196 | 1178 | } |
|---|
| .. | .. |
|---|
| 1208 | 1190 | } |
|---|
| 1209 | 1191 | |
|---|
| 1210 | 1192 | line->state = PBLK_LINESTATE_OPEN; |
|---|
| 1193 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
|---|
| 1194 | + line->state); |
|---|
| 1211 | 1195 | |
|---|
| 1212 | 1196 | atomic_set(&line->left_eblks, blk_to_erase); |
|---|
| 1213 | 1197 | atomic_set(&line->left_seblks, blk_to_erase); |
|---|
| .. | .. |
|---|
| 1216 | 1200 | spin_unlock(&line->lock); |
|---|
| 1217 | 1201 | |
|---|
| 1218 | 1202 | kref_init(&line->ref); |
|---|
| 1203 | + atomic_set(&line->sec_to_update, 0); |
|---|
| 1219 | 1204 | |
|---|
| 1220 | 1205 | return 0; |
|---|
| 1221 | 1206 | } |
|---|
| 1222 | 1207 | |
|---|
| 1208 | +/* Line allocations in the recovery path are always single threaded */ |
|---|
| 1223 | 1209 | int pblk_line_recov_alloc(struct pblk *pblk, struct pblk_line *line) |
|---|
| 1224 | 1210 | { |
|---|
| 1225 | 1211 | struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
|---|
| .. | .. |
|---|
| 1259 | 1245 | |
|---|
| 1260 | 1246 | void pblk_line_recov_close(struct pblk *pblk, struct pblk_line *line) |
|---|
| 1261 | 1247 | { |
|---|
| 1262 | | - kfree(line->map_bitmap); |
|---|
| 1248 | + struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
|---|
| 1249 | + |
|---|
| 1250 | + mempool_free(line->map_bitmap, l_mg->bitmap_pool); |
|---|
| 1263 | 1251 | line->map_bitmap = NULL; |
|---|
| 1264 | 1252 | line->smeta = NULL; |
|---|
| 1265 | 1253 | line->emeta = NULL; |
|---|
| .. | .. |
|---|
| 1277 | 1265 | |
|---|
| 1278 | 1266 | void pblk_line_free(struct pblk_line *line) |
|---|
| 1279 | 1267 | { |
|---|
| 1280 | | - kfree(line->map_bitmap); |
|---|
| 1281 | | - kfree(line->invalid_bitmap); |
|---|
| 1268 | + struct pblk *pblk = line->pblk; |
|---|
| 1269 | + struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
|---|
| 1270 | + |
|---|
| 1271 | + mempool_free(line->map_bitmap, l_mg->bitmap_pool); |
|---|
| 1272 | + mempool_free(line->invalid_bitmap, l_mg->bitmap_pool); |
|---|
| 1282 | 1273 | |
|---|
| 1283 | 1274 | pblk_line_reinit(line); |
|---|
| 1284 | 1275 | } |
|---|
| .. | .. |
|---|
| 1306 | 1297 | if (unlikely(bit >= lm->blk_per_line)) { |
|---|
| 1307 | 1298 | spin_lock(&line->lock); |
|---|
| 1308 | 1299 | line->state = PBLK_LINESTATE_BAD; |
|---|
| 1300 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
|---|
| 1301 | + line->state); |
|---|
| 1309 | 1302 | spin_unlock(&line->lock); |
|---|
| 1310 | 1303 | |
|---|
| 1311 | 1304 | list_add_tail(&line->list, &l_mg->bad_list); |
|---|
| .. | .. |
|---|
| 1440 | 1433 | return line; |
|---|
| 1441 | 1434 | } |
|---|
| 1442 | 1435 | |
|---|
| 1436 | +void pblk_ppa_to_line_put(struct pblk *pblk, struct ppa_addr ppa) |
|---|
| 1437 | +{ |
|---|
| 1438 | + struct pblk_line *line; |
|---|
| 1439 | + |
|---|
| 1440 | + line = pblk_ppa_to_line(pblk, ppa); |
|---|
| 1441 | + kref_put(&line->ref, pblk_line_put_wq); |
|---|
| 1442 | +} |
|---|
| 1443 | + |
|---|
| 1444 | +void pblk_rq_to_line_put(struct pblk *pblk, struct nvm_rq *rqd) |
|---|
| 1445 | +{ |
|---|
| 1446 | + struct ppa_addr *ppa_list = nvm_rq_to_ppa_list(rqd); |
|---|
| 1447 | + int i; |
|---|
| 1448 | + |
|---|
| 1449 | + for (i = 0; i < rqd->nr_ppas; i++) |
|---|
| 1450 | + pblk_ppa_to_line_put(pblk, ppa_list[i]); |
|---|
| 1451 | +} |
|---|
| 1452 | + |
|---|
| 1443 | 1453 | static void pblk_stop_writes(struct pblk *pblk, struct pblk_line *line) |
|---|
| 1444 | 1454 | { |
|---|
| 1445 | 1455 | lockdep_assert_held(&pblk->l_mg.free_lock); |
|---|
| 1446 | 1456 | |
|---|
| 1447 | 1457 | pblk_set_space_limit(pblk); |
|---|
| 1448 | 1458 | pblk->state = PBLK_STATE_STOPPING; |
|---|
| 1459 | + trace_pblk_state(pblk_disk_name(pblk), pblk->state); |
|---|
| 1449 | 1460 | } |
|---|
| 1450 | 1461 | |
|---|
| 1451 | 1462 | static void pblk_line_close_meta_sync(struct pblk *pblk) |
|---|
| .. | .. |
|---|
| 1495 | 1506 | return; |
|---|
| 1496 | 1507 | } |
|---|
| 1497 | 1508 | pblk->state = PBLK_STATE_RECOVERING; |
|---|
| 1509 | + trace_pblk_state(pblk_disk_name(pblk), pblk->state); |
|---|
| 1498 | 1510 | spin_unlock(&l_mg->free_lock); |
|---|
| 1499 | 1511 | |
|---|
| 1500 | 1512 | pblk_flush_writer(pblk); |
|---|
| .. | .. |
|---|
| 1516 | 1528 | |
|---|
| 1517 | 1529 | spin_lock(&l_mg->free_lock); |
|---|
| 1518 | 1530 | pblk->state = PBLK_STATE_STOPPED; |
|---|
| 1531 | + trace_pblk_state(pblk_disk_name(pblk), pblk->state); |
|---|
| 1519 | 1532 | l_mg->data_line = NULL; |
|---|
| 1520 | 1533 | l_mg->data_next = NULL; |
|---|
| 1521 | 1534 | spin_unlock(&l_mg->free_lock); |
|---|
| .. | .. |
|---|
| 1606 | 1619 | |
|---|
| 1607 | 1620 | spin_lock(&line->lock); |
|---|
| 1608 | 1621 | WARN_ON(line->state != PBLK_LINESTATE_GC); |
|---|
| 1622 | + if (line->w_err_gc->has_gc_err) { |
|---|
| 1623 | + spin_unlock(&line->lock); |
|---|
| 1624 | + pblk_err(pblk, "line %d had errors during GC\n", line->id); |
|---|
| 1625 | + pblk_put_line_back(pblk, line); |
|---|
| 1626 | + line->w_err_gc->has_gc_err = 0; |
|---|
| 1627 | + return; |
|---|
| 1628 | + } |
|---|
| 1629 | + |
|---|
| 1609 | 1630 | line->state = PBLK_LINESTATE_FREE; |
|---|
| 1631 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
|---|
| 1632 | + line->state); |
|---|
| 1610 | 1633 | line->gc_group = PBLK_LINEGC_NONE; |
|---|
| 1611 | 1634 | pblk_line_free(line); |
|---|
| 1612 | 1635 | |
|---|
| .. | .. |
|---|
| 1675 | 1698 | rqd->end_io = pblk_end_io_erase; |
|---|
| 1676 | 1699 | rqd->private = pblk; |
|---|
| 1677 | 1700 | |
|---|
| 1701 | + trace_pblk_chunk_reset(pblk_disk_name(pblk), |
|---|
| 1702 | + &ppa, PBLK_CHUNK_RESET_START); |
|---|
| 1703 | + |
|---|
| 1678 | 1704 | /* The write thread schedules erases so that it minimizes disturbances |
|---|
| 1679 | 1705 | * with writes. Thus, there is no need to take the LUN semaphore. |
|---|
| 1680 | 1706 | */ |
|---|
| 1681 | | - err = pblk_submit_io(pblk, rqd); |
|---|
| 1707 | + err = pblk_submit_io(pblk, rqd, NULL); |
|---|
| 1682 | 1708 | if (err) { |
|---|
| 1683 | 1709 | struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 1684 | 1710 | struct nvm_geo *geo = &dev->geo; |
|---|
| 1685 | 1711 | |
|---|
| 1686 | 1712 | pblk_err(pblk, "could not async erase line:%d,blk:%d\n", |
|---|
| 1687 | | - pblk_ppa_to_line(ppa), |
|---|
| 1713 | + pblk_ppa_to_line_id(ppa), |
|---|
| 1688 | 1714 | pblk_ppa_to_pos(geo, ppa)); |
|---|
| 1689 | 1715 | } |
|---|
| 1690 | 1716 | |
|---|
| .. | .. |
|---|
| 1736 | 1762 | WARN_ON(line->state != PBLK_LINESTATE_OPEN); |
|---|
| 1737 | 1763 | line->state = PBLK_LINESTATE_CLOSED; |
|---|
| 1738 | 1764 | move_list = pblk_line_gc_list(pblk, line); |
|---|
| 1739 | | - |
|---|
| 1740 | 1765 | list_add_tail(&line->list, move_list); |
|---|
| 1741 | 1766 | |
|---|
| 1742 | | - kfree(line->map_bitmap); |
|---|
| 1767 | + mempool_free(line->map_bitmap, l_mg->bitmap_pool); |
|---|
| 1743 | 1768 | line->map_bitmap = NULL; |
|---|
| 1744 | 1769 | line->smeta = NULL; |
|---|
| 1745 | 1770 | line->emeta = NULL; |
|---|
| .. | .. |
|---|
| 1755 | 1780 | |
|---|
| 1756 | 1781 | spin_unlock(&line->lock); |
|---|
| 1757 | 1782 | spin_unlock(&l_mg->gc_lock); |
|---|
| 1783 | + |
|---|
| 1784 | + trace_pblk_line_state(pblk_disk_name(pblk), line->id, |
|---|
| 1785 | + line->state); |
|---|
| 1758 | 1786 | } |
|---|
| 1759 | 1787 | |
|---|
| 1760 | 1788 | void pblk_line_close_meta(struct pblk *pblk, struct pblk_line *line) |
|---|
| .. | .. |
|---|
| 1775 | 1803 | |
|---|
| 1776 | 1804 | if (le32_to_cpu(emeta_buf->header.identifier) != PBLK_MAGIC) { |
|---|
| 1777 | 1805 | emeta_buf->header.identifier = cpu_to_le32(PBLK_MAGIC); |
|---|
| 1778 | | - memcpy(emeta_buf->header.uuid, pblk->instance_uuid, 16); |
|---|
| 1806 | + guid_copy((guid_t *)&emeta_buf->header.uuid, |
|---|
| 1807 | + &pblk->instance_uuid); |
|---|
| 1779 | 1808 | emeta_buf->header.id = cpu_to_le32(line->id); |
|---|
| 1780 | 1809 | emeta_buf->header.type = cpu_to_le16(line->type); |
|---|
| 1781 | 1810 | emeta_buf->header.version_major = EMETA_VERSION_MAJOR; |
|---|
| .. | .. |
|---|
| 1806 | 1835 | static void pblk_save_lba_list(struct pblk *pblk, struct pblk_line *line) |
|---|
| 1807 | 1836 | { |
|---|
| 1808 | 1837 | struct pblk_line_meta *lm = &pblk->lm; |
|---|
| 1809 | | - struct pblk_line_mgmt *l_mg = &pblk->l_mg; |
|---|
| 1810 | 1838 | unsigned int lba_list_size = lm->emeta_len[2]; |
|---|
| 1811 | 1839 | struct pblk_w_err_gc *w_err_gc = line->w_err_gc; |
|---|
| 1812 | 1840 | struct pblk_emeta *emeta = line->emeta; |
|---|
| 1813 | 1841 | |
|---|
| 1814 | | - w_err_gc->lba_list = pblk_malloc(lba_list_size, |
|---|
| 1815 | | - l_mg->emeta_alloc_type, GFP_KERNEL); |
|---|
| 1842 | + w_err_gc->lba_list = kvmalloc(lba_list_size, GFP_KERNEL); |
|---|
| 1816 | 1843 | memcpy(w_err_gc->lba_list, emeta_to_lbas(pblk, emeta->buf), |
|---|
| 1817 | 1844 | lba_list_size); |
|---|
| 1818 | 1845 | } |
|---|
| .. | .. |
|---|
| 1851 | 1878 | queue_work(wq, &line_ws->ws); |
|---|
| 1852 | 1879 | } |
|---|
| 1853 | 1880 | |
|---|
| 1854 | | -static void __pblk_down_page(struct pblk *pblk, struct ppa_addr *ppa_list, |
|---|
| 1855 | | - int nr_ppas, int pos) |
|---|
| 1881 | +static void __pblk_down_chunk(struct pblk *pblk, int pos) |
|---|
| 1856 | 1882 | { |
|---|
| 1857 | 1883 | struct pblk_lun *rlun = &pblk->luns[pos]; |
|---|
| 1858 | 1884 | int ret; |
|---|
| .. | .. |
|---|
| 1861 | 1887 | * Only send one inflight I/O per LUN. Since we map at a page |
|---|
| 1862 | 1888 | * granurality, all ppas in the I/O will map to the same LUN |
|---|
| 1863 | 1889 | */ |
|---|
| 1864 | | -#ifdef CONFIG_NVM_PBLK_DEBUG |
|---|
| 1865 | | - int i; |
|---|
| 1866 | | - |
|---|
| 1867 | | - for (i = 1; i < nr_ppas; i++) |
|---|
| 1868 | | - WARN_ON(ppa_list[0].a.lun != ppa_list[i].a.lun || |
|---|
| 1869 | | - ppa_list[0].a.ch != ppa_list[i].a.ch); |
|---|
| 1870 | | -#endif |
|---|
| 1871 | 1890 | |
|---|
| 1872 | 1891 | ret = down_timeout(&rlun->wr_sem, msecs_to_jiffies(30000)); |
|---|
| 1873 | 1892 | if (ret == -ETIME || ret == -EINTR) |
|---|
| .. | .. |
|---|
| 1875 | 1894 | -ret); |
|---|
| 1876 | 1895 | } |
|---|
| 1877 | 1896 | |
|---|
| 1878 | | -void pblk_down_page(struct pblk *pblk, struct ppa_addr *ppa_list, int nr_ppas) |
|---|
| 1897 | +void pblk_down_chunk(struct pblk *pblk, struct ppa_addr ppa) |
|---|
| 1879 | 1898 | { |
|---|
| 1880 | 1899 | struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 1881 | 1900 | struct nvm_geo *geo = &dev->geo; |
|---|
| 1882 | | - int pos = pblk_ppa_to_pos(geo, ppa_list[0]); |
|---|
| 1901 | + int pos = pblk_ppa_to_pos(geo, ppa); |
|---|
| 1883 | 1902 | |
|---|
| 1884 | | - __pblk_down_page(pblk, ppa_list, nr_ppas, pos); |
|---|
| 1903 | + __pblk_down_chunk(pblk, pos); |
|---|
| 1885 | 1904 | } |
|---|
| 1886 | 1905 | |
|---|
| 1887 | | -void pblk_down_rq(struct pblk *pblk, struct ppa_addr *ppa_list, int nr_ppas, |
|---|
| 1906 | +void pblk_down_rq(struct pblk *pblk, struct ppa_addr ppa, |
|---|
| 1888 | 1907 | unsigned long *lun_bitmap) |
|---|
| 1889 | 1908 | { |
|---|
| 1890 | 1909 | struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 1891 | 1910 | struct nvm_geo *geo = &dev->geo; |
|---|
| 1892 | | - int pos = pblk_ppa_to_pos(geo, ppa_list[0]); |
|---|
| 1911 | + int pos = pblk_ppa_to_pos(geo, ppa); |
|---|
| 1893 | 1912 | |
|---|
| 1894 | 1913 | /* If the LUN has been locked for this same request, do no attempt to |
|---|
| 1895 | 1914 | * lock it again |
|---|
| .. | .. |
|---|
| 1897 | 1916 | if (test_and_set_bit(pos, lun_bitmap)) |
|---|
| 1898 | 1917 | return; |
|---|
| 1899 | 1918 | |
|---|
| 1900 | | - __pblk_down_page(pblk, ppa_list, nr_ppas, pos); |
|---|
| 1919 | + __pblk_down_chunk(pblk, pos); |
|---|
| 1901 | 1920 | } |
|---|
| 1902 | 1921 | |
|---|
| 1903 | | -void pblk_up_page(struct pblk *pblk, struct ppa_addr *ppa_list, int nr_ppas) |
|---|
| 1922 | +void pblk_up_chunk(struct pblk *pblk, struct ppa_addr ppa) |
|---|
| 1904 | 1923 | { |
|---|
| 1905 | 1924 | struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 1906 | 1925 | struct nvm_geo *geo = &dev->geo; |
|---|
| 1907 | 1926 | struct pblk_lun *rlun; |
|---|
| 1908 | | - int pos = pblk_ppa_to_pos(geo, ppa_list[0]); |
|---|
| 1909 | | - |
|---|
| 1910 | | -#ifdef CONFIG_NVM_PBLK_DEBUG |
|---|
| 1911 | | - int i; |
|---|
| 1912 | | - |
|---|
| 1913 | | - for (i = 1; i < nr_ppas; i++) |
|---|
| 1914 | | - WARN_ON(ppa_list[0].a.lun != ppa_list[i].a.lun || |
|---|
| 1915 | | - ppa_list[0].a.ch != ppa_list[i].a.ch); |
|---|
| 1916 | | -#endif |
|---|
| 1927 | + int pos = pblk_ppa_to_pos(geo, ppa); |
|---|
| 1917 | 1928 | |
|---|
| 1918 | 1929 | rlun = &pblk->luns[pos]; |
|---|
| 1919 | 1930 | up(&rlun->wr_sem); |
|---|
| 1920 | 1931 | } |
|---|
| 1921 | 1932 | |
|---|
| 1922 | | -void pblk_up_rq(struct pblk *pblk, struct ppa_addr *ppa_list, int nr_ppas, |
|---|
| 1923 | | - unsigned long *lun_bitmap) |
|---|
| 1933 | +void pblk_up_rq(struct pblk *pblk, unsigned long *lun_bitmap) |
|---|
| 1924 | 1934 | { |
|---|
| 1925 | 1935 | struct nvm_tgt_dev *dev = pblk->dev; |
|---|
| 1926 | 1936 | struct nvm_geo *geo = &dev->geo; |
|---|
| .. | .. |
|---|
| 1939 | 1949 | struct ppa_addr ppa_l2p; |
|---|
| 1940 | 1950 | |
|---|
| 1941 | 1951 | /* logic error: lba out-of-bounds. Ignore update */ |
|---|
| 1942 | | - if (!(lba < pblk->rl.nr_secs)) { |
|---|
| 1952 | + if (!(lba < pblk->capacity)) { |
|---|
| 1943 | 1953 | WARN(1, "pblk: corrupted L2P map request\n"); |
|---|
| 1944 | 1954 | return; |
|---|
| 1945 | 1955 | } |
|---|
| .. | .. |
|---|
| 1979 | 1989 | #endif |
|---|
| 1980 | 1990 | |
|---|
| 1981 | 1991 | /* logic error: lba out-of-bounds. Ignore update */ |
|---|
| 1982 | | - if (!(lba < pblk->rl.nr_secs)) { |
|---|
| 1992 | + if (!(lba < pblk->capacity)) { |
|---|
| 1983 | 1993 | WARN(1, "pblk: corrupted L2P map request\n"); |
|---|
| 1984 | 1994 | return 0; |
|---|
| 1985 | 1995 | } |
|---|
| .. | .. |
|---|
| 2025 | 2035 | } |
|---|
| 2026 | 2036 | |
|---|
| 2027 | 2037 | /* logic error: lba out-of-bounds. Ignore update */ |
|---|
| 2028 | | - if (!(lba < pblk->rl.nr_secs)) { |
|---|
| 2038 | + if (!(lba < pblk->capacity)) { |
|---|
| 2029 | 2039 | WARN(1, "pblk: corrupted L2P map request\n"); |
|---|
| 2030 | 2040 | return; |
|---|
| 2031 | 2041 | } |
|---|
| .. | .. |
|---|
| 2051 | 2061 | spin_unlock(&pblk->trans_lock); |
|---|
| 2052 | 2062 | } |
|---|
| 2053 | 2063 | |
|---|
| 2054 | | -void pblk_lookup_l2p_seq(struct pblk *pblk, struct ppa_addr *ppas, |
|---|
| 2055 | | - sector_t blba, int nr_secs) |
|---|
| 2064 | +int pblk_lookup_l2p_seq(struct pblk *pblk, struct ppa_addr *ppas, |
|---|
| 2065 | + sector_t blba, int nr_secs, bool *from_cache) |
|---|
| 2056 | 2066 | { |
|---|
| 2057 | 2067 | int i; |
|---|
| 2058 | 2068 | |
|---|
| .. | .. |
|---|
| 2064 | 2074 | |
|---|
| 2065 | 2075 | /* If the L2P entry maps to a line, the reference is valid */ |
|---|
| 2066 | 2076 | if (!pblk_ppa_empty(ppa) && !pblk_addr_in_cache(ppa)) { |
|---|
| 2067 | | - int line_id = pblk_ppa_to_line(ppa); |
|---|
| 2068 | | - struct pblk_line *line = &pblk->lines[line_id]; |
|---|
| 2077 | + struct pblk_line *line = pblk_ppa_to_line(pblk, ppa); |
|---|
| 2078 | + |
|---|
| 2079 | + if (i > 0 && *from_cache) |
|---|
| 2080 | + break; |
|---|
| 2081 | + *from_cache = false; |
|---|
| 2069 | 2082 | |
|---|
| 2070 | 2083 | kref_get(&line->ref); |
|---|
| 2084 | + } else { |
|---|
| 2085 | + if (i > 0 && !*from_cache) |
|---|
| 2086 | + break; |
|---|
| 2087 | + *from_cache = true; |
|---|
| 2071 | 2088 | } |
|---|
| 2072 | 2089 | } |
|---|
| 2073 | 2090 | spin_unlock(&pblk->trans_lock); |
|---|
| 2091 | + return i; |
|---|
| 2074 | 2092 | } |
|---|
| 2075 | 2093 | |
|---|
| 2076 | 2094 | void pblk_lookup_l2p_rand(struct pblk *pblk, struct ppa_addr *ppas, |
|---|
| .. | .. |
|---|
| 2084 | 2102 | lba = lba_list[i]; |
|---|
| 2085 | 2103 | if (lba != ADDR_EMPTY) { |
|---|
| 2086 | 2104 | /* logic error: lba out-of-bounds. Ignore update */ |
|---|
| 2087 | | - if (!(lba < pblk->rl.nr_secs)) { |
|---|
| 2105 | + if (!(lba < pblk->capacity)) { |
|---|
| 2088 | 2106 | WARN(1, "pblk: corrupted L2P map request\n"); |
|---|
| 2089 | 2107 | continue; |
|---|
| 2090 | 2108 | } |
|---|
| .. | .. |
|---|
| 2093 | 2111 | } |
|---|
| 2094 | 2112 | spin_unlock(&pblk->trans_lock); |
|---|
| 2095 | 2113 | } |
|---|
| 2114 | + |
|---|
| 2115 | +void *pblk_get_meta_for_writes(struct pblk *pblk, struct nvm_rq *rqd) |
|---|
| 2116 | +{ |
|---|
| 2117 | + void *buffer; |
|---|
| 2118 | + |
|---|
| 2119 | + if (pblk_is_oob_meta_supported(pblk)) { |
|---|
| 2120 | + /* Just use OOB metadata buffer as always */ |
|---|
| 2121 | + buffer = rqd->meta_list; |
|---|
| 2122 | + } else { |
|---|
| 2123 | + /* We need to reuse last page of request (packed metadata) |
|---|
| 2124 | + * in similar way as traditional oob metadata |
|---|
| 2125 | + */ |
|---|
| 2126 | + buffer = page_to_virt( |
|---|
| 2127 | + rqd->bio->bi_io_vec[rqd->bio->bi_vcnt - 1].bv_page); |
|---|
| 2128 | + } |
|---|
| 2129 | + |
|---|
| 2130 | + return buffer; |
|---|
| 2131 | +} |
|---|
| 2132 | + |
|---|
| 2133 | +void pblk_get_packed_meta(struct pblk *pblk, struct nvm_rq *rqd) |
|---|
| 2134 | +{ |
|---|
| 2135 | + void *meta_list = rqd->meta_list; |
|---|
| 2136 | + void *page; |
|---|
| 2137 | + int i = 0; |
|---|
| 2138 | + |
|---|
| 2139 | + if (pblk_is_oob_meta_supported(pblk)) |
|---|
| 2140 | + return; |
|---|
| 2141 | + |
|---|
| 2142 | + page = page_to_virt(rqd->bio->bi_io_vec[rqd->bio->bi_vcnt - 1].bv_page); |
|---|
| 2143 | + /* We need to fill oob meta buffer with data from packed metadata */ |
|---|
| 2144 | + for (; i < rqd->nr_ppas; i++) |
|---|
| 2145 | + memcpy(pblk_get_meta(pblk, meta_list, i), |
|---|
| 2146 | + page + (i * sizeof(struct pblk_sec_meta)), |
|---|
| 2147 | + sizeof(struct pblk_sec_meta)); |
|---|
| 2148 | +} |
|---|