| .. | .. |
|---|
| 49 | 49 | * we simply install a virtual mapping for it only for the |
|---|
| 50 | 50 | * TLB lookup to occur, hence no need to flush the untouched |
|---|
| 51 | 51 | * memory mapping afterwards (note: a cache flush may happen |
|---|
| 52 | | - * in some circumstances depending on the path taken in kunmap_atomic). |
|---|
| 52 | + * in some circumstances depending on the path taken in kunmap_local). |
|---|
| 53 | 53 | */ |
|---|
| 54 | | - void *vaddr = kmap_atomic_pfn(paddr >> PAGE_SHIFT); |
|---|
| 54 | + void *vaddr = kmap_local_pfn(paddr >> PAGE_SHIFT); |
|---|
| 55 | 55 | return (unsigned long)vaddr + (paddr & ~PAGE_MASK); |
|---|
| 56 | 56 | #else |
|---|
| 57 | 57 | return __phys_to_virt(paddr); |
|---|
| .. | .. |
|---|
| 61 | 61 | static inline void l2_put_va(unsigned long vaddr) |
|---|
| 62 | 62 | { |
|---|
| 63 | 63 | #ifdef CONFIG_HIGHMEM |
|---|
| 64 | | - kunmap_atomic((void *)vaddr); |
|---|
| 64 | + kunmap_local((void *)vaddr); |
|---|
| 65 | 65 | #endif |
|---|
| 66 | 66 | } |
|---|
| 67 | 67 | |
|---|