.. | .. |
---|
361 | 361 | |
---|
362 | 362 | list_for_each_entry(m_page, &os_mem->pages, list) { |
---|
363 | 363 | /* We should use vm_insert_page, but it does a dcache |
---|
364 | | - * flush which makes it way slower than remap_pfn_range or vm_insert_pfn. |
---|
| 364 | + * flush which makes it way slower than remap_pfn_range or vmf_insert_pfn. |
---|
365 | 365 | ret = vm_insert_page(vma, addr, page); |
---|
366 | 366 | */ |
---|
367 | 367 | page = m_page->page; |
---|
368 | | - ret = vm_insert_pfn(vma, addr, page_to_pfn(page)); |
---|
| 368 | + ret = vmf_insert_pfn(vma, addr, page_to_pfn(page)); |
---|
369 | 369 | |
---|
370 | | - if (unlikely(0 != ret)) { |
---|
| 370 | + if (unlikely(VM_FAULT_NOPAGE != ret)) { |
---|
371 | 371 | return -EFAULT; |
---|
372 | 372 | } |
---|
373 | 373 | addr += _MALI_OSK_MALI_PAGE_SIZE; |
---|
.. | .. |
---|
403 | 403 | |
---|
404 | 404 | vm_end -= _MALI_OSK_MALI_PAGE_SIZE; |
---|
405 | 405 | if (mapping_page_num > 0) { |
---|
406 | | - ret = vm_insert_pfn(vma, vm_end, page_to_pfn(m_page->page)); |
---|
| 406 | + ret = vmf_insert_pfn(vma, vm_end, page_to_pfn(m_page->page)); |
---|
407 | 407 | |
---|
408 | | - if (unlikely(0 != ret)) { |
---|
| 408 | + if (unlikely(VM_FAULT_NOPAGE != ret)) { |
---|
409 | 409 | /*will return -EBUSY If the page has already been mapped into table, but it's OK*/ |
---|
410 | 410 | if (-EBUSY == ret) { |
---|
411 | 411 | break; |
---|
.. | .. |
---|
426 | 426 | list_for_each_entry(m_page, &os_mem->pages, list) { |
---|
427 | 427 | if (count >= offset) { |
---|
428 | 428 | |
---|
429 | | - ret = vm_insert_pfn(vma, vstart, page_to_pfn(m_page->page)); |
---|
| 429 | + ret = vmf_insert_pfn(vma, vstart, page_to_pfn(m_page->page)); |
---|
430 | 430 | |
---|
431 | | - if (unlikely(0 != ret)) { |
---|
| 431 | + if (unlikely(VM_FAULT_NOPAGE != ret)) { |
---|
432 | 432 | /*will return -EBUSY If the page has already been mapped into table, but it's OK*/ |
---|
433 | 433 | if (-EBUSY == ret) { |
---|
434 | 434 | break; |
---|