| /* |
| * Copyright (c) 2014-2016, Intel Corporation. |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms and conditions of the GNU General Public License, |
| * version 2, as published by the Free Software Foundation. |
| * |
| * This program is distributed in the hope it will be useful, but WITHOUT |
| * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or |
| * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for |
| * more details. |
| */ |
| #include "test/nfit_test.h" |
| #include <linux/blkdev.h> |
| #include <pmem.h> |
| #include <nd.h> |
| |
| long __pmem_direct_access(struct pmem_device *pmem, pgoff_t pgoff, |
| long nr_pages, void **kaddr, pfn_t *pfn) |
| { |
| resource_size_t offset = PFN_PHYS(pgoff) + pmem->data_offset; |
| |
| if (unlikely(is_bad_pmem(&pmem->bb, PFN_PHYS(pgoff) / 512, |
| PFN_PHYS(nr_pages)))) |
| return -EIO; |
| |
| /* |
| * Limit dax to a single page at a time given vmalloc()-backed |
| * in the nfit_test case. |
| */ |
| if (get_nfit_res(pmem->phys_addr + offset)) { |
| struct page *page; |
| |
| if (kaddr) |
| *kaddr = pmem->virt_addr + offset; |
| page = vmalloc_to_page(pmem->virt_addr + offset); |
| if (pfn) |
| *pfn = page_to_pfn_t(page); |
| pr_debug_ratelimited("%s: pmem: %p pgoff: %#lx pfn: %#lx\n", |
| __func__, pmem, pgoff, page_to_pfn(page)); |
| |
| return 1; |
| } |
| |
| if (kaddr) |
| *kaddr = pmem->virt_addr + offset; |
| if (pfn) |
| *pfn = phys_to_pfn_t(pmem->phys_addr + offset, pmem->pfn_flags); |
| |
| /* |
| * If badblocks are present, limit known good range to the |
| * requested range. |
| */ |
| if (unlikely(pmem->bb.count)) |
| return nr_pages; |
| return PHYS_PFN(pmem->size - pmem->pfn_pad - offset); |
| } |