| /* SPDX-License-Identifier: GPL-2.0-only */ |
| /* |
| * Copyright © 2008 Keith Packard <keithp@keithp.com> |
| */ |
| |
| #ifndef _LINUX_IO_MAPPING_H |
| #define _LINUX_IO_MAPPING_H |
| |
| #include <linux/types.h> |
| #include <linux/slab.h> |
| #include <linux/bug.h> |
| #include <linux/io.h> |
| #include <linux/pgtable.h> |
| #include <asm/page.h> |
| |
| /* |
| * The io_mapping mechanism provides an abstraction for mapping |
| * individual pages from an io device to the CPU in an efficient fashion. |
| * |
| * See Documentation/driver-api/io-mapping.rst |
| */ |
| |
| struct io_mapping { |
| resource_size_t base; |
| unsigned long size; |
| pgprot_t prot; |
| void __iomem *iomem; |
| }; |
| |
| #ifdef CONFIG_HAVE_ATOMIC_IOMAP |
| |
| #include <linux/pfn.h> |
| #include <asm/iomap.h> |
| /* |
| * For small address space machines, mapping large objects |
| * into the kernel virtual space isn't practical. Where |
| * available, use fixmap support to dynamically map pages |
| * of the object at run time. |
| */ |
| |
| static inline struct io_mapping * |
| io_mapping_init_wc(struct io_mapping *iomap, |
| resource_size_t base, |
| unsigned long size) |
| { |
| pgprot_t prot; |
| |
| if (iomap_create_wc(base, size, &prot)) |
| return NULL; |
| |
| iomap->base = base; |
| iomap->size = size; |
| iomap->prot = prot; |
| return iomap; |
| } |
| |
| static inline void |
| io_mapping_fini(struct io_mapping *mapping) |
| { |
| iomap_free(mapping->base, mapping->size); |
| } |
| |
| /* Atomic map/unmap */ |
| static inline void __iomem * |
| io_mapping_map_atomic_wc(struct io_mapping *mapping, |
| unsigned long offset) |
| { |
| resource_size_t phys_addr; |
| |
| BUG_ON(offset >= mapping->size); |
| phys_addr = mapping->base + offset; |
| preempt_disable(); |
| pagefault_disable(); |
| return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); |
| } |
| |
| static inline void |
| io_mapping_unmap_atomic(void __iomem *vaddr) |
| { |
| kunmap_local_indexed((void __force *)vaddr); |
| pagefault_enable(); |
| preempt_enable(); |
| } |
| |
| static inline void __iomem * |
| io_mapping_map_local_wc(struct io_mapping *mapping, unsigned long offset) |
| { |
| resource_size_t phys_addr; |
| |
| BUG_ON(offset >= mapping->size); |
| phys_addr = mapping->base + offset; |
| return __iomap_local_pfn_prot(PHYS_PFN(phys_addr), mapping->prot); |
| } |
| |
| static inline void io_mapping_unmap_local(void __iomem *vaddr) |
| { |
| kunmap_local_indexed((void __force *)vaddr); |
| } |
| |
| static inline void __iomem * |
| io_mapping_map_wc(struct io_mapping *mapping, |
| unsigned long offset, |
| unsigned long size) |
| { |
| resource_size_t phys_addr; |
| |
| BUG_ON(offset >= mapping->size); |
| phys_addr = mapping->base + offset; |
| |
| return ioremap_wc(phys_addr, size); |
| } |
| |
| static inline void |
| io_mapping_unmap(void __iomem *vaddr) |
| { |
| iounmap(vaddr); |
| } |
| |
| #else /* HAVE_ATOMIC_IOMAP */ |
| |
| #include <linux/uaccess.h> |
| |
| /* Create the io_mapping object*/ |
| static inline struct io_mapping * |
| io_mapping_init_wc(struct io_mapping *iomap, |
| resource_size_t base, |
| unsigned long size) |
| { |
| iomap->iomem = ioremap_wc(base, size); |
| if (!iomap->iomem) |
| return NULL; |
| |
| iomap->base = base; |
| iomap->size = size; |
| #if defined(pgprot_noncached_wc) /* archs can't agree on a name ... */ |
| iomap->prot = pgprot_noncached_wc(PAGE_KERNEL); |
| #elif defined(pgprot_writecombine) |
| iomap->prot = pgprot_writecombine(PAGE_KERNEL); |
| #else |
| iomap->prot = pgprot_noncached(PAGE_KERNEL); |
| #endif |
| |
| return iomap; |
| } |
| |
| static inline void |
| io_mapping_fini(struct io_mapping *mapping) |
| { |
| iounmap(mapping->iomem); |
| } |
| |
| /* Non-atomic map/unmap */ |
| static inline void __iomem * |
| io_mapping_map_wc(struct io_mapping *mapping, |
| unsigned long offset, |
| unsigned long size) |
| { |
| return mapping->iomem + offset; |
| } |
| |
| static inline void |
| io_mapping_unmap(void __iomem *vaddr) |
| { |
| } |
| |
| /* Atomic map/unmap */ |
| static inline void __iomem * |
| io_mapping_map_atomic_wc(struct io_mapping *mapping, |
| unsigned long offset) |
| { |
| preempt_disable(); |
| pagefault_disable(); |
| return io_mapping_map_wc(mapping, offset, PAGE_SIZE); |
| } |
| |
| static inline void |
| io_mapping_unmap_atomic(void __iomem *vaddr) |
| { |
| io_mapping_unmap(vaddr); |
| pagefault_enable(); |
| preempt_enable(); |
| } |
| |
| static inline void __iomem * |
| io_mapping_map_local_wc(struct io_mapping *mapping, unsigned long offset) |
| { |
| return io_mapping_map_wc(mapping, offset, PAGE_SIZE); |
| } |
| |
| static inline void io_mapping_unmap_local(void __iomem *vaddr) |
| { |
| io_mapping_unmap(vaddr); |
| } |
| |
| #endif /* !HAVE_ATOMIC_IOMAP */ |
| |
| static inline struct io_mapping * |
| io_mapping_create_wc(resource_size_t base, |
| unsigned long size) |
| { |
| struct io_mapping *iomap; |
| |
| iomap = kmalloc(sizeof(*iomap), GFP_KERNEL); |
| if (!iomap) |
| return NULL; |
| |
| if (!io_mapping_init_wc(iomap, base, size)) { |
| kfree(iomap); |
| return NULL; |
| } |
| |
| return iomap; |
| } |
| |
| static inline void |
| io_mapping_free(struct io_mapping *iomap) |
| { |
| io_mapping_fini(iomap); |
| kfree(iomap); |
| } |
| |
| #endif /* _LINUX_IO_MAPPING_H */ |
| |
| int io_mapping_map_user(struct io_mapping *iomap, struct vm_area_struct *vma, |
| unsigned long addr, unsigned long pfn, unsigned long size); |