| /* SPDX-License-Identifier: GPL-2.0 */ |
| #ifndef __NVKM_MMU_H__ |
| #define __NVKM_MMU_H__ |
| #include <core/subdev.h> |
| |
| struct nvkm_vma { |
| struct list_head head; |
| struct rb_node tree; |
| u64 addr; |
| u64 size:50; |
| bool mapref:1; /* PTs (de)referenced on (un)map (vs pre-allocated). */ |
| bool sparse:1; /* Unmapped PDEs/PTEs will not trigger MMU faults. */ |
| #define NVKM_VMA_PAGE_NONE 7 |
| u8 page:3; /* Requested page type (index, or NONE for automatic). */ |
| u8 refd:3; /* Current page type (index, or NONE for unreferenced). */ |
| bool used:1; /* Region allocated. */ |
| bool part:1; /* Region was split from an allocated region by map(). */ |
| bool user:1; /* Region user-allocated. */ |
| bool busy:1; /* Region busy (for temporarily preventing user access). */ |
| bool mapped:1; /* Region contains valid pages. */ |
| struct nvkm_memory *memory; /* Memory currently mapped into VMA. */ |
| struct nvkm_tags *tags; /* Compression tag reference. */ |
| }; |
| |
| struct nvkm_vmm { |
| const struct nvkm_vmm_func *func; |
| struct nvkm_mmu *mmu; |
| const char *name; |
| u32 debug; |
| struct kref kref; |
| struct mutex mutex; |
| |
| u64 start; |
| u64 limit; |
| |
| struct nvkm_vmm_pt *pd; |
| struct list_head join; |
| |
| struct list_head list; |
| struct rb_root free; |
| struct rb_root root; |
| |
| bool bootstrapped; |
| atomic_t engref[NVKM_SUBDEV_NR]; |
| |
| dma_addr_t null; |
| void *nullp; |
| }; |
| |
| int nvkm_vmm_new(struct nvkm_device *, u64 addr, u64 size, void *argv, u32 argc, |
| struct lock_class_key *, const char *name, struct nvkm_vmm **); |
| struct nvkm_vmm *nvkm_vmm_ref(struct nvkm_vmm *); |
| void nvkm_vmm_unref(struct nvkm_vmm **); |
| int nvkm_vmm_boot(struct nvkm_vmm *); |
| int nvkm_vmm_join(struct nvkm_vmm *, struct nvkm_memory *inst); |
| void nvkm_vmm_part(struct nvkm_vmm *, struct nvkm_memory *inst); |
| int nvkm_vmm_get(struct nvkm_vmm *, u8 page, u64 size, struct nvkm_vma **); |
| void nvkm_vmm_put(struct nvkm_vmm *, struct nvkm_vma **); |
| |
| struct nvkm_vmm_map { |
| struct nvkm_memory *memory; |
| u64 offset; |
| |
| struct nvkm_mm_node *mem; |
| struct scatterlist *sgl; |
| dma_addr_t *dma; |
| u64 *pfn; |
| u64 off; |
| |
| const struct nvkm_vmm_page *page; |
| |
| struct nvkm_tags *tags; |
| u64 next; |
| u64 type; |
| u64 ctag; |
| }; |
| |
| int nvkm_vmm_map(struct nvkm_vmm *, struct nvkm_vma *, void *argv, u32 argc, |
| struct nvkm_vmm_map *); |
| void nvkm_vmm_unmap(struct nvkm_vmm *, struct nvkm_vma *); |
| |
| struct nvkm_memory *nvkm_umem_search(struct nvkm_client *, u64); |
| struct nvkm_vmm *nvkm_uvmm_search(struct nvkm_client *, u64 handle); |
| |
| struct nvkm_mmu { |
| const struct nvkm_mmu_func *func; |
| struct nvkm_subdev subdev; |
| |
| u8 dma_bits; |
| |
| int heap_nr; |
| struct { |
| #define NVKM_MEM_VRAM 0x01 |
| #define NVKM_MEM_HOST 0x02 |
| #define NVKM_MEM_COMP 0x04 |
| #define NVKM_MEM_DISP 0x08 |
| u8 type; |
| u64 size; |
| } heap[4]; |
| |
| int type_nr; |
| struct { |
| #define NVKM_MEM_KIND 0x10 |
| #define NVKM_MEM_MAPPABLE 0x20 |
| #define NVKM_MEM_COHERENT 0x40 |
| #define NVKM_MEM_UNCACHED 0x80 |
| u8 type; |
| u8 heap; |
| } type[16]; |
| |
| struct nvkm_vmm *vmm; |
| |
| struct { |
| struct mutex mutex; |
| struct list_head list; |
| } ptc, ptp; |
| |
| struct nvkm_device_oclass user; |
| }; |
| |
| int nv04_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int nv41_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int nv44_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int nv50_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int g84_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int mcp77_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int gf100_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int gk104_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int gk20a_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int gm200_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int gm20b_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int gp100_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int gp10b_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int gv100_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| int tu102_mmu_new(struct nvkm_device *, int, struct nvkm_mmu **); |
| #endif |