333 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
			
		
		
	
	
			333 lines
		
	
	
		
			14 KiB
		
	
	
	
		
			C
		
	
	
	
	
	
| /* SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note */
 | |
| /*
 | |
|  *
 | |
|  * (C) COPYRIGHT 2019-2023 ARM Limited. All rights reserved.
 | |
|  *
 | |
|  * This program is free software and is provided to you under the terms of the
 | |
|  * GNU General Public License version 2 as published by the Free Software
 | |
|  * Foundation, and any use by you of this program is subject to the terms
 | |
|  * of such GNU license.
 | |
|  *
 | |
|  * This program is distributed in the hope that it will be useful,
 | |
|  * but WITHOUT ANY WARRANTY; without even the implied warranty of
 | |
|  * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
 | |
|  * GNU General Public License for more details.
 | |
|  *
 | |
|  * You should have received a copy of the GNU General Public License
 | |
|  * along with this program; if not, you can access it online at
 | |
|  * http://www.gnu.org/licenses/gpl-2.0.html.
 | |
|  *
 | |
|  */
 | |
| 
 | |
| #ifndef _KBASE_MMU_H_
 | |
| #define _KBASE_MMU_H_
 | |
| 
 | |
| #include <uapi/gpu/arm/bifrost/mali_base_kernel.h>
 | |
| #include <mali_kbase_debug.h>
 | |
| 
 | |
| #define KBASE_MMU_PAGE_ENTRIES 512
 | |
| 
 | |
| struct kbase_context;
 | |
| struct kbase_device;
 | |
| struct kbase_mmu_table;
 | |
| struct kbase_va_region;
 | |
| struct tagged_addr;
 | |
| 
 | |
| /**
 | |
|  * enum kbase_caller_mmu_sync_info - MMU-synchronous caller info.
 | |
|  * A pointer to this type is passed down from the outer-most callers in the kbase
 | |
|  * module - where the information resides as to the synchronous / asynchronous
 | |
|  * nature of the call flow, with respect to MMU operations. ie - does the call flow relate to
 | |
|  * existing GPU work or does it come from requests (like ioctl) from user-space, power
 | |
|  * management, etc.
 | |
|  *
 | |
|  * @CALLER_MMU_UNSET_SYNCHRONICITY: default value must be invalid to avoid accidental choice
 | |
|  *                                  of a 'valid' value
 | |
|  * @CALLER_MMU_SYNC: Arbitrary value for 'synchronous that isn't easy to choose by accident
 | |
|  * @CALLER_MMU_ASYNC: Also hard to choose by accident
 | |
|  */
 | |
| enum kbase_caller_mmu_sync_info {
 | |
| 	CALLER_MMU_UNSET_SYNCHRONICITY,
 | |
| 	CALLER_MMU_SYNC = 0x02,
 | |
| 	CALLER_MMU_ASYNC
 | |
| };
 | |
| 
 | |
| /**
 | |
|  * enum kbase_mmu_op_type - enum for MMU operations
 | |
|  * @KBASE_MMU_OP_NONE:        To help catch uninitialized struct
 | |
|  * @KBASE_MMU_OP_FIRST:       The lower boundary of enum
 | |
|  * @KBASE_MMU_OP_LOCK:        Lock memory region
 | |
|  * @KBASE_MMU_OP_UNLOCK:      Unlock memory region
 | |
|  * @KBASE_MMU_OP_FLUSH_PT:    Flush page table (CLN+INV L2 only)
 | |
|  * @KBASE_MMU_OP_FLUSH_MEM:   Flush memory (CLN+INV L2+LSC)
 | |
|  * @KBASE_MMU_OP_COUNT:       The upper boundary of enum
 | |
|  */
 | |
| enum kbase_mmu_op_type {
 | |
| 	KBASE_MMU_OP_NONE = 0, /* Must be zero */
 | |
| 	KBASE_MMU_OP_FIRST, /* Must be the first non-zero op */
 | |
| 	KBASE_MMU_OP_LOCK = KBASE_MMU_OP_FIRST,
 | |
| 	KBASE_MMU_OP_UNLOCK,
 | |
| 	KBASE_MMU_OP_FLUSH_PT,
 | |
| 	KBASE_MMU_OP_FLUSH_MEM,
 | |
| 	KBASE_MMU_OP_COUNT /* Must be the last in enum */
 | |
| };
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_as_init() - Initialising GPU address space object.
 | |
|  *
 | |
|  * @kbdev: The kbase device structure for the device (must be a valid pointer).
 | |
|  * @i:     Array index of address space object.
 | |
|  *
 | |
|  * This is called from device probe to initialise an address space object
 | |
|  * of the device.
 | |
|  *
 | |
|  * Return: 0 on success and non-zero value on failure.
 | |
|  */
 | |
| int kbase_mmu_as_init(struct kbase_device *kbdev, unsigned int i);
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_as_term() - Terminate address space object.
 | |
|  *
 | |
|  * @kbdev: The kbase device structure for the device (must be a valid pointer).
 | |
|  * @i:     Array index of address space object.
 | |
|  *
 | |
|  * This is called upon device termination to destroy
 | |
|  * the address space object of the device.
 | |
|  */
 | |
| void kbase_mmu_as_term(struct kbase_device *kbdev, unsigned int i);
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_init - Initialise an object representing GPU page tables
 | |
|  *
 | |
|  * @kbdev:    Instance of GPU platform device, allocated from the probe method.
 | |
|  * @mmut:     GPU page tables to be initialized.
 | |
|  * @kctx:     Optional kbase context, may be NULL if this set of MMU tables
 | |
|  *            is not associated with a context.
 | |
|  * @group_id: The physical group ID from which to allocate GPU page tables.
 | |
|  *            Valid range is 0..(MEMORY_GROUP_MANAGER_NR_GROUPS-1).
 | |
|  *
 | |
|  * The structure should be terminated using kbase_mmu_term()
 | |
|  *
 | |
|  * Return:    0 if successful, otherwise a negative error code.
 | |
|  */
 | |
| int kbase_mmu_init(struct kbase_device *kbdev, struct kbase_mmu_table *mmut,
 | |
| 		   struct kbase_context *kctx, int group_id);
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_interrupt - Process an MMU interrupt.
 | |
|  *
 | |
|  * @kbdev:       Pointer to the kbase device for which the interrupt happened.
 | |
|  * @irq_stat:    Value of the MMU_IRQ_STATUS register.
 | |
|  *
 | |
|  * Process the MMU interrupt that was reported by the &kbase_device.
 | |
|  */
 | |
| void kbase_mmu_interrupt(struct kbase_device *kbdev, u32 irq_stat);
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_term - Terminate an object representing GPU page tables
 | |
|  *
 | |
|  * @kbdev: Instance of GPU platform device, allocated from the probe method.
 | |
|  * @mmut:  GPU page tables to be destroyed.
 | |
|  *
 | |
|  * This will free any page tables that have been allocated
 | |
|  */
 | |
| void kbase_mmu_term(struct kbase_device *kbdev, struct kbase_mmu_table *mmut);
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_create_ate - Create an address translation entry
 | |
|  *
 | |
|  * @kbdev:    Instance of GPU platform device, allocated from the probe method.
 | |
|  * @phy:      Physical address of the page to be mapped for GPU access.
 | |
|  * @flags:    Bitmask of attributes of the GPU memory region being mapped.
 | |
|  * @level:    Page table level for which to build an address translation entry.
 | |
|  * @group_id: The physical memory group in which the page was allocated.
 | |
|  *            Valid range is 0..(MEMORY_GROUP_MANAGER_NR_GROUPS-1).
 | |
|  *
 | |
|  * This function creates an address translation entry to encode the physical
 | |
|  * address of a page to be mapped for access by the GPU, along with any extra
 | |
|  * attributes required for the GPU memory region.
 | |
|  *
 | |
|  * Return: An address translation entry, either in LPAE or AArch64 format
 | |
|  *         (depending on the driver's configuration).
 | |
|  */
 | |
| u64 kbase_mmu_create_ate(struct kbase_device *kbdev, struct tagged_addr phy, unsigned long flags,
 | |
| 			 int level, int group_id);
 | |
| 
 | |
| int kbase_mmu_insert_pages_no_flush(struct kbase_device *kbdev, struct kbase_mmu_table *mmut,
 | |
| 				    u64 vpfn, struct tagged_addr *phys, size_t nr,
 | |
| 				    unsigned long flags, int group_id, u64 *dirty_pgds,
 | |
| 				    struct kbase_va_region *reg);
 | |
| int kbase_mmu_insert_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn,
 | |
| 			   struct tagged_addr *phys, size_t nr, unsigned long flags, int as_nr,
 | |
| 			   int group_id, enum kbase_caller_mmu_sync_info mmu_sync_info,
 | |
| 			   struct kbase_va_region *reg);
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_insert_pages_skip_status_update - Map 'nr' pages pointed to by 'phys'
 | |
|  * at GPU PFN 'vpfn' for GPU address space number 'as_nr'.
 | |
|  *
 | |
|  * @kbdev:         Instance of GPU platform device, allocated from the probe method.
 | |
|  * @mmut:          GPU page tables.
 | |
|  * @vpfn:          Start page frame number (in PAGE_SIZE units) of the GPU virtual pages to map.
 | |
|  * @phys:          Physical address of the page to be mapped.
 | |
|  * @nr:            The number of pages (in PAGE_SIZE units) to map.
 | |
|  * @flags:         Bitmask of attributes of the GPU memory region being mapped.
 | |
|  * @as_nr:         The GPU address space number.
 | |
|  * @group_id:      The physical memory group in which the page was allocated.
 | |
|  * @mmu_sync_info: MMU-synchronous caller info.
 | |
|  * @reg:           The region whose physical allocation is to be mapped.
 | |
|  *
 | |
|  * Similar to kbase_mmu_insert_pages() but skips updating each pages metadata
 | |
|  * for page migration.
 | |
|  *
 | |
|  * Return: 0 if successful, otherwise a negative error code.
 | |
|  */
 | |
| int kbase_mmu_insert_pages_skip_status_update(struct kbase_device *kbdev,
 | |
| 					      struct kbase_mmu_table *mmut, u64 vpfn,
 | |
| 					      struct tagged_addr *phys, size_t nr,
 | |
| 					      unsigned long flags, int as_nr, int group_id,
 | |
| 					      enum kbase_caller_mmu_sync_info mmu_sync_info,
 | |
| 					      struct kbase_va_region *reg);
 | |
| int kbase_mmu_insert_aliased_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut,
 | |
| 				   u64 vpfn, struct tagged_addr *phys, size_t nr,
 | |
| 				   unsigned long flags, int as_nr, int group_id,
 | |
| 				   enum kbase_caller_mmu_sync_info mmu_sync_info,
 | |
| 				   struct kbase_va_region *reg);
 | |
| int kbase_mmu_insert_single_imported_page(struct kbase_context *kctx, u64 vpfn,
 | |
| 					  struct tagged_addr phys, size_t nr, unsigned long flags,
 | |
| 					  int group_id,
 | |
| 					  enum kbase_caller_mmu_sync_info mmu_sync_info);
 | |
| int kbase_mmu_insert_single_aliased_page(struct kbase_context *kctx, u64 vpfn,
 | |
| 					 struct tagged_addr phys, size_t nr, unsigned long flags,
 | |
| 					 int group_id,
 | |
| 					 enum kbase_caller_mmu_sync_info mmu_sync_info);
 | |
| 
 | |
| int kbase_mmu_teardown_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut, u64 vpfn,
 | |
| 			     struct tagged_addr *phys, size_t nr_phys_pages, size_t nr_virt_pages,
 | |
| 			     int as_nr);
 | |
| int kbase_mmu_teardown_imported_pages(struct kbase_device *kbdev, struct kbase_mmu_table *mmut,
 | |
| 				      u64 vpfn, struct tagged_addr *phys, size_t nr_phys_pages,
 | |
| 				      size_t nr_virt_pages, int as_nr);
 | |
| #define kbase_mmu_teardown_firmware_pages(kbdev, mmut, vpfn, phys, nr_phys_pages, nr_virt_pages, \
 | |
| 					  as_nr)                                                 \
 | |
| 	kbase_mmu_teardown_imported_pages(kbdev, mmut, vpfn, phys, nr_phys_pages, nr_virt_pages, \
 | |
| 					  as_nr)
 | |
| 
 | |
| int kbase_mmu_update_pages(struct kbase_context *kctx, u64 vpfn, struct tagged_addr *phys,
 | |
| 			   size_t nr, unsigned long flags, int const group_id);
 | |
| #if MALI_USE_CSF
 | |
| /**
 | |
|  * kbase_mmu_update_csf_mcu_pages - Update MCU mappings with changes of phys and flags
 | |
|  *
 | |
|  * @kbdev:    Pointer to kbase device.
 | |
|  * @vpfn:     GPU Virtual PFN (Page Frame Number), in PAGE_SIZE units, of the first page to update
 | |
|  * @phys:     Pointer to the array of tagged physical addresses of the physical
 | |
|  *            pages that are pointed to by the page table entries (that need to
 | |
|  *            be updated).
 | |
|  * @nr:       Number of pages (in PAGE_SIZE units) to update
 | |
|  * @flags:    Flags
 | |
|  * @group_id: The physical memory group in which the page was allocated.
 | |
|  *            Valid range is 0..(MEMORY_GROUP_MANAGER_NR_GROUPS-1).
 | |
|  *
 | |
|  * Return: 0 on success, otherwise an error code.
 | |
|  */
 | |
| int kbase_mmu_update_csf_mcu_pages(struct kbase_device *kbdev, u64 vpfn, struct tagged_addr *phys,
 | |
| 				   size_t nr, unsigned long flags, int const group_id);
 | |
| #endif
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_migrate_page - Migrate GPU mappings and content between memory pages
 | |
|  *
 | |
|  * @old_phys:     Old physical page to be replaced.
 | |
|  * @new_phys:     New physical page used to replace old physical page.
 | |
|  * @old_dma_addr: DMA address of the old page.
 | |
|  * @new_dma_addr: DMA address of the new page.
 | |
|  * @level:        MMU page table level of the provided PGD.
 | |
|  *
 | |
|  * The page migration process is made of 2 big steps:
 | |
|  *
 | |
|  * 1) Copy the content of the old page to the new page.
 | |
|  * 2) Remap the virtual page, that is: replace either the ATE (if the old page
 | |
|  *    was a regular page) or the PTE (if the old page was used as a PGD) in the
 | |
|  *    MMU page table with the new page.
 | |
|  *
 | |
|  * During the process, the MMU region is locked to prevent GPU access to the
 | |
|  * virtual memory page that is being remapped.
 | |
|  *
 | |
|  * Before copying the content of the old page to the new page and while the
 | |
|  * MMU region is locked, a GPU cache flush is performed to make sure that
 | |
|  * pending GPU writes are finalized to the old page before copying.
 | |
|  * That is necessary because otherwise there's a risk that GPU writes might
 | |
|  * be finalized to the old page, and not new page, after migration.
 | |
|  * The MMU region is unlocked only at the end of the migration operation.
 | |
|  *
 | |
|  * Return: 0 on success, otherwise an error code.
 | |
|  */
 | |
| int kbase_mmu_migrate_page(struct tagged_addr old_phys, struct tagged_addr new_phys,
 | |
| 			   dma_addr_t old_dma_addr, dma_addr_t new_dma_addr, int level);
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_flush_pa_range() - Flush physical address range from the GPU caches
 | |
|  *
 | |
|  * @kbdev:    Instance of GPU platform device, allocated from the probe method.
 | |
|  * @kctx:     Pointer to kbase context, it can be NULL if the physical address
 | |
|  *            range is not associated with User created context.
 | |
|  * @phys:     Starting address of the physical range to start the operation on.
 | |
|  * @size:     Number of bytes to work on.
 | |
|  * @flush_op: Type of cache flush operation to perform.
 | |
|  *
 | |
|  * Issue a cache flush physical range command. This function won't perform any
 | |
|  * flush if the GPU doesn't support FLUSH_PA_RANGE command. The flush would be
 | |
|  * performed only if the context has a JASID assigned to it.
 | |
|  * This function is basically a wrapper for kbase_gpu_cache_flush_pa_range_and_busy_wait().
 | |
|  */
 | |
| void kbase_mmu_flush_pa_range(struct kbase_device *kbdev, struct kbase_context *kctx,
 | |
| 			      phys_addr_t phys, size_t size, enum kbase_mmu_op_type flush_op);
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_bus_fault_interrupt - Process a bus fault interrupt.
 | |
|  *
 | |
|  * @kbdev:       Pointer to the kbase device for which bus fault was reported.
 | |
|  * @status:      Value of the GPU_FAULTSTATUS register.
 | |
|  * @as_nr:       GPU address space for which the bus fault occurred.
 | |
|  *
 | |
|  * Process the bus fault interrupt that was reported for a particular GPU
 | |
|  * address space.
 | |
|  *
 | |
|  * Return: zero if the operation was successful, non-zero otherwise.
 | |
|  */
 | |
| int kbase_mmu_bus_fault_interrupt(struct kbase_device *kbdev, u32 status, u32 as_nr);
 | |
| 
 | |
| /**
 | |
|  * kbase_mmu_gpu_fault_interrupt() - Report a GPU fault.
 | |
|  *
 | |
|  * @kbdev:    Kbase device pointer
 | |
|  * @status:   GPU fault status
 | |
|  * @as_nr:    Faulty address space
 | |
|  * @address:  GPU fault address
 | |
|  * @as_valid: true if address space is valid
 | |
|  *
 | |
|  * This function builds GPU fault information to submit a work
 | |
|  * for reporting the details of the fault.
 | |
|  */
 | |
| void kbase_mmu_gpu_fault_interrupt(struct kbase_device *kbdev, u32 status, u32 as_nr, u64 address,
 | |
| 				   bool as_valid);
 | |
| 
 | |
| /**
 | |
|  * kbase_context_mmu_group_id_get - Decode a memory group ID from
 | |
|  *                                 base_context_create_flags
 | |
|  *
 | |
|  * @flags: Bitmask of flags to pass to base_context_init.
 | |
|  *
 | |
|  * Memory allocated for GPU page tables will come from the returned group.
 | |
|  *
 | |
|  * Return: Physical memory group ID. Valid range is 0..(BASE_MEM_GROUP_COUNT-1).
 | |
|  */
 | |
| static inline int kbase_context_mmu_group_id_get(base_context_create_flags const flags)
 | |
| {
 | |
| 	KBASE_DEBUG_ASSERT(flags == (flags & BASEP_CONTEXT_CREATE_ALLOWED_FLAGS));
 | |
| 	return (int)BASE_CONTEXT_MMU_GROUP_ID_GET(flags);
 | |
| }
 | |
| 
 | |
| #endif /* _KBASE_MMU_H_ */
 |