// SPDX-License-Identifier: GPL-2.0 WITH Linux-syscall-note /* * * (C) COPYRIGHT 2019-2023 ARM Limited. All rights reserved. * * This program is free software and is provided to you under the terms of the * GNU General Public License version 2 as published by the Free Software * Foundation, and any use by you of this program is subject to the terms * of such GNU license. * * This program is distributed in the hope that it will be useful, * but WITHOUT ANY WARRANTY; without even the implied warranty of * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the * GNU General Public License for more details. * * You should have received a copy of the GNU General Public License * along with this program; if not, you can access it online at * http://www.gnu.org/licenses/gpl-2.0.html. * */ #include #include "mali_kbase_csf_heap_context_alloc.h" /* Size of one heap context structure, in bytes. */ #define HEAP_CTX_SIZE ((u32)32) /** * sub_alloc - Sub-allocate a heap context from a GPU memory region * * @ctx_alloc: Pointer to the heap context allocator. * * Return: GPU virtual address of the allocated heap context or 0 on failure. */ static u64 sub_alloc(struct kbase_csf_heap_context_allocator *const ctx_alloc) { struct kbase_context *const kctx = ctx_alloc->kctx; unsigned long heap_nr = 0; u32 ctx_offset = 0; u64 heap_gpu_va = 0; struct kbase_vmap_struct mapping; void *ctx_ptr = NULL; lockdep_assert_held(&ctx_alloc->lock); heap_nr = find_first_zero_bit(ctx_alloc->in_use, MAX_TILER_HEAPS); if (unlikely(heap_nr >= MAX_TILER_HEAPS)) { dev_dbg(kctx->kbdev->dev, "No free tiler heap contexts in the pool"); return 0; } ctx_offset = heap_nr * ctx_alloc->heap_context_size_aligned; heap_gpu_va = ctx_alloc->gpu_va + ctx_offset; ctx_ptr = kbase_vmap_prot(kctx, heap_gpu_va, ctx_alloc->heap_context_size_aligned, KBASE_REG_CPU_WR, &mapping); if (unlikely(!ctx_ptr)) { dev_err(kctx->kbdev->dev, "Failed to map tiler heap context %lu (0x%llX)\n", heap_nr, heap_gpu_va); return 0; } memset(ctx_ptr, 0, ctx_alloc->heap_context_size_aligned); kbase_vunmap(ctx_ptr, &mapping); bitmap_set(ctx_alloc->in_use, heap_nr, 1); dev_dbg(kctx->kbdev->dev, "Allocated tiler heap context %lu (0x%llX)\n", heap_nr, heap_gpu_va); return heap_gpu_va; } /** * evict_heap_context - Evict the data of heap context from GPU's L2 cache. * * @ctx_alloc: Pointer to the heap context allocator. * @heap_gpu_va: The GPU virtual address of a heap context structure to free. * * This function is called when memory for the heap context is freed. It uses the * FLUSH_PA_RANGE command to evict the data of heap context, so on older CSF GPUs * there is nothing done. The whole GPU cache is anyways expected to be flushed * on older GPUs when initial chunks of the heap are freed just before the memory * for heap context is freed. */ static void evict_heap_context(struct kbase_csf_heap_context_allocator *const ctx_alloc, u64 const heap_gpu_va) { struct kbase_context *const kctx = ctx_alloc->kctx; u32 offset_in_bytes = (u32)(heap_gpu_va - ctx_alloc->gpu_va); u32 offset_within_page = offset_in_bytes & ~PAGE_MASK; u32 page_index = offset_in_bytes >> PAGE_SHIFT; struct tagged_addr page = kbase_get_gpu_phy_pages(ctx_alloc->region)[page_index]; phys_addr_t heap_context_pa = as_phys_addr_t(page) + offset_within_page; lockdep_assert_held(&ctx_alloc->lock); /* There is no need to take vm_lock here as the ctx_alloc region is protected * via a nonzero no_user_free_count. The region and the backing page can't * disappear whilst this function is executing. Flush type is passed as FLUSH_PT * to CLN+INV L2 only. */ kbase_mmu_flush_pa_range(kctx->kbdev, kctx, heap_context_pa, ctx_alloc->heap_context_size_aligned, KBASE_MMU_OP_FLUSH_PT); } /** * sub_free - Free a heap context sub-allocated from a GPU memory region * * @ctx_alloc: Pointer to the heap context allocator. * @heap_gpu_va: The GPU virtual address of a heap context structure to free. */ static void sub_free(struct kbase_csf_heap_context_allocator *const ctx_alloc, u64 const heap_gpu_va) { struct kbase_context *const kctx = ctx_alloc->kctx; u32 ctx_offset = 0; unsigned int heap_nr = 0; lockdep_assert_held(&ctx_alloc->lock); if (WARN_ON(!ctx_alloc->region)) return; if (WARN_ON(heap_gpu_va < ctx_alloc->gpu_va)) return; ctx_offset = (u32)(heap_gpu_va - ctx_alloc->gpu_va); if (WARN_ON(ctx_offset >= (ctx_alloc->region->nr_pages << PAGE_SHIFT)) || WARN_ON(ctx_offset % ctx_alloc->heap_context_size_aligned)) return; evict_heap_context(ctx_alloc, heap_gpu_va); heap_nr = ctx_offset / ctx_alloc->heap_context_size_aligned; dev_dbg(kctx->kbdev->dev, "Freed tiler heap context %d (0x%llX)\n", heap_nr, heap_gpu_va); bitmap_clear(ctx_alloc->in_use, heap_nr, 1); } int kbase_csf_heap_context_allocator_init(struct kbase_csf_heap_context_allocator *const ctx_alloc, struct kbase_context *const kctx) { const u32 gpu_cache_line_size = (1U << kctx->kbdev->gpu_props.log2_line_size); /* We cannot pre-allocate GPU memory here because the * custom VA zone may not have been created yet. */ ctx_alloc->kctx = kctx; ctx_alloc->heap_context_size_aligned = (HEAP_CTX_SIZE + gpu_cache_line_size - 1) & ~(gpu_cache_line_size - 1); mutex_init(&ctx_alloc->lock); dev_dbg(kctx->kbdev->dev, "Initialized a tiler heap context allocator\n"); return 0; } void kbase_csf_heap_context_allocator_term(struct kbase_csf_heap_context_allocator *const ctx_alloc) { struct kbase_context *const kctx = ctx_alloc->kctx; dev_dbg(kctx->kbdev->dev, "Terminating tiler heap context allocator\n"); if (ctx_alloc->region) { kbase_gpu_vm_lock(kctx); WARN_ON(!kbase_va_region_is_no_user_free(ctx_alloc->region)); kbase_va_region_no_user_free_dec(ctx_alloc->region); kbase_mem_free_region(kctx, ctx_alloc->region); kbase_gpu_vm_unlock(kctx); } mutex_destroy(&ctx_alloc->lock); } u64 kbase_csf_heap_context_allocator_alloc(struct kbase_csf_heap_context_allocator *const ctx_alloc) { struct kbase_context *const kctx = ctx_alloc->kctx; u64 flags = BASE_MEM_PROT_GPU_RD | BASE_MEM_PROT_GPU_WR | BASE_MEM_PROT_CPU_WR | BASEP_MEM_NO_USER_FREE | BASE_MEM_PROT_CPU_RD; u64 nr_pages = PFN_UP(MAX_TILER_HEAPS * ctx_alloc->heap_context_size_aligned); u64 heap_gpu_va = 0; /* Calls to this function are inherently asynchronous, with respect to * MMU operations. */ const enum kbase_caller_mmu_sync_info mmu_sync_info = CALLER_MMU_ASYNC; mutex_lock(&ctx_alloc->lock); /* If the pool of heap contexts wasn't already allocated then * allocate it. */ if (!ctx_alloc->region) { ctx_alloc->region = kbase_mem_alloc(kctx, nr_pages, nr_pages, 0, &flags, &ctx_alloc->gpu_va, mmu_sync_info); } /* If the pool still isn't allocated then an error occurred. */ if (unlikely(!ctx_alloc->region)) dev_dbg(kctx->kbdev->dev, "Failed to allocate a pool of tiler heap contexts"); else heap_gpu_va = sub_alloc(ctx_alloc); mutex_unlock(&ctx_alloc->lock); return heap_gpu_va; } void kbase_csf_heap_context_allocator_free(struct kbase_csf_heap_context_allocator *const ctx_alloc, u64 const heap_gpu_va) { mutex_lock(&ctx_alloc->lock); sub_free(ctx_alloc, heap_gpu_va); mutex_unlock(&ctx_alloc->lock); }