/* * * (C) COPYRIGHT 2014-2017 ARM Limited. All rights reserved. * * This program is free software and is provided to you under the terms of the * GNU General Public License version 2 as published by the Free Software * Foundation, and any use by you of this program is subject to the terms * of such GNU licence. * * A copy of the licence is included with the program, and can also be obtained * from Free Software Foundation, Inc., 51 Franklin Street, Fifth Floor, * Boston, MA 02110-1301, USA. * */ /* * HW access job manager common APIs */ #ifndef _KBASE_HWACCESS_JM_H_ #define _KBASE_HWACCESS_JM_H_ /** * kbase_backend_run_atom() - Run an atom on the GPU * @kbdev: Device pointer * @atom: Atom to run * * Caller must hold the HW access lock */ void kbase_backend_run_atom(struct kbase_device *kbdev, struct kbase_jd_atom *katom); /** * kbase_backend_slot_update - Update state based on slot ringbuffers * * @kbdev: Device pointer * * Inspect the jobs in the slot ringbuffers and update state. * * This will cause jobs to be submitted to hardware if they are unblocked */ void kbase_backend_slot_update(struct kbase_device *kbdev); /** * kbase_backend_find_and_release_free_address_space() - Release a free AS * @kbdev: Device pointer * @kctx: Context pointer * * This function can evict an idle context from the runpool, freeing up the * address space it was using. * * The address space is marked as in use. The caller must either assign a * context using kbase_gpu_use_ctx(), or release it using * kbase_ctx_sched_release() * * Return: Number of free address space, or KBASEP_AS_NR_INVALID if none * available */ int kbase_backend_find_and_release_free_address_space( struct kbase_device *kbdev, struct kbase_context *kctx); /** * kbase_backend_use_ctx() - Activate a currently unscheduled context, using the * provided address space. * @kbdev: Device pointer * @kctx: Context pointer. May be NULL * @as_nr: Free address space to use * * kbase_gpu_next_job() will pull atoms from the active context. * * Return: true if successful, false if ASID not assigned. */ bool kbase_backend_use_ctx(struct kbase_device *kbdev, struct kbase_context *kctx, int as_nr); /** * kbase_backend_use_ctx_sched() - Activate a context. * @kbdev: Device pointer * @kctx: Context pointer * * kbase_gpu_next_job() will pull atoms from the active context. * * The context must already be scheduled and assigned to an address space. If * the context is not scheduled, then kbase_gpu_use_ctx() should be used * instead. * * Caller must hold hwaccess_lock * * Return: true if context is now active, false otherwise (ie if context does * not have an address space assigned) */ bool kbase_backend_use_ctx_sched(struct kbase_device *kbdev, struct kbase_context *kctx); /** * kbase_backend_release_ctx_irq - Release a context from the GPU. This will * de-assign the assigned address space. * @kbdev: Device pointer * @kctx: Context pointer * * Caller must hold kbase_device->mmu_hw_mutex and hwaccess_lock */ void kbase_backend_release_ctx_irq(struct kbase_device *kbdev, struct kbase_context *kctx); /** * kbase_backend_release_ctx_noirq - Release a context from the GPU. This will * de-assign the assigned address space. * @kbdev: Device pointer * @kctx: Context pointer * * Caller must hold kbase_device->mmu_hw_mutex * * This function must perform any operations that could not be performed in IRQ * context by kbase_backend_release_ctx_irq(). */ void kbase_backend_release_ctx_noirq(struct kbase_device *kbdev, struct kbase_context *kctx); /** * kbase_backend_cacheclean - Perform a cache clean if the given atom requires * one * @kbdev: Device pointer * @katom: Pointer to the failed atom * * On some GPUs, the GPU cache must be cleaned following a failed atom. This * function performs a clean if it is required by @katom. */ void kbase_backend_cacheclean(struct kbase_device *kbdev, struct kbase_jd_atom *katom); /** * kbase_backend_complete_wq() - Perform backend-specific actions required on * completing an atom. * @kbdev: Device pointer * @katom: Pointer to the atom to complete * * This function should only be called from kbase_jd_done_worker() or * js_return_worker(). * * Return: true if atom has completed, false if atom should be re-submitted */ void kbase_backend_complete_wq(struct kbase_device *kbdev, struct kbase_jd_atom *katom); /** * kbase_backend_complete_wq_post_sched - Perform backend-specific actions * required on completing an atom, after * any scheduling has taken place. * @kbdev: Device pointer * @core_req: Core requirements of atom * @affinity: Affinity of atom * @coreref_state: Coreref state of atom * * This function should only be called from kbase_jd_done_worker() or * js_return_worker(). */ void kbase_backend_complete_wq_post_sched(struct kbase_device *kbdev, base_jd_core_req core_req, u64 affinity, enum kbase_atom_coreref_state coreref_state); /** * kbase_backend_reset() - The GPU is being reset. Cancel all jobs on the GPU * and remove any others from the ringbuffers. * @kbdev: Device pointer * @end_timestamp: Timestamp of reset */ void kbase_backend_reset(struct kbase_device *kbdev, ktime_t *end_timestamp); /** * kbase_backend_inspect_head() - Return the atom currently at the head of slot * @js * @kbdev: Device pointer * @js: Job slot to inspect * * Return : Atom currently at the head of slot @js, or NULL */ struct kbase_jd_atom *kbase_backend_inspect_head(struct kbase_device *kbdev, int js); /** * kbase_backend_inspect_tail - Return the atom currently at the tail of slot * @js * @kbdev: Device pointer * @js: Job slot to inspect * * Return : Atom currently at the head of slot @js, or NULL */ struct kbase_jd_atom *kbase_backend_inspect_tail(struct kbase_device *kbdev, int js); /** * kbase_backend_nr_atoms_on_slot() - Return the number of atoms currently on a * slot. * @kbdev: Device pointer * @js: Job slot to inspect * * Return : Number of atoms currently on slot */ int kbase_backend_nr_atoms_on_slot(struct kbase_device *kbdev, int js); /** * kbase_backend_nr_atoms_submitted() - Return the number of atoms on a slot * that are currently on the GPU. * @kbdev: Device pointer * @js: Job slot to inspect * * Return : Number of atoms currently on slot @js that are currently on the GPU. */ int kbase_backend_nr_atoms_submitted(struct kbase_device *kbdev, int js); /** * kbase_backend_ctx_count_changed() - Number of contexts ready to submit jobs * has changed. * @kbdev: Device pointer * * Perform any required backend-specific actions (eg starting/stopping * scheduling timers). */ void kbase_backend_ctx_count_changed(struct kbase_device *kbdev); /** * kbase_backend_timeouts_changed() - Job Scheduler timeouts have changed. * @kbdev: Device pointer * * Perform any required backend-specific actions (eg updating timeouts of * currently running atoms). */ void kbase_backend_timeouts_changed(struct kbase_device *kbdev); /** * kbase_backend_slot_free() - Return the number of jobs that can be currently * submitted to slot @js. * @kbdev: Device pointer * @js: Job slot to inspect * * Return : Number of jobs that can be submitted. */ int kbase_backend_slot_free(struct kbase_device *kbdev, int js); /** * kbase_job_check_enter_disjoint - potentially leave disjoint state * @kbdev: kbase device * @target_katom: atom which is finishing * * Work out whether to leave disjoint state when finishing an atom that was * originated by kbase_job_check_enter_disjoint(). */ void kbase_job_check_leave_disjoint(struct kbase_device *kbdev, struct kbase_jd_atom *target_katom); /** * kbase_backend_jm_kill_jobs_from_kctx - Kill all jobs that are currently * running from a context * @kctx: Context pointer * * This is used in response to a page fault to remove all jobs from the faulting * context from the hardware. */ void kbase_backend_jm_kill_jobs_from_kctx(struct kbase_context *kctx); /** * kbase_jm_wait_for_zero_jobs - Wait for context to have zero jobs running, and * to be descheduled. * @kctx: Context pointer * * This should be called following kbase_js_zap_context(), to ensure the context * can be safely destroyed. */ void kbase_jm_wait_for_zero_jobs(struct kbase_context *kctx); /** * kbase_backend_get_current_flush_id - Return the current flush ID * * @kbdev: Device pointer * * Return: the current flush ID to be recorded for each job chain */ u32 kbase_backend_get_current_flush_id(struct kbase_device *kbdev); #if KBASE_GPU_RESET_EN /** * kbase_prepare_to_reset_gpu - Prepare for resetting the GPU. * @kbdev: Device pointer * * This function just soft-stops all the slots to ensure that as many jobs as * possible are saved. * * Return: a boolean which should be interpreted as follows: * - true - Prepared for reset, kbase_reset_gpu should be called. * - false - Another thread is performing a reset, kbase_reset_gpu should * not be called. */ bool kbase_prepare_to_reset_gpu(struct kbase_device *kbdev); /** * kbase_reset_gpu - Reset the GPU * @kbdev: Device pointer * * This function should be called after kbase_prepare_to_reset_gpu if it returns * true. It should never be called without a corresponding call to * kbase_prepare_to_reset_gpu. * * After this function is called (or not called if kbase_prepare_to_reset_gpu * returned false), the caller should wait for kbdev->reset_waitq to be * signalled to know when the reset has completed. */ void kbase_reset_gpu(struct kbase_device *kbdev); /** * kbase_prepare_to_reset_gpu_locked - Prepare for resetting the GPU. * @kbdev: Device pointer * * This function just soft-stops all the slots to ensure that as many jobs as * possible are saved. * * Return: a boolean which should be interpreted as follows: * - true - Prepared for reset, kbase_reset_gpu should be called. * - false - Another thread is performing a reset, kbase_reset_gpu should * not be called. */ bool kbase_prepare_to_reset_gpu_locked(struct kbase_device *kbdev); /** * kbase_reset_gpu_locked - Reset the GPU * @kbdev: Device pointer * * This function should be called after kbase_prepare_to_reset_gpu if it * returns true. It should never be called without a corresponding call to * kbase_prepare_to_reset_gpu. * * After this function is called (or not called if kbase_prepare_to_reset_gpu * returned false), the caller should wait for kbdev->reset_waitq to be * signalled to know when the reset has completed. */ void kbase_reset_gpu_locked(struct kbase_device *kbdev); /** * kbase_reset_gpu_silent - Reset the GPU silently * @kbdev: Device pointer * * Reset the GPU without trying to cancel jobs and don't emit messages into * the kernel log while doing the reset. * * This function should be used in cases where we are doing a controlled reset * of the GPU as part of normal processing (e.g. exiting protected mode) where * the driver will have ensured the scheduler has been idled and all other * users of the GPU (e.g. instrumentation) have been suspended. */ void kbase_reset_gpu_silent(struct kbase_device *kbdev); /** * kbase_reset_gpu_active - Reports if the GPU is being reset * @kbdev: Device pointer * * Return: True if the GPU is in the process of being reset. */ bool kbase_reset_gpu_active(struct kbase_device *kbdev); #endif /** * kbase_job_slot_hardstop - Hard-stop the specified job slot * @kctx: The kbase context that contains the job(s) that should * be hard-stopped * @js: The job slot to hard-stop * @target_katom: The job that should be hard-stopped (or NULL for all * jobs from the context) * Context: * The job slot lock must be held when calling this function. */ void kbase_job_slot_hardstop(struct kbase_context *kctx, int js, struct kbase_jd_atom *target_katom); extern struct protected_mode_ops kbase_native_protected_ops; #endif /* _KBASE_HWACCESS_JM_H_ */