// SPDX-License-Identifier: GPL-2.0-or-later /* * Copyright (C) 2012 ARM Ltd. * Author: Catalin Marinas * Copyright (C) 2017 Linaro Ltd. * Copyright (C) 2021 SiFive */ #include #include #include #include #include #include #include #include #ifdef CONFIG_RISCV_ISA_V_PREEMPTIVE #include #endif static inline void riscv_v_flags_set(u32 flags) { WRITE_ONCE(current->thread.riscv_v_flags, flags); } static inline void riscv_v_start(u32 flags) { int orig; orig = riscv_v_flags(); BUG_ON((orig & flags) != 0); riscv_v_flags_set(orig | flags); barrier(); } static inline void riscv_v_stop(u32 flags) { int orig; barrier(); orig = riscv_v_flags(); BUG_ON((orig & flags) == 0); riscv_v_flags_set(orig & ~flags); } /* * Claim ownership of the CPU vector context for use by the calling context. * * The caller may freely manipulate the vector context metadata until * put_cpu_vector_context() is called. */ void get_cpu_vector_context(void) { /* * disable softirqs so it is impossible for softirqs to nest * get_cpu_vector_context() when kernel is actively using Vector. */ if (!IS_ENABLED(CONFIG_PREEMPT_RT)) local_bh_disable(); else preempt_disable(); riscv_v_start(RISCV_KERNEL_MODE_V); } /* * Release the CPU vector context. * * Must be called from a context in which get_cpu_vector_context() was * previously called, with no call to put_cpu_vector_context() in the * meantime. */ void put_cpu_vector_context(void) { riscv_v_stop(RISCV_KERNEL_MODE_V); if (!IS_ENABLED(CONFIG_PREEMPT_RT)) local_bh_enable(); else preempt_enable(); } #ifdef CONFIG_RISCV_ISA_V_PREEMPTIVE static __always_inline u32 *riscv_v_flags_ptr(void) { return ¤t->thread.riscv_v_flags; } static inline void riscv_preempt_v_set_dirty(void) { *riscv_v_flags_ptr() |= RISCV_PREEMPT_V_DIRTY; } static inline void riscv_preempt_v_reset_flags(void) { *riscv_v_flags_ptr() &= ~(RISCV_PREEMPT_V_DIRTY | RISCV_PREEMPT_V_NEED_RESTORE); } static inline void riscv_v_ctx_depth_inc(void) { *riscv_v_flags_ptr() += RISCV_V_CTX_UNIT_DEPTH; } static inline void riscv_v_ctx_depth_dec(void) { *riscv_v_flags_ptr() -= RISCV_V_CTX_UNIT_DEPTH; } static inline u32 riscv_v_ctx_get_depth(void) { return *riscv_v_flags_ptr() & RISCV_V_CTX_DEPTH_MASK; } static int riscv_v_stop_kernel_context(void) { if (riscv_v_ctx_get_depth() != 0 || !riscv_preempt_v_started(current)) return 1; riscv_preempt_v_clear_dirty(current); riscv_v_stop(RISCV_PREEMPT_V); return 0; } static int riscv_v_start_kernel_context(bool *is_nested) { struct __riscv_v_ext_state *kvstate, *uvstate; kvstate = ¤t->thread.kernel_vstate; if (!kvstate->datap) return -ENOENT; if (riscv_preempt_v_started(current)) { WARN_ON(riscv_v_ctx_get_depth() == 0); *is_nested = true; get_cpu_vector_context(); if (riscv_preempt_v_dirty(current)) { __riscv_v_vstate_save(kvstate, kvstate->datap); riscv_preempt_v_clear_dirty(current); } riscv_preempt_v_set_restore(current); return 0; } /* Transfer the ownership of V from user to kernel, then save */ riscv_v_start(RISCV_PREEMPT_V | RISCV_PREEMPT_V_DIRTY); if ((task_pt_regs(current)->status & SR_VS) == SR_VS_DIRTY) { uvstate = ¤t->thread.vstate; __riscv_v_vstate_save(uvstate, uvstate->datap); } riscv_preempt_v_clear_dirty(current); return 0; } /* low-level V context handling code, called with irq disabled */ asmlinkage void riscv_v_context_nesting_start(struct pt_regs *regs) { int depth; if (!riscv_preempt_v_started(current)) return; depth = riscv_v_ctx_get_depth(); if (depth == 0 && (regs->status & SR_VS) == SR_VS_DIRTY) riscv_preempt_v_set_dirty(); riscv_v_ctx_depth_inc(); } asmlinkage void riscv_v_context_nesting_end(struct pt_regs *regs) { struct __riscv_v_ext_state *vstate = ¤t->thread.kernel_vstate; u32 depth; WARN_ON(!irqs_disabled()); if (!riscv_preempt_v_started(current)) return; riscv_v_ctx_depth_dec(); depth = riscv_v_ctx_get_depth(); if (depth == 0) { if (riscv_preempt_v_restore(current)) { __riscv_v_vstate_restore(vstate, vstate->datap); __riscv_v_vstate_clean(regs); riscv_preempt_v_reset_flags(); } } } #else #define riscv_v_start_kernel_context(nested) (-ENOENT) #define riscv_v_stop_kernel_context() (-ENOENT) #endif /* CONFIG_RISCV_ISA_V_PREEMPTIVE */ /* * kernel_vector_begin(): obtain the CPU vector registers for use by the calling * context * * Must not be called unless may_use_simd() returns true. * Task context in the vector registers is saved back to memory as necessary. * * A matching call to kernel_vector_end() must be made before returning from the * calling context. * * The caller may freely use the vector registers until kernel_vector_end() is * called. */ void kernel_vector_begin(void) { bool nested = false; if (WARN_ON(!has_vector())) return; BUG_ON(!may_use_simd()); if (riscv_v_start_kernel_context(&nested)) { get_cpu_vector_context(); riscv_v_vstate_save(¤t->thread.vstate, task_pt_regs(current)); } if (!nested) riscv_v_vstate_set_restore(current, task_pt_regs(current)); riscv_v_enable(); } EXPORT_SYMBOL_GPL(kernel_vector_begin); /* * kernel_vector_end(): give the CPU vector registers back to the current task * * Must be called from a context in which kernel_vector_begin() was previously * called, with no call to kernel_vector_end() in the meantime. * * The caller must not use the vector registers after this function is called, * unless kernel_vector_begin() is called again in the meantime. */ void kernel_vector_end(void) { if (WARN_ON(!has_vector())) return; riscv_v_disable(); if (riscv_v_stop_kernel_context()) put_cpu_vector_context(); } EXPORT_SYMBOL_GPL(kernel_vector_end);