Support new prctl with key PR_RISCV_SET_ICACHE_FLUSH_CTX to enable optimization of cross modifying code. This prctl enables userspace code to use icache flushing instructions such as fence.i with the guarantee that the icache will continue to be clean after thread migration. Signed-off-by: Charlie Jenkins <charlie@rivosinc.com> Reviewed-by: Atish Patra <atishp@rivosinc.com> Reviewed-by: Alexandre Ghiti <alexghiti@rivosinc.com> Reviewed-by: Samuel Holland <samuel.holland@sifive.com> Link: https://lore.kernel.org/r/20240312-fencei-v13-2-4b6bdc2bbf32@rivosinc.com Signed-off-by: Palmer Dabbelt <palmer@rivosinc.com>
109 lines
2.8 KiB
C
109 lines
2.8 KiB
C
/* SPDX-License-Identifier: GPL-2.0-only */
|
|
/*
|
|
* Copyright (C) 2012 Regents of the University of California
|
|
*/
|
|
|
|
#ifndef _ASM_RISCV_SWITCH_TO_H
|
|
#define _ASM_RISCV_SWITCH_TO_H
|
|
|
|
#include <linux/jump_label.h>
|
|
#include <linux/sched/task_stack.h>
|
|
#include <linux/mm_types.h>
|
|
#include <asm/vector.h>
|
|
#include <asm/cpufeature.h>
|
|
#include <asm/processor.h>
|
|
#include <asm/ptrace.h>
|
|
#include <asm/csr.h>
|
|
|
|
#ifdef CONFIG_FPU
|
|
extern void __fstate_save(struct task_struct *save_to);
|
|
extern void __fstate_restore(struct task_struct *restore_from);
|
|
|
|
static inline void __fstate_clean(struct pt_regs *regs)
|
|
{
|
|
regs->status = (regs->status & ~SR_FS) | SR_FS_CLEAN;
|
|
}
|
|
|
|
static inline void fstate_off(struct task_struct *task,
|
|
struct pt_regs *regs)
|
|
{
|
|
regs->status = (regs->status & ~SR_FS) | SR_FS_OFF;
|
|
}
|
|
|
|
static inline void fstate_save(struct task_struct *task,
|
|
struct pt_regs *regs)
|
|
{
|
|
if ((regs->status & SR_FS) == SR_FS_DIRTY) {
|
|
__fstate_save(task);
|
|
__fstate_clean(regs);
|
|
}
|
|
}
|
|
|
|
static inline void fstate_restore(struct task_struct *task,
|
|
struct pt_regs *regs)
|
|
{
|
|
if ((regs->status & SR_FS) != SR_FS_OFF) {
|
|
__fstate_restore(task);
|
|
__fstate_clean(regs);
|
|
}
|
|
}
|
|
|
|
static inline void __switch_to_fpu(struct task_struct *prev,
|
|
struct task_struct *next)
|
|
{
|
|
struct pt_regs *regs;
|
|
|
|
regs = task_pt_regs(prev);
|
|
fstate_save(prev, regs);
|
|
fstate_restore(next, task_pt_regs(next));
|
|
}
|
|
|
|
static __always_inline bool has_fpu(void)
|
|
{
|
|
return riscv_has_extension_likely(RISCV_ISA_EXT_f) ||
|
|
riscv_has_extension_likely(RISCV_ISA_EXT_d);
|
|
}
|
|
#else
|
|
static __always_inline bool has_fpu(void) { return false; }
|
|
#define fstate_save(task, regs) do { } while (0)
|
|
#define fstate_restore(task, regs) do { } while (0)
|
|
#define __switch_to_fpu(__prev, __next) do { } while (0)
|
|
#endif
|
|
|
|
extern struct task_struct *__switch_to(struct task_struct *,
|
|
struct task_struct *);
|
|
|
|
static inline bool switch_to_should_flush_icache(struct task_struct *task)
|
|
{
|
|
#ifdef CONFIG_SMP
|
|
bool stale_mm = task->mm && task->mm->context.force_icache_flush;
|
|
bool stale_thread = task->thread.force_icache_flush;
|
|
bool thread_migrated = smp_processor_id() != task->thread.prev_cpu;
|
|
|
|
return thread_migrated && (stale_mm || stale_thread);
|
|
#else
|
|
return false;
|
|
#endif
|
|
}
|
|
|
|
#ifdef CONFIG_SMP
|
|
#define __set_prev_cpu(thread) ((thread).prev_cpu = smp_processor_id())
|
|
#else
|
|
#define __set_prev_cpu(thread)
|
|
#endif
|
|
|
|
#define switch_to(prev, next, last) \
|
|
do { \
|
|
struct task_struct *__prev = (prev); \
|
|
struct task_struct *__next = (next); \
|
|
__set_prev_cpu(__prev->thread); \
|
|
if (has_fpu()) \
|
|
__switch_to_fpu(__prev, __next); \
|
|
if (has_vector()) \
|
|
__switch_to_vector(__prev, __next); \
|
|
if (switch_to_should_flush_icache(__next)) \
|
|
local_flush_icache_all(); \
|
|
((last) = __switch_to(__prev, __next)); \
|
|
} while (0)
|
|
|
|
#endif /* _ASM_RISCV_SWITCH_TO_H */
|