void user_disable_single_step(struct task_struct *child) { /* * Make sure block stepping (BTF) is disabled. */ if (test_tsk_thread_flag(child, TIF_BLOCKSTEP)) set_task_blockstep(child, false); /* Always clear TIF_SINGLESTEP... */ clear_tsk_thread_flag(child, TIF_SINGLESTEP); /* But touch TF only if it was set by us.. */ if (test_and_clear_tsk_thread_flag(child, TIF_FORCED_TF)) task_pt_regs(child)->flags &= ~X86_EFLAGS_TF; }
static int ssb_prctl_set(struct task_struct *task, unsigned long ctrl) { bool update; if (ssb_mode != SPEC_STORE_BYPASS_PRCTL && ssb_mode != SPEC_STORE_BYPASS_SECCOMP) return -ENXIO; switch (ctrl) { case PR_SPEC_ENABLE: /* If speculation is force disabled, enable is not allowed */ if (task_spec_ssb_force_disable(task)) return -EPERM; task_clear_spec_ssb_disable(task); update = test_and_clear_tsk_thread_flag(task, TIF_SSBD); break; case PR_SPEC_DISABLE: task_set_spec_ssb_disable(task); update = !test_and_set_tsk_thread_flag(task, TIF_SSBD); break; case PR_SPEC_FORCE_DISABLE: task_set_spec_ssb_disable(task); task_set_spec_ssb_force_disable(task); update = !test_and_set_tsk_thread_flag(task, TIF_SSBD); break; default: return -ERANGE; } /* * If being set on non-current task, delay setting the CPU * mitigation until it is next scheduled. */ if (task == current && update) speculative_store_bypass_update_current(); return 0; }
void ocd_disable(struct task_struct *child) { u32 dc; if (!child) pr_debug("ocd_disable (no child)\n"); else if (test_tsk_thread_flag(child, TIF_DEBUG)) pr_debug("ocd_disable: child=%s [%u]\n", child->comm, child->pid); if (!child || test_and_clear_tsk_thread_flag(child, TIF_DEBUG)) { spin_lock(&ocd_lock); ocd_count--; WARN_ON(ocd_count < 0); if (ocd_count <= 0) { dc = ocd_read(DC); dc &= ~((1 << OCD_DC_MM_BIT) | (1 << OCD_DC_DBE_BIT)); ocd_write(DC, dc); } spin_unlock(&ocd_lock); } }
int sve_set_vector_length(struct task_struct *task, unsigned long vl, unsigned long flags) { if (flags & ~(unsigned long)(PR_SVE_VL_INHERIT | PR_SVE_SET_VL_ONEXEC)) return -EINVAL; if (!sve_vl_valid(vl)) return -EINVAL; /* * Clamp to the maximum vector length that VL-agnostic SVE code can * work with. A flag may be assigned in the future to allow setting * of larger vector lengths without confusing older software. */ if (vl > SVE_VL_ARCH_MAX) vl = SVE_VL_ARCH_MAX; vl = find_supported_vector_length(vl); if (flags & (PR_SVE_VL_INHERIT | PR_SVE_SET_VL_ONEXEC)) task->thread.sve_vl_onexec = vl; else /* Reset VL to system default on next exec: */ task->thread.sve_vl_onexec = 0; /* Only actually set the VL if not deferred: */ if (flags & PR_SVE_SET_VL_ONEXEC) goto out; if (vl == task->thread.sve_vl) goto out; /* * To ensure the FPSIMD bits of the SVE vector registers are preserved, * write any live register state back to task_struct, and convert to a * non-SVE thread. */ if (task == current) { local_bh_disable(); fpsimd_save(); set_thread_flag(TIF_FOREIGN_FPSTATE); } fpsimd_flush_task_state(task); if (test_and_clear_tsk_thread_flag(task, TIF_SVE)) sve_to_fpsimd(task); if (task == current) local_bh_enable(); /* * Force reallocation of task SVE state to the correct size * on next use: */ sve_free(task); task->thread.sve_vl = vl; out: update_tsk_thread_flag(task, TIF_SVE_VL_INHERIT, flags & PR_SVE_VL_INHERIT); return 0; }