// SPDX-License-Identifier: GPL-2.0 #include #include #include #include #include #include #include #define MODULE_NAME "nomiss" #include #include #include #include #include #define RV_MON_TYPE RV_MON_PER_OBJ #define HA_TIMER_TYPE HA_TIMER_WHEEL /* The start condition is on sched_switch, it's dangerous to allocate there */ #define DA_SKIP_AUTO_ALLOC typedef struct sched_dl_entity *monitor_target; #include "nomiss.h" #include #include /* * User configurable deadline threshold. If the total utilisation of deadline * tasks is larger than 1, they are only guaranteed bounded tardiness. See * Documentation/scheduler/sched-deadline.rst for more details. * The minimum tardiness without sched_feat(HRTICK_DL) is 1 tick to accommodate * for throttle enforced on the next tick. */ static u64 deadline_thresh = TICK_NSEC; module_param(deadline_thresh, ullong, 0644); #define DEADLINE_NS(ha_mon) (ha_get_target(ha_mon)->dl_deadline + deadline_thresh) static u64 ha_get_env(struct ha_monitor *ha_mon, enum envs_nomiss env, u64 time_ns) { if (env == clk_nomiss) return ha_get_clk_ns(ha_mon, env, time_ns); else if (env == is_constr_dl_nomiss) return !dl_is_implicit(ha_get_target(ha_mon)); else if (env == is_defer_nomiss) return ha_get_target(ha_mon)->dl_defer; return ENV_INVALID_VALUE; } static void ha_reset_env(struct ha_monitor *ha_mon, enum envs_nomiss env, u64 time_ns) { if (env == clk_nomiss) ha_reset_clk_ns(ha_mon, env, time_ns); } static inline bool ha_verify_invariants(struct ha_monitor *ha_mon, enum states curr_state, enum events event, enum states next_state, u64 time_ns) { if (curr_state == ready_nomiss) return ha_check_invariant_ns(ha_mon, clk_nomiss, time_ns); else if (curr_state == running_nomiss) return ha_check_invariant_ns(ha_mon, clk_nomiss, time_ns); return true; } static inline void ha_convert_inv_guard(struct ha_monitor *ha_mon, enum states curr_state, enum events event, enum states next_state, u64 time_ns) { if (curr_state == next_state) return; if (curr_state == ready_nomiss) ha_inv_to_guard(ha_mon, clk_nomiss, DEADLINE_NS(ha_mon), time_ns); else if (curr_state == running_nomiss) ha_inv_to_guard(ha_mon, clk_nomiss, DEADLINE_NS(ha_mon), time_ns); } static inline bool ha_verify_guards(struct ha_monitor *ha_mon, enum states curr_state, enum events event, enum states next_state, u64 time_ns) { bool res = true; if (curr_state == ready_nomiss && event == dl_replenish_nomiss) ha_reset_env(ha_mon, clk_nomiss, time_ns); else if (curr_state == ready_nomiss && event == dl_throttle_nomiss) res = ha_get_env(ha_mon, is_defer_nomiss, time_ns) == 1ull; else if (curr_state == idle_nomiss && event == dl_replenish_nomiss) ha_reset_env(ha_mon, clk_nomiss, time_ns); else if (curr_state == running_nomiss && event == dl_replenish_nomiss) ha_reset_env(ha_mon, clk_nomiss, time_ns); else if (curr_state == sleeping_nomiss && event == dl_replenish_nomiss) ha_reset_env(ha_mon, clk_nomiss, time_ns); else if (curr_state == sleeping_nomiss && event == dl_throttle_nomiss) res = ha_get_env(ha_mon, is_constr_dl_nomiss, time_ns) == 1ull || ha_get_env(ha_mon, is_defer_nomiss, time_ns) == 1ull; else if (curr_state == throttled_nomiss && event == dl_replenish_nomiss) ha_reset_env(ha_mon, clk_nomiss, time_ns); return res; } static inline void ha_setup_invariants(struct ha_monitor *ha_mon, enum states curr_state, enum events event, enum states next_state, u64 time_ns) { if (next_state == curr_state && event != dl_replenish_nomiss) return; if (next_state == ready_nomiss) ha_start_timer_ns(ha_mon, clk_nomiss, DEADLINE_NS(ha_mon), time_ns); else if (next_state == running_nomiss) ha_start_timer_ns(ha_mon, clk_nomiss, DEADLINE_NS(ha_mon), time_ns); else if (curr_state == ready_nomiss) ha_cancel_timer(ha_mon); else if (curr_state == running_nomiss) ha_cancel_timer(ha_mon); } static bool ha_verify_constraint(struct ha_monitor *ha_mon, enum states curr_state, enum events event, enum states next_state, u64 time_ns) { if (!ha_verify_invariants(ha_mon, curr_state, event, next_state, time_ns)) return false; ha_convert_inv_guard(ha_mon, curr_state, event, next_state, time_ns); if (!ha_verify_guards(ha_mon, curr_state, event, next_state, time_ns)) return false; ha_setup_invariants(ha_mon, curr_state, event, next_state, time_ns); return true; } static void handle_dl_replenish(void *data, struct sched_dl_entity *dl_se, int cpu, u8 type) { if (is_supported_type(type)) da_handle_event(EXPAND_ID(dl_se, cpu, type), dl_replenish_nomiss); } static void handle_dl_throttle(void *data, struct sched_dl_entity *dl_se, int cpu, u8 type) { if (is_supported_type(type)) da_handle_event(EXPAND_ID(dl_se, cpu, type), dl_throttle_nomiss); } static void handle_dl_server_stop(void *data, struct sched_dl_entity *dl_se, int cpu, u8 type) { /* * This isn't the standard use of da_handle_start_run_event since this * event cannot only occur from the initial state. * It is fine to use here because it always brings to a known state and * the fact we "pretend" the transition starts from the initial state * has no side effect. */ if (is_supported_type(type)) da_handle_start_run_event(EXPAND_ID(dl_se, cpu, type), dl_server_stop_nomiss); } static inline void handle_server_switch(struct task_struct *next, int cpu, u8 type) { struct sched_dl_entity *dl_se = get_server(next, type); if (dl_se && is_idle_task(next)) da_handle_event(EXPAND_ID(dl_se, cpu, type), dl_server_idle_nomiss); } static void handle_sched_switch(void *data, bool preempt, struct task_struct *prev, struct task_struct *next, unsigned int prev_state) { int cpu = task_cpu(next); if (prev_state != TASK_RUNNING && !preempt && prev->policy == SCHED_DEADLINE) da_handle_event(EXPAND_ID_TASK(prev), sched_switch_suspend_nomiss); if (next->policy == SCHED_DEADLINE) da_handle_start_run_event(EXPAND_ID_TASK(next), sched_switch_in_nomiss); /* * The server is available in next only if the next task is boosted, * otherwise we need to retrieve it. * Here the server continues in the state running/armed until actually * stopped, this works since we continue expecting a throttle. */ if (next->dl_server) da_handle_start_event(EXPAND_ID(next->dl_server, cpu, get_server_type(next)), sched_switch_in_nomiss); else { handle_server_switch(next, cpu, DL_SERVER_FAIR); if (IS_ENABLED(CONFIG_SCHED_CLASS_EXT)) handle_server_switch(next, cpu, DL_SERVER_EXT); } } static void handle_sys_enter(void *data, struct pt_regs *regs, long id) { struct task_struct *p; int new_policy = -1; pid_t pid = 0; new_policy = extract_params(regs, id, &pid); if (new_policy < 0) return; guard(rcu)(); p = pid ? find_task_by_vpid(pid) : current; if (unlikely(!p) || new_policy == p->policy) return; if (p->policy == SCHED_DEADLINE) da_reset(EXPAND_ID_TASK(p)); else if (new_policy == SCHED_DEADLINE) da_create_or_get(EXPAND_ID_TASK(p)); } static void handle_sched_wakeup(void *data, struct task_struct *tsk) { if (tsk->policy == SCHED_DEADLINE) da_handle_event(EXPAND_ID_TASK(tsk), sched_wakeup_nomiss); } static int enable_nomiss(void) { int retval; retval = da_monitor_init(); if (retval) return retval; retval = init_storage(false); if (retval) return retval; rv_attach_trace_probe("nomiss", sched_dl_replenish_tp, handle_dl_replenish); rv_attach_trace_probe("nomiss", sched_dl_throttle_tp, handle_dl_throttle); rv_attach_trace_probe("nomiss", sched_dl_server_stop_tp, handle_dl_server_stop); rv_attach_trace_probe("nomiss", sched_switch, handle_sched_switch); rv_attach_trace_probe("nomiss", sched_wakeup, handle_sched_wakeup); if (!should_skip_syscall_handle()) rv_attach_trace_probe("nomiss", sys_enter, handle_sys_enter); rv_attach_trace_probe("nomiss", task_newtask, handle_newtask); rv_attach_trace_probe("nomiss", sched_process_exit, handle_exit); return 0; } static void disable_nomiss(void) { rv_this.enabled = 0; /* Those are RCU writers, detach earlier hoping to close a bit faster */ rv_detach_trace_probe("nomiss", task_newtask, handle_newtask); rv_detach_trace_probe("nomiss", sched_process_exit, handle_exit); if (!should_skip_syscall_handle()) rv_detach_trace_probe("nomiss", sys_enter, handle_sys_enter); rv_detach_trace_probe("nomiss", sched_dl_replenish_tp, handle_dl_replenish); rv_detach_trace_probe("nomiss", sched_dl_throttle_tp, handle_dl_throttle); rv_detach_trace_probe("nomiss", sched_dl_server_stop_tp, handle_dl_server_stop); rv_detach_trace_probe("nomiss", sched_switch, handle_sched_switch); rv_detach_trace_probe("nomiss", sched_wakeup, handle_sched_wakeup); da_monitor_destroy(); } static struct rv_monitor rv_this = { .name = "nomiss", .description = "dl entities run to completion before their deadline.", .enable = enable_nomiss, .disable = disable_nomiss, .reset = da_monitor_reset_all, .enabled = 0, }; static int __init register_nomiss(void) { return rv_register_monitor(&rv_this, &rv_deadline); } static void __exit unregister_nomiss(void) { rv_unregister_monitor(&rv_this); } module_init(register_nomiss); module_exit(unregister_nomiss); MODULE_LICENSE("GPL"); MODULE_AUTHOR("Gabriele Monaco "); MODULE_DESCRIPTION("nomiss: dl entities run to completion before their deadline.");