1079 lines
34 KiB
C
1079 lines
34 KiB
C
/* SPDX-License-Identifier: GPL-2.0 */
|
|
/*
|
|
* BPF extensible scheduler class: Documentation/scheduler/sched-ext.rst
|
|
*
|
|
* Copyright (c) 2025 Meta Platforms, Inc. and affiliates.
|
|
* Copyright (c) 2025 Tejun Heo <tj@kernel.org>
|
|
*/
|
|
#define SCX_OP_IDX(op) (offsetof(struct sched_ext_ops, op) / sizeof(void (*)(void)))
|
|
|
|
enum scx_consts {
|
|
SCX_DSP_DFL_MAX_BATCH = 32,
|
|
SCX_DSP_MAX_LOOPS = 32,
|
|
SCX_WATCHDOG_MAX_TIMEOUT = 30 * HZ,
|
|
|
|
SCX_EXIT_BT_LEN = 64,
|
|
SCX_EXIT_MSG_LEN = 1024,
|
|
SCX_EXIT_DUMP_DFL_LEN = 32768,
|
|
|
|
SCX_CPUPERF_ONE = SCHED_CAPACITY_SCALE,
|
|
|
|
/*
|
|
* Iterating all tasks may take a while. Periodically drop
|
|
* scx_tasks_lock to avoid causing e.g. CSD and RCU stalls.
|
|
*/
|
|
SCX_TASK_ITER_BATCH = 32,
|
|
};
|
|
|
|
enum scx_exit_kind {
|
|
SCX_EXIT_NONE,
|
|
SCX_EXIT_DONE,
|
|
|
|
SCX_EXIT_UNREG = 64, /* user-space initiated unregistration */
|
|
SCX_EXIT_UNREG_BPF, /* BPF-initiated unregistration */
|
|
SCX_EXIT_UNREG_KERN, /* kernel-initiated unregistration */
|
|
SCX_EXIT_SYSRQ, /* requested by 'S' sysrq */
|
|
|
|
SCX_EXIT_ERROR = 1024, /* runtime error, error msg contains details */
|
|
SCX_EXIT_ERROR_BPF, /* ERROR but triggered through scx_bpf_error() */
|
|
SCX_EXIT_ERROR_STALL, /* watchdog detected stalled runnable tasks */
|
|
};
|
|
|
|
/*
|
|
* An exit code can be specified when exiting with scx_bpf_exit() or scx_exit(),
|
|
* corresponding to exit_kind UNREG_BPF and UNREG_KERN respectively. The codes
|
|
* are 64bit of the format:
|
|
*
|
|
* Bits: [63 .. 48 47 .. 32 31 .. 0]
|
|
* [ SYS ACT ] [ SYS RSN ] [ USR ]
|
|
*
|
|
* SYS ACT: System-defined exit actions
|
|
* SYS RSN: System-defined exit reasons
|
|
* USR : User-defined exit codes and reasons
|
|
*
|
|
* Using the above, users may communicate intention and context by ORing system
|
|
* actions and/or system reasons with a user-defined exit code.
|
|
*/
|
|
enum scx_exit_code {
|
|
/* Reasons */
|
|
SCX_ECODE_RSN_HOTPLUG = 1LLU << 32,
|
|
|
|
/* Actions */
|
|
SCX_ECODE_ACT_RESTART = 1LLU << 48,
|
|
};
|
|
|
|
enum scx_exit_flags {
|
|
/*
|
|
* ops.exit() may be called even if the loading failed before ops.init()
|
|
* finishes successfully. This is because ops.exit() allows rich exit
|
|
* info communication. The following flag indicates whether ops.init()
|
|
* finished successfully.
|
|
*/
|
|
SCX_EFLAG_INITIALIZED,
|
|
};
|
|
|
|
/*
|
|
* scx_exit_info is passed to ops.exit() to describe why the BPF scheduler is
|
|
* being disabled.
|
|
*/
|
|
struct scx_exit_info {
|
|
/* %SCX_EXIT_* - broad category of the exit reason */
|
|
enum scx_exit_kind kind;
|
|
|
|
/* exit code if gracefully exiting */
|
|
s64 exit_code;
|
|
|
|
/* %SCX_EFLAG_* */
|
|
u64 flags;
|
|
|
|
/* textual representation of the above */
|
|
const char *reason;
|
|
|
|
/* backtrace if exiting due to an error */
|
|
unsigned long *bt;
|
|
u32 bt_len;
|
|
|
|
/* informational message */
|
|
char *msg;
|
|
|
|
/* debug dump */
|
|
char *dump;
|
|
};
|
|
|
|
/* sched_ext_ops.flags */
|
|
enum scx_ops_flags {
|
|
/*
|
|
* Keep built-in idle tracking even if ops.update_idle() is implemented.
|
|
*/
|
|
SCX_OPS_KEEP_BUILTIN_IDLE = 1LLU << 0,
|
|
|
|
/*
|
|
* By default, if there are no other task to run on the CPU, ext core
|
|
* keeps running the current task even after its slice expires. If this
|
|
* flag is specified, such tasks are passed to ops.enqueue() with
|
|
* %SCX_ENQ_LAST. See the comment above %SCX_ENQ_LAST for more info.
|
|
*/
|
|
SCX_OPS_ENQ_LAST = 1LLU << 1,
|
|
|
|
/*
|
|
* An exiting task may schedule after PF_EXITING is set. In such cases,
|
|
* bpf_task_from_pid() may not be able to find the task and if the BPF
|
|
* scheduler depends on pid lookup for dispatching, the task will be
|
|
* lost leading to various issues including RCU grace period stalls.
|
|
*
|
|
* To mask this problem, by default, unhashed tasks are automatically
|
|
* dispatched to the local DSQ on enqueue. If the BPF scheduler doesn't
|
|
* depend on pid lookups and wants to handle these tasks directly, the
|
|
* following flag can be used.
|
|
*/
|
|
SCX_OPS_ENQ_EXITING = 1LLU << 2,
|
|
|
|
/*
|
|
* If set, only tasks with policy set to SCHED_EXT are attached to
|
|
* sched_ext. If clear, SCHED_NORMAL tasks are also included.
|
|
*/
|
|
SCX_OPS_SWITCH_PARTIAL = 1LLU << 3,
|
|
|
|
/*
|
|
* A migration disabled task can only execute on its current CPU. By
|
|
* default, such tasks are automatically put on the CPU's local DSQ with
|
|
* the default slice on enqueue. If this ops flag is set, they also go
|
|
* through ops.enqueue().
|
|
*
|
|
* A migration disabled task never invokes ops.select_cpu() as it can
|
|
* only select the current CPU. Also, p->cpus_ptr will only contain its
|
|
* current CPU while p->nr_cpus_allowed keeps tracking p->user_cpus_ptr
|
|
* and thus may disagree with cpumask_weight(p->cpus_ptr).
|
|
*/
|
|
SCX_OPS_ENQ_MIGRATION_DISABLED = 1LLU << 4,
|
|
|
|
/*
|
|
* Queued wakeup (ttwu_queue) is a wakeup optimization that invokes
|
|
* ops.enqueue() on the ops.select_cpu() selected or the wakee's
|
|
* previous CPU via IPI (inter-processor interrupt) to reduce cacheline
|
|
* transfers. When this optimization is enabled, ops.select_cpu() is
|
|
* skipped in some cases (when racing against the wakee switching out).
|
|
* As the BPF scheduler may depend on ops.select_cpu() being invoked
|
|
* during wakeups, queued wakeup is disabled by default.
|
|
*
|
|
* If this ops flag is set, queued wakeup optimization is enabled and
|
|
* the BPF scheduler must be able to handle ops.enqueue() invoked on the
|
|
* wakee's CPU without preceding ops.select_cpu() even for tasks which
|
|
* may be executed on multiple CPUs.
|
|
*/
|
|
SCX_OPS_ALLOW_QUEUED_WAKEUP = 1LLU << 5,
|
|
|
|
/*
|
|
* If set, enable per-node idle cpumasks. If clear, use a single global
|
|
* flat idle cpumask.
|
|
*/
|
|
SCX_OPS_BUILTIN_IDLE_PER_NODE = 1LLU << 6,
|
|
|
|
/*
|
|
* CPU cgroup support flags
|
|
*/
|
|
SCX_OPS_HAS_CGROUP_WEIGHT = 1LLU << 16, /* DEPRECATED, will be removed on 6.18 */
|
|
|
|
SCX_OPS_ALL_FLAGS = SCX_OPS_KEEP_BUILTIN_IDLE |
|
|
SCX_OPS_ENQ_LAST |
|
|
SCX_OPS_ENQ_EXITING |
|
|
SCX_OPS_ENQ_MIGRATION_DISABLED |
|
|
SCX_OPS_ALLOW_QUEUED_WAKEUP |
|
|
SCX_OPS_SWITCH_PARTIAL |
|
|
SCX_OPS_BUILTIN_IDLE_PER_NODE |
|
|
SCX_OPS_HAS_CGROUP_WEIGHT,
|
|
|
|
/* high 8 bits are internal, don't include in SCX_OPS_ALL_FLAGS */
|
|
__SCX_OPS_INTERNAL_MASK = 0xffLLU << 56,
|
|
|
|
SCX_OPS_HAS_CPU_PREEMPT = 1LLU << 56,
|
|
};
|
|
|
|
/* argument container for ops.init_task() */
|
|
struct scx_init_task_args {
|
|
/*
|
|
* Set if ops.init_task() is being invoked on the fork path, as opposed
|
|
* to the scheduler transition path.
|
|
*/
|
|
bool fork;
|
|
#ifdef CONFIG_EXT_GROUP_SCHED
|
|
/* the cgroup the task is joining */
|
|
struct cgroup *cgroup;
|
|
#endif
|
|
};
|
|
|
|
/* argument container for ops.exit_task() */
|
|
struct scx_exit_task_args {
|
|
/* Whether the task exited before running on sched_ext. */
|
|
bool cancelled;
|
|
};
|
|
|
|
/* argument container for ops->cgroup_init() */
|
|
struct scx_cgroup_init_args {
|
|
/* the weight of the cgroup [1..10000] */
|
|
u32 weight;
|
|
|
|
/* bandwidth control parameters from cpu.max and cpu.max.burst */
|
|
u64 bw_period_us;
|
|
u64 bw_quota_us;
|
|
u64 bw_burst_us;
|
|
};
|
|
|
|
enum scx_cpu_preempt_reason {
|
|
/* next task is being scheduled by &sched_class_rt */
|
|
SCX_CPU_PREEMPT_RT,
|
|
/* next task is being scheduled by &sched_class_dl */
|
|
SCX_CPU_PREEMPT_DL,
|
|
/* next task is being scheduled by &sched_class_stop */
|
|
SCX_CPU_PREEMPT_STOP,
|
|
/* unknown reason for SCX being preempted */
|
|
SCX_CPU_PREEMPT_UNKNOWN,
|
|
};
|
|
|
|
/*
|
|
* Argument container for ops->cpu_acquire(). Currently empty, but may be
|
|
* expanded in the future.
|
|
*/
|
|
struct scx_cpu_acquire_args {};
|
|
|
|
/* argument container for ops->cpu_release() */
|
|
struct scx_cpu_release_args {
|
|
/* the reason the CPU was preempted */
|
|
enum scx_cpu_preempt_reason reason;
|
|
|
|
/* the task that's going to be scheduled on the CPU */
|
|
struct task_struct *task;
|
|
};
|
|
|
|
/*
|
|
* Informational context provided to dump operations.
|
|
*/
|
|
struct scx_dump_ctx {
|
|
enum scx_exit_kind kind;
|
|
s64 exit_code;
|
|
const char *reason;
|
|
u64 at_ns;
|
|
u64 at_jiffies;
|
|
};
|
|
|
|
/**
|
|
* struct sched_ext_ops - Operation table for BPF scheduler implementation
|
|
*
|
|
* A BPF scheduler can implement an arbitrary scheduling policy by
|
|
* implementing and loading operations in this table. Note that a userland
|
|
* scheduling policy can also be implemented using the BPF scheduler
|
|
* as a shim layer.
|
|
*/
|
|
struct sched_ext_ops {
|
|
/**
|
|
* @select_cpu: Pick the target CPU for a task which is being woken up
|
|
* @p: task being woken up
|
|
* @prev_cpu: the cpu @p was on before sleeping
|
|
* @wake_flags: SCX_WAKE_*
|
|
*
|
|
* Decision made here isn't final. @p may be moved to any CPU while it
|
|
* is getting dispatched for execution later. However, as @p is not on
|
|
* the rq at this point, getting the eventual execution CPU right here
|
|
* saves a small bit of overhead down the line.
|
|
*
|
|
* If an idle CPU is returned, the CPU is kicked and will try to
|
|
* dispatch. While an explicit custom mechanism can be added,
|
|
* select_cpu() serves as the default way to wake up idle CPUs.
|
|
*
|
|
* @p may be inserted into a DSQ directly by calling
|
|
* scx_bpf_dsq_insert(). If so, the ops.enqueue() will be skipped.
|
|
* Directly inserting into %SCX_DSQ_LOCAL will put @p in the local DSQ
|
|
* of the CPU returned by this operation.
|
|
*
|
|
* Note that select_cpu() is never called for tasks that can only run
|
|
* on a single CPU or tasks with migration disabled, as they don't have
|
|
* the option to select a different CPU. See select_task_rq() for
|
|
* details.
|
|
*/
|
|
s32 (*select_cpu)(struct task_struct *p, s32 prev_cpu, u64 wake_flags);
|
|
|
|
/**
|
|
* @enqueue: Enqueue a task on the BPF scheduler
|
|
* @p: task being enqueued
|
|
* @enq_flags: %SCX_ENQ_*
|
|
*
|
|
* @p is ready to run. Insert directly into a DSQ by calling
|
|
* scx_bpf_dsq_insert() or enqueue on the BPF scheduler. If not directly
|
|
* inserted, the bpf scheduler owns @p and if it fails to dispatch @p,
|
|
* the task will stall.
|
|
*
|
|
* If @p was inserted into a DSQ from ops.select_cpu(), this callback is
|
|
* skipped.
|
|
*/
|
|
void (*enqueue)(struct task_struct *p, u64 enq_flags);
|
|
|
|
/**
|
|
* @dequeue: Remove a task from the BPF scheduler
|
|
* @p: task being dequeued
|
|
* @deq_flags: %SCX_DEQ_*
|
|
*
|
|
* Remove @p from the BPF scheduler. This is usually called to isolate
|
|
* the task while updating its scheduling properties (e.g. priority).
|
|
*
|
|
* The ext core keeps track of whether the BPF side owns a given task or
|
|
* not and can gracefully ignore spurious dispatches from BPF side,
|
|
* which makes it safe to not implement this method. However, depending
|
|
* on the scheduling logic, this can lead to confusing behaviors - e.g.
|
|
* scheduling position not being updated across a priority change.
|
|
*/
|
|
void (*dequeue)(struct task_struct *p, u64 deq_flags);
|
|
|
|
/**
|
|
* @dispatch: Dispatch tasks from the BPF scheduler and/or user DSQs
|
|
* @cpu: CPU to dispatch tasks for
|
|
* @prev: previous task being switched out
|
|
*
|
|
* Called when a CPU's local dsq is empty. The operation should dispatch
|
|
* one or more tasks from the BPF scheduler into the DSQs using
|
|
* scx_bpf_dsq_insert() and/or move from user DSQs into the local DSQ
|
|
* using scx_bpf_dsq_move_to_local().
|
|
*
|
|
* The maximum number of times scx_bpf_dsq_insert() can be called
|
|
* without an intervening scx_bpf_dsq_move_to_local() is specified by
|
|
* ops.dispatch_max_batch. See the comments on top of the two functions
|
|
* for more details.
|
|
*
|
|
* When not %NULL, @prev is an SCX task with its slice depleted. If
|
|
* @prev is still runnable as indicated by set %SCX_TASK_QUEUED in
|
|
* @prev->scx.flags, it is not enqueued yet and will be enqueued after
|
|
* ops.dispatch() returns. To keep executing @prev, return without
|
|
* dispatching or moving any tasks. Also see %SCX_OPS_ENQ_LAST.
|
|
*/
|
|
void (*dispatch)(s32 cpu, struct task_struct *prev);
|
|
|
|
/**
|
|
* @tick: Periodic tick
|
|
* @p: task running currently
|
|
*
|
|
* This operation is called every 1/HZ seconds on CPUs which are
|
|
* executing an SCX task. Setting @p->scx.slice to 0 will trigger an
|
|
* immediate dispatch cycle on the CPU.
|
|
*/
|
|
void (*tick)(struct task_struct *p);
|
|
|
|
/**
|
|
* @runnable: A task is becoming runnable on its associated CPU
|
|
* @p: task becoming runnable
|
|
* @enq_flags: %SCX_ENQ_*
|
|
*
|
|
* This and the following three functions can be used to track a task's
|
|
* execution state transitions. A task becomes ->runnable() on a CPU,
|
|
* and then goes through one or more ->running() and ->stopping() pairs
|
|
* as it runs on the CPU, and eventually becomes ->quiescent() when it's
|
|
* done running on the CPU.
|
|
*
|
|
* @p is becoming runnable on the CPU because it's
|
|
*
|
|
* - waking up (%SCX_ENQ_WAKEUP)
|
|
* - being moved from another CPU
|
|
* - being restored after temporarily taken off the queue for an
|
|
* attribute change.
|
|
*
|
|
* This and ->enqueue() are related but not coupled. This operation
|
|
* notifies @p's state transition and may not be followed by ->enqueue()
|
|
* e.g. when @p is being dispatched to a remote CPU, or when @p is
|
|
* being enqueued on a CPU experiencing a hotplug event. Likewise, a
|
|
* task may be ->enqueue()'d without being preceded by this operation
|
|
* e.g. after exhausting its slice.
|
|
*/
|
|
void (*runnable)(struct task_struct *p, u64 enq_flags);
|
|
|
|
/**
|
|
* @running: A task is starting to run on its associated CPU
|
|
* @p: task starting to run
|
|
*
|
|
* Note that this callback may be called from a CPU other than the
|
|
* one the task is going to run on. This can happen when a task
|
|
* property is changed (i.e., affinity), since scx_next_task_scx(),
|
|
* which triggers this callback, may run on a CPU different from
|
|
* the task's assigned CPU.
|
|
*
|
|
* Therefore, always use scx_bpf_task_cpu(@p) to determine the
|
|
* target CPU the task is going to use.
|
|
*
|
|
* See ->runnable() for explanation on the task state notifiers.
|
|
*/
|
|
void (*running)(struct task_struct *p);
|
|
|
|
/**
|
|
* @stopping: A task is stopping execution
|
|
* @p: task stopping to run
|
|
* @runnable: is task @p still runnable?
|
|
*
|
|
* Note that this callback may be called from a CPU other than the
|
|
* one the task was running on. This can happen when a task
|
|
* property is changed (i.e., affinity), since dequeue_task_scx(),
|
|
* which triggers this callback, may run on a CPU different from
|
|
* the task's assigned CPU.
|
|
*
|
|
* Therefore, always use scx_bpf_task_cpu(@p) to retrieve the CPU
|
|
* the task was running on.
|
|
*
|
|
* See ->runnable() for explanation on the task state notifiers. If
|
|
* !@runnable, ->quiescent() will be invoked after this operation
|
|
* returns.
|
|
*/
|
|
void (*stopping)(struct task_struct *p, bool runnable);
|
|
|
|
/**
|
|
* @quiescent: A task is becoming not runnable on its associated CPU
|
|
* @p: task becoming not runnable
|
|
* @deq_flags: %SCX_DEQ_*
|
|
*
|
|
* See ->runnable() for explanation on the task state notifiers.
|
|
*
|
|
* @p is becoming quiescent on the CPU because it's
|
|
*
|
|
* - sleeping (%SCX_DEQ_SLEEP)
|
|
* - being moved to another CPU
|
|
* - being temporarily taken off the queue for an attribute change
|
|
* (%SCX_DEQ_SAVE)
|
|
*
|
|
* This and ->dequeue() are related but not coupled. This operation
|
|
* notifies @p's state transition and may not be preceded by ->dequeue()
|
|
* e.g. when @p is being dispatched to a remote CPU.
|
|
*/
|
|
void (*quiescent)(struct task_struct *p, u64 deq_flags);
|
|
|
|
/**
|
|
* @yield: Yield CPU
|
|
* @from: yielding task
|
|
* @to: optional yield target task
|
|
*
|
|
* If @to is NULL, @from is yielding the CPU to other runnable tasks.
|
|
* The BPF scheduler should ensure that other available tasks are
|
|
* dispatched before the yielding task. Return value is ignored in this
|
|
* case.
|
|
*
|
|
* If @to is not-NULL, @from wants to yield the CPU to @to. If the bpf
|
|
* scheduler can implement the request, return %true; otherwise, %false.
|
|
*/
|
|
bool (*yield)(struct task_struct *from, struct task_struct *to);
|
|
|
|
/**
|
|
* @core_sched_before: Task ordering for core-sched
|
|
* @a: task A
|
|
* @b: task B
|
|
*
|
|
* Used by core-sched to determine the ordering between two tasks. See
|
|
* Documentation/admin-guide/hw-vuln/core-scheduling.rst for details on
|
|
* core-sched.
|
|
*
|
|
* Both @a and @b are runnable and may or may not currently be queued on
|
|
* the BPF scheduler. Should return %true if @a should run before @b.
|
|
* %false if there's no required ordering or @b should run before @a.
|
|
*
|
|
* If not specified, the default is ordering them according to when they
|
|
* became runnable.
|
|
*/
|
|
bool (*core_sched_before)(struct task_struct *a, struct task_struct *b);
|
|
|
|
/**
|
|
* @set_weight: Set task weight
|
|
* @p: task to set weight for
|
|
* @weight: new weight [1..10000]
|
|
*
|
|
* Update @p's weight to @weight.
|
|
*/
|
|
void (*set_weight)(struct task_struct *p, u32 weight);
|
|
|
|
/**
|
|
* @set_cpumask: Set CPU affinity
|
|
* @p: task to set CPU affinity for
|
|
* @cpumask: cpumask of cpus that @p can run on
|
|
*
|
|
* Update @p's CPU affinity to @cpumask.
|
|
*/
|
|
void (*set_cpumask)(struct task_struct *p,
|
|
const struct cpumask *cpumask);
|
|
|
|
/**
|
|
* @update_idle: Update the idle state of a CPU
|
|
* @cpu: CPU to update the idle state for
|
|
* @idle: whether entering or exiting the idle state
|
|
*
|
|
* This operation is called when @rq's CPU goes or leaves the idle
|
|
* state. By default, implementing this operation disables the built-in
|
|
* idle CPU tracking and the following helpers become unavailable:
|
|
*
|
|
* - scx_bpf_select_cpu_dfl()
|
|
* - scx_bpf_select_cpu_and()
|
|
* - scx_bpf_test_and_clear_cpu_idle()
|
|
* - scx_bpf_pick_idle_cpu()
|
|
*
|
|
* The user also must implement ops.select_cpu() as the default
|
|
* implementation relies on scx_bpf_select_cpu_dfl().
|
|
*
|
|
* Specify the %SCX_OPS_KEEP_BUILTIN_IDLE flag to keep the built-in idle
|
|
* tracking.
|
|
*/
|
|
void (*update_idle)(s32 cpu, bool idle);
|
|
|
|
/**
|
|
* @cpu_acquire: A CPU is becoming available to the BPF scheduler
|
|
* @cpu: The CPU being acquired by the BPF scheduler.
|
|
* @args: Acquire arguments, see the struct definition.
|
|
*
|
|
* A CPU that was previously released from the BPF scheduler is now once
|
|
* again under its control.
|
|
*/
|
|
void (*cpu_acquire)(s32 cpu, struct scx_cpu_acquire_args *args);
|
|
|
|
/**
|
|
* @cpu_release: A CPU is taken away from the BPF scheduler
|
|
* @cpu: The CPU being released by the BPF scheduler.
|
|
* @args: Release arguments, see the struct definition.
|
|
*
|
|
* The specified CPU is no longer under the control of the BPF
|
|
* scheduler. This could be because it was preempted by a higher
|
|
* priority sched_class, though there may be other reasons as well. The
|
|
* caller should consult @args->reason to determine the cause.
|
|
*/
|
|
void (*cpu_release)(s32 cpu, struct scx_cpu_release_args *args);
|
|
|
|
/**
|
|
* @init_task: Initialize a task to run in a BPF scheduler
|
|
* @p: task to initialize for BPF scheduling
|
|
* @args: init arguments, see the struct definition
|
|
*
|
|
* Either we're loading a BPF scheduler or a new task is being forked.
|
|
* Initialize @p for BPF scheduling. This operation may block and can
|
|
* be used for allocations, and is called exactly once for a task.
|
|
*
|
|
* Return 0 for success, -errno for failure. An error return while
|
|
* loading will abort loading of the BPF scheduler. During a fork, it
|
|
* will abort that specific fork.
|
|
*/
|
|
s32 (*init_task)(struct task_struct *p, struct scx_init_task_args *args);
|
|
|
|
/**
|
|
* @exit_task: Exit a previously-running task from the system
|
|
* @p: task to exit
|
|
* @args: exit arguments, see the struct definition
|
|
*
|
|
* @p is exiting or the BPF scheduler is being unloaded. Perform any
|
|
* necessary cleanup for @p.
|
|
*/
|
|
void (*exit_task)(struct task_struct *p, struct scx_exit_task_args *args);
|
|
|
|
/**
|
|
* @enable: Enable BPF scheduling for a task
|
|
* @p: task to enable BPF scheduling for
|
|
*
|
|
* Enable @p for BPF scheduling. enable() is called on @p any time it
|
|
* enters SCX, and is always paired with a matching disable().
|
|
*/
|
|
void (*enable)(struct task_struct *p);
|
|
|
|
/**
|
|
* @disable: Disable BPF scheduling for a task
|
|
* @p: task to disable BPF scheduling for
|
|
*
|
|
* @p is exiting, leaving SCX or the BPF scheduler is being unloaded.
|
|
* Disable BPF scheduling for @p. A disable() call is always matched
|
|
* with a prior enable() call.
|
|
*/
|
|
void (*disable)(struct task_struct *p);
|
|
|
|
/**
|
|
* @dump: Dump BPF scheduler state on error
|
|
* @ctx: debug dump context
|
|
*
|
|
* Use scx_bpf_dump() to generate BPF scheduler specific debug dump.
|
|
*/
|
|
void (*dump)(struct scx_dump_ctx *ctx);
|
|
|
|
/**
|
|
* @dump_cpu: Dump BPF scheduler state for a CPU on error
|
|
* @ctx: debug dump context
|
|
* @cpu: CPU to generate debug dump for
|
|
* @idle: @cpu is currently idle without any runnable tasks
|
|
*
|
|
* Use scx_bpf_dump() to generate BPF scheduler specific debug dump for
|
|
* @cpu. If @idle is %true and this operation doesn't produce any
|
|
* output, @cpu is skipped for dump.
|
|
*/
|
|
void (*dump_cpu)(struct scx_dump_ctx *ctx, s32 cpu, bool idle);
|
|
|
|
/**
|
|
* @dump_task: Dump BPF scheduler state for a runnable task on error
|
|
* @ctx: debug dump context
|
|
* @p: runnable task to generate debug dump for
|
|
*
|
|
* Use scx_bpf_dump() to generate BPF scheduler specific debug dump for
|
|
* @p.
|
|
*/
|
|
void (*dump_task)(struct scx_dump_ctx *ctx, struct task_struct *p);
|
|
|
|
#ifdef CONFIG_EXT_GROUP_SCHED
|
|
/**
|
|
* @cgroup_init: Initialize a cgroup
|
|
* @cgrp: cgroup being initialized
|
|
* @args: init arguments, see the struct definition
|
|
*
|
|
* Either the BPF scheduler is being loaded or @cgrp created, initialize
|
|
* @cgrp for sched_ext. This operation may block.
|
|
*
|
|
* Return 0 for success, -errno for failure. An error return while
|
|
* loading will abort loading of the BPF scheduler. During cgroup
|
|
* creation, it will abort the specific cgroup creation.
|
|
*/
|
|
s32 (*cgroup_init)(struct cgroup *cgrp,
|
|
struct scx_cgroup_init_args *args);
|
|
|
|
/**
|
|
* @cgroup_exit: Exit a cgroup
|
|
* @cgrp: cgroup being exited
|
|
*
|
|
* Either the BPF scheduler is being unloaded or @cgrp destroyed, exit
|
|
* @cgrp for sched_ext. This operation my block.
|
|
*/
|
|
void (*cgroup_exit)(struct cgroup *cgrp);
|
|
|
|
/**
|
|
* @cgroup_prep_move: Prepare a task to be moved to a different cgroup
|
|
* @p: task being moved
|
|
* @from: cgroup @p is being moved from
|
|
* @to: cgroup @p is being moved to
|
|
*
|
|
* Prepare @p for move from cgroup @from to @to. This operation may
|
|
* block and can be used for allocations.
|
|
*
|
|
* Return 0 for success, -errno for failure. An error return aborts the
|
|
* migration.
|
|
*/
|
|
s32 (*cgroup_prep_move)(struct task_struct *p,
|
|
struct cgroup *from, struct cgroup *to);
|
|
|
|
/**
|
|
* @cgroup_move: Commit cgroup move
|
|
* @p: task being moved
|
|
* @from: cgroup @p is being moved from
|
|
* @to: cgroup @p is being moved to
|
|
*
|
|
* Commit the move. @p is dequeued during this operation.
|
|
*/
|
|
void (*cgroup_move)(struct task_struct *p,
|
|
struct cgroup *from, struct cgroup *to);
|
|
|
|
/**
|
|
* @cgroup_cancel_move: Cancel cgroup move
|
|
* @p: task whose cgroup move is being canceled
|
|
* @from: cgroup @p was being moved from
|
|
* @to: cgroup @p was being moved to
|
|
*
|
|
* @p was cgroup_prep_move()'d but failed before reaching cgroup_move().
|
|
* Undo the preparation.
|
|
*/
|
|
void (*cgroup_cancel_move)(struct task_struct *p,
|
|
struct cgroup *from, struct cgroup *to);
|
|
|
|
/**
|
|
* @cgroup_set_weight: A cgroup's weight is being changed
|
|
* @cgrp: cgroup whose weight is being updated
|
|
* @weight: new weight [1..10000]
|
|
*
|
|
* Update @cgrp's weight to @weight.
|
|
*/
|
|
void (*cgroup_set_weight)(struct cgroup *cgrp, u32 weight);
|
|
|
|
/**
|
|
* @cgroup_set_bandwidth: A cgroup's bandwidth is being changed
|
|
* @cgrp: cgroup whose bandwidth is being updated
|
|
* @period_us: bandwidth control period
|
|
* @quota_us: bandwidth control quota
|
|
* @burst_us: bandwidth control burst
|
|
*
|
|
* Update @cgrp's bandwidth control parameters. This is from the cpu.max
|
|
* cgroup interface.
|
|
*
|
|
* @quota_us / @period_us determines the CPU bandwidth @cgrp is entitled
|
|
* to. For example, if @period_us is 1_000_000 and @quota_us is
|
|
* 2_500_000. @cgrp is entitled to 2.5 CPUs. @burst_us can be
|
|
* interpreted in the same fashion and specifies how much @cgrp can
|
|
* burst temporarily. The specific control mechanism and thus the
|
|
* interpretation of @period_us and burstiness is upto to the BPF
|
|
* scheduler.
|
|
*/
|
|
void (*cgroup_set_bandwidth)(struct cgroup *cgrp,
|
|
u64 period_us, u64 quota_us, u64 burst_us);
|
|
|
|
#endif /* CONFIG_EXT_GROUP_SCHED */
|
|
|
|
/*
|
|
* All online ops must come before ops.cpu_online().
|
|
*/
|
|
|
|
/**
|
|
* @cpu_online: A CPU became online
|
|
* @cpu: CPU which just came up
|
|
*
|
|
* @cpu just came online. @cpu will not call ops.enqueue() or
|
|
* ops.dispatch(), nor run tasks associated with other CPUs beforehand.
|
|
*/
|
|
void (*cpu_online)(s32 cpu);
|
|
|
|
/**
|
|
* @cpu_offline: A CPU is going offline
|
|
* @cpu: CPU which is going offline
|
|
*
|
|
* @cpu is going offline. @cpu will not call ops.enqueue() or
|
|
* ops.dispatch(), nor run tasks associated with other CPUs afterwards.
|
|
*/
|
|
void (*cpu_offline)(s32 cpu);
|
|
|
|
/*
|
|
* All CPU hotplug ops must come before ops.init().
|
|
*/
|
|
|
|
/**
|
|
* @init: Initialize the BPF scheduler
|
|
*/
|
|
s32 (*init)(void);
|
|
|
|
/**
|
|
* @exit: Clean up after the BPF scheduler
|
|
* @info: Exit info
|
|
*
|
|
* ops.exit() is also called on ops.init() failure, which is a bit
|
|
* unusual. This is to allow rich reporting through @info on how
|
|
* ops.init() failed.
|
|
*/
|
|
void (*exit)(struct scx_exit_info *info);
|
|
|
|
/**
|
|
* @dispatch_max_batch: Max nr of tasks that dispatch() can dispatch
|
|
*/
|
|
u32 dispatch_max_batch;
|
|
|
|
/**
|
|
* @flags: %SCX_OPS_* flags
|
|
*/
|
|
u64 flags;
|
|
|
|
/**
|
|
* @timeout_ms: The maximum amount of time, in milliseconds, that a
|
|
* runnable task should be able to wait before being scheduled. The
|
|
* maximum timeout may not exceed the default timeout of 30 seconds.
|
|
*
|
|
* Defaults to the maximum allowed timeout value of 30 seconds.
|
|
*/
|
|
u32 timeout_ms;
|
|
|
|
/**
|
|
* @exit_dump_len: scx_exit_info.dump buffer length. If 0, the default
|
|
* value of 32768 is used.
|
|
*/
|
|
u32 exit_dump_len;
|
|
|
|
/**
|
|
* @hotplug_seq: A sequence number that may be set by the scheduler to
|
|
* detect when a hotplug event has occurred during the loading process.
|
|
* If 0, no detection occurs. Otherwise, the scheduler will fail to
|
|
* load if the sequence number does not match @scx_hotplug_seq on the
|
|
* enable path.
|
|
*/
|
|
u64 hotplug_seq;
|
|
|
|
/**
|
|
* @name: BPF scheduler's name
|
|
*
|
|
* Must be a non-zero valid BPF object name including only isalnum(),
|
|
* '_' and '.' chars. Shows up in kernel.sched_ext_ops sysctl while the
|
|
* BPF scheduler is enabled.
|
|
*/
|
|
char name[SCX_OPS_NAME_LEN];
|
|
|
|
/* internal use only, must be NULL */
|
|
void *priv;
|
|
};
|
|
|
|
enum scx_opi {
|
|
SCX_OPI_BEGIN = 0,
|
|
SCX_OPI_NORMAL_BEGIN = 0,
|
|
SCX_OPI_NORMAL_END = SCX_OP_IDX(cpu_online),
|
|
SCX_OPI_CPU_HOTPLUG_BEGIN = SCX_OP_IDX(cpu_online),
|
|
SCX_OPI_CPU_HOTPLUG_END = SCX_OP_IDX(init),
|
|
SCX_OPI_END = SCX_OP_IDX(init),
|
|
};
|
|
|
|
/*
|
|
* Collection of event counters. Event types are placed in descending order.
|
|
*/
|
|
struct scx_event_stats {
|
|
/*
|
|
* If ops.select_cpu() returns a CPU which can't be used by the task,
|
|
* the core scheduler code silently picks a fallback CPU.
|
|
*/
|
|
s64 SCX_EV_SELECT_CPU_FALLBACK;
|
|
|
|
/*
|
|
* When dispatching to a local DSQ, the CPU may have gone offline in
|
|
* the meantime. In this case, the task is bounced to the global DSQ.
|
|
*/
|
|
s64 SCX_EV_DISPATCH_LOCAL_DSQ_OFFLINE;
|
|
|
|
/*
|
|
* If SCX_OPS_ENQ_LAST is not set, the number of times that a task
|
|
* continued to run because there were no other tasks on the CPU.
|
|
*/
|
|
s64 SCX_EV_DISPATCH_KEEP_LAST;
|
|
|
|
/*
|
|
* If SCX_OPS_ENQ_EXITING is not set, the number of times that a task
|
|
* is dispatched to a local DSQ when exiting.
|
|
*/
|
|
s64 SCX_EV_ENQ_SKIP_EXITING;
|
|
|
|
/*
|
|
* If SCX_OPS_ENQ_MIGRATION_DISABLED is not set, the number of times a
|
|
* migration disabled task skips ops.enqueue() and is dispatched to its
|
|
* local DSQ.
|
|
*/
|
|
s64 SCX_EV_ENQ_SKIP_MIGRATION_DISABLED;
|
|
|
|
/*
|
|
* Total number of times a task's time slice was refilled with the
|
|
* default value (SCX_SLICE_DFL).
|
|
*/
|
|
s64 SCX_EV_REFILL_SLICE_DFL;
|
|
|
|
/*
|
|
* The total duration of bypass modes in nanoseconds.
|
|
*/
|
|
s64 SCX_EV_BYPASS_DURATION;
|
|
|
|
/*
|
|
* The number of tasks dispatched in the bypassing mode.
|
|
*/
|
|
s64 SCX_EV_BYPASS_DISPATCH;
|
|
|
|
/*
|
|
* The number of times the bypassing mode has been activated.
|
|
*/
|
|
s64 SCX_EV_BYPASS_ACTIVATE;
|
|
};
|
|
|
|
struct scx_sched_pcpu {
|
|
/*
|
|
* The event counters are in a per-CPU variable to minimize the
|
|
* accounting overhead. A system-wide view on the event counter is
|
|
* constructed when requested by scx_bpf_events().
|
|
*/
|
|
struct scx_event_stats event_stats;
|
|
};
|
|
|
|
struct scx_sched {
|
|
struct sched_ext_ops ops;
|
|
DECLARE_BITMAP(has_op, SCX_OPI_END);
|
|
|
|
/*
|
|
* Dispatch queues.
|
|
*
|
|
* The global DSQ (%SCX_DSQ_GLOBAL) is split per-node for scalability.
|
|
* This is to avoid live-locking in bypass mode where all tasks are
|
|
* dispatched to %SCX_DSQ_GLOBAL and all CPUs consume from it. If
|
|
* per-node split isn't sufficient, it can be further split.
|
|
*/
|
|
struct rhashtable dsq_hash;
|
|
struct scx_dispatch_q **global_dsqs;
|
|
struct scx_sched_pcpu __percpu *pcpu;
|
|
|
|
bool warned_zero_slice:1;
|
|
bool warned_deprecated_rq:1;
|
|
|
|
atomic_t exit_kind;
|
|
struct scx_exit_info *exit_info;
|
|
|
|
struct kobject kobj;
|
|
|
|
struct kthread_worker *helper;
|
|
struct irq_work error_irq_work;
|
|
struct kthread_work disable_work;
|
|
struct rcu_work rcu_work;
|
|
};
|
|
|
|
enum scx_wake_flags {
|
|
/* expose select WF_* flags as enums */
|
|
SCX_WAKE_FORK = WF_FORK,
|
|
SCX_WAKE_TTWU = WF_TTWU,
|
|
SCX_WAKE_SYNC = WF_SYNC,
|
|
};
|
|
|
|
enum scx_enq_flags {
|
|
/* expose select ENQUEUE_* flags as enums */
|
|
SCX_ENQ_WAKEUP = ENQUEUE_WAKEUP,
|
|
SCX_ENQ_HEAD = ENQUEUE_HEAD,
|
|
SCX_ENQ_CPU_SELECTED = ENQUEUE_RQ_SELECTED,
|
|
|
|
/* high 32bits are SCX specific */
|
|
|
|
/*
|
|
* Set the following to trigger preemption when calling
|
|
* scx_bpf_dsq_insert() with a local dsq as the target. The slice of the
|
|
* current task is cleared to zero and the CPU is kicked into the
|
|
* scheduling path. Implies %SCX_ENQ_HEAD.
|
|
*/
|
|
SCX_ENQ_PREEMPT = 1LLU << 32,
|
|
|
|
/*
|
|
* The task being enqueued was previously enqueued on the current CPU's
|
|
* %SCX_DSQ_LOCAL, but was removed from it in a call to the
|
|
* scx_bpf_reenqueue_local() kfunc. If scx_bpf_reenqueue_local() was
|
|
* invoked in a ->cpu_release() callback, and the task is again
|
|
* dispatched back to %SCX_LOCAL_DSQ by this current ->enqueue(), the
|
|
* task will not be scheduled on the CPU until at least the next invocation
|
|
* of the ->cpu_acquire() callback.
|
|
*/
|
|
SCX_ENQ_REENQ = 1LLU << 40,
|
|
|
|
/*
|
|
* The task being enqueued is the only task available for the cpu. By
|
|
* default, ext core keeps executing such tasks but when
|
|
* %SCX_OPS_ENQ_LAST is specified, they're ops.enqueue()'d with the
|
|
* %SCX_ENQ_LAST flag set.
|
|
*
|
|
* The BPF scheduler is responsible for triggering a follow-up
|
|
* scheduling event. Otherwise, Execution may stall.
|
|
*/
|
|
SCX_ENQ_LAST = 1LLU << 41,
|
|
|
|
/* high 8 bits are internal */
|
|
__SCX_ENQ_INTERNAL_MASK = 0xffLLU << 56,
|
|
|
|
SCX_ENQ_CLEAR_OPSS = 1LLU << 56,
|
|
SCX_ENQ_DSQ_PRIQ = 1LLU << 57,
|
|
};
|
|
|
|
enum scx_deq_flags {
|
|
/* expose select DEQUEUE_* flags as enums */
|
|
SCX_DEQ_SLEEP = DEQUEUE_SLEEP,
|
|
|
|
/* high 32bits are SCX specific */
|
|
|
|
/*
|
|
* The generic core-sched layer decided to execute the task even though
|
|
* it hasn't been dispatched yet. Dequeue from the BPF side.
|
|
*/
|
|
SCX_DEQ_CORE_SCHED_EXEC = 1LLU << 32,
|
|
};
|
|
|
|
enum scx_pick_idle_cpu_flags {
|
|
SCX_PICK_IDLE_CORE = 1LLU << 0, /* pick a CPU whose SMT siblings are also idle */
|
|
SCX_PICK_IDLE_IN_NODE = 1LLU << 1, /* pick a CPU in the same target NUMA node */
|
|
};
|
|
|
|
enum scx_kick_flags {
|
|
/*
|
|
* Kick the target CPU if idle. Guarantees that the target CPU goes
|
|
* through at least one full scheduling cycle before going idle. If the
|
|
* target CPU can be determined to be currently not idle and going to go
|
|
* through a scheduling cycle before going idle, noop.
|
|
*/
|
|
SCX_KICK_IDLE = 1LLU << 0,
|
|
|
|
/*
|
|
* Preempt the current task and execute the dispatch path. If the
|
|
* current task of the target CPU is an SCX task, its ->scx.slice is
|
|
* cleared to zero before the scheduling path is invoked so that the
|
|
* task expires and the dispatch path is invoked.
|
|
*/
|
|
SCX_KICK_PREEMPT = 1LLU << 1,
|
|
|
|
/*
|
|
* Wait for the CPU to be rescheduled. The scx_bpf_kick_cpu() call will
|
|
* return after the target CPU finishes picking the next task.
|
|
*/
|
|
SCX_KICK_WAIT = 1LLU << 2,
|
|
};
|
|
|
|
enum scx_tg_flags {
|
|
SCX_TG_ONLINE = 1U << 0,
|
|
SCX_TG_INITED = 1U << 1,
|
|
};
|
|
|
|
enum scx_enable_state {
|
|
SCX_ENABLING,
|
|
SCX_ENABLED,
|
|
SCX_DISABLING,
|
|
SCX_DISABLED,
|
|
};
|
|
|
|
static const char *scx_enable_state_str[] = {
|
|
[SCX_ENABLING] = "enabling",
|
|
[SCX_ENABLED] = "enabled",
|
|
[SCX_DISABLING] = "disabling",
|
|
[SCX_DISABLED] = "disabled",
|
|
};
|
|
|
|
/*
|
|
* sched_ext_entity->ops_state
|
|
*
|
|
* Used to track the task ownership between the SCX core and the BPF scheduler.
|
|
* State transitions look as follows:
|
|
*
|
|
* NONE -> QUEUEING -> QUEUED -> DISPATCHING
|
|
* ^ | |
|
|
* | v v
|
|
* \-------------------------------/
|
|
*
|
|
* QUEUEING and DISPATCHING states can be waited upon. See wait_ops_state() call
|
|
* sites for explanations on the conditions being waited upon and why they are
|
|
* safe. Transitions out of them into NONE or QUEUED must store_release and the
|
|
* waiters should load_acquire.
|
|
*
|
|
* Tracking scx_ops_state enables sched_ext core to reliably determine whether
|
|
* any given task can be dispatched by the BPF scheduler at all times and thus
|
|
* relaxes the requirements on the BPF scheduler. This allows the BPF scheduler
|
|
* to try to dispatch any task anytime regardless of its state as the SCX core
|
|
* can safely reject invalid dispatches.
|
|
*/
|
|
enum scx_ops_state {
|
|
SCX_OPSS_NONE, /* owned by the SCX core */
|
|
SCX_OPSS_QUEUEING, /* in transit to the BPF scheduler */
|
|
SCX_OPSS_QUEUED, /* owned by the BPF scheduler */
|
|
SCX_OPSS_DISPATCHING, /* in transit back to the SCX core */
|
|
|
|
/*
|
|
* QSEQ brands each QUEUED instance so that, when dispatch races
|
|
* dequeue/requeue, the dispatcher can tell whether it still has a claim
|
|
* on the task being dispatched.
|
|
*
|
|
* As some 32bit archs can't do 64bit store_release/load_acquire,
|
|
* p->scx.ops_state is atomic_long_t which leaves 30 bits for QSEQ on
|
|
* 32bit machines. The dispatch race window QSEQ protects is very narrow
|
|
* and runs with IRQ disabled. 30 bits should be sufficient.
|
|
*/
|
|
SCX_OPSS_QSEQ_SHIFT = 2,
|
|
};
|
|
|
|
/* Use macros to ensure that the type is unsigned long for the masks */
|
|
#define SCX_OPSS_STATE_MASK ((1LU << SCX_OPSS_QSEQ_SHIFT) - 1)
|
|
#define SCX_OPSS_QSEQ_MASK (~SCX_OPSS_STATE_MASK)
|
|
|
|
DECLARE_PER_CPU(struct rq *, scx_locked_rq_state);
|
|
|
|
/*
|
|
* Return the rq currently locked from an scx callback, or NULL if no rq is
|
|
* locked.
|
|
*/
|
|
static inline struct rq *scx_locked_rq(void)
|
|
{
|
|
return __this_cpu_read(scx_locked_rq_state);
|
|
}
|
|
|
|
static inline bool scx_kf_allowed_if_unlocked(void)
|
|
{
|
|
return !current->scx.kf_mask;
|
|
}
|
|
|
|
static inline bool scx_rq_bypassing(struct rq *rq)
|
|
{
|
|
return unlikely(rq->scx.flags & SCX_RQ_BYPASSING);
|
|
}
|