sched: Add sched tracepoints for RV task model
Add the following tracepoints:
* sched_entry(bool preempt, ip)
Called while entering __schedule
* sched_exit(bool is_switch, ip)
Called while exiting __schedule
* sched_set_state(task, curr_state, state)
Called when a task changes its state (to and from running)
These tracepoints are useful to describe the Linux task model and are
adapted from the patches by Daniel Bristot de Oliveira
(https://bristot.me/linux-task-model/).
Cc: Ingo Molnar <mingo@redhat.com>
Cc: Masami Hiramatsu <mhiramat@kernel.org>
Cc: Juri Lelli <juri.lelli@redhat.com>
Link: https://lore.kernel.org/20250305140406.350227-2-gmonaco@redhat.com
Signed-off-by: Gabriele Monaco <gmonaco@redhat.com>
Acked-by: Peter Zijlstra (Intel) <peterz@infradead.org>
Signed-off-by: Steven Rostedt (Google) <rostedt@goodmis.org>
pull/1188/head
parent
41a4d2d3e3
commit
26f80681a0
|
|
@ -7,7 +7,7 @@
|
||||||
#ifndef _LINUX_RV_H
|
#ifndef _LINUX_RV_H
|
||||||
#define _LINUX_RV_H
|
#define _LINUX_RV_H
|
||||||
|
|
||||||
#define MAX_DA_NAME_LEN 24
|
#define MAX_DA_NAME_LEN 32
|
||||||
|
|
||||||
#ifdef CONFIG_RV
|
#ifdef CONFIG_RV
|
||||||
/*
|
/*
|
||||||
|
|
|
||||||
|
|
@ -46,6 +46,7 @@
|
||||||
#include <linux/rv.h>
|
#include <linux/rv.h>
|
||||||
#include <linux/livepatch_sched.h>
|
#include <linux/livepatch_sched.h>
|
||||||
#include <linux/uidgid_types.h>
|
#include <linux/uidgid_types.h>
|
||||||
|
#include <linux/tracepoint-defs.h>
|
||||||
#include <asm/kmap_size.h>
|
#include <asm/kmap_size.h>
|
||||||
|
|
||||||
/* task_struct member predeclarations (sorted alphabetically): */
|
/* task_struct member predeclarations (sorted alphabetically): */
|
||||||
|
|
@ -186,6 +187,12 @@ struct user_event_mm;
|
||||||
# define debug_rtlock_wait_restore_state() do { } while (0)
|
# define debug_rtlock_wait_restore_state() do { } while (0)
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
|
#define trace_set_current_state(state_value) \
|
||||||
|
do { \
|
||||||
|
if (tracepoint_enabled(sched_set_state_tp)) \
|
||||||
|
__trace_set_current_state(state_value); \
|
||||||
|
} while (0)
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* set_current_state() includes a barrier so that the write of current->__state
|
* set_current_state() includes a barrier so that the write of current->__state
|
||||||
* is correctly serialised wrt the caller's subsequent test of whether to
|
* is correctly serialised wrt the caller's subsequent test of whether to
|
||||||
|
|
@ -226,12 +233,14 @@ struct user_event_mm;
|
||||||
#define __set_current_state(state_value) \
|
#define __set_current_state(state_value) \
|
||||||
do { \
|
do { \
|
||||||
debug_normal_state_change((state_value)); \
|
debug_normal_state_change((state_value)); \
|
||||||
|
trace_set_current_state(state_value); \
|
||||||
WRITE_ONCE(current->__state, (state_value)); \
|
WRITE_ONCE(current->__state, (state_value)); \
|
||||||
} while (0)
|
} while (0)
|
||||||
|
|
||||||
#define set_current_state(state_value) \
|
#define set_current_state(state_value) \
|
||||||
do { \
|
do { \
|
||||||
debug_normal_state_change((state_value)); \
|
debug_normal_state_change((state_value)); \
|
||||||
|
trace_set_current_state(state_value); \
|
||||||
smp_store_mb(current->__state, (state_value)); \
|
smp_store_mb(current->__state, (state_value)); \
|
||||||
} while (0)
|
} while (0)
|
||||||
|
|
||||||
|
|
@ -247,6 +256,7 @@ struct user_event_mm;
|
||||||
\
|
\
|
||||||
raw_spin_lock_irqsave(¤t->pi_lock, flags); \
|
raw_spin_lock_irqsave(¤t->pi_lock, flags); \
|
||||||
debug_special_state_change((state_value)); \
|
debug_special_state_change((state_value)); \
|
||||||
|
trace_set_current_state(state_value); \
|
||||||
WRITE_ONCE(current->__state, (state_value)); \
|
WRITE_ONCE(current->__state, (state_value)); \
|
||||||
raw_spin_unlock_irqrestore(¤t->pi_lock, flags); \
|
raw_spin_unlock_irqrestore(¤t->pi_lock, flags); \
|
||||||
} while (0)
|
} while (0)
|
||||||
|
|
@ -282,6 +292,7 @@ struct user_event_mm;
|
||||||
raw_spin_lock(¤t->pi_lock); \
|
raw_spin_lock(¤t->pi_lock); \
|
||||||
current->saved_state = current->__state; \
|
current->saved_state = current->__state; \
|
||||||
debug_rtlock_wait_set_state(); \
|
debug_rtlock_wait_set_state(); \
|
||||||
|
trace_set_current_state(TASK_RTLOCK_WAIT); \
|
||||||
WRITE_ONCE(current->__state, TASK_RTLOCK_WAIT); \
|
WRITE_ONCE(current->__state, TASK_RTLOCK_WAIT); \
|
||||||
raw_spin_unlock(¤t->pi_lock); \
|
raw_spin_unlock(¤t->pi_lock); \
|
||||||
} while (0);
|
} while (0);
|
||||||
|
|
@ -291,6 +302,7 @@ struct user_event_mm;
|
||||||
lockdep_assert_irqs_disabled(); \
|
lockdep_assert_irqs_disabled(); \
|
||||||
raw_spin_lock(¤t->pi_lock); \
|
raw_spin_lock(¤t->pi_lock); \
|
||||||
debug_rtlock_wait_restore_state(); \
|
debug_rtlock_wait_restore_state(); \
|
||||||
|
trace_set_current_state(current->saved_state); \
|
||||||
WRITE_ONCE(current->__state, current->saved_state); \
|
WRITE_ONCE(current->__state, current->saved_state); \
|
||||||
current->saved_state = TASK_RUNNING; \
|
current->saved_state = TASK_RUNNING; \
|
||||||
raw_spin_unlock(¤t->pi_lock); \
|
raw_spin_unlock(¤t->pi_lock); \
|
||||||
|
|
@ -327,6 +339,10 @@ extern void io_schedule_finish(int token);
|
||||||
extern long io_schedule_timeout(long timeout);
|
extern long io_schedule_timeout(long timeout);
|
||||||
extern void io_schedule(void);
|
extern void io_schedule(void);
|
||||||
|
|
||||||
|
/* wrapper function to trace from this header file */
|
||||||
|
DECLARE_TRACEPOINT(sched_set_state_tp);
|
||||||
|
extern void __trace_set_current_state(int state_value);
|
||||||
|
|
||||||
/**
|
/**
|
||||||
* struct prev_cputime - snapshot of system and user cputime
|
* struct prev_cputime - snapshot of system and user cputime
|
||||||
* @utime: time spent in user mode
|
* @utime: time spent in user mode
|
||||||
|
|
|
||||||
|
|
@ -824,6 +824,19 @@ DECLARE_TRACE(sched_compute_energy_tp,
|
||||||
unsigned long max_util, unsigned long busy_time),
|
unsigned long max_util, unsigned long busy_time),
|
||||||
TP_ARGS(p, dst_cpu, energy, max_util, busy_time));
|
TP_ARGS(p, dst_cpu, energy, max_util, busy_time));
|
||||||
|
|
||||||
|
DECLARE_TRACE(sched_entry_tp,
|
||||||
|
TP_PROTO(bool preempt, unsigned long ip),
|
||||||
|
TP_ARGS(preempt, ip));
|
||||||
|
|
||||||
|
DECLARE_TRACE(sched_exit_tp,
|
||||||
|
TP_PROTO(bool is_switch, unsigned long ip),
|
||||||
|
TP_ARGS(is_switch, ip));
|
||||||
|
|
||||||
|
DECLARE_TRACE_CONDITION(sched_set_state_tp,
|
||||||
|
TP_PROTO(struct task_struct *tsk, int state),
|
||||||
|
TP_ARGS(tsk, state),
|
||||||
|
TP_CONDITION(!!(tsk->__state) != !!state));
|
||||||
|
|
||||||
#endif /* _TRACE_SCHED_H */
|
#endif /* _TRACE_SCHED_H */
|
||||||
|
|
||||||
/* This part must be outside protection */
|
/* This part must be outside protection */
|
||||||
|
|
|
||||||
|
|
@ -491,6 +491,16 @@ sched_core_dequeue(struct rq *rq, struct task_struct *p, int flags) { }
|
||||||
|
|
||||||
#endif /* CONFIG_SCHED_CORE */
|
#endif /* CONFIG_SCHED_CORE */
|
||||||
|
|
||||||
|
/* need a wrapper since we may need to trace from modules */
|
||||||
|
EXPORT_TRACEPOINT_SYMBOL(sched_set_state_tp);
|
||||||
|
|
||||||
|
/* Call via the helper macro trace_set_current_state. */
|
||||||
|
void __trace_set_current_state(int state_value)
|
||||||
|
{
|
||||||
|
trace_sched_set_state_tp(current, state_value);
|
||||||
|
}
|
||||||
|
EXPORT_SYMBOL(__trace_set_current_state);
|
||||||
|
|
||||||
/*
|
/*
|
||||||
* Serialization rules:
|
* Serialization rules:
|
||||||
*
|
*
|
||||||
|
|
@ -5307,6 +5317,12 @@ asmlinkage __visible void schedule_tail(struct task_struct *prev)
|
||||||
*/
|
*/
|
||||||
|
|
||||||
finish_task_switch(prev);
|
finish_task_switch(prev);
|
||||||
|
/*
|
||||||
|
* This is a special case: the newly created task has just
|
||||||
|
* switched the context for the first time. It is returning from
|
||||||
|
* schedule for the first time in this path.
|
||||||
|
*/
|
||||||
|
trace_sched_exit_tp(true, CALLER_ADDR0);
|
||||||
preempt_enable();
|
preempt_enable();
|
||||||
|
|
||||||
if (current->set_child_tid)
|
if (current->set_child_tid)
|
||||||
|
|
@ -6650,12 +6666,15 @@ static void __sched notrace __schedule(int sched_mode)
|
||||||
* as a preemption by schedule_debug() and RCU.
|
* as a preemption by schedule_debug() and RCU.
|
||||||
*/
|
*/
|
||||||
bool preempt = sched_mode > SM_NONE;
|
bool preempt = sched_mode > SM_NONE;
|
||||||
|
bool is_switch = false;
|
||||||
unsigned long *switch_count;
|
unsigned long *switch_count;
|
||||||
unsigned long prev_state;
|
unsigned long prev_state;
|
||||||
struct rq_flags rf;
|
struct rq_flags rf;
|
||||||
struct rq *rq;
|
struct rq *rq;
|
||||||
int cpu;
|
int cpu;
|
||||||
|
|
||||||
|
trace_sched_entry_tp(preempt, CALLER_ADDR0);
|
||||||
|
|
||||||
cpu = smp_processor_id();
|
cpu = smp_processor_id();
|
||||||
rq = cpu_rq(cpu);
|
rq = cpu_rq(cpu);
|
||||||
prev = rq->curr;
|
prev = rq->curr;
|
||||||
|
|
@ -6723,7 +6742,8 @@ picked:
|
||||||
rq->last_seen_need_resched_ns = 0;
|
rq->last_seen_need_resched_ns = 0;
|
||||||
#endif
|
#endif
|
||||||
|
|
||||||
if (likely(prev != next)) {
|
is_switch = prev != next;
|
||||||
|
if (likely(is_switch)) {
|
||||||
rq->nr_switches++;
|
rq->nr_switches++;
|
||||||
/*
|
/*
|
||||||
* RCU users of rcu_dereference(rq->curr) may not see
|
* RCU users of rcu_dereference(rq->curr) may not see
|
||||||
|
|
@ -6768,6 +6788,7 @@ picked:
|
||||||
__balance_callbacks(rq);
|
__balance_callbacks(rq);
|
||||||
raw_spin_rq_unlock_irq(rq);
|
raw_spin_rq_unlock_irq(rq);
|
||||||
}
|
}
|
||||||
|
trace_sched_exit_tp(is_switch, CALLER_ADDR0);
|
||||||
}
|
}
|
||||||
|
|
||||||
void __noreturn do_task_dead(void)
|
void __noreturn do_task_dead(void)
|
||||||
|
|
|
||||||
|
|
@ -1,7 +1,7 @@
|
||||||
// SPDX-License-Identifier: GPL-2.0
|
// SPDX-License-Identifier: GPL-2.0
|
||||||
|
|
||||||
#define MAX_DESCRIPTION 1024
|
#define MAX_DESCRIPTION 1024
|
||||||
#define MAX_DA_NAME_LEN 24
|
#define MAX_DA_NAME_LEN 32
|
||||||
|
|
||||||
struct monitor {
|
struct monitor {
|
||||||
char name[MAX_DA_NAME_LEN];
|
char name[MAX_DA_NAME_LEN];
|
||||||
|
|
|
||||||
Loading…
Reference in New Issue