Lines Matching refs:task_struct

185 static inline int task_has_idle_policy(struct task_struct *p)  in task_has_idle_policy()
190 static inline int task_has_rt_policy(struct task_struct *p) in task_has_rt_policy()
195 static inline int task_has_dl_policy(struct task_struct *p) in task_has_dl_policy()
267 void __dl_clear_params(struct task_struct *p);
335 static inline bool dl_task_fits_capacity(struct task_struct *p, int cpu) in dl_task_fits_capacity()
345 extern int sched_dl_overflow(struct task_struct *p, int policy, const struct sched_attr *attr);
346 extern void __setparam_dl(struct task_struct *p, const struct sched_attr *attr);
347 extern void __getparam_dl(struct task_struct *p, struct sched_attr *attr);
349 extern bool dl_param_changed(struct task_struct *p, const struct sched_attr *attr);
350 extern int dl_task_can_attach(struct task_struct *p, const struct cpumask *cs_cpus_allowed);
501 extern int sched_rt_can_attach(struct task_group *tg, struct task_struct *tsk);
509 extern void sched_move_task(struct task_struct *tsk);
980 struct task_struct __rcu *curr;
981 struct task_struct *idle;
982 struct task_struct *stop;
1105 struct task_struct *core_pick;
1146 static inline bool is_migration_disabled(struct task_struct *p) in is_migration_disabled()
1191 bool cfs_prio_less(struct task_struct *a, struct task_struct *b, bool fi);
1199 static inline bool sched_cpu_cookie_match(struct rq *rq, struct task_struct *p) in sched_cpu_cookie_match()
1208 static inline bool sched_core_cookie_match(struct rq *rq, struct task_struct *p) in sched_core_cookie_match()
1232 struct task_struct *p, in sched_group_cookie_match()
1250 static inline bool sched_core_enqueued(struct task_struct *p) in sched_core_enqueued()
1255 extern void sched_core_enqueue(struct rq *rq, struct task_struct *p);
1256 extern void sched_core_dequeue(struct rq *rq, struct task_struct *p);
1287 static inline bool sched_cpu_cookie_match(struct rq *rq, struct task_struct *p) in sched_cpu_cookie_match()
1292 static inline bool sched_core_cookie_match(struct rq *rq, struct task_struct *p) in sched_core_cookie_match()
1298 struct task_struct *p, in sched_group_cookie_match()
1372 static inline struct task_struct *task_of(struct sched_entity *se) in task_of()
1375 return container_of(se, struct task_struct, se); in task_of()
1378 static inline struct cfs_rq *task_cfs_rq(struct task_struct *p) in task_cfs_rq()
1397 static inline struct task_struct *task_of(struct sched_entity *se) in task_of()
1399 return container_of(se, struct task_struct, se); in task_of()
1402 static inline struct cfs_rq *task_cfs_rq(struct task_struct *p) in task_cfs_rq()
1409 struct task_struct *p = task_of(se); in cfs_rq_of()
1570 struct rq *__task_rq_lock(struct task_struct *p, struct rq_flags *rf)
1573 struct rq *task_rq_lock(struct task_struct *p, struct rq_flags *rf)
1585 task_rq_unlock(struct rq *rq, struct task_struct *p, struct rq_flags *rf) in task_rq_unlock()
1685 extern void sched_setnuma(struct task_struct *p, int node);
1686 extern int migrate_task_to(struct task_struct *p, int cpu);
1687 extern int migrate_swap(struct task_struct *p, struct task_struct *t,
1689 extern void init_numa_balancing(unsigned long clone_flags, struct task_struct *p);
1692 init_numa_balancing(unsigned long clone_flags, struct task_struct *p) in init_numa_balancing()
1872 static inline struct task_group *task_group(struct task_struct *p) in task_group()
1878 static inline void set_task_rq(struct task_struct *p, unsigned int cpu) in set_task_rq()
1898 static inline void set_task_rq(struct task_struct *p, unsigned int cpu) { } in set_task_rq()
1899 static inline struct task_group *task_group(struct task_struct *p) in task_group()
1906 static inline void __set_task_cpu(struct task_struct *p, unsigned int cpu) in __set_task_cpu()
2002 static inline int task_current(struct rq *rq, struct task_struct *p) in task_current()
2007 static inline int task_running(struct rq *rq, struct task_struct *p) in task_running()
2016 static inline int task_on_rq_queued(struct task_struct *p) in task_on_rq_queued()
2021 static inline int task_on_rq_migrating(struct task_struct *p) in task_on_rq_migrating()
2101 void (*enqueue_task) (struct rq *rq, struct task_struct *p, int flags);
2102 void (*dequeue_task) (struct rq *rq, struct task_struct *p, int flags);
2104 bool (*yield_to_task)(struct rq *rq, struct task_struct *p);
2106 void (*check_preempt_curr)(struct rq *rq, struct task_struct *p, int flags);
2108 struct task_struct *(*pick_next_task)(struct rq *rq);
2110 void (*put_prev_task)(struct rq *rq, struct task_struct *p);
2111 void (*set_next_task)(struct rq *rq, struct task_struct *p, bool first);
2114 int (*balance)(struct rq *rq, struct task_struct *prev, struct rq_flags *rf);
2115 int (*select_task_rq)(struct task_struct *p, int task_cpu, int flags);
2117 struct task_struct * (*pick_task)(struct rq *rq);
2119 void (*migrate_task_rq)(struct task_struct *p, int new_cpu);
2121 void (*task_woken)(struct rq *this_rq, struct task_struct *task);
2123 void (*set_cpus_allowed)(struct task_struct *p,
2130 struct rq *(*find_lock_rq)(struct task_struct *p, struct rq *rq);
2133 void (*task_tick)(struct rq *rq, struct task_struct *p, int queued);
2134 void (*task_fork)(struct task_struct *p);
2135 void (*task_dead)(struct task_struct *p);
2142 void (*switched_from)(struct rq *this_rq, struct task_struct *task);
2143 void (*switched_to) (struct rq *this_rq, struct task_struct *task);
2144 void (*prio_changed) (struct rq *this_rq, struct task_struct *task,
2148 struct task_struct *task);
2156 void (*task_change_group)(struct task_struct *p, int type);
2160 static inline void put_prev_task(struct rq *rq, struct task_struct *prev) in put_prev_task()
2166 static inline void set_next_task(struct rq *rq, struct task_struct *next) in set_next_task()
2224 extern struct task_struct *pick_next_task_fair(struct rq *rq, struct task_struct *prev, struct rq_f…
2225 extern struct task_struct *pick_next_task_idle(struct rq *rq);
2238 extern void set_cpus_allowed_common(struct task_struct *p, const struct cpumask *new_mask, u32 flag…
2240 static inline struct task_struct *get_push_task(struct rq *rq) in get_push_task()
2242 struct task_struct *p = rq->curr; in get_push_task()
2298 extern void reweight_task(struct task_struct *p, int prio);
2319 extern void post_init_entity_util_avg(struct task_struct *p);
2377 extern void activate_task(struct rq *rq, struct task_struct *p, int flags);
2378 extern void deactivate_task(struct rq *rq, struct task_struct *p, int flags);
2380 extern void check_preempt_curr(struct rq *rq, struct task_struct *p, int flags);
2676 show_numa_stats(struct task_struct *p, struct seq_file *m);
2817 unsigned long uclamp_eff_value(struct task_struct *p, enum uclamp_id clamp_id);
2838 struct task_struct *p) in uclamp_rq_util_with()
2887 struct task_struct *p) in uclamp_rq_util_with()
2929 struct task_struct *p);
3036 static inline bool is_per_cpu_kthread(struct task_struct *p) in is_per_cpu_kthread()