OXIESEC PANEL
- Current Dir:
/
/
usr
/
src
/
linux-headers-4.15.0-213
/
include
/
linux
/
sched
Server IP: 139.59.38.164
Upload:
Create Dir:
Name
Size
Modified
Perms
📁
..
-
05/09/2024 07:14:16 AM
rwxr-xr-x
📄
autogroup.h
1.2 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
clock.h
2.45 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
coredump.h
2.75 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
cpufreq.h
887 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
cputime.h
5.13 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
deadline.h
597 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
debug.h
1.41 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
hotplug.h
578 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
idle.h
1.77 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
init.h
240 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
isolation.h
1.33 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
jobctl.h
1.46 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
loadavg.h
1.17 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
mm.h
7.63 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
nohz.h
1.22 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
numa_balancing.h
1.26 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
prio.h
1.71 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
rt.h
1.41 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
signal.h
17.79 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
smt.h
415 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
stat.h
1.02 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
sysctl.h
2.5 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
task.h
3.94 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
task_stack.h
3.03 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
topology.h
6.18 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
user.h
1.7 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
wake_q.h
1.79 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
xacct.h
854 bytes
01/28/2018 09:20:33 PM
rw-r--r--
Editing: idle.h
Close
/* SPDX-License-Identifier: GPL-2.0 */ #ifndef _LINUX_SCHED_IDLE_H #define _LINUX_SCHED_IDLE_H #include <linux/sched.h> enum cpu_idle_type { CPU_IDLE, CPU_NOT_IDLE, CPU_NEWLY_IDLE, CPU_MAX_IDLE_TYPES }; extern void wake_up_if_idle(int cpu); /* * Idle thread specific functions to determine the need_resched * polling state. */ #ifdef TIF_POLLING_NRFLAG static inline void __current_set_polling(void) { set_thread_flag(TIF_POLLING_NRFLAG); } static inline bool __must_check current_set_polling_and_test(void) { __current_set_polling(); /* * Polling state must be visible before we test NEED_RESCHED, * paired by resched_curr() */ smp_mb__after_atomic(); return unlikely(tif_need_resched()); } static inline void __current_clr_polling(void) { clear_thread_flag(TIF_POLLING_NRFLAG); } static inline bool __must_check current_clr_polling_and_test(void) { __current_clr_polling(); /* * Polling state must be visible before we test NEED_RESCHED, * paired by resched_curr() */ smp_mb__after_atomic(); return unlikely(tif_need_resched()); } #else static inline void __current_set_polling(void) { } static inline void __current_clr_polling(void) { } static inline bool __must_check current_set_polling_and_test(void) { return unlikely(tif_need_resched()); } static inline bool __must_check current_clr_polling_and_test(void) { return unlikely(tif_need_resched()); } #endif static inline void current_clr_polling(void) { __current_clr_polling(); /* * Ensure we check TIF_NEED_RESCHED after we clear the polling bit. * Once the bit is cleared, we'll get IPIs with every new * TIF_NEED_RESCHED and the IPI handler, scheduler_ipi(), will also * fold. */ smp_mb(); /* paired with resched_curr() */ preempt_fold_need_resched(); } #endif /* _LINUX_SCHED_IDLE_H */