OXIESEC PANEL
- Current Dir:
/
/
usr
/
src
/
linux-headers-4.15.0-213
/
arch
/
x86
/
include
/
asm
Server IP: 139.59.38.164
Upload:
Create Dir:
Name
Size
Modified
Perms
📁
..
-
05/09/2024 07:14:13 AM
rwxr-xr-x
📄
Kbuild
294 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
a.out-core.h
1.89 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
acenv.h
1.56 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
acpi.h
4.76 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
agp.h
1.04 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
alternative-asm.h
2.43 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
alternative.h
8.28 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
amd_nb.h
2.98 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
apb_timer.h
1.43 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
apic.h
14.53 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
apic_flat_64.h
151 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
apicdef.h
11.26 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
apm.h
1.8 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
arch_hweight.h
1.28 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
archrandom.h
3.03 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
asm-offsets.h
35 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
asm-prototypes.h
946 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
asm.h
4.97 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
atomic.h
6.02 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
atomic64_32.h
8.71 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
atomic64_64.h
6.31 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
barrier.h
3.6 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
bios_ebda.h
914 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
bitops.h
13.78 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
boot.h
1.53 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
bootparam_utils.h
2.86 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
bug.h
2.07 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
bugs.h
493 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
cache.h
641 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
cacheflush.h
306 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
cacheinfo.h
209 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
calgary.h
2.31 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
ce4100.h
121 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
checksum.h
133 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
checksum_32.h
4.86 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
checksum_64.h
5.41 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
clocksource.h
488 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
cmdline.h
302 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
cmpxchg.h
7.68 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
cmpxchg_32.h
3.15 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
cmpxchg_64.h
543 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
compat.h
7.37 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
cpu.h
975 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
cpu_device_id.h
7.39 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
cpu_entry_area.h
2.27 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
cpufeature.h
7.75 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
cpufeatures.h
25.59 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
cpumask.h
408 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
crash.h
320 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📁
crypto
-
05/09/2024 07:14:16 AM
rwxr-xr-x
📄
current.h
443 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
debugreg.h
2.67 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
delay.h
208 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
desc.h
11.42 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
desc_defs.h
3.16 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
device.h
568 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
disabled-features.h
2.31 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
div64.h
1.79 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
dma-mapping.h
2.4 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
dma.h
9.58 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
dmi.h
556 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
dwarf2.h
2.43 KB
01/28/2018 09:20:33 PM
rw-r--r--
📁
e820
-
05/09/2024 07:14:16 AM
rwxr-xr-x
📄
edac.h
474 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
efi.h
6.9 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
elf.h
10.82 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
emergency-restart.h
202 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
entry_arch.h
1.88 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
espfix.h
426 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
exec.h
37 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
export.h
120 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
extable.h
1.27 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
fb.h
540 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
fixmap.h
6.04 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
floppy.h
6.59 KB
01/28/2018 09:20:33 PM
rw-r--r--
📁
fpu
-
05/09/2024 07:14:16 AM
rwxr-xr-x
📄
frame.h
815 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
ftrace.h
1.8 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
futex.h
2.2 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
gart.h
2.64 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
genapic.h
22 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
geode.h
842 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
hardirq.h
2.3 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
highmem.h
2.6 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
hpet.h
3.38 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
hugetlb.h
2.15 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
hw_breakpoint.h
1.96 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
hw_irq.h
3.85 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
hypervisor.h
1.84 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
i8259.h
1.93 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
ia32.h
1.46 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
ia32_unistd.h
313 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
imr.h
1.81 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
inat.h
6.58 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
inat_types.h
1013 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
init.h
632 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
insn-eval.h
837 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
insn.h
7.46 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
inst.h
5.07 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
intel-family.h
3.47 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
intel-mid.h
4.91 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
intel_ds.h
793 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
intel_mid_vrtc.h
326 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
intel_pmc_ipc.h
2.08 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
intel_pt.h
292 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
intel_punit_ipc.h
4.56 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
intel_rdt_sched.h
2.59 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
intel_scu_ipc.h
2.3 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
intel_telemetry.h
3.96 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
invpcid.h
1.57 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
io.h
12.21 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
io_apic.h
5.63 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
iomap.h
1.22 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
iommu.h
392 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
iommu_table.h
3.82 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
iosf_mbi.h
5.74 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
ipi.h
2.84 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
irq.h
1.12 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
irq_regs.h
679 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
irq_remapping.h
2.96 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
irq_vectors.h
4.12 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
irq_work.h
397 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
irqdomain.h
1.61 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
irqflags.h
4.38 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
ist.h
735 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
jump_label.h
2.44 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
kasan.h
966 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
kaslr.h
424 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
kbdleds.h
454 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
kdebug.h
752 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
kexec-bzimage64.h
189 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
kexec.h
6.69 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
kgdb.h
2.09 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
kmap_types.h
289 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
kprobes.h
3.82 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
kvm_emulate.h
15.23 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
kvm_guest.h
172 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
kvm_host.h
42.72 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
kvm_page_track.h
2.48 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
kvm_para.h
3 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
kvmclock.h
170 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
linkage.h
581 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
livepatch.h
1.12 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
local.h
3.83 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
local64.h
33 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
mach_timer.h
1.55 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
mach_traps.h
1013 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
math_emu.h
395 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
mc146818rtc.h
2.76 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
mce.h
12.54 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
mem_encrypt.h
2.83 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
microcode.h
4.14 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
microcode_amd.h
1.41 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
microcode_intel.h
2.46 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
misc.h
143 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
mmconfig.h
374 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
mmu.h
1.57 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
mmu_context.h
10.27 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
mmx.h
337 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
mmzone.h
129 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
mmzone_32.h
1.16 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
mmzone_64.h
430 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
module.h
2.05 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
mpspec.h
3.93 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
mpspec_def.h
3.93 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
mpx.h
2.97 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
mshyperv.h
10.69 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
msi.h
392 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
msidef.h
1.77 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
msr-index.h
30.92 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
msr-trace.h
1.35 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
msr.h
10.85 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
mtrr.h
4.62 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
mwait.h
3.74 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
nmi.h
1.39 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
nops.h
4.31 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
nospec-branch.h
11.49 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
numa.h
2.18 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
numa_32.h
256 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📁
numachip
-
05/09/2024 07:14:16 AM
rwxr-xr-x
📄
olpc.h
3.16 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
olpc_ofw.h
1.1 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
orc_lookup.h
1.63 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
orc_types.h
3.47 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
page.h
2.18 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
page_32.h
1.01 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
page_32_types.h
1.7 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
page_64.h
1.42 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
page_64_types.h
2.34 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
page_types.h
2.29 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
paravirt.h
23.31 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
paravirt_types.h
22.15 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
parport.h
314 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
pat.h
768 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
pci-direct.h
995 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
pci-functions.h
654 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
pci.h
3.51 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
pci_64.h
684 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
pci_x86.h
5.71 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
percpu.h
18.97 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
perf_event.h
8.82 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
perf_event_p4.h
26.1 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
pgalloc.h
5.57 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
pgtable-2level.h
2.75 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
pgtable-2level_types.h
867 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
pgtable-3level.h
10.24 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
pgtable-3level_types.h
1.06 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
pgtable-invert.h
1.07 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
pgtable.h
33.91 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
pgtable_32.h
3.1 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
pgtable_32_types.h
2.01 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
pgtable_64.h
7.37 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
pgtable_64_types.h
3.67 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
pgtable_types.h
16.75 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
pkeys.h
3.17 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
platform_sst_audio.h
3.18 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
pm-trace.h
611 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
posix_types.h
144 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
preempt.h
3.04 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
probe_roms.h
273 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
processor-cyrix.h
879 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
processor-flags.h
1.71 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
processor.h
24.54 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
prom.h
1 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
proto.h
1003 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
pti.h
369 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
ptrace.h
8.52 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
purgatory.h
571 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
pvclock-abi.h
1.49 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
pvclock.h
2.64 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
qrwlock.h
199 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
qspinlock.h
2.54 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
qspinlock_paravirt.h
1.85 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
realmode.h
1.76 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
reboot.h
850 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
reboot_fixups.h
183 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
refcount.h
2.83 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
required-features.h
2.81 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
rio.h
2.57 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
rmwcc.h
1.62 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
rwsem.h
7.02 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
seccomp.h
510 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
sections.h
465 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
segment.h
9.47 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
serial.h
1.11 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
set_memory.h
3.86 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
setup.h
3.4 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
setup_arch.h
77 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
shmparam.h
193 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
sigcontext.h
261 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
sigframe.h
2.25 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
sighandling.h
649 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
signal.h
2.37 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
simd.h
287 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
smap.h
1.71 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
smp.h
4.73 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
sparsemem.h
1.01 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
spec-ctrl.h
2.81 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
special_insns.h
5.04 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
spinlock.h
1.19 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
spinlock_types.h
719 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
sta2x11.h
352 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
stackprotector.h
4.13 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
stacktrace.h
2.57 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
string.h
129 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
string_32.h
7.74 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
string_64.h
3.56 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
suspend.h
131 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
suspend_32.h
822 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
suspend_64.h
1.79 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
svm.h
7.08 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
swiotlb.h
991 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
switch_to.h
2.98 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
sync_bitops.h
3.42 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
sys_ia32.h
1.8 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
syscall.h
5.14 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
syscalls.h
1.39 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
sysfb.h
2.54 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
tce.h
1.68 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
text-patching.h
2.3 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
thread_info.h
9.33 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
time.h
331 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
timer.h
1 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
timex.h
546 bytes
06/16/2023 05:32:39 PM
rw-r--r--
📄
tlb.h
1.05 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
tlbbatch.h
332 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
tlbflush.h
17.09 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
topology.h
4.84 KB
06/16/2023 05:32:39 PM
rw-r--r--
📁
trace
-
05/09/2024 07:14:16 AM
rwxr-xr-x
📄
trace_clock.h
406 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
traps.h
5.74 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
tsc.h
1.93 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
uaccess.h
21.69 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
uaccess_32.h
1.54 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
uaccess_64.h
5.32 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
umip.h
329 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
unaligned.h
345 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
unistd.h
1.45 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
unwind.h
3.13 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
unwind_hints.h
3.01 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
uprobes.h
1.57 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
user.h
2.2 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
user32.h
2.11 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
user_32.h
4.92 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
user_64.h
5.21 KB
01/28/2018 09:20:33 PM
rw-r--r--
📁
uv
-
05/09/2024 07:14:16 AM
rwxr-xr-x
📄
vdso.h
1.09 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
vga.h
740 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
vgtod.h
2.13 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
virtext.h
2.62 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
vm86.h
2.16 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
vmx.h
23.5 KB
06/16/2023 05:32:39 PM
rw-r--r--
📄
vsyscall.h
635 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
vvar.h
1.38 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
word-at-a-time.h
2.54 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
x86_init.h
9.25 KB
06/16/2023 05:32:39 PM
rw-r--r--
📁
xen
-
05/09/2024 07:14:16 AM
rwxr-xr-x
📄
xor.h
10.26 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
xor_32.h
14.4 KB
01/28/2018 09:20:33 PM
rw-r--r--
📄
xor_64.h
716 bytes
01/28/2018 09:20:33 PM
rw-r--r--
📄
xor_avx.h
4.5 KB
01/28/2018 09:20:33 PM
rw-r--r--
Editing: tlbflush.h
Close
/* SPDX-License-Identifier: GPL-2.0 */ #ifndef _ASM_X86_TLBFLUSH_H #define _ASM_X86_TLBFLUSH_H #include <linux/mm.h> #include <linux/sched.h> #include <asm/processor.h> #include <asm/cpufeature.h> #include <asm/special_insns.h> #include <asm/smp.h> #include <asm/invpcid.h> #include <asm/pti.h> #include <asm/processor-flags.h> /* * The x86 feature is called PCID (Process Context IDentifier). It is similar * to what is traditionally called ASID on the RISC processors. * * We don't use the traditional ASID implementation, where each process/mm gets * its own ASID and flush/restart when we run out of ASID space. * * Instead we have a small per-cpu array of ASIDs and cache the last few mm's * that came by on this CPU, allowing cheaper switch_mm between processes on * this CPU. * * We end up with different spaces for different things. To avoid confusion we * use different names for each of them: * * ASID - [0, TLB_NR_DYN_ASIDS-1] * the canonical identifier for an mm * * kPCID - [1, TLB_NR_DYN_ASIDS] * the value we write into the PCID part of CR3; corresponds to the * ASID+1, because PCID 0 is special. * * uPCID - [2048 + 1, 2048 + TLB_NR_DYN_ASIDS] * for KPTI each mm has two address spaces and thus needs two * PCID values, but we can still do with a single ASID denomination * for each mm. Corresponds to kPCID + 2048. * */ /* There are 12 bits of space for ASIDS in CR3 */ #define CR3_HW_ASID_BITS 12 /* * When enabled, PAGE_TABLE_ISOLATION consumes a single bit for * user/kernel switches */ #ifdef CONFIG_PAGE_TABLE_ISOLATION # define PTI_CONSUMED_PCID_BITS 1 #else # define PTI_CONSUMED_PCID_BITS 0 #endif #define CR3_AVAIL_PCID_BITS (X86_CR3_PCID_BITS - PTI_CONSUMED_PCID_BITS) /* * ASIDs are zero-based: 0->MAX_AVAIL_ASID are valid. -1 below to account * for them being zero-based. Another -1 is because PCID 0 is reserved for * use by non-PCID-aware users. */ #define MAX_ASID_AVAILABLE ((1 << CR3_AVAIL_PCID_BITS) - 2) /* * 6 because 6 should be plenty and struct tlb_state will fit in two cache * lines. */ #define TLB_NR_DYN_ASIDS 6 /* * Given @asid, compute kPCID */ static inline u16 kern_pcid(u16 asid) { VM_WARN_ON_ONCE(asid > MAX_ASID_AVAILABLE); #ifdef CONFIG_PAGE_TABLE_ISOLATION /* * Make sure that the dynamic ASID space does not confict with the * bit we are using to switch between user and kernel ASIDs. */ BUILD_BUG_ON(TLB_NR_DYN_ASIDS >= (1 << X86_CR3_PTI_PCID_USER_BIT)); /* * The ASID being passed in here should have respected the * MAX_ASID_AVAILABLE and thus never have the switch bit set. */ VM_WARN_ON_ONCE(asid & (1 << X86_CR3_PTI_PCID_USER_BIT)); #endif /* * The dynamically-assigned ASIDs that get passed in are small * (<TLB_NR_DYN_ASIDS). They never have the high switch bit set, * so do not bother to clear it. * * If PCID is on, ASID-aware code paths put the ASID+1 into the * PCID bits. This serves two purposes. It prevents a nasty * situation in which PCID-unaware code saves CR3, loads some other * value (with PCID == 0), and then restores CR3, thus corrupting * the TLB for ASID 0 if the saved ASID was nonzero. It also means * that any bugs involving loading a PCID-enabled CR3 with * CR4.PCIDE off will trigger deterministically. */ return asid + 1; } /* * Given @asid, compute uPCID */ static inline u16 user_pcid(u16 asid) { u16 ret = kern_pcid(asid); #ifdef CONFIG_PAGE_TABLE_ISOLATION ret |= 1 << X86_CR3_PTI_PCID_USER_BIT; #endif return ret; } struct pgd_t; static inline unsigned long build_cr3(pgd_t *pgd, u16 asid) { if (static_cpu_has(X86_FEATURE_PCID)) { return __sme_pa(pgd) | kern_pcid(asid); } else { VM_WARN_ON_ONCE(asid != 0); return __sme_pa(pgd); } } static inline unsigned long build_cr3_noflush(pgd_t *pgd, u16 asid) { VM_WARN_ON_ONCE(asid > MAX_ASID_AVAILABLE); /* * Use boot_cpu_has() instead of this_cpu_has() as this function * might be called during early boot. This should work even after * boot because all CPU's the have same capabilities: */ VM_WARN_ON_ONCE(!boot_cpu_has(X86_FEATURE_PCID)); return __sme_pa(pgd) | kern_pcid(asid) | CR3_NOFLUSH; } #ifdef CONFIG_PARAVIRT #include <asm/paravirt.h> #else #define __flush_tlb() __native_flush_tlb() #define __flush_tlb_global() __native_flush_tlb_global() #define __flush_tlb_one_user(addr) __native_flush_tlb_one_user(addr) #endif static inline bool tlb_defer_switch_to_init_mm(void) { /* * If we have PCID, then switching to init_mm is reasonably * fast. If we don't have PCID, then switching to init_mm is * quite slow, so we try to defer it in the hopes that we can * avoid it entirely. The latter approach runs the risk of * receiving otherwise unnecessary IPIs. * * This choice is just a heuristic. The tlb code can handle this * function returning true or false regardless of whether we have * PCID. */ return !static_cpu_has(X86_FEATURE_PCID); } struct tlb_context { u64 ctx_id; u64 tlb_gen; }; struct tlb_state { /* * cpu_tlbstate.loaded_mm should match CR3 whenever interrupts * are on. This means that it may not match current->active_mm, * which will contain the previous user mm when we're in lazy TLB * mode even if we've already switched back to swapper_pg_dir. * * During switch_mm_irqs_off(), loaded_mm will be set to * LOADED_MM_SWITCHING during the brief interrupts-off window * when CR3 and loaded_mm would otherwise be inconsistent. This * is for nmi_uaccess_okay()'s benefit. */ struct mm_struct *loaded_mm; #define LOADED_MM_SWITCHING ((struct mm_struct *)1) /* Last user mm for optimizing IBPB */ union { struct mm_struct *last_user_mm; unsigned long last_user_mm_ibpb; }; u16 loaded_mm_asid; u16 next_asid; /* * We can be in one of several states: * * - Actively using an mm. Our CPU's bit will be set in * mm_cpumask(loaded_mm) and is_lazy == false; * * - Not using a real mm. loaded_mm == &init_mm. Our CPU's bit * will not be set in mm_cpumask(&init_mm) and is_lazy == false. * * - Lazily using a real mm. loaded_mm != &init_mm, our bit * is set in mm_cpumask(loaded_mm), but is_lazy == true. * We're heuristically guessing that the CR3 load we * skipped more than makes up for the overhead added by * lazy mode. */ bool is_lazy; /* * If set we changed the page tables in such a way that we * needed an invalidation of all contexts (aka. PCIDs / ASIDs). * This tells us to go invalidate all the non-loaded ctxs[] * on the next context switch. * * The current ctx was kept up-to-date as it ran and does not * need to be invalidated. */ bool invalidate_other; /* * Mask that contains TLB_NR_DYN_ASIDS+1 bits to indicate * the corresponding user PCID needs a flush next time we * switch to it; see SWITCH_TO_USER_CR3. */ unsigned short user_pcid_flush_mask; /* * Access to this CR4 shadow and to H/W CR4 is protected by * disabling interrupts when modifying either one. */ unsigned long cr4; /* * This is a list of all contexts that might exist in the TLB. * There is one per ASID that we use, and the ASID (what the * CPU calls PCID) is the index into ctxts. * * For each context, ctx_id indicates which mm the TLB's user * entries came from. As an invariant, the TLB will never * contain entries that are out-of-date as when that mm reached * the tlb_gen in the list. * * To be clear, this means that it's legal for the TLB code to * flush the TLB without updating tlb_gen. This can happen * (for now, at least) due to paravirt remote flushes. * * NB: context 0 is a bit special, since it's also used by * various bits of init code. This is fine -- code that * isn't aware of PCID will end up harmlessly flushing * context 0. */ struct tlb_context ctxs[TLB_NR_DYN_ASIDS]; }; DECLARE_PER_CPU_SHARED_ALIGNED(struct tlb_state, cpu_tlbstate); /* * Blindly accessing user memory from NMI context can be dangerous * if we're in the middle of switching the current user task or * switching the loaded mm. It can also be dangerous if we * interrupted some kernel code that was temporarily using a * different mm. */ static inline bool nmi_uaccess_okay(void) { struct mm_struct *loaded_mm = this_cpu_read(cpu_tlbstate.loaded_mm); struct mm_struct *current_mm = current->mm; VM_WARN_ON_ONCE(!loaded_mm); /* * The condition we want to check is * current_mm->pgd == __va(read_cr3_pa()). This may be slow, though, * if we're running in a VM with shadow paging, and nmi_uaccess_okay() * is supposed to be reasonably fast. * * Instead, we check the almost equivalent but somewhat conservative * condition below, and we rely on the fact that switch_mm_irqs_off() * sets loaded_mm to LOADED_MM_SWITCHING before writing to CR3. */ if (loaded_mm != current_mm) return false; VM_WARN_ON_ONCE(current_mm->pgd != __va(read_cr3_pa())); return true; } /* Initialize cr4 shadow for this CPU. */ static inline void cr4_init_shadow(void) { this_cpu_write(cpu_tlbstate.cr4, __read_cr4()); } static inline void __cr4_set(unsigned long cr4) { lockdep_assert_irqs_disabled(); this_cpu_write(cpu_tlbstate.cr4, cr4); __write_cr4(cr4); } /* Set in this cpu's CR4. */ static inline void cr4_set_bits(unsigned long mask) { unsigned long cr4, flags; local_irq_save(flags); cr4 = this_cpu_read(cpu_tlbstate.cr4); if ((cr4 | mask) != cr4) __cr4_set(cr4 | mask); local_irq_restore(flags); } /* Clear in this cpu's CR4. */ static inline void cr4_clear_bits(unsigned long mask) { unsigned long cr4, flags; local_irq_save(flags); cr4 = this_cpu_read(cpu_tlbstate.cr4); if ((cr4 & ~mask) != cr4) __cr4_set(cr4 & ~mask); local_irq_restore(flags); } static inline void cr4_toggle_bits_irqsoff(unsigned long mask) { unsigned long cr4; cr4 = this_cpu_read(cpu_tlbstate.cr4); __cr4_set(cr4 ^ mask); } /* Read the CR4 shadow. */ static inline unsigned long cr4_read_shadow(void) { return this_cpu_read(cpu_tlbstate.cr4); } /* * Mark all other ASIDs as invalid, preserves the current. */ static inline void invalidate_other_asid(void) { this_cpu_write(cpu_tlbstate.invalidate_other, true); } /* * Save some of cr4 feature set we're using (e.g. Pentium 4MB * enable and PPro Global page enable), so that any CPU's that boot * up after us can get the correct flags. This should only be used * during boot on the boot cpu. */ extern unsigned long mmu_cr4_features; extern u32 *trampoline_cr4_features; static inline void cr4_set_bits_and_update_boot(unsigned long mask) { mmu_cr4_features |= mask; if (trampoline_cr4_features) *trampoline_cr4_features = mmu_cr4_features; cr4_set_bits(mask); } extern void initialize_tlbstate_and_flush(void); /* * Given an ASID, flush the corresponding user ASID. We can delay this * until the next time we switch to it. * * See SWITCH_TO_USER_CR3. */ static inline void invalidate_user_asid(u16 asid) { /* There is no user ASID if address space separation is off */ if (!IS_ENABLED(CONFIG_PAGE_TABLE_ISOLATION)) return; /* * We only have a single ASID if PCID is off and the CR3 * write will have flushed it. */ if (!cpu_feature_enabled(X86_FEATURE_PCID)) return; if (!static_cpu_has(X86_FEATURE_PTI)) return; __set_bit(kern_pcid(asid), (unsigned long *)this_cpu_ptr(&cpu_tlbstate.user_pcid_flush_mask)); } /* * flush the entire current user mapping */ static inline void __native_flush_tlb(void) { /* * Preemption or interrupts must be disabled to protect the access * to the per CPU variable and to prevent being preempted between * read_cr3() and write_cr3(). */ WARN_ON_ONCE(preemptible()); invalidate_user_asid(this_cpu_read(cpu_tlbstate.loaded_mm_asid)); /* If current->mm == NULL then the read_cr3() "borrows" an mm */ native_write_cr3(__native_read_cr3()); } /* * flush everything */ static inline void __native_flush_tlb_global(void) { unsigned long cr4, flags; if (static_cpu_has(X86_FEATURE_INVPCID)) { /* * Using INVPCID is considerably faster than a pair of writes * to CR4 sandwiched inside an IRQ flag save/restore. * * Note, this works with CR4.PCIDE=0 or 1. */ invpcid_flush_all(); return; } /* * Read-modify-write to CR4 - protect it from preemption and * from interrupts. (Use the raw variant because this code can * be called from deep inside debugging code.) */ raw_local_irq_save(flags); cr4 = this_cpu_read(cpu_tlbstate.cr4); /* toggle PGE */ native_write_cr4(cr4 ^ X86_CR4_PGE); /* write old PGE again and flush TLBs */ native_write_cr4(cr4); raw_local_irq_restore(flags); } /* * flush one page in the user mapping */ static inline void __native_flush_tlb_one_user(unsigned long addr) { u32 loaded_mm_asid = this_cpu_read(cpu_tlbstate.loaded_mm_asid); asm volatile("invlpg (%0)" ::"r" (addr) : "memory"); if (!static_cpu_has(X86_FEATURE_PTI)) return; /* * Some platforms #GP if we call invpcid(type=1/2) before CR4.PCIDE=1. * Just use invalidate_user_asid() in case we are called early. */ if (!this_cpu_has(X86_FEATURE_INVPCID_SINGLE)) invalidate_user_asid(loaded_mm_asid); else invpcid_flush_one(user_pcid(loaded_mm_asid), addr); } /* * flush everything */ static inline void __flush_tlb_all(void) { /* * This is to catch users with enabled preemption and the PGE feature * and don't trigger the warning in __native_flush_tlb(). */ VM_WARN_ON_ONCE(preemptible()); if (boot_cpu_has(X86_FEATURE_PGE)) { __flush_tlb_global(); } else { /* * !PGE -> !PCID (setup_pcid()), thus every flush is total. */ __flush_tlb(); } } /* * flush one page in the kernel mapping */ static inline void __flush_tlb_one_kernel(unsigned long addr) { count_vm_tlb_event(NR_TLB_LOCAL_FLUSH_ONE); /* * If PTI is off, then __flush_tlb_one_user() is just INVLPG or its * paravirt equivalent. Even with PCID, this is sufficient: we only * use PCID if we also use global PTEs for the kernel mapping, and * INVLPG flushes global translations across all address spaces. * * If PTI is on, then the kernel is mapped with non-global PTEs, and * __flush_tlb_one_user() will flush the given address for the current * kernel address space and for its usermode counterpart, but it does * not flush it for other address spaces. */ __flush_tlb_one_user(addr); if (!static_cpu_has(X86_FEATURE_PTI)) return; /* * See above. We need to propagate the flush to all other address * spaces. In principle, we only need to propagate it to kernelmode * address spaces, but the extra bookkeeping we would need is not * worth it. */ invalidate_other_asid(); } #define TLB_FLUSH_ALL -1UL /* * TLB flushing: * * - flush_tlb_all() flushes all processes TLBs * - flush_tlb_mm(mm) flushes the specified mm context TLB's * - flush_tlb_page(vma, vmaddr) flushes one page * - flush_tlb_range(vma, start, end) flushes a range of pages * - flush_tlb_kernel_range(start, end) flushes a range of kernel pages * - flush_tlb_others(cpumask, info) flushes TLBs on other cpus * * ..but the i386 has somewhat limited tlb flushing capabilities, * and page-granular flushes are available only on i486 and up. */ struct flush_tlb_info { /* * We support several kinds of flushes. * * - Fully flush a single mm. .mm will be set, .end will be * TLB_FLUSH_ALL, and .new_tlb_gen will be the tlb_gen to * which the IPI sender is trying to catch us up. * * - Partially flush a single mm. .mm will be set, .start and * .end will indicate the range, and .new_tlb_gen will be set * such that the changes between generation .new_tlb_gen-1 and * .new_tlb_gen are entirely contained in the indicated range. * * - Fully flush all mms whose tlb_gens have been updated. .mm * will be NULL, .end will be TLB_FLUSH_ALL, and .new_tlb_gen * will be zero. */ struct mm_struct *mm; unsigned long start; unsigned long end; u64 new_tlb_gen; }; #define local_flush_tlb() __flush_tlb() #define flush_tlb_mm(mm) flush_tlb_mm_range(mm, 0UL, TLB_FLUSH_ALL, 0UL) #define flush_tlb_range(vma, start, end) \ flush_tlb_mm_range(vma->vm_mm, start, end, vma->vm_flags) extern void flush_tlb_all(void); extern void flush_tlb_mm_range(struct mm_struct *mm, unsigned long start, unsigned long end, unsigned long vmflag); extern void flush_tlb_kernel_range(unsigned long start, unsigned long end); static inline void flush_tlb_page(struct vm_area_struct *vma, unsigned long a) { flush_tlb_mm_range(vma->vm_mm, a, a + PAGE_SIZE, VM_NONE); } void native_flush_tlb_others(const struct cpumask *cpumask, const struct flush_tlb_info *info); static inline u64 inc_mm_tlb_gen(struct mm_struct *mm) { /* * Bump the generation count. This also serves as a full barrier * that synchronizes with switch_mm(): callers are required to order * their read of mm_cpumask after their writes to the paging * structures. */ return atomic64_inc_return(&mm->context.tlb_gen); } static inline void arch_tlbbatch_add_mm(struct arch_tlbflush_unmap_batch *batch, struct mm_struct *mm) { inc_mm_tlb_gen(mm); cpumask_or(&batch->cpumask, &batch->cpumask, mm_cpumask(mm)); } extern void arch_tlbbatch_flush(struct arch_tlbflush_unmap_batch *batch); #ifndef CONFIG_PARAVIRT #define flush_tlb_others(mask, info) \ native_flush_tlb_others(mask, info) #endif #endif /* _ASM_X86_TLBFLUSH_H */