Blame view
arch/Kconfig
29.5 KB
b24413180
|
1 |
# SPDX-License-Identifier: GPL-2.0 |
fb32e03fd
|
2 3 4 |
# # General architecture dependent options # |
125e56458
|
5 |
|
692f66f26
|
6 7 |
config CRASH_CORE bool |
2965faa5e
|
8 |
config KEXEC_CORE |
692f66f26
|
9 |
select CRASH_CORE |
2965faa5e
|
10 |
bool |
467d27824
|
11 12 |
config HAVE_IMA_KEXEC bool |
c5ac43ee8
|
13 14 |
config HOTPLUG_SMT bool |
125e56458
|
15 |
config OPROFILE |
b309a294e
|
16 |
tristate "OProfile system profiling" |
125e56458
|
17 18 |
depends on PROFILING depends on HAVE_OPROFILE |
d69d59f49
|
19 |
select RING_BUFFER |
9a5963eb4
|
20 |
select RING_BUFFER_ALLOW_SWAP |
125e56458
|
21 22 23 24 25 26 |
help OProfile is a profiling system capable of profiling the whole system, include the kernel, kernel modules, libraries, and applications. If unsure, say N. |
4d4036e0e
|
27 28 29 30 31 32 33 34 |
config OPROFILE_EVENT_MULTIPLEX bool "OProfile multiplexing support (EXPERIMENTAL)" default n depends on OPROFILE && X86 help The number of hardware counters is limited. The multiplexing feature enables OProfile to gather more events than counters are provided by the hardware. This is realized by switching |
9332ef9db
|
35 |
between events at a user specified time interval. |
4d4036e0e
|
36 37 |
If unsure, say N. |
125e56458
|
38 |
config HAVE_OPROFILE |
9ba16087d
|
39 |
bool |
125e56458
|
40 |
|
dcfce4a09
|
41 42 |
config OPROFILE_NMI_TIMER def_bool y |
af9feebe6
|
43 |
depends on PERF_EVENTS && HAVE_PERF_EVENTS_NMI && !PPC64 |
dcfce4a09
|
44 |
|
125e56458
|
45 46 |
config KPROBES bool "Kprobes" |
05ed160e8
|
47 |
depends on MODULES |
125e56458
|
48 |
depends on HAVE_KPROBES |
05ed160e8
|
49 |
select KALLSYMS |
125e56458
|
50 51 52 53 54 55 |
help Kprobes allows you to trap at almost any kernel address and execute a callback function. register_kprobe() establishes a probepoint and specifies the callback. Kprobes is useful for kernel debugging, non-intrusive instrumentation and testing. If in doubt, say "N". |
45f81b1c9
|
56 |
config JUMP_LABEL |
c5905afb0
|
57 |
bool "Optimize very unlikely/likely branches" |
45f81b1c9
|
58 59 |
depends on HAVE_ARCH_JUMP_LABEL help |
c5905afb0
|
60 61 62 63 64 65 66 |
This option enables a transparent branch optimization that makes certain almost-always-true or almost-always-false branch conditions even cheaper to execute within the kernel. Certain performance-sensitive kernel code, such as trace points, scheduler functionality, networking code and KVM have such branches and include support for this optimization technique. |
45f81b1c9
|
67 |
If it is detected that the compiler has support for "asm goto", |
c5905afb0
|
68 69 70 71 72 73 74 75 |
the kernel will compile such branches with just a nop instruction. When the condition flag is toggled to true, the nop will be converted to a jump instruction to execute the conditional block of instructions. This technique lowers overhead and stress on the branch prediction of the processor and generally makes the kernel faster. The update of the condition is slower, but those are always very rare. |
45f81b1c9
|
76 |
|
c5905afb0
|
77 78 |
( On 32-bit x86, the necessary options added to the compiler flags may increase the size of the kernel slightly. ) |
45f81b1c9
|
79 |
|
1987c947d
|
80 81 82 83 84 |
config STATIC_KEYS_SELFTEST bool "Static key selftest" depends on JUMP_LABEL help Boot time self-test of the branch patching code. |
afd66255b
|
85 |
config OPTPROBES |
5cc718b9d
|
86 87 |
def_bool y depends on KPROBES && HAVE_OPTPROBES |
5715de464
|
88 |
select TASKS_RCU if PREEMPT |
afd66255b
|
89 |
|
e7dbfe349
|
90 91 92 93 94 95 96 97 |
config KPROBES_ON_FTRACE def_bool y depends on KPROBES && HAVE_KPROBES_ON_FTRACE depends on DYNAMIC_FTRACE_WITH_REGS help If function tracer is enabled and the arch supports full passing of pt_regs to function tracing, then kprobes can optimize on top of function tracing. |
2b1444983
|
98 |
config UPROBES |
09294e31b
|
99 |
def_bool n |
e8f4aa608
|
100 |
depends on ARCH_SUPPORTS_UPROBES |
2b1444983
|
101 |
help |
7b2d81d48
|
102 103 104 105 106 107 108 109 110 |
Uprobes is the user-space counterpart to kprobes: they enable instrumentation applications (such as 'perf probe') to establish unintrusive probes in user-space binaries and libraries, by executing handler functions when the probes are hit by user-space applications. ( These probes come in the form of single-byte breakpoints, managed by the kernel and kept transparent to the probed application. ) |
2b1444983
|
111 |
|
c19fa94a8
|
112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 |
config HAVE_64BIT_ALIGNED_ACCESS def_bool 64BIT && !HAVE_EFFICIENT_UNALIGNED_ACCESS help Some architectures require 64 bit accesses to be 64 bit aligned, which also requires structs containing 64 bit values to be 64 bit aligned too. This includes some 32 bit architectures which can do 64 bit accesses, as well as 64 bit architectures without unaligned access. This symbol should be selected by an architecture if 64 bit accesses are required to be 64 bit aligned in this way even though it is not a 64 bit architecture. See Documentation/unaligned-memory-access.txt for more information on the topic of unaligned memory accesses. |
58340a07c
|
127 |
config HAVE_EFFICIENT_UNALIGNED_ACCESS |
9ba16087d
|
128 |
bool |
58340a07c
|
129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 |
help Some architectures are unable to perform unaligned accesses without the use of get_unaligned/put_unaligned. Others are unable to perform such accesses efficiently (e.g. trap on unaligned access and require fixing it up in the exception handler.) This symbol should be selected by an architecture if it can perform unaligned accesses efficiently to allow different code paths to be selected for these cases. Some network drivers, for example, could opt to not fix up alignment problems with received packets if doing so would not help much. See Documentation/unaligned-memory-access.txt for more information on the topic of unaligned memory accesses. |
cf66bb93e
|
145 146 147 148 149 150 151 152 153 154 155 156 157 158 159 160 161 162 |
config ARCH_USE_BUILTIN_BSWAP bool help Modern versions of GCC (since 4.4) have builtin functions for handling byte-swapping. Using these, instead of the old inline assembler that the architecture code provides in the __arch_bswapXX() macros, allows the compiler to see what's happening and offers more opportunity for optimisation. In particular, the compiler will be able to combine the byteswap with a nearby load or store and use load-and-swap or store-and-swap instructions if the architecture has them. It should almost *never* result in code which is worse than the hand-coded assembler in <asm/swab.h>. But just in case it does, the use of the builtins is optional. Any architecture with load-and-swap or store-and-swap instructions should set this. And it shouldn't hurt to set it on architectures that don't have such instructions. |
9edddaa20
|
163 164 165 |
config KRETPROBES def_bool y depends on KPROBES && HAVE_KRETPROBES |
7c68af6e3
|
166 167 168 169 170 171 |
config USER_RETURN_NOTIFIER bool depends on HAVE_USER_RETURN_NOTIFIER help Provide a kernel-internal notification when a cpu is about to switch to user mode. |
28b2ee20c
|
172 |
config HAVE_IOREMAP_PROT |
9ba16087d
|
173 |
bool |
28b2ee20c
|
174 |
|
125e56458
|
175 |
config HAVE_KPROBES |
9ba16087d
|
176 |
bool |
9edddaa20
|
177 178 |
config HAVE_KRETPROBES |
9ba16087d
|
179 |
bool |
74bc7ceeb
|
180 |
|
afd66255b
|
181 182 |
config HAVE_OPTPROBES bool |
d314d74c6
|
183 |
|
e7dbfe349
|
184 185 |
config HAVE_KPROBES_ON_FTRACE bool |
42a0bb3f7
|
186 187 |
config HAVE_NMI bool |
1f5a4ad97
|
188 189 190 191 192 193 |
# # An arch should select this if it provides all these things: # # task_pt_regs() in asm/processor.h or asm/ptrace.h # arch_has_single_step() if there is hardware single-step support # arch_has_block_step() if there is hardware block-step support |
1f5a4ad97
|
194 195 196 197 198 199 200 201 |
# asm/syscall.h supplying asm-generic/syscall.h interface # linux/regset.h user_regset interfaces # CORE_DUMP_USE_REGSET #define'd in linux/elf.h # TIF_SYSCALL_TRACE calls tracehook_report_syscall_{entry,exit} # TIF_NOTIFY_RESUME calls tracehook_notify_resume() # signal delivery calls tracehook_signal_handler() # config HAVE_ARCH_TRACEHOOK |
9ba16087d
|
202 |
bool |
1f5a4ad97
|
203 |
|
c64be2bb1
|
204 205 |
config HAVE_DMA_CONTIGUOUS bool |
29d5e0476
|
206 207 |
config GENERIC_SMP_IDLE_THREAD bool |
485cf5dac
|
208 209 |
config GENERIC_IDLE_POLL_SETUP bool |
6974f0c45
|
210 211 212 213 214 |
config ARCH_HAS_FORTIFY_SOURCE bool help An architecture should select this when it can successfully build and run with CONFIG_FORTIFY_SOURCE. |
d2852a224
|
215 216 217 |
# Select if arch has all set_memory_ro/rw/x/nx() functions in asm/cacheflush.h config ARCH_HAS_SET_MEMORY bool |
a6359d1ee
|
218 219 |
# Select if arch init_task initializer is different to init/init_task.c config ARCH_INIT_TASK |
a4a2eb490
|
220 |
bool |
f5e102873
|
221 222 223 |
# Select if arch has its private alloc_task_struct() function config ARCH_TASK_STRUCT_ALLOCATOR bool |
b235beea9
|
224 225 |
# Select if arch has its private alloc_thread_stack() function config ARCH_THREAD_STACK_ALLOCATOR |
f5e102873
|
226 |
bool |
5aaeb5c01
|
227 228 229 |
# Select if arch wants to size task_struct dynamically via arch_task_struct_size: config ARCH_WANTS_DYNAMIC_TASK_STRUCT bool |
f850c30c8
|
230 231 |
config HAVE_REGS_AND_STACK_ACCESS_API bool |
e01292b1f
|
232 233 234 235 236 |
help This symbol should be selected by an architecure if it supports the API needed to access registers and stack entries from pt_regs, declared in asm/ptrace.h For example the kprobes-based event tracer needs this API. |
f850c30c8
|
237 |
|
9483a578d
|
238 |
config HAVE_CLK |
9ba16087d
|
239 |
bool |
9483a578d
|
240 241 242 |
help The <linux/clk.h> calls support software clock gating and thus are a key power management tool on many systems. |
5ee00bd46
|
243 244 |
config HAVE_DMA_API_DEBUG bool |
36cd3c9f9
|
245 |
|
62a038d34
|
246 247 |
config HAVE_HW_BREAKPOINT bool |
99e8c5a3b
|
248 |
depends on PERF_EVENTS |
62a038d34
|
249 |
|
0102752e4
|
250 251 252 253 254 255 256 257 258 259 |
config HAVE_MIXED_BREAKPOINTS_REGS bool depends on HAVE_HW_BREAKPOINT help Depending on the arch implementation of hardware breakpoints, some of them have separate registers for data and instruction breakpoints addresses, others have mixed registers to store them but define the access type in a control register. Select this option if your arch implements breakpoints under the latter fashion. |
7c68af6e3
|
260 261 |
config HAVE_USER_RETURN_NOTIFIER bool |
a1922ed66
|
262 |
|
c01d43233
|
263 264 |
config HAVE_PERF_EVENTS_NMI bool |
23637d477
|
265 266 267 268 |
help System hardware can generate an NMI using the perf event subsystem. Also has support for calculating CPU cycle events to determine how many clock cycles in a given period. |
c01d43233
|
269 |
|
05a4a9527
|
270 271 272 273 274 275 276 277 278 279 280 281 282 283 284 285 286 287 288 289 290 |
config HAVE_HARDLOCKUP_DETECTOR_PERF bool depends on HAVE_PERF_EVENTS_NMI help The arch chooses to use the generic perf-NMI-based hardlockup detector. Must define HAVE_PERF_EVENTS_NMI. config HAVE_NMI_WATCHDOG depends on HAVE_NMI bool help The arch provides a low level NMI watchdog. It provides asm/nmi.h, and defines its own arch_touch_nmi_watchdog(). config HAVE_HARDLOCKUP_DETECTOR_ARCH bool select HAVE_NMI_WATCHDOG help The arch chooses to provide its own hardlockup detector, which is a superset of the HAVE_NMI_WATCHDOG. It also conforms to config interfaces and parameters provided by hardlockup detector subsystem. |
c5e63197d
|
291 292 293 294 295 |
config HAVE_PERF_REGS bool help Support selective register dumps for perf events. This includes bit-mapping of each registers and a unique architecture id. |
c5ebcedb5
|
296 297 298 299 300 301 |
config HAVE_PERF_USER_STACK_DUMP bool help Support user stack dumps for perf event samples. This needs access to the user stack pointer which is not unified across architectures. |
bf5438fca
|
302 303 |
config HAVE_ARCH_JUMP_LABEL bool |
267239116
|
304 305 |
config HAVE_RCU_TABLE_FREE bool |
e9afa7c1e
|
306 307 |
config HAVE_RCU_TABLE_INVALIDATE bool |
df013ffb8
|
308 309 |
config ARCH_HAVE_NMI_SAFE_CMPXCHG bool |
43570fd2f
|
310 311 312 313 314 315 316 |
config HAVE_ALIGNED_STRUCT_PAGE bool help This makes sure that struct pages are double word aligned and that e.g. the SLUB allocator can perform double word atomic operations on a struct page for better performance. However selecting this might increase the size of a struct page by a word. |
4156153c4
|
317 318 |
config HAVE_CMPXCHG_LOCAL bool |
2565409fc
|
319 320 |
config HAVE_CMPXCHG_DOUBLE bool |
77e584968
|
321 322 |
config ARCH_WEAK_RELEASE_ACQUIRE bool |
c1d7e01d7
|
323 324 325 326 327 |
config ARCH_WANT_IPC_PARSE_VERSION bool config ARCH_WANT_COMPAT_IPC_PARSE_VERSION bool |
48b25c43e
|
328 |
config ARCH_WANT_OLD_COMPAT_IPC |
c1d7e01d7
|
329 |
select ARCH_WANT_COMPAT_IPC_PARSE_VERSION |
48b25c43e
|
330 |
bool |
e2cfabdfd
|
331 332 333 |
config HAVE_ARCH_SECCOMP_FILTER bool help |
fb0fadf9b
|
334 |
An arch should select this symbol if it provides all of these things: |
bb6ea4301
|
335 336 337 338 |
- syscall_get_arch() - syscall_get_arguments() - syscall_rollback() - syscall_set_return_value() |
fb0fadf9b
|
339 340 341 342 |
- SIGSYS siginfo_t support - secure_computing is called from a ptrace_event()-safe context - secure_computing return value is checked and a return value of -1 results in the system call being skipped immediately. |
48dc92b9f
|
343 |
- seccomp syscall wired up |
e2cfabdfd
|
344 345 346 347 348 349 350 351 352 353 |
config SECCOMP_FILTER def_bool y depends on HAVE_ARCH_SECCOMP_FILTER && SECCOMP && NET help Enable tasks to build secure computing environments defined in terms of Berkeley Packet Filter programs which implement task-defined system call filtering polices. See Documentation/prctl/seccomp_filter.txt for details. |
6b90bd4ba
|
354 355 356 357 358 359 360 361 362 |
config HAVE_GCC_PLUGINS bool help An arch should select this symbol if it supports building with GCC plugins. menuconfig GCC_PLUGINS bool "GCC plugins" depends on HAVE_GCC_PLUGINS |
a519167e7
|
363 |
depends on !COMPILE_TEST |
6b90bd4ba
|
364 365 366 367 368 |
help GCC plugins are loadable modules that provide extra features to the compiler. They are useful for runtime instrumentation and static analysis. See Documentation/gcc-plugins.txt for details. |
0dae776c6
|
369 |
config GCC_PLUGIN_CYC_COMPLEXITY |
215e2aa6c
|
370 |
bool "Compute the cyclomatic complexity of a function" if EXPERT |
0dae776c6
|
371 |
depends on GCC_PLUGINS |
215e2aa6c
|
372 |
depends on !COMPILE_TEST |
0dae776c6
|
373 374 375 376 377 378 379 380 |
help The complexity M of a function's control flow graph is defined as: M = E - N + 2P where E = the number of edges N = the number of nodes P = the number of connected components (exit nodes). |
215e2aa6c
|
381 382 383 |
Enabling this plugin reports the complexity to stderr during the build. It mainly serves as a simple example of how to create a gcc plugin for the kernel. |
543c37cb1
|
384 385 386 387 388 389 390 391 |
config GCC_PLUGIN_SANCOV bool depends on GCC_PLUGINS help This plugin inserts a __sanitizer_cov_trace_pc() call at the start of basic blocks. It supports all gcc versions with plugin support (from gcc-4.5 on). It is based on the commit "Add fuzzing coverage support" by Dmitry Vyukov <dvyukov@google.com>. |
38addce8b
|
392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 |
config GCC_PLUGIN_LATENT_ENTROPY bool "Generate some entropy during boot and runtime" depends on GCC_PLUGINS help By saying Y here the kernel will instrument some kernel code to extract some entropy from both original and artificially created program state. This will help especially embedded systems where there is little 'natural' source of entropy normally. The cost is some slowdown of the boot process (about 0.5%) and fork and irq processing. Note that entropy extracted this way is not cryptographically secure! This plugin was ported from grsecurity/PaX. More information at: * https://grsecurity.net/ * https://pax.grsecurity.net/ |
c61f13eaa
|
409 410 411 412 |
config GCC_PLUGIN_STRUCTLEAK bool "Force initialization of variables containing userspace addresses" depends on GCC_PLUGINS help |
f136e090c
|
413 |
This plugin zero-initializes any structures containing a |
c61f13eaa
|
414 415 416 417 418 419 |
__user attribute. This can prevent some classes of information exposures. This plugin was ported from grsecurity/PaX. More information at: * https://grsecurity.net/ * https://pax.grsecurity.net/ |
f7dd25078
|
420 421 422 423 424 425 |
config GCC_PLUGIN_STRUCTLEAK_BYREF_ALL bool "Force initialize all struct type variables passed by reference" depends on GCC_PLUGIN_STRUCTLEAK help Zero initialize any struct type local variable that may be passed by reference without having been initialized. |
c61f13eaa
|
426 427 428 429 430 431 432 433 434 |
config GCC_PLUGIN_STRUCTLEAK_VERBOSE bool "Report forcefully initialized variables" depends on GCC_PLUGIN_STRUCTLEAK depends on !COMPILE_TEST help This option will cause a warning to be printed each time the structleak plugin finds a variable it thinks needs to be initialized. Since not all existing initializers are detected by the plugin, this can produce false positive warnings. |
313dd1b62
|
435 436 437 438 439 |
config GCC_PLUGIN_RANDSTRUCT bool "Randomize layout of sensitive kernel structures" depends on GCC_PLUGINS select MODVERSIONS if MODULES help |
9225331b3
|
440 441 442 443 444 445 446 |
If you say Y here, the layouts of structures that are entirely function pointers (and have not been manually annotated with __no_randomize_layout), or structures that have been explicitly marked with __randomize_layout, will be randomized at compile-time. This can introduce the requirement of an additional information exposure vulnerability for exploits targeting these structure types. |
313dd1b62
|
447 448 449 450 451 452 453 454 455 456 457 458 459 460 461 462 463 464 465 466 467 468 469 470 471 472 473 474 |
Enabling this feature will introduce some performance impact, slightly increase memory usage, and prevent the use of forensic tools like Volatility against the system (unless the kernel source tree isn't cleaned after kernel installation). The seed used for compilation is located at scripts/gcc-plgins/randomize_layout_seed.h. It remains after a make clean to allow for external modules to be compiled with the existing seed and will be removed by a make mrproper or make distclean. Note that the implementation requires gcc 4.7 or newer. This plugin was ported from grsecurity/PaX. More information at: * https://grsecurity.net/ * https://pax.grsecurity.net/ config GCC_PLUGIN_RANDSTRUCT_PERFORMANCE bool "Use cacheline-aware structure randomization" depends on GCC_PLUGIN_RANDSTRUCT depends on !COMPILE_TEST help If you say Y here, the RANDSTRUCT randomization will make a best effort at restricting randomization to cacheline-sized groups of elements. It will further not randomize bitfields in structures. This reduces the performance hit of RANDSTRUCT at the cost of weakened randomization. |
19952a920
|
475 476 477 478 479 480 481 482 |
config HAVE_CC_STACKPROTECTOR bool help An arch should select this symbol if: - its compiler supports the -fstack-protector option - it has implemented a stack canary (e.g. __stack_chk_guard) config CC_STACKPROTECTOR |
8779657d2
|
483 484 485 486 487 488 489 |
def_bool n help Set when a stack-protector mode is enabled, so that the build can enable kernel-side support for the GCC feature. choice prompt "Stack Protector buffer overflow detection" |
19952a920
|
490 |
depends on HAVE_CC_STACKPROTECTOR |
8779657d2
|
491 |
default CC_STACKPROTECTOR_NONE |
19952a920
|
492 |
help |
8779657d2
|
493 |
This option turns on the "stack-protector" GCC feature. This |
19952a920
|
494 495 496 497 498 499 |
feature puts, at the beginning of functions, a canary value on the stack just before the return address, and validates the value just before actually returning. Stack based buffer overflows (that need to overwrite this return address) now also overwrite the canary, which gets detected and the attack is then neutralized via a kernel panic. |
8779657d2
|
500 501 502 503 504 505 506 507 508 509 510 |
config CC_STACKPROTECTOR_NONE bool "None" help Disable "stack-protector" GCC feature. config CC_STACKPROTECTOR_REGULAR bool "Regular" select CC_STACKPROTECTOR help Functions will have the stack-protector canary logic added if they have an 8-byte or larger character array on the stack. |
19952a920
|
511 |
This feature requires gcc version 4.2 or above, or a distribution |
8779657d2
|
512 513 514 515 516 517 518 519 520 521 522 523 524 525 526 527 528 529 530 531 532 533 534 535 536 537 538 |
gcc with the feature backported ("-fstack-protector"). On an x86 "defconfig" build, this feature adds canary checks to about 3% of all kernel functions, which increases kernel code size by about 0.3%. config CC_STACKPROTECTOR_STRONG bool "Strong" select CC_STACKPROTECTOR help Functions will have the stack-protector canary logic added in any of the following conditions: - local variable's address used as part of the right hand side of an assignment or function argument - local variable is an array (or union containing an array), regardless of array type or length - uses register local variables This feature requires gcc version 4.9 or above, or a distribution gcc with the feature backported ("-fstack-protector-strong"). On an x86 "defconfig" build, this feature adds canary checks to about 20% of all kernel functions, which increases the kernel code size by about 2%. endchoice |
19952a920
|
539 |
|
a5967db9a
|
540 |
config THIN_ARCHIVES |
799c43415
|
541 |
def_bool y |
a5967db9a
|
542 543 544 |
help Select this if the architecture wants to use thin archives instead of ld -r to create the built-in.o files. |
b67067f11
|
545 546 547 548 549 550 551 552 553 554 555 |
config LD_DEAD_CODE_DATA_ELIMINATION bool help Select this if the architecture wants to do dead code and data elimination with the linker by compiling with -ffunction-sections -fdata-sections and linking with --gc-sections. This requires that the arch annotates or otherwise protects its external entry points from being discarded. Linker scripts must also merge .text.*, .data.*, and .bss.* correctly into |
0f4c4af06
|
556 557 558 |
output sections. Care must be taken not to pull in unrelated sections (e.g., '.text.init'). Typically '.' in section names is used to distinguish them from label names / C identifiers. |
b67067f11
|
559 |
|
0f60a8efe
|
560 561 562 563 564 565 566 567 |
config HAVE_ARCH_WITHIN_STACK_FRAMES bool help An architecture should select this if it can walk the kernel stack frames to determine if an object is part of either the arguments or local variables (i.e. that it excludes saved return addresses, and similar) by implementing an inline arch_within_stack_frames(), which is used by CONFIG_HARDENED_USERCOPY. |
91d1aa43d
|
568 |
config HAVE_CONTEXT_TRACKING |
2b1d5024e
|
569 570 |
bool help |
91d1aa43d
|
571 572 573 574 575 576 577 |
Provide kernel/user boundaries probes necessary for subsystems that need it, such as userspace RCU extended quiescent state. Syscalls need to be wrapped inside user_exit()-user_enter() through the slow path using TIF_NOHZ flag. Exceptions handlers must be wrapped as well. Irqs are already protected inside rcu_irq_enter/rcu_irq_exit() but preemption or signal handling on irq exit still need to be protected. |
2b1d5024e
|
578 |
|
b952741c8
|
579 580 |
config HAVE_VIRT_CPU_ACCOUNTING bool |
40565b5ae
|
581 582 |
config ARCH_HAS_SCALED_CPUTIME bool |
554b0004d
|
583 584 585 586 587 588 589 590 591 592 |
config HAVE_VIRT_CPU_ACCOUNTING_GEN bool default y if 64BIT help With VIRT_CPU_ACCOUNTING_GEN, cputime_t becomes 64-bit. Before enabling this option, arch code must be audited to ensure there are no races in concurrent read/write of cputime_t. For example, reading/writing 64-bit cputime_t on some 32-bit arches may require multiple accesses, so proper locking is needed to protect against concurrent accesses. |
fdf9c3565
|
593 594 595 596 597 |
config HAVE_IRQ_TIME_ACCOUNTING bool help Archs need to ensure they use a high enough resolution clock to support irq time accounting and then call enable_sched_clock_irqtime(). |
15626062f
|
598 599 |
config HAVE_ARCH_TRANSPARENT_HUGEPAGE bool |
a00cc7d9d
|
600 601 |
config HAVE_ARCH_TRANSPARENT_HUGEPAGE_PUD bool |
0ddab1d2e
|
602 603 |
config HAVE_ARCH_HUGE_VMAP bool |
0f8975ec4
|
604 605 |
config HAVE_ARCH_SOFT_DIRTY bool |
786d35d45
|
606 607 608 609 610 611 612 613 614 615 616 617 618 619 620 621 622 623 |
config HAVE_MOD_ARCH_SPECIFIC bool help The arch uses struct mod_arch_specific to store data. Many arches just need a simple module loader without arch specific data - those should not enable this. config MODULES_USE_ELF_RELA bool help Modules only use ELF RELA relocations. Modules with ELF REL relocations will give an error. config MODULES_USE_ELF_REL bool help Modules only use ELF REL relocations. Modules with ELF RELA relocations will give an error. |
b92021b09
|
624 625 626 627 628 |
config HAVE_UNDERSCORE_SYMBOL_PREFIX bool help Some architectures generate an _ in front of C symbols; things like module loading and assembly files need to know about this. |
cc1f02745
|
629 630 631 632 633 634 635 636 637 |
config HAVE_IRQ_EXIT_ON_IRQ_STACK bool help Architecture doesn't only execute the irq handler on the irq stack but also irq_exit(). This way we can process softirqs on this irq stack instead of switching to a new one when we call __do_softirq() in the end of an hardirq. This spares a stack switch and improves cache usage on softirq processing. |
235a8f028
|
638 639 640 |
config PGTABLE_LEVELS int default 2 |
2b68f6cae
|
641 642 643 644 645 646 |
config ARCH_HAS_ELF_RANDOMIZE bool help An architecture supports choosing randomized locations for stack, mmap, brk, and ET_DYN. Defined functions: - arch_mmap_rnd() |
204db6ed1
|
647 |
- arch_randomize_brk() |
2b68f6cae
|
648 |
|
d07e22597
|
649 650 651 652 653 654 655 656 |
config HAVE_ARCH_MMAP_RND_BITS bool help An arch should select this symbol if it supports setting a variable number of bits for use in establishing the base address for mmap allocations, has MMU enabled and provides values for both: - ARCH_MMAP_RND_BITS_MIN - ARCH_MMAP_RND_BITS_MAX |
5f56a5dfd
|
657 658 659 660 |
config HAVE_EXIT_THREAD bool help An architecture implements exit_thread. |
d07e22597
|
661 662 663 664 665 666 667 668 669 670 671 672 673 674 675 676 677 678 679 680 681 682 683 684 685 686 687 688 689 690 691 692 693 694 695 696 697 698 699 700 701 702 703 704 705 706 707 708 709 710 711 712 713 714 715 716 717 718 |
config ARCH_MMAP_RND_BITS_MIN int config ARCH_MMAP_RND_BITS_MAX int config ARCH_MMAP_RND_BITS_DEFAULT int config ARCH_MMAP_RND_BITS int "Number of bits to use for ASLR of mmap base address" if EXPERT range ARCH_MMAP_RND_BITS_MIN ARCH_MMAP_RND_BITS_MAX default ARCH_MMAP_RND_BITS_DEFAULT if ARCH_MMAP_RND_BITS_DEFAULT default ARCH_MMAP_RND_BITS_MIN depends on HAVE_ARCH_MMAP_RND_BITS help This value can be used to select the number of bits to use to determine the random offset to the base address of vma regions resulting from mmap allocations. This value will be bounded by the architecture's minimum and maximum supported values. This value can be changed after boot using the /proc/sys/vm/mmap_rnd_bits tunable config HAVE_ARCH_MMAP_RND_COMPAT_BITS bool help An arch should select this symbol if it supports running applications in compatibility mode, supports setting a variable number of bits for use in establishing the base address for mmap allocations, has MMU enabled and provides values for both: - ARCH_MMAP_RND_COMPAT_BITS_MIN - ARCH_MMAP_RND_COMPAT_BITS_MAX config ARCH_MMAP_RND_COMPAT_BITS_MIN int config ARCH_MMAP_RND_COMPAT_BITS_MAX int config ARCH_MMAP_RND_COMPAT_BITS_DEFAULT int config ARCH_MMAP_RND_COMPAT_BITS int "Number of bits to use for ASLR of mmap base address for compatible applications" if EXPERT range ARCH_MMAP_RND_COMPAT_BITS_MIN ARCH_MMAP_RND_COMPAT_BITS_MAX default ARCH_MMAP_RND_COMPAT_BITS_DEFAULT if ARCH_MMAP_RND_COMPAT_BITS_DEFAULT default ARCH_MMAP_RND_COMPAT_BITS_MIN depends on HAVE_ARCH_MMAP_RND_COMPAT_BITS help This value can be used to select the number of bits to use to determine the random offset to the base address of vma regions resulting from mmap allocations for compatible applications This value will be bounded by the architecture's minimum and maximum supported values. This value can be changed after boot using the /proc/sys/vm/mmap_rnd_compat_bits tunable |
1b028f784
|
719 720 721 722 723 724 |
config HAVE_ARCH_COMPAT_MMAP_BASES bool help This allows 64bit applications to invoke 32-bit mmap() syscall and vice-versa 32-bit applications to call 64-bit mmap(). Required for applications doing different bitness syscalls. |
3033f14ab
|
725 726 727 728 729 730 |
config HAVE_COPY_THREAD_TLS bool help Architecture provides copy_thread_tls to accept tls argument via normal C parameter passing, rather than extracting the syscall argument from pt_regs. |
b9ab5ebb1
|
731 732 733 734 735 |
config HAVE_STACK_VALIDATION bool help Architecture supports the 'objtool check' host tool command, which performs compile-time stack metadata validation. |
af085d908
|
736 737 738 739 740 |
config HAVE_RELIABLE_STACKTRACE bool help Architecture has a save_stack_trace_tsk_reliable() function which only returns a stack trace if it can guarantee the trace is reliable. |
468a94285
|
741 742 743 744 745 746 747 |
config HAVE_ARCH_HASH bool default n help If this is set, the architecture provides an <asm/hash.h> file which provides platform-specific implementations of some functions in <linux/hash.h> or fs/namei.c. |
3a4955111
|
748 749 |
config ISA_BUS_API def_bool ISA |
d2125043a
|
750 751 752 753 754 755 756 757 758 759 760 761 762 |
# # ABI hall of shame # config CLONE_BACKWARDS bool help Architecture has tls passed as the 4th argument of clone(2), not the 5th one. config CLONE_BACKWARDS2 bool help Architecture has the first two arguments of clone(2) swapped. |
dfa9771a7
|
763 764 765 766 767 |
config CLONE_BACKWARDS3 bool help Architecture has tls passed as the 3rd argument of clone(2), not the 5th one. |
eaca6eae3
|
768 769 770 771 |
config ODD_RT_SIGACTION bool help Architecture has unusual rt_sigaction(2) arguments |
0a0e8cdf7
|
772 773 774 775 776 777 778 779 780 |
config OLD_SIGSUSPEND bool help Architecture has old sigsuspend(2) syscall, of one-argument variety config OLD_SIGSUSPEND3 bool help Even weirder antique ABI - three-argument sigsuspend(2) |
495dfbf76
|
781 782 783 784 785 786 787 788 789 790 |
config OLD_SIGACTION bool help Architecture has old sigaction(2) syscall. Nope, not the same as OLD_SIGSUSPEND | OLD_SIGSUSPEND3 - alpha has sigsuspend(2), but fairly different variant of sigaction(2), thanks to OSF/1 compatibility... config COMPAT_OLD_SIGACTION bool |
0d4a619b6
|
791 792 |
config ARCH_NO_COHERENT_DMA_MMAP bool |
fff7fb0b2
|
793 794 |
config CPU_NO_EFFICIENT_FFS def_bool n |
ba14a194a
|
795 796 797 798 799 800 801 802 803 804 805 806 807 808 809 810 811 812 813 814 815 816 817 818 819 820 821 822 823 824 825 826 827 |
config HAVE_ARCH_VMAP_STACK def_bool n help An arch should select this symbol if it can support kernel stacks in vmalloc space. This means: - vmalloc space must be large enough to hold many kernel stacks. This may rule out many 32-bit architectures. - Stacks in vmalloc space need to work reliably. For example, if vmap page tables are created on demand, either this mechanism needs to work while the stack points to a virtual address with unpopulated page tables or arch code (switch_to() and switch_mm(), most likely) needs to ensure that the stack's page table entries are populated before running on a possibly unpopulated stack. - If the stack overflows into a guard page, something reasonable should happen. The definition of "reasonable" is flexible, but instantly rebooting without logging anything would be unfriendly. config VMAP_STACK default y bool "Use a virtually-mapped stack" depends on HAVE_ARCH_VMAP_STACK && !KASAN ---help--- Enable this if you want the use virtually-mapped kernel stacks with guard pages. This causes kernel stack overflows to be caught immediately rather than causing difficult-to-diagnose corruption. This is presently incompatible with KASAN because KASAN expects the stack to map directly to the KASAN shadow map using a formula that is incorrect if the stack is in vmalloc space. |
ad21fc4fa
|
828 829 830 831 832 833 834 835 |
config ARCH_OPTIONAL_KERNEL_RWX def_bool n config ARCH_OPTIONAL_KERNEL_RWX_DEFAULT def_bool n config ARCH_HAS_STRICT_KERNEL_RWX def_bool n |
0f5bf6d0a
|
836 |
config STRICT_KERNEL_RWX |
ad21fc4fa
|
837 838 839 840 841 842 843 844 845 846 847 848 849 850 |
bool "Make kernel text and rodata read-only" if ARCH_OPTIONAL_KERNEL_RWX depends on ARCH_HAS_STRICT_KERNEL_RWX default !ARCH_OPTIONAL_KERNEL_RWX || ARCH_OPTIONAL_KERNEL_RWX_DEFAULT help If this is set, kernel text and rodata memory will be made read-only, and non-text memory will be made non-executable. This provides protection against certain security exploits (e.g. executing the heap or modifying text) These features are considered standard security practice these days. You should say Y here in almost all cases. config ARCH_HAS_STRICT_MODULE_RWX def_bool n |
0f5bf6d0a
|
851 |
config STRICT_MODULE_RWX |
ad21fc4fa
|
852 853 854 855 856 857 858 |
bool "Set loadable kernel module data as NX and text as RO" if ARCH_OPTIONAL_KERNEL_RWX depends on ARCH_HAS_STRICT_MODULE_RWX && MODULES default !ARCH_OPTIONAL_KERNEL_RWX || ARCH_OPTIONAL_KERNEL_RWX_DEFAULT help If this is set, module text and rodata memory will be made read-only, and non-text memory will be made non-executable. This provides protection against certain security exploits (e.g. writing to text) |
7a46ec0e2
|
859 860 861 862 863 864 865 866 867 868 869 |
config ARCH_HAS_REFCOUNT bool help An architecture selects this when it has implemented refcount_t using open coded assembly primitives that provide an optimized refcount_t implementation, possibly at the expense of some full refcount state checks of CONFIG_REFCOUNT_FULL=y. The refcount overflow check behavior, however, must be retained. Catching overflows is the primary security concern for protecting against bugs in reference counts. |
fd25d19f6
|
870 871 872 873 874 875 876 877 |
config REFCOUNT_FULL bool "Perform full reference count validation at the expense of speed" help Enabling this switches the refcounting infrastructure from a fast unchecked atomic_t implementation to a fully state checked implementation, which can be (slightly) slower but provides protections against various use-after-free conditions that can be used in security flaw exploits. |
2521f2c22
|
878 |
source "kernel/gcov/Kconfig" |