| /* SPDX-License-Identifier: GPL-2.0 */ |
| #ifndef _LINUX_MMU_CONTEXT_H |
| #define _LINUX_MMU_CONTEXT_H |
| |
| #include <asm/mmu_context.h> |
| #include <asm/mmu.h> |
| |
| /* Architectures that care about IRQ state in switch_mm can override this. */ |
| #ifndef switch_mm_irqs_off |
| # define switch_mm_irqs_off switch_mm |
| #endif |
| |
| #ifndef leave_mm |
| static inline void leave_mm(int cpu) { } |
| #endif |
| |
| /* |
| * CPUs that are capable of running user task @p. Must contain at least one |
| * active CPU. It is assumed that the kernel can run on all CPUs, so calling |
| * this for a kernel thread is pointless. |
| * |
| * By default, we assume a sane, homogeneous system. |
| */ |
| #ifndef task_cpu_possible_mask |
| # define task_cpu_possible_mask(p) cpu_possible_mask |
| # define task_cpu_possible(cpu, p) true |
| #else |
| # define task_cpu_possible(cpu, p) cpumask_test_cpu((cpu), task_cpu_possible_mask(p)) |
| #endif |
| |
| #ifndef mm_untag_mask |
| static inline unsigned long mm_untag_mask(struct mm_struct *mm) |
| { |
| return -1UL; |
| } |
| #endif |
| |
| #ifndef arch_pgtable_dma_compat |
| static inline bool arch_pgtable_dma_compat(struct mm_struct *mm) |
| { |
| return true; |
| } |
| #endif |
| |
| #endif |