mirror of
https://github.com/torvalds/linux.git
synced 2024-11-21 19:46:16 +00:00
mm: implement update_mmu_tlb() using update_mmu_tlb_range()
Let's make update_mmu_tlb() simply a generic wrapper around update_mmu_tlb_range(). Only the latter can now be overridden by the architecture. We can now remove __HAVE_ARCH_UPDATE_MMU_TLB as well. Link: https://lkml.kernel.org/r/20240522061204.117421-3-libang.li@antgroup.com Signed-off-by: Bang Li <libang.li@antgroup.com> Acked-by: David Hildenbrand <david@redhat.com> Cc: Chris Zankel <chris@zankel.net> Cc: Huacai Chen <chenhuacai@kernel.org> Cc: Lance Yang <ioworker0@gmail.com> Cc: Max Filippov <jcmvbkbc@gmail.com> Cc: Palmer Dabbelt <palmer@dabbelt.com> Cc: Paul Walmsley <paul.walmsley@sifive.com> Cc: Ryan Roberts <ryan.roberts@arm.com> Cc: Thomas Bogendoerfer <tsbogend@alpha.franken.de> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
parent
23b1b44e6c
commit
8f65aa3223
@ -467,8 +467,6 @@ static inline void update_mmu_cache_range(struct vm_fault *vmf,
|
||||
#define update_mmu_cache(vma, addr, ptep) \
|
||||
update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
||||
|
||||
#define __HAVE_ARCH_UPDATE_MMU_TLB
|
||||
#define update_mmu_tlb update_mmu_cache
|
||||
#define update_mmu_tlb_range(vma, addr, ptep, nr) \
|
||||
update_mmu_cache_range(NULL, vma, addr, ptep, nr)
|
||||
|
||||
|
@ -594,8 +594,6 @@ static inline void update_mmu_cache_range(struct vm_fault *vmf,
|
||||
#define update_mmu_cache(vma, address, ptep) \
|
||||
update_mmu_cache_range(NULL, vma, address, ptep, 1)
|
||||
|
||||
#define __HAVE_ARCH_UPDATE_MMU_TLB
|
||||
#define update_mmu_tlb update_mmu_cache
|
||||
#define update_mmu_tlb_range(vma, address, ptep, nr) \
|
||||
update_mmu_cache_range(NULL, vma, address, ptep, nr)
|
||||
|
||||
|
@ -489,8 +489,6 @@ static inline void update_mmu_cache_range(struct vm_fault *vmf,
|
||||
#define update_mmu_cache(vma, addr, ptep) \
|
||||
update_mmu_cache_range(NULL, vma, addr, ptep, 1)
|
||||
|
||||
#define __HAVE_ARCH_UPDATE_MMU_TLB
|
||||
#define update_mmu_tlb update_mmu_cache
|
||||
#define update_mmu_tlb_range(vma, addr, ptep, nr) \
|
||||
update_mmu_cache_range(NULL, vma, addr, ptep, nr)
|
||||
|
||||
|
@ -410,9 +410,6 @@ void update_mmu_cache_range(struct vm_fault *vmf, struct vm_area_struct *vma,
|
||||
|
||||
typedef pte_t *pte_addr_t;
|
||||
|
||||
void update_mmu_tlb(struct vm_area_struct *vma,
|
||||
unsigned long address, pte_t *ptep);
|
||||
#define __HAVE_ARCH_UPDATE_MMU_TLB
|
||||
void update_mmu_tlb_range(struct vm_area_struct *vma,
|
||||
unsigned long address, pte_t *ptep, unsigned int nr);
|
||||
#define update_mmu_tlb_range update_mmu_tlb_range
|
||||
|
@ -163,12 +163,6 @@ void local_flush_tlb_kernel_range(unsigned long start, unsigned long end)
|
||||
}
|
||||
}
|
||||
|
||||
void update_mmu_tlb(struct vm_area_struct *vma,
|
||||
unsigned long address, pte_t *ptep)
|
||||
{
|
||||
local_flush_tlb_page(vma, address);
|
||||
}
|
||||
|
||||
void update_mmu_tlb_range(struct vm_area_struct *vma,
|
||||
unsigned long address, pte_t *ptep, unsigned int nr)
|
||||
{
|
||||
|
@ -736,13 +736,11 @@ static inline void update_mmu_tlb_range(struct vm_area_struct *vma,
|
||||
}
|
||||
#endif
|
||||
|
||||
#ifndef __HAVE_ARCH_UPDATE_MMU_TLB
|
||||
static inline void update_mmu_tlb(struct vm_area_struct *vma,
|
||||
unsigned long address, pte_t *ptep)
|
||||
{
|
||||
update_mmu_tlb_range(vma, address, ptep, 1);
|
||||
}
|
||||
#define __HAVE_ARCH_UPDATE_MMU_TLB
|
||||
#endif
|
||||
|
||||
/*
|
||||
* Some architectures may be able to avoid expensive synchronization
|
||||
|
Loading…
Reference in New Issue
Block a user