mirror of
https://git.kernel.org/pub/scm/linux/kernel/git/torvalds/linux.git
synced 2026-04-08 16:25:44 -04:00
s390: pgtable: consolidate PxD and PTE TLB free paths
Call pagetable_dtor() for PMD|PUD|P4D tables just before ptdesc is freed - same as it is done for PTE tables. That allows consolidating TLB free paths for all table types. Link: https://lkml.kernel.org/r/ac69360a5f3350ebb2f63cd14b7b45316a130ee4.1736317725.git.zhengqi.arch@bytedance.com Signed-off-by: Qi Zheng <zhengqi.arch@bytedance.com> Suggested-by: Peter Zijlstra (Intel) <peterz@infradead.org> Reviewed-by: Kevin Brodsky <kevin.brodsky@arm.com> Acked-by: Alexander Gordeev <agordeev@linux.ibm.com> Cc: Alexandre Ghiti <alex@ghiti.fr> Cc: Alexandre Ghiti <alexghiti@rivosinc.com> Cc: Andreas Larsson <andreas@gaisler.com> Cc: Aneesh Kumar K.V (Arm) <aneesh.kumar@kernel.org> Cc: Arnd Bergmann <arnd@arndb.de> Cc: Dave Hansen <dave.hansen@linux.intel.com> Cc: David Hildenbrand <david@redhat.com> Cc: David Rientjes <rientjes@google.com> Cc: Hugh Dickins <hughd@google.com> Cc: Jann Horn <jannh@google.com> Cc: Lorenzo Stoakes <lorenzo.stoakes@oracle.com> Cc: Matthew Wilcox (Oracle) <willy@infradead.org> Cc: Mike Rapoport (Microsoft) <rppt@kernel.org> Cc: Muchun Song <muchun.song@linux.dev> Cc: Nicholas Piggin <npiggin@gmail.com> Cc: Palmer Dabbelt <palmer@dabbelt.com> Cc: Ryan Roberts <ryan.roberts@arm.com> Cc: Thomas Gleixner <tglx@linutronix.de> Cc: Vishal Moola (Oracle) <vishal.moola@gmail.com> Cc: Will Deacon <will@kernel.org> Cc: Yu Zhao <yuzhao@google.com> Signed-off-by: Andrew Morton <akpm@linux-foundation.org>
This commit is contained in:
@@ -102,7 +102,6 @@ static inline void pmd_free_tlb(struct mmu_gather *tlb, pmd_t *pmd,
|
||||
{
|
||||
if (mm_pmd_folded(tlb->mm))
|
||||
return;
|
||||
pagetable_dtor(virt_to_ptdesc(pmd));
|
||||
__tlb_adjust_range(tlb, address, PAGE_SIZE);
|
||||
tlb->mm->context.flush_mm = 1;
|
||||
tlb->freed_tables = 1;
|
||||
@@ -122,7 +121,6 @@ static inline void p4d_free_tlb(struct mmu_gather *tlb, p4d_t *p4d,
|
||||
{
|
||||
if (mm_p4d_folded(tlb->mm))
|
||||
return;
|
||||
pagetable_dtor(virt_to_ptdesc(p4d));
|
||||
__tlb_adjust_range(tlb, address, PAGE_SIZE);
|
||||
tlb->mm->context.flush_mm = 1;
|
||||
tlb->freed_tables = 1;
|
||||
@@ -141,7 +139,6 @@ static inline void pud_free_tlb(struct mmu_gather *tlb, pud_t *pud,
|
||||
{
|
||||
if (mm_pud_folded(tlb->mm))
|
||||
return;
|
||||
pagetable_dtor(virt_to_ptdesc(pud));
|
||||
tlb->mm->context.flush_mm = 1;
|
||||
tlb->freed_tables = 1;
|
||||
tlb->cleared_p4ds = 1;
|
||||
|
||||
@@ -180,7 +180,7 @@ unsigned long *page_table_alloc(struct mm_struct *mm)
|
||||
return table;
|
||||
}
|
||||
|
||||
static void pagetable_pte_dtor_free(struct ptdesc *ptdesc)
|
||||
static void pagetable_dtor_free(struct ptdesc *ptdesc)
|
||||
{
|
||||
pagetable_dtor(ptdesc);
|
||||
pagetable_free(ptdesc);
|
||||
@@ -190,20 +190,14 @@ void page_table_free(struct mm_struct *mm, unsigned long *table)
|
||||
{
|
||||
struct ptdesc *ptdesc = virt_to_ptdesc(table);
|
||||
|
||||
pagetable_pte_dtor_free(ptdesc);
|
||||
pagetable_dtor_free(ptdesc);
|
||||
}
|
||||
|
||||
void __tlb_remove_table(void *table)
|
||||
{
|
||||
struct ptdesc *ptdesc = virt_to_ptdesc(table);
|
||||
struct page *page = ptdesc_page(ptdesc);
|
||||
|
||||
if (compound_order(page) == CRST_ALLOC_ORDER) {
|
||||
/* pmd, pud, or p4d */
|
||||
pagetable_free(ptdesc);
|
||||
return;
|
||||
}
|
||||
pagetable_pte_dtor_free(ptdesc);
|
||||
pagetable_dtor_free(ptdesc);
|
||||
}
|
||||
|
||||
#ifdef CONFIG_TRANSPARENT_HUGEPAGE
|
||||
@@ -211,7 +205,7 @@ static void pte_free_now(struct rcu_head *head)
|
||||
{
|
||||
struct ptdesc *ptdesc = container_of(head, struct ptdesc, pt_rcu_head);
|
||||
|
||||
pagetable_pte_dtor_free(ptdesc);
|
||||
pagetable_dtor_free(ptdesc);
|
||||
}
|
||||
|
||||
void pte_free_defer(struct mm_struct *mm, pgtable_t pgtable)
|
||||
|
||||
Reference in New Issue
Block a user