| /linux-5.15/drivers/dma-buf/heaps/ |
| A D | system_heap.c | 295 __free_pages(page, compound_order(page)); in system_heap_dma_buf_release() 377 max_order = compound_order(page); in system_heap_allocate() 409 __free_pages(p, compound_order(p)); in system_heap_allocate() 414 __free_pages(page, compound_order(page)); in system_heap_allocate()
|
| /linux-5.15/include/linux/ |
| H A D | hugetlb_cgroup.h | 67 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in __hugetlb_cgroup_from_page() 91 if (compound_order(page) < HUGETLB_CGROUP_MIN_ORDER) in __set_hugetlb_cgroup()
|
| H A D | mm.h | 947 static inline unsigned int compound_order(struct page *page) in compound_order() function 951 return page[1].compound_order; in compound_order() 962 return PageCompound(page) && compound_order(page) > 1; in hpage_pincount_available() 979 page[1].compound_order = order; in set_compound_order() 994 return PAGE_SIZE << compound_order(page); in page_size() 1000 return PAGE_SHIFT + compound_order(page); in page_shift()
|
| H A D | mm_types.h | 169 unsigned char compound_order; member
|
| /linux-5.15/mm/ |
| A D | debug.c | 95 head, compound_order(head), in __dump_page() 100 head, compound_order(head), in __dump_page()
|
| H A D | compaction.c | 250 if (compound_order(page) >= pageblock_order) in pageblock_skip_persistent() 594 const unsigned int order = compound_order(page); in isolate_freepages_block() 919 low_pfn += (1UL << compound_order(page)) - 1; in isolate_migratepages_block() 970 const unsigned int order = compound_order(page); in isolate_migratepages_block()
|
| H A D | slob.c | 561 unsigned int order = compound_order(sp); in kfree()
|
| H A D | swap_state.c | 105 XA_STATE_ORDER(xas, &address_space->i_pages, idx, compound_order(page)); in add_to_swap_cache()
|
| A D | page_alloc.c | 730 free_the_page(page, compound_order(page)); in free_compound_page() 1326 VM_BUG_ON_PAGE(compound && compound_order(page) != order, page); in free_pages_prepare() 5660 free_the_page(page, compound_order(page)); in __page_frag_cache_drain() 5701 free_the_page(page, compound_order(page)); in page_frag_alloc_align() 5733 free_the_page(page, compound_order(page)); in page_frag_free()
|
| H A D | slub.c | 1097 maxobj = order_objects(compound_order(page), s->size); in check_slab() 1147 max_objects = order_objects(compound_order(page), s->size); in on_freelist() 1976 int order = compound_order(page); in __free_slab() 3523 unsigned int order = compound_order(page); in free_nonslab_page()
|
| H A D | huge_memory.c | 108 __free_pages(zero_page, compound_order(zero_page)); in get_huge_zero_page() 162 __free_pages(zero_page, compound_order(zero_page)); in shrink_huge_zero_page_scan()
|
| H A D | shmem.c | 697 XA_STATE_ORDER(xas, &mapping->i_pages, index, compound_order(page)); in shmem_add_to_page_cache() 1944 inode->i_blocks += BLOCKS_PER_PAGE << compound_order(page); in shmem_getpage_gfp()
|
| A D | gup.c | 244 if (PageCompound(page) && compound_order(page) >= 1) in compound_range_next()
|
| H A D | migrate.c | 2143 VM_BUG_ON_PAGE(compound_order(page) && !PageTransHuge(page), page); in numamigrate_isolate_page()
|
| H A D | hugetlb.c | 1786 if (compound_order(page_head) >= MAX_ORDER) in hugetlb_basepage_index() 1791 return (index << compound_order(page_head)) + compound_idx; in hugetlb_basepage_index()
|
| H A D | filemap.c | 134 xas_set_order(&xas, page->index, compound_order(page)); in page_cache_delete()
|
| H A D | memory.c | 3953 if (compound_order(page) != HPAGE_PMD_ORDER) in do_set_pmd()
|
| /linux-5.15/drivers/virt/acrn/ |
| A D | mm.c | 217 order = compound_order(page); in acrn_vm_ram_map() 243 order = compound_order(page); in acrn_vm_ram_map()
|
| /linux-5.15/kernel/ |
| A D | crash_core.c | 437 VMCOREINFO_OFFSET(page, compound_order); in crash_save_vmcoreinfo_init()
|
| /linux-5.15/Documentation/admin-guide/kdump/ |
| A D | vmcoreinfo.rst | 144 (page, flags|_refcount|mapping|lru|_mapcount|private|compound_dtor|compound_order|compound_head)
|