| /linux-5.15/arch/arm64/kvm/hyp/nvhe/ ! |
| A D | page_alloc.c | 35 unsigned short order) in __find_buddy_nocheck() argument 39 addr ^= (PAGE_SIZE << order); in __find_buddy_nocheck() 54 unsigned short order) in __find_buddy_avail() argument 56 struct hyp_page *buddy = __find_buddy_nocheck(pool, p, order); in __find_buddy_avail() 58 if (!buddy || buddy->order != order || buddy->refcount) in __find_buddy_avail() 96 unsigned short order = p->order; in __hyp_attach_page() local 99 memset(hyp_page_to_virt(p), 0, PAGE_SIZE << p->order); in __hyp_attach_page() 107 p->order = HYP_NO_ORDER; in __hyp_attach_page() 108 for (; (order + 1) < pool->max_order; order++) { in __hyp_attach_page() 109 buddy = __find_buddy_avail(pool, p, order); in __hyp_attach_page() [all …]
|
| /linux-5.15/include/trace/events/ ! |
| H A D | compaction.h | 173 int order, 177 TP_ARGS(order, gfp_mask, prio), 180 __field(int, order) 186 __entry->order = order; 192 __entry->order, 200 int order, 203 TP_ARGS(zone, order, ret), 208 __field(int, order) 215 __entry->order = order; 222 __entry->order, [all …]
|
| H A D | vmscan.h | 54 TP_PROTO(int nid, int zid, int order), 56 TP_ARGS(nid, zid, order), 61 __field( int, order ) 67 __entry->order = order; 72 __entry->order) 77 TP_PROTO(int nid, int zid, int order, gfp_t gfp_flags), 79 TP_ARGS(nid, zid, order, gfp_flags), 84 __field( int, order ) 91 __entry->order = order; 97 __entry->order, [all …]
|
| H A D | kmem.h | 162 TP_PROTO(struct page *page, unsigned int order), 164 TP_ARGS(page, order), 168 __field( unsigned int, order ) 173 __entry->order = order; 179 __entry->order) 203 TP_PROTO(struct page *page, unsigned int order, 206 TP_ARGS(page, order, gfp_flags, migratetype), 210 __field( unsigned int, order ) 217 __entry->order = order; 225 __entry->order, [all …]
|
| H A D | oom.h | 35 int order, 42 TP_ARGS(zoneref, order, reclaimable, available, min_wmark, no_progress_loops, wmark_check), 47 __field( int, order) 58 __entry->order = order; 68 __entry->order, 157 TP_PROTO(int order, 164 TP_ARGS(order, priority, result, retries, max_retries, ret), 167 __field( int, order) 176 __entry->order = order; 185 __entry->order,
|
| /linux-5.15/drivers/gpu/drm/ttm/ ! |
| A D | ttm_pool.c | 79 unsigned int order) in ttm_pool_alloc_page() argument 90 if (order) in ttm_pool_alloc_page() 95 p = alloc_pages(gfp_flags, order); in ttm_pool_alloc_page() 97 p->private = order; in ttm_pool_alloc_page() 105 if (order) in ttm_pool_alloc_page() 108 vaddr = dma_alloc_attrs(pool->dev, (1ULL << order) * PAGE_SIZE, in ttm_pool_alloc_page() 121 dma->vaddr = (unsigned long)vaddr | order; in ttm_pool_alloc_page() 132 unsigned int order, struct page *p) in ttm_pool_free_page() argument 143 set_pages_wb(p, 1 << order); in ttm_pool_free_page() 147 __free_pages(p, order); in ttm_pool_free_page() [all …]
|
| /linux-5.15/drivers/gpu/drm/lib/ ! |
| A D | drm_random.c | 15 void drm_random_reorder(unsigned int *order, unsigned int count, in drm_random_reorder() argument 23 swap(order[i], order[j]); in drm_random_reorder() 30 unsigned int *order, i; in drm_random_order() local 32 order = kmalloc_array(count, sizeof(*order), GFP_KERNEL); in drm_random_order() 33 if (!order) in drm_random_order() 34 return order; in drm_random_order() 37 order[i] = i; in drm_random_order() 39 drm_random_reorder(order, count, state); in drm_random_order() 40 return order; in drm_random_order()
|
| /linux-5.15/drivers/gpu/drm/i915/selftests/ ! |
| A D | i915_buddy.c | 200 unsigned int order; in igt_check_mm() local 216 order = i915_buddy_block_order(root); in igt_check_mm() 219 if (order != mm->max_order) { in igt_check_mm() 240 block = list_first_entry_or_null(&mm->free_list[order], in igt_check_mm() 244 pr_err("root mismatch at order=%u\n", order); in igt_check_mm() 305 int *order; in igt_buddy_alloc_smoke() local 318 order = i915_random_order(mm.max_order + 1, &prng); in igt_buddy_alloc_smoke() 319 if (!order) in igt_buddy_alloc_smoke() 324 int max_order = order[i]; in igt_buddy_alloc_smoke() 327 int order; in igt_buddy_alloc_smoke() local [all …]
|
| A D | i915_random.c | 70 void i915_random_reorder(unsigned int *order, unsigned int count, in i915_random_reorder() argument 73 i915_prandom_shuffle(order, sizeof(*order), count, state); in i915_random_reorder() 78 unsigned int *order, i; in i915_random_order() local 80 order = kmalloc_array(count, sizeof(*order), in i915_random_order() 82 if (!order) in i915_random_order() 83 return order; in i915_random_order() 86 order[i] = i; in i915_random_order() 88 i915_random_reorder(order, count, state); in i915_random_order() 89 return order; in i915_random_order()
|
| A D | i915_syncmap.c | 274 unsigned int pass, order; in igt_syncmap_join_above() local 296 for (order = 0; order < 64; order += SHIFT) { in igt_syncmap_join_above() 297 u64 context = BIT_ULL(order); in igt_syncmap_join_above() 335 unsigned int step, order, idx; in igt_syncmap_join_below() local 345 for (order = 64 - SHIFT; order > 0; order -= SHIFT) { in igt_syncmap_join_below() 346 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below() 354 context, order, step, sync->height, sync->prefix); in igt_syncmap_join_below() 362 for (order = SHIFT; order < 64; order += SHIFT) { in igt_syncmap_join_below() 363 u64 context = step * BIT_ULL(order); in igt_syncmap_join_below() 367 context, order, step); in igt_syncmap_join_below() [all …]
|
| /linux-5.15/drivers/media/pci/cx18/ ! |
| H A D | cx18-mailbox.c | 230 static void epu_dma_done(struct cx18 *cx, struct cx18_in_work_order *order) in epu_dma_done() argument 239 mb = &order->mb; in epu_dma_done() 246 (order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) ? in epu_dma_done() 252 mdl_ack = order->mdl_ack; in epu_dma_done() 276 if ((order->flags & CX18_F_EWO_MB_STALE_UPON_RECEIPT) && in epu_dma_done() 323 static void epu_debug(struct cx18 *cx, struct cx18_in_work_order *order) in epu_debug() argument 326 char *str = order->str; in epu_debug() 328 CX18_DEBUG_INFO("%x %s\n", order->mb.args[0], str); in epu_debug() 334 static void epu_cmd(struct cx18 *cx, struct cx18_in_work_order *order) in epu_cmd() argument 336 switch (order->rpu) { in epu_cmd() [all …]
|
| /linux-5.15/mm/ ! |
| H A D | compaction.c | 48 #define block_start_pfn(pfn, order) round_down(pfn, 1UL << (order)) argument 49 #define block_end_pfn(pfn, order) ALIGN((pfn) + 1, 1UL << (order)) argument 89 unsigned int i, order, nr_pages; in split_map_pages() local 96 order = page_private(page); in split_map_pages() 97 nr_pages = 1 << order; in split_map_pages() 99 post_alloc_hook(page, order, __GFP_MOVABLE); in split_map_pages() 100 if (order) in split_map_pages() 101 split_page(page, order); in split_map_pages() 159 static void defer_compaction(struct zone *zone, int order) in defer_compaction() argument 164 if (order < zone->compact_order_failed) in defer_compaction() [all …]
|
| A D | page_alloc.c | 270 static void __free_pages_ok(struct page *page, unsigned int order, 659 static inline unsigned int order_to_pindex(int migratetype, int order) in order_to_pindex() argument 661 int base = order; in order_to_pindex() 664 if (order > PAGE_ALLOC_COSTLY_ORDER) { in order_to_pindex() 665 VM_BUG_ON(order != pageblock_order); in order_to_pindex() 669 VM_BUG_ON(order > PAGE_ALLOC_COSTLY_ORDER); in order_to_pindex() 677 int order = pindex / MIGRATE_PCPTYPES; in pindex_to_order() local 680 if (order > PAGE_ALLOC_COSTLY_ORDER) { in pindex_to_order() 681 order = pageblock_order; in pindex_to_order() 682 VM_BUG_ON(order != pageblock_order); in pindex_to_order() [all …]
|
| /linux-5.15/tools/testing/radix-tree/ ! |
| A D | multiorder.c | 16 unsigned order) in item_insert_order() argument 18 XA_STATE_ORDER(xas, xa, index, order); in item_insert_order() 19 struct item *item = item_create(index, order); in item_insert_order() 42 int order[NUM_ENTRIES] = {1, 1, 2, 3, 4, 1, 0, 1, 3, 0, 7}; in multiorder_iteration() local 47 err = item_insert_order(xa, index[i], order[i]); in multiorder_iteration() 53 if (j <= (index[i] | ((1 << order[i]) - 1))) in multiorder_iteration() 58 int height = order[i] / XA_CHUNK_SHIFT; in multiorder_iteration() 60 unsigned long mask = (1UL << order[i]) - 1; in multiorder_iteration() 66 assert(item->order == order[i]); in multiorder_iteration() 82 int order[MT_NUM_ENTRIES] = {1, 0, 2, 4, 3, 1, 3, 0, 7}; in multiorder_tagged_iteration() local [all …]
|
| A D | iteration_check.c | 25 int order; in my_item_insert() local 29 for (order = max_order; order >= 0; order--) { in my_item_insert() 30 xas_set_order(&xas, index, order); in my_item_insert() 31 item->order = order; in my_item_insert() 41 if (order < 0) in my_item_insert() 165 void iteration_test(unsigned order, unsigned test_duration) in iteration_test() argument 170 order > 0 ? "multiorder " : "", test_duration); in iteration_test() 172 max_order = order; in iteration_test()
|
| /linux-5.15/Documentation/trace/postprocess/ ! |
| H A D | trace-vmscan-postprocess.pl | 317 my $order = $1; 318 $perprocesspid{$process_pid}->{MM_VMSCAN_DIRECT_RECLAIM_BEGIN_PERORDER}[$order]++; 319 $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER} = $order; 328 my $order = $perprocesspid{$process_pid}->{STATE_DIRECT_ORDER}; 330 $perprocesspid{$process_pid}->{HIGH_DIRECT_RECLAIM_LATENCY}[$index] = "$order-$latency"; 341 my $order = $2; 342 $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER} = $order; 347 $perprocesspid{$process_pid}->{MM_VMSCAN_KSWAPD_WAKE_PERORDER}[$order]++; 350 $perprocesspid{$process_pid}->{HIGH_KSWAPD_REWAKEUP_PERORDER}[$order]++; 360 my $order = $perprocesspid{$process_pid}->{STATE_KSWAPD_ORDER}; [all …]
|
| /linux-5.15/drivers/gpu/drm/nouveau/nvkm/subdev/therm/ ! |
| A D | gk104.c | 34 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_enable() local 38 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable() 39 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_enable() 42 nvkm_mask(dev, 0x20200 + order[i].offset, 0xff00, 0x4500); in gk104_clkgate_enable() 50 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_enable() 51 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_enable() 54 nvkm_mask(dev, 0x20200 + order[i].offset, 0x00ff, 0x0045); in gk104_clkgate_enable() 63 const struct gk104_clkgate_engine_info *order = therm->clkgate_order; in gk104_clkgate_fini() local 67 for (i = 0; order[i].type != NVKM_SUBDEV_NR; i++) { in gk104_clkgate_fini() 68 if (!nvkm_device_subdev(dev, order[i].type, order[i].inst)) in gk104_clkgate_fini() [all …]
|
| /linux-5.15/drivers/net/ethernet/mellanox/mlx5/core/steering/ ! |
| A D | dr_buddy.c | 77 unsigned int *order) in dr_buddy_find_free_seg() argument 101 *order = order_iter; in dr_buddy_find_free_seg() 122 unsigned int order, in mlx5dr_buddy_alloc_mem() argument 128 err = dr_buddy_find_free_seg(buddy, order, &seg, &order_iter); in mlx5dr_buddy_alloc_mem() 139 while (order_iter > order) { in mlx5dr_buddy_alloc_mem() 146 seg <<= order; in mlx5dr_buddy_alloc_mem() 153 unsigned int seg, unsigned int order) in mlx5dr_buddy_free_mem() argument 155 seg >>= order; in mlx5dr_buddy_free_mem() 160 while (test_bit(seg ^ 1, buddy->bitmap[order])) { in mlx5dr_buddy_free_mem() 161 bitmap_clear(buddy->bitmap[order], seg ^ 1, 1); in mlx5dr_buddy_free_mem() [all …]
|
| /linux-5.15/lib/ ! |
| A D | test_xarray.c | 72 unsigned order, void *entry, gfp_t gfp) in xa_store_order() argument 74 XA_STATE_ORDER(xas, xa, index, order); in xa_store_order() 177 unsigned int order; in check_xa_mark_1() local 207 for (order = 2; order < max_order; order++) { in check_xa_mark_1() 208 unsigned long base = round_down(index, 1UL << order); in check_xa_mark_1() 209 unsigned long next = base + (1UL << order); in check_xa_mark_1() 217 xa_store_order(xa, index, order, xa_mk_index(index), in check_xa_mark_1() 328 unsigned int order; in check_xa_shrink() local 353 for (order = 0; order < max_order; order++) { in check_xa_shrink() 354 unsigned long max = (1UL << order) - 1; in check_xa_shrink() [all …]
|
| /linux-5.15/include/linux/ ! |
| H A D | compaction.h | 65 static inline unsigned long compact_gap(unsigned int order) in compact_gap() argument 80 return 2UL << order; in compact_gap() 92 extern unsigned int extfrag_for_order(struct zone *zone, unsigned int order); 93 extern int fragmentation_index(struct zone *zone, unsigned int order); 95 unsigned int order, unsigned int alloc_flags, 99 extern enum compact_result compaction_suitable(struct zone *zone, int order, 102 extern void compaction_defer_reset(struct zone *zone, int order, 177 bool compaction_zonelist_suitable(struct alloc_context *ac, int order, 182 extern void wakeup_kcompactd(pg_data_t *pgdat, int order, int highest_zoneidx); 189 static inline enum compact_result compaction_suitable(struct zone *zone, int order, in compaction_suitable() argument [all …]
|
| H A D | gfp.h | 518 static inline void arch_free_page(struct page *page, int order) { } in arch_free_page() argument 521 static inline void arch_alloc_page(struct page *page, int order) { } in arch_alloc_page() argument 530 struct page *__alloc_pages(gfp_t gfp, unsigned int order, int preferred_nid, 565 __alloc_pages_node(int nid, gfp_t gfp_mask, unsigned int order) in __alloc_pages_node() argument 570 return __alloc_pages(gfp_mask, order, nid, NULL); in __alloc_pages_node() 579 unsigned int order) in alloc_pages_node() argument 584 return __alloc_pages_node(nid, gfp_mask, order); in alloc_pages_node() 588 struct page *alloc_pages(gfp_t gfp, unsigned int order); 589 extern struct page *alloc_pages_vma(gfp_t gfp_mask, int order, 592 #define alloc_hugepage_vma(gfp_mask, vma, addr, order) \ argument [all …]
|
| A D | page_owner.h | 11 extern void __reset_page_owner(struct page *page, unsigned int order); 13 unsigned int order, gfp_t gfp_mask); 21 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument 24 __reset_page_owner(page, order); in reset_page_owner() 28 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument 31 __set_page_owner(page, order, gfp_mask); in set_page_owner() 55 static inline void reset_page_owner(struct page *page, unsigned int order) in reset_page_owner() argument 59 unsigned int order, gfp_t gfp_mask) in set_page_owner() argument 63 unsigned int order) in split_page_owner() argument
|
| /linux-5.15/arch/arm/lib/ ! |
| H A D | lib1funcs.S | 106 .macro ARM_DIV2_ORDER divisor, order argument 110 clz \order, \divisor 111 rsb \order, \order, #31 117 movhs \order, #16 118 movlo \order, #0 122 addhs \order, \order, #8 126 addhs \order, \order, #4 129 addhi \order, \order, #3 130 addls \order, \order, \divisor, lsr #1 137 .macro ARM_MOD_BODY dividend, divisor, order, spare [all …]
|
| /linux-5.15/scripts/atomic/ ! |
| A D | gen-atomic-fallback.sh | 16 local order="$1"; shift 20 local atomicname="arch_${atomic}_${pfx}${name}${sfx}${order}" 42 local order="$1"; shift 44 local tmpl="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")" 45 gen_template_fallback "${tmpl}" "${meta}" "${pfx}" "${name}" "${sfx}" "${order}" "$@" 65 local order="$1"; shift 70 printf "#define ${basename}${order} ${basename}${order}\n" 84 local template="$(find_fallback_template "${pfx}" "${name}" "${sfx}" "${order}")" 167 local order="$1"; shift; 174 ___r = arch_cmpxchg${order}((_ptr), ___o, (_new)); \\ [all …]
|
| /linux-5.15/arch/s390/mm/ ! |
| H A D | page-states.c | 71 static inline void set_page_unused(struct page *page, int order) in set_page_unused() argument 75 for (i = 0; i < (1 << order); i++) in set_page_unused() 82 static inline void set_page_stable_dat(struct page *page, int order) in set_page_stable_dat() argument 86 for (i = 0; i < (1 << order); i++) in set_page_stable_dat() 93 static inline void set_page_stable_nodat(struct page *page, int order) in set_page_stable_nodat() argument 97 for (i = 0; i < (1 << order); i++) in set_page_stable_nodat() 208 void arch_free_page(struct page *page, int order) in arch_free_page() argument 212 set_page_unused(page, order); in arch_free_page() 215 void arch_alloc_page(struct page *page, int order) in arch_alloc_page() argument 220 set_page_stable_dat(page, order); in arch_alloc_page() [all …]
|