/linux/mm/ |
A D | mmu_gather.c | 30 if (!batch) in tlb_next_batch() 48 for (batch = &tlb->local; batch && batch->nr; batch = batch->next) { in tlb_batch_pages_flush() 49 free_pages_and_swap_cache(batch->pages, batch->nr); in tlb_batch_pages_flush() 50 batch->nr = 0; in tlb_batch_pages_flush() 59 for (batch = tlb->local.next; batch; batch = next) { in tlb_batch_list_free() 81 batch->pages[batch->nr++] = page; in __tlb_remove_page_size() 82 if (batch->nr == batch->max) { in __tlb_remove_page_size() 87 VM_BUG_ON_PAGE(batch->nr > batch->max, page); in __tlb_remove_page_size() 198 struct mmu_table_batch **batch = &tlb->batch; in tlb_table_flush() local 209 struct mmu_table_batch **batch = &tlb->batch; in tlb_remove_table() local [all …]
|
/linux/arch/powerpc/mm/book3s64/ |
A D | hash_tlb.c | 51 i = batch->index; in hpte_need_flush() 103 if (!batch->active) { in hpte_need_flush() 119 if (i != 0 && (mm != batch->mm || batch->psize != psize || in hpte_need_flush() 125 batch->mm = mm; in hpte_need_flush() 129 batch->pte[i] = rpte; in hpte_need_flush() 130 batch->vpn[i] = vpn; in hpte_need_flush() 131 batch->index = ++i; in hpte_need_flush() 148 i = batch->index; in __flush_tlb_pending() 151 flush_hash_page(batch->vpn[0], batch->pte[0], in __flush_tlb_pending() 152 batch->psize, batch->ssize, local); in __flush_tlb_pending() [all …]
|
/linux/drivers/gpu/drm/i915/selftests/ |
A D | igt_spinner.c | 96 if (!spin->batch) { in igt_spinner_pin() 104 spin->batch = vaddr; in igt_spinner_pin() 146 u32 *batch; in igt_spinner_create_request() local 154 if (!spin->batch) { in igt_spinner_create_request() 175 batch = spin->batch; in igt_spinner_create_request() 183 *batch++ = 0; in igt_spinner_create_request() 187 *batch++ = 0; in igt_spinner_create_request() 193 *batch++ = rq->fence.seqno; in igt_spinner_create_request() 195 *batch++ = arbitration_command; in igt_spinner_create_request() 241 if (!spin->batch) in igt_spinner_end() [all …]
|
A D | i915_request.c | 899 if (IS_ERR(batch)) in live_empty_request() 958 i915_vma_unpin(batch); in live_empty_request() 959 i915_vma_put(batch); in live_empty_request() 1060 if (IS_ERR(batch)) { in live_all_engines() 1061 err = PTR_ERR(batch); in live_all_engines() 1066 i915_vma_lock(batch); in live_all_engines() 1088 request[idx]->batch = batch; in live_all_engines() 1142 i915_vma_unpin(batch); in live_all_engines() 1143 i915_vma_put(batch); in live_all_engines() 1180 if (IS_ERR(batch)) { in live_sequential_engines() [all …]
|
/linux/drivers/gpu/drm/i915/gt/ |
A D | gen7_renderclear.c | 256 batch_advance(batch, cs); in gen7_emit_state_base_address() 285 batch_advance(batch, cs); in gen7_emit_vfe_state() 293 u32 *cs = batch_alloc_items(batch, 8, 4); in gen7_emit_interface_descriptor_load() 304 batch_advance(batch, cs); in gen7_emit_interface_descriptor_load() 316 cs = batch_alloc_items(batch, 8, pkt); in gen7_emit_media_object() 336 batch_advance(batch, cs); in gen7_emit_media_object() 351 batch_advance(batch, cs); in gen7_emit_pipeline_flush() 372 batch_advance(batch, cs); in gen7_emit_pipeline_invalidate() 434 u32 *batch; in gen7_setup_clear_gpr_bb() local 443 if (IS_ERR(batch)) in gen7_setup_clear_gpr_bb() [all …]
|
A D | intel_lrc.c | 1377 *batch++ = 0; in gen8_emit_flush_coherentl3_wa() 1383 batch = gen8_emit_pipe_control(batch, in gen8_emit_flush_coherentl3_wa() 1392 *batch++ = 0; in gen8_emit_flush_coherentl3_wa() 1394 return batch; in gen8_emit_flush_coherentl3_wa() 1419 batch = gen8_emit_flush_coherentl3_wa(engine, batch); in gen8_init_indirectctx_bb() 1423 batch = gen8_emit_pipe_control(batch, in gen8_init_indirectctx_bb() 1442 return batch; in gen8_init_indirectctx_bb() 1461 return batch; in emit_lri() 1492 batch = gen8_emit_flush_coherentl3_wa(engine, batch); in gen9_init_indirectctx_bb() 1495 batch = gen8_emit_pipe_control(batch, in gen9_init_indirectctx_bb() [all …]
|
A D | gen8_engine_cs.h | 42 __gen8_emit_pipe_control(u32 *batch, u32 flags0, u32 flags1, u32 offset) in __gen8_emit_pipe_control() argument 44 memset(batch, 0, 6 * sizeof(u32)); in __gen8_emit_pipe_control() 46 batch[0] = GFX_OP_PIPE_CONTROL(6) | flags0; in __gen8_emit_pipe_control() 47 batch[1] = flags1; in __gen8_emit_pipe_control() 48 batch[2] = offset; in __gen8_emit_pipe_control() 50 return batch + 6; in __gen8_emit_pipe_control() 53 static inline u32 *gen8_emit_pipe_control(u32 *batch, u32 flags, u32 offset) in gen8_emit_pipe_control() argument 55 return __gen8_emit_pipe_control(batch, 0, flags, offset); in gen8_emit_pipe_control() 58 static inline u32 *gen12_emit_pipe_control(u32 *batch, u32 flags0, u32 flags1, u32 offset) in gen12_emit_pipe_control() argument 60 return __gen8_emit_pipe_control(batch, flags0, flags1, offset); in gen12_emit_pipe_control()
|
A D | selftest_hangcheck.c | 36 u32 *batch; member 79 h->batch = vaddr; in hang_init() 126 u32 *batch; in hang_create_request() local 146 h->batch = vaddr; in hang_create_request() 184 batch = h->batch; in hang_create_request() 193 batch += 1024 / sizeof(*batch); in hang_create_request() 201 *batch++ = 0; in hang_create_request() 207 batch += 1024 / sizeof(*batch); in hang_create_request() 214 *batch++ = 0; in hang_create_request() 220 batch += 1024 / sizeof(*batch); in hang_create_request() [all …]
|
A D | selftest_workarounds.c | 507 struct i915_vma *batch; in check_dirty_whitelist() local 516 batch = create_batch(ce->vm); in check_dirty_whitelist() 517 if (IS_ERR(batch)) { in check_dirty_whitelist() 518 err = PTR_ERR(batch); in check_dirty_whitelist() 648 batch->node.start, PAGE_SIZE, in check_dirty_whitelist() 901 struct i915_vma *batch; in scrub_whitelisted_registers() local 905 batch = create_batch(ce->vm); in scrub_whitelisted_registers() 906 if (IS_ERR(batch)) in scrub_whitelisted_registers() 907 return PTR_ERR(batch); in scrub_whitelisted_registers() 945 i915_vma_lock(batch); in scrub_whitelisted_registers() [all …]
|
A D | selftest_lrc.c | 920 if (IS_ERR(batch)) in store_context() 921 return batch; in store_context() 925 i915_vma_put(batch); in store_context() 932 i915_vma_put(batch); in store_context() 974 return batch; in store_context() 1084 if (IS_ERR(batch)) in load_context() 1085 return batch; in load_context() 1089 i915_vma_put(batch); in load_context() 1134 return batch; in load_context() 1145 if (IS_ERR(batch)) in poison_registers() [all …]
|
A D | selftest_engine_cs.c | 134 struct i915_vma *batch; in perf_mi_bb_start() local 140 batch = create_empty_batch(ce); in perf_mi_bb_start() 141 if (IS_ERR(batch)) { in perf_mi_bb_start() 142 err = PTR_ERR(batch); in perf_mi_bb_start() 147 err = i915_vma_sync(batch); in perf_mi_bb_start() 150 i915_vma_put(batch); in perf_mi_bb_start() 168 batch->node.start, 8, in perf_mi_bb_start() 189 i915_vma_put(batch); in perf_mi_bb_start()
|
/linux/arch/powerpc/include/asm/book3s/64/ |
A D | tlbflush-hash.h | 25 extern void __flush_tlb_pending(struct ppc64_tlb_batch *batch); 31 struct ppc64_tlb_batch *batch; in arch_enter_lazy_mmu_mode() local 35 batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_enter_lazy_mmu_mode() 36 batch->active = 1; in arch_enter_lazy_mmu_mode() 41 struct ppc64_tlb_batch *batch; in arch_leave_lazy_mmu_mode() local 45 batch = this_cpu_ptr(&ppc64_tlb_batch); in arch_leave_lazy_mmu_mode() 47 if (batch->index) in arch_leave_lazy_mmu_mode() 48 __flush_tlb_pending(batch); in arch_leave_lazy_mmu_mode() 49 batch->active = 0; in arch_leave_lazy_mmu_mode()
|
/linux/drivers/gpu/drm/i915/gem/selftests/ |
A D | igt_gem_utils.c | 115 struct i915_vma *batch; in igt_gpu_fill_dw() local 122 batch = igt_emit_store_dw(vma, offset, count, val); in igt_gpu_fill_dw() 123 if (IS_ERR(batch)) in igt_gpu_fill_dw() 124 return PTR_ERR(batch); in igt_gpu_fill_dw() 132 i915_vma_lock(batch); in igt_gpu_fill_dw() 133 err = i915_request_await_object(rq, batch->obj, false); in igt_gpu_fill_dw() 135 err = i915_vma_move_to_active(batch, rq, 0); in igt_gpu_fill_dw() 136 i915_vma_unlock(batch); in igt_gpu_fill_dw() 153 batch->node.start, batch->node.size, in igt_gpu_fill_dw() 161 i915_vma_unpin_and_release(&batch, 0); in igt_gpu_fill_dw()
|
A D | i915_gem_client_blt.c | 40 struct i915_vma *batch; member 49 struct drm_i915_gem_object *batch) in prepare_blit() argument 108 i915_gem_object_flush_map(batch); in prepare_blit() 109 i915_gem_object_unpin_map(batch); in prepare_blit() 122 i915_vma_put(t->batch); in tiled_blits_destroy_buffers() 162 if (IS_ERR(t->batch)) in tiled_blits_create_buffers() 163 return PTR_ERR(t->batch); in tiled_blits_create_buffers() 167 i915_vma_put(t->batch); in tiled_blits_create_buffers() 376 t->batch->node.start, in tiled_blit() 377 t->batch->node.size, in tiled_blit() [all …]
|
/linux/drivers/xen/ |
A D | gntdev.c | 746 batch->pages[batch->nr_pages++] = page; in gntdev_get_page() 756 unpin_user_pages_dirty_lock(batch->pages, batch->nr_pages, batch->writeable); in gntdev_put_pages() 757 batch->nr_pages = 0; in gntdev_put_pages() 765 gnttab_batch_copy(batch->ops, batch->nr_ops); in gntdev_copy() 789 batch->nr_ops = 0; in gntdev_copy() 835 op = &batch->ops[batch->nr_ops]; in gntdev_grant_copy_seg() 881 batch->status[batch->nr_ops] = status; in gntdev_grant_copy_seg() 882 batch->nr_ops++; in gntdev_grant_copy_seg() 898 batch.nr_ops = 0; in gntdev_ioctl_grant_copy() 899 batch.nr_pages = 0; in gntdev_ioctl_grant_copy() [all …]
|
/linux/drivers/gpu/drm/vmwgfx/ |
A D | vmwgfx_mob.c | 246 struct vmw_otable_batch *batch) in vmw_otable_batch_setup() argument 269 if (!batch->otables[i].enabled) in vmw_otable_batch_setup() 284 if (batch->otables[i].enabled) in vmw_otable_batch_setup() 286 &batch->otables[i]); in vmw_otable_batch_setup() 289 vmw_bo_unpin_unlocked(batch->otable_bo); in vmw_otable_batch_setup() 290 ttm_bo_put(batch->otable_bo); in vmw_otable_batch_setup() 291 batch->otable_bo = NULL; in vmw_otable_batch_setup() 345 if (batch->otables[i].enabled) in vmw_otable_batch_takedown() 347 &batch->otables[i]); in vmw_otable_batch_takedown() 356 ttm_bo_put(batch->otable_bo); in vmw_otable_batch_takedown() [all …]
|
/linux/tools/virtio/ |
A D | virtio_test.c | 169 bool delayed, int batch, int reset_n, int bufs) in run_test() argument 177 const bool random_batch = batch == RANDOM_BATCH; in run_test() 192 batch = (random() % vq->vring.num) + 1; in run_test() 195 (started - completed) < batch) { in run_test() 346 long batch = 1, reset = 0; in main() local 375 batch = RANDOM_BATCH; in main() 377 batch = strtol(optarg, NULL, 10); in main() 378 assert(batch > 0); in main() 379 assert(batch < (long)INT_MAX + 1); in main() 400 run_test(&dev, &dev.vqs[0], delayed, batch, reset, 0x100000); in main()
|
/linux/tools/testing/selftests/bpf/map_tests/ |
A D | htab_map_batch_ops.c | 78 __u32 batch, count, total, total_success; in __test_map_lookup_and_delete_batch() local 115 err = bpf_map_lookup_and_delete_batch(map_fd, NULL, &batch, keys, in __test_map_lookup_and_delete_batch() 125 err = bpf_map_lookup_and_delete_batch(map_fd, NULL, &batch, keys, in __test_map_lookup_and_delete_batch() 133 err = bpf_map_lookup_and_delete_batch(map_fd, NULL, &batch, keys, in __test_map_lookup_and_delete_batch() 159 total ? &batch : NULL, in __test_map_lookup_and_delete_batch() 160 &batch, keys + total, in __test_map_lookup_and_delete_batch() 222 total ? &batch : NULL, in __test_map_lookup_and_delete_batch() 223 &batch, keys + total, in __test_map_lookup_and_delete_batch()
|
/linux/drivers/net/ethernet/netronome/nfp/flower/ |
A D | lag_conf.c | 204 unsigned int member_cnt, enum nfp_fl_lag_batch *batch) in nfp_fl_lag_config_group() argument 224 if (*batch == NFP_FL_LAG_BATCH_FIRST) { in nfp_fl_lag_config_group() 227 *batch = NFP_FL_LAG_BATCH_MEMBER; in nfp_fl_lag_config_group() 233 *batch = NFP_FL_LAG_BATCH_FINISHED; in nfp_fl_lag_config_group() 239 if (*batch == NFP_FL_LAG_BATCH_FINISHED) { in nfp_fl_lag_config_group() 266 enum nfp_fl_lag_batch batch = NFP_FL_LAG_BATCH_FIRST; in nfp_fl_lag_do_work() local 288 &batch); in nfp_fl_lag_do_work() 357 active_count, &batch); in nfp_fl_lag_do_work() 371 if (batch == NFP_FL_LAG_BATCH_MEMBER) { in nfp_fl_lag_do_work() 372 batch = NFP_FL_LAG_BATCH_FINISHED; in nfp_fl_lag_do_work() [all …]
|
/linux/drivers/vfio/ |
A D | vfio_iommu_type1.c | 470 batch->size = 0; in vfio_batch_init() 471 batch->offset = 0; in vfio_batch_init() 484 batch->pages = &batch->fallback_page; in vfio_batch_init() 491 unsigned long pfn = page_to_pfn(batch->pages[batch->offset]); in vfio_batch_unpin() 494 batch->offset++; in vfio_batch_unpin() 495 batch->size--; in vfio_batch_unpin() 662 *pfn_base = page_to_pfn(batch->pages[batch->offset]); in vfio_pin_pages_remote() 722 batch->size--; in vfio_pin_pages_remote() 727 pfn = page_to_pfn(batch->pages[batch->offset]); in vfio_pin_pages_remote() 738 if (batch->size == 1 && !batch->offset) { in vfio_pin_pages_remote() [all …]
|
/linux/tools/virtio/ringtest/ |
A D | main.c | 22 int batch = 1; variable 116 int tokick = batch; in run_guest() 129 tokick = batch; in run_guest() 348 batch = c; in main() 372 if (batch > max_outstanding) in main() 373 batch = max_outstanding; in main()
|
/linux/arch/x86/include/asm/ |
A D | tlbflush.h | 253 static inline void arch_tlbbatch_add_mm(struct arch_tlbflush_unmap_batch *batch, in arch_tlbbatch_add_mm() argument 257 cpumask_or(&batch->cpumask, &batch->cpumask, mm_cpumask(mm)); in arch_tlbbatch_add_mm() 260 extern void arch_tlbbatch_flush(struct arch_tlbflush_unmap_batch *batch);
|
/linux/net/core/ |
A D | netclassid_cgroup.c | 64 unsigned int batch; member 76 if (--ctx->batch == 0) { in update_classid_sock() 77 ctx->batch = UPDATE_CLASSID_BATCH; in update_classid_sock() 87 .batch = UPDATE_CLASSID_BATCH in update_classid_task()
|
/linux/include/linux/ |
A D | ptr_ring.h | 39 int batch; /* number of entries to consume in a batch */ member 272 if (unlikely(consumer_head - r->consumer_tail >= r->batch || in __ptr_ring_discard_one() 477 r->batch = SMP_CACHE_BYTES * 2 / sizeof(*(r->queue)); in __ptr_ring_set_size() 483 if (r->batch > r->size / 2 || !r->batch) in __ptr_ring_set_size() 484 r->batch = 1; in __ptr_ring_set_size() 511 static inline void ptr_ring_unconsume(struct ptr_ring *r, void **batch, int n, in ptr_ring_unconsume() argument 544 r->queue[head] = batch[--n]; in ptr_ring_unconsume() 553 destroy(batch[--n]); in ptr_ring_unconsume()
|
/linux/tools/testing/selftests/bpf/prog_tests/ |
A D | bpf_tcp_ca.c | 39 char batch[1500]; in server() local 55 nr_sent = send(fd, &batch, in server() 56 min(total_bytes - bytes, sizeof(batch)), 0); in server() 87 char batch[1500]; in do_test() local 148 nr_recv = recv(fd, &batch, in do_test() 149 min(total_bytes - bytes, sizeof(batch)), 0); in do_test()
|