/linux/drivers/gpu/drm/i915/selftests/ |
A D | i915_gem_evict.c | 56 I915_GTT_PAGE_SIZE); in populate_ggtt() 121 I915_GTT_PAGE_SIZE, 0, 0, in igt_evict_something() 136 I915_GTT_PAGE_SIZE, 0, 0, in igt_evict_something() 168 obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE); in igt_overcommit() 245 .start = I915_GTT_PAGE_SIZE * 2, in igt_evict_for_cache_color() 246 .size = I915_GTT_PAGE_SIZE, in igt_evict_for_cache_color() 263 obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE); in igt_evict_for_cache_color() 272 I915_GTT_PAGE_SIZE | flags); in igt_evict_for_cache_color() 279 obj = i915_gem_object_create_internal(gt->i915, I915_GTT_PAGE_SIZE); in igt_evict_for_cache_color() 289 (I915_GTT_PAGE_SIZE * 2) | flags); in igt_evict_for_cache_color()
|
A D | i915_gem_gtt.c | 717 for (addr = round_up(hole_start + I915_GTT_PAGE_SIZE, step) - I915_GTT_PAGE_SIZE; in pot_hole() 718 addr <= round_down(hole_end - 2*I915_GTT_PAGE_SIZE, step) - I915_GTT_PAGE_SIZE; in pot_hole() 1406 for (total = I915_GTT_PAGE_SIZE; in igt_gtt_reserve() 1476 2 * I915_GTT_PAGE_SIZE, in igt_gtt_reserve() 1527 2*I915_GTT_PAGE_SIZE, 0, in igt_gtt_insert() 1528 0, I915_GTT_PAGE_SIZE, in igt_gtt_insert() 1531 -(u64)I915_GTT_PAGE_SIZE, 0, in igt_gtt_insert() 1532 0, 4*I915_GTT_PAGE_SIZE, in igt_gtt_insert() 1535 -(u64)2*I915_GTT_PAGE_SIZE, 2*I915_GTT_PAGE_SIZE, in igt_gtt_insert() 1536 0, 4*I915_GTT_PAGE_SIZE, in igt_gtt_insert() [all …]
|
/linux/drivers/gpu/drm/i915/display/ |
A D | intel_dpt.c | 46 gen8_set_pte(base + offset / I915_GTT_PAGE_SIZE, in dpt_insert_page() 67 i = vma->node.start / I915_GTT_PAGE_SIZE; in dpt_insert_entries() 95 vma->page_sizes.gtt = I915_GTT_PAGE_SIZE; in dpt_bind_vma() 184 size = DIV_ROUND_UP_ULL(obj->size, I915_GTT_PAGE_SIZE); in intel_dpt_create() 186 size = round_up(size * sizeof(gen8_pte_t), I915_GTT_PAGE_SIZE); in intel_dpt_create() 212 vm->total = (size / sizeof(gen8_pte_t)) * I915_GTT_PAGE_SIZE; in intel_dpt_create()
|
/linux/drivers/gpu/drm/i915/gt/ |
A D | intel_ggtt.c | 32 *start += I915_GTT_PAGE_SIZE; in i915_ggtt_color_adjust() 42 *end -= I915_GTT_PAGE_SIZE; in i915_ggtt_color_adjust() 241 gte += vma->node.start / I915_GTT_PAGE_SIZE; in gen8_ggtt_insert_entries() 242 end = gte + vma->node.size / I915_GTT_PAGE_SIZE; in gen8_ggtt_insert_entries() 292 gte += vma->node.start / I915_GTT_PAGE_SIZE; in gen6_ggtt_insert_entries() 293 end = gte + vma->node.size / I915_GTT_PAGE_SIZE; in gen6_ggtt_insert_entries() 474 vma->page_sizes.gtt = I915_GTT_PAGE_SIZE; in ggtt_bind_vma() 539 ggtt->pin_bias = max_t(u32, I915_GTT_PAGE_SIZE, in init_ggtt() 566 ggtt->error_capture.size = I915_GTT_PAGE_SIZE; in init_ggtt() 1321 sg_dma_len(sg) = I915_GTT_PAGE_SIZE; in rotate_pages() [all …]
|
A D | gen6_ppgtt.c | 74 const unsigned int first_entry = start / I915_GTT_PAGE_SIZE; in gen6_ppgtt_clear_range() 78 unsigned int num_entries = length / I915_GTT_PAGE_SIZE; in gen6_ppgtt_clear_range() 113 unsigned int first_entry = vma->node.start / I915_GTT_PAGE_SIZE; in gen6_ppgtt_insert_entries() 124 GEM_BUG_ON(sg_dma_len(iter.sg) < I915_GTT_PAGE_SIZE); in gen6_ppgtt_insert_entries() 127 iter.dma += I915_GTT_PAGE_SIZE; in gen6_ppgtt_insert_entries() 143 vma->page_sizes.gtt = I915_GTT_PAGE_SIZE; in gen6_ppgtt_insert_entries() 297 u32 ggtt_offset = i915_ggtt_offset(vma) / I915_GTT_PAGE_SIZE; in pd_vma_bind() 339 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in pd_vma_create()
|
A D | selftest_context.c | 111 vaddr += engine->context_size - I915_GTT_PAGE_SIZE; in __live_context_size() 112 memset(vaddr, POISON_INUSE, I915_GTT_PAGE_SIZE); in __live_context_size() 135 if (memchr_inv(vaddr, POISON_INUSE, I915_GTT_PAGE_SIZE)) { in __live_context_size() 177 engine->context_size += I915_GTT_PAGE_SIZE; in live_context_size() 181 engine->context_size -= I915_GTT_PAGE_SIZE; in live_context_size()
|
A D | gen8_ppgtt.c | 424 GEM_BUG_ON(sg_dma_len(iter->sg) < I915_GTT_PAGE_SIZE); in gen8_ppgtt_insert_pte() 427 iter->dma += I915_GTT_PAGE_SIZE; in gen8_ppgtt_insert_pte() 493 page_size = I915_GTT_PAGE_SIZE; in gen8_ppgtt_insert_huge() 499 rem >= (I915_PDES - index) * I915_GTT_PAGE_SIZE)) in gen8_ppgtt_insert_huge() 527 rem >= (I915_PDES - index) * I915_GTT_PAGE_SIZE))) in gen8_ppgtt_insert_huge() 588 if (vma->page_sizes.sg > I915_GTT_PAGE_SIZE) { in gen8_ppgtt_insert() 601 vma->page_sizes.gtt = I915_GTT_PAGE_SIZE; in gen8_ppgtt_insert()
|
A D | intel_gtt.h | 46 #define I915_GTT_PAGE_SIZE I915_GTT_PAGE_SIZE_4K macro 49 #define I915_GTT_PAGE_MASK -I915_GTT_PAGE_SIZE 51 #define I915_GTT_MIN_ALIGNMENT I915_GTT_PAGE_SIZE 149 __for_each_sgt_daddr(__dp, __iter, __sgt, I915_GTT_PAGE_SIZE)
|
A D | intel_lrc.c | 886 memset(vaddr, CONTEXT_REDZONE, I915_GTT_PAGE_SIZE); in set_redzone() 897 if (memchr_inv(vaddr, CONTEXT_REDZONE, I915_GTT_PAGE_SIZE)) in check_redzone() 935 context_size = round_up(engine->context_size, I915_GTT_PAGE_SIZE); in __lrc_alloc_state() 938 context_size += I915_GTT_PAGE_SIZE; /* for redzone */ in __lrc_alloc_state() 1209 GEM_BUG_ON(cs - start > I915_GTT_PAGE_SIZE / sizeof(*cs)); in setup_indirect_ctx_bb()
|
A D | intel_migrate.c | 328 total += I915_GTT_PAGE_SIZE; in emit_pte() 330 it->dma += I915_GTT_PAGE_SIZE; in emit_pte()
|
A D | intel_engine_cs.c | 177 BUILD_BUG_ON(I915_GTT_PAGE_SIZE != PAGE_SIZE); in intel_engine_context_size()
|
/linux/drivers/gpu/drm/i915/ |
A D | i915_gem_gtt.c | 103 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in i915_gem_gtt_reserve() 198 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in i915_gem_gtt_insert() 202 GEM_BUG_ON(start > 0 && !IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); in i915_gem_gtt_insert() 203 GEM_BUG_ON(end < U64_MAX && !IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); in i915_gem_gtt_insert() 225 BUILD_BUG_ON(I915_GTT_MIN_ALIGNMENT > I915_GTT_PAGE_SIZE); in i915_gem_gtt_insert()
|
A D | i915_gem_evict.c | 275 GEM_BUG_ON(!IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); in i915_gem_evict_for_node() 276 GEM_BUG_ON(!IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); in i915_gem_evict_for_node() 291 start -= I915_GTT_PAGE_SIZE; in i915_gem_evict_for_node() 294 end += I915_GTT_PAGE_SIZE; in i915_gem_evict_for_node()
|
A D | i915_vma.c | 153 GEM_BUG_ON(!IS_ALIGNED(vma->size, I915_GTT_PAGE_SIZE)); in vma_create() 644 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in i915_vma_insert() 649 GEM_BUG_ON(!IS_ALIGNED(start, I915_GTT_PAGE_SIZE)); in i915_vma_insert() 655 end = min_t(u64, end, (1ULL << 32) - I915_GTT_PAGE_SIZE); in i915_vma_insert() 656 GEM_BUG_ON(!IS_ALIGNED(end, I915_GTT_PAGE_SIZE)); in i915_vma_insert() 694 vma->page_sizes.sg > I915_GTT_PAGE_SIZE) { in i915_vma_insert()
|
A D | i915_perf.c | 1872 config_length = ALIGN(sizeof(u32) * config_length, I915_GTT_PAGE_SIZE); in alloc_oa_config_buffer()
|
/linux/drivers/gpu/drm/i915/gvt/ |
A D | aperture_gm.c | 53 start = ALIGN(gvt_hidden_gmadr_base(gvt), I915_GTT_PAGE_SIZE); in alloc_gm() 54 end = ALIGN(gvt_hidden_gmadr_end(gvt), I915_GTT_PAGE_SIZE); in alloc_gm() 59 start = ALIGN(gvt_aperture_gmadr_base(gvt), I915_GTT_PAGE_SIZE); in alloc_gm() 60 end = ALIGN(gvt_aperture_gmadr_end(gvt), I915_GTT_PAGE_SIZE); in alloc_gm() 67 size, I915_GTT_PAGE_SIZE, in alloc_gm() 262 vgpu_aperture_sz(vgpu) = ALIGN(request, I915_GTT_PAGE_SIZE); in alloc_resource() 273 vgpu_hidden_sz(vgpu) = ALIGN(request, I915_GTT_PAGE_SIZE); in alloc_resource()
|
A D | reg.h | 116 I915_GTT_PAGE_SIZE)
|
A D | scheduler.c | 184 I915_GTT_PAGE_SIZE - RING_CTX_SIZE); in populate_shadow_context() 238 gpa_size += I915_GTT_PAGE_SIZE; in populate_shadow_context() 248 gpa_size = I915_GTT_PAGE_SIZE; in populate_shadow_context() 1000 gpa_size += I915_GTT_PAGE_SIZE; in update_guest_context() 1010 gpa_size = I915_GTT_PAGE_SIZE; in update_guest_context() 1042 I915_GTT_PAGE_SIZE - sizeof(*shadow_ring_context)); in update_guest_context()
|
A D | cmd_parser.c | 1566 if (guest_gma >= I915_GTT_PAGE_SIZE) { in cmd_address_audit() 1771 offset = gma & (I915_GTT_PAGE_SIZE - 1); in copy_gma_to_hva() 1773 copy_len = (end_gma - gma) >= (I915_GTT_PAGE_SIZE - offset) ? in copy_gma_to_hva() 1774 I915_GTT_PAGE_SIZE - offset : end_gma - gma; in copy_gma_to_hva() 2837 if (WARN_ON(!IS_ALIGNED(workload->rb_start, I915_GTT_PAGE_SIZE))) in scan_workload() 2882 I915_GTT_PAGE_SIZE))) in scan_wa_ctx()
|
A D | gtt.c | 909 (I915_GTT_PAGE_SIZE >> pt_entry_size_shift(spt)) 1447 for (index = 0; index < (I915_GTT_PAGE_SIZE >> in sync_oos_page() 1500 oos_page->mem, I915_GTT_PAGE_SIZE); in attach_oos_page() 2395 int page_entry_num = I915_GTT_PAGE_SIZE >> in alloc_scratch_pages()
|
A D | handlers.c | 1742 !intel_gvt_ggtt_validate_range(vgpu, value, I915_GTT_PAGE_SIZE)) { in hws_pga_write()
|
/linux/drivers/gpu/drm/i915/gem/selftests/ |
A D | huge_gem_object.c | 107 GEM_BUG_ON(!IS_ALIGNED(dma_size, I915_GTT_PAGE_SIZE)); in huge_gem_object()
|
A D | i915_gem_context.c | 1494 GEM_BUG_ON(offset < I915_GTT_PAGE_SIZE); in write_to_scratch() 1591 GEM_BUG_ON(offset < I915_GTT_PAGE_SIZE); in read_from_scratch() 1832 I915_GTT_PAGE_SIZE, vm_total, in igt_vm_isolation()
|
A D | huge_pages.c | 311 GEM_BUG_ON(!IS_ALIGNED(size, I915_GTT_PAGE_SIZE)); in fake_huge_pages_object()
|