Lines Matching refs:pt

17 			   const struct i915_page_table *pt)  in gen6_write_pde()  argument
19 dma_addr_t addr = pt ? px_dma(pt) : px_dma(ppgtt->base.vm.scratch[1]); in gen6_write_pde()
81 struct i915_page_table * const pt = in gen6_ppgtt_clear_range() local
88 GEM_BUG_ON(count > atomic_read(&pt->used)); in gen6_ppgtt_clear_range()
89 if (!atomic_sub_return(count, &pt->used)) in gen6_ppgtt_clear_range()
99 vaddr = px_vaddr(pt); in gen6_ppgtt_clear_range()
149 struct i915_page_table *pt; in gen6_flush_pd() local
157 gen6_for_each_pde(pt, pd, start, end, pde) in gen6_flush_pd()
158 gen6_write_pde(ppgtt, pde, pt); in gen6_flush_pd()
174 struct i915_page_table *pt; in gen6_alloc_va_range() local
180 gen6_for_each_pde(pt, pd, start, length, pde) { in gen6_alloc_va_range()
183 if (!pt) { in gen6_alloc_va_range()
186 pt = stash->pt[0]; in gen6_alloc_va_range()
187 __i915_gem_object_pin_pages(pt->base); in gen6_alloc_va_range()
188 i915_gem_object_make_unshrinkable(pt->base); in gen6_alloc_va_range()
190 fill32_px(pt, vm->scratch[0]->encode); in gen6_alloc_va_range()
194 stash->pt[0] = pt->stash; in gen6_alloc_va_range()
195 atomic_set(&pt->used, 0); in gen6_alloc_va_range()
196 pd->entry[pde] = pt; in gen6_alloc_va_range()
198 pt = pd->entry[pde]; in gen6_alloc_va_range()
204 atomic_add(count, &pt->used); in gen6_alloc_va_range()
253 struct i915_page_table *pt; in gen6_ppgtt_free_pd() local
256 gen6_for_all_pdes(pt, pd, pde) in gen6_ppgtt_free_pd()
257 if (pt) in gen6_ppgtt_free_pd()
258 free_pt(&ppgtt->base.vm, pt); in gen6_ppgtt_free_pd()
309 struct i915_page_table *pt; in pd_vma_unbind() local
316 gen6_for_all_pdes(pt, ppgtt->base.pd, pde) { in pd_vma_unbind()
317 if (!pt || atomic_read(&pt->used)) in pd_vma_unbind()
320 free_pt(&ppgtt->base.vm, pt); in pd_vma_unbind()