/linux/drivers/gpu/drm/i915/gvt/ |
A D | gtt.c | 720 spt = kzalloc(sizeof(*spt), gfp_mask); in alloc_spt() 745 trace_spt_free(spt->vgpu->id, spt, spt->guest_page.type); in ppgtt_free_spt() 1008 trace_spt_change(spt->vgpu->id, "die", spt, in ppgtt_invalidate_spt() 1403 trace_spt_guest_change(spt->vgpu->id, "add", spt, spt->shadow_page.type, in ppgtt_handle_guest_entry_add() 1442 spt, spt->guest_page.type); in sync_oos_page() 1458 spt, spt->guest_page.type, in sync_oos_page() 1480 spt, spt->guest_page.type); in detach_oos_page() 1504 oos_page->spt = spt; in attach_oos_page() 1510 spt, spt->guest_page.type); in attach_oos_page() 1524 spt, spt->guest_page.type); in ppgtt_set_guest_page_sync() [all …]
|
A D | trace.h | 46 TP_ARGS(id, spt, type, mfn, gpt_gfn), 50 __field(void *, spt) 58 __entry->spt = spt; 66 __entry->spt, 75 TP_ARGS(id, spt, type), 79 __field(void *, spt) 85 __entry->spt = spt; 91 __entry->spt, 146 id, action, spt, before, after); 165 id, action, spt, gfn, type); [all …]
|
A D | gtt.h | 235 struct intel_vgpu_ppgtt_spt *spt; member
|
/linux/arch/powerpc/mm/book3s64/ |
A D | subpage_prot.c | 28 if (!spt) in subpage_prot_free() 32 if (spt->low_prot[i]) { in subpage_prot_free() 39 p = spt->protptrs[i]; in subpage_prot_free() 49 spt->maxaddr = 0; in subpage_prot_free() 50 kfree(spt); in subpage_prot_free() 100 if (!spt) in subpage_prot_clear() 105 limit = spt->maxaddr; in subpage_prot_clear() 109 spm = spt->low_prot; in subpage_prot_clear() 225 if (!spt) { in SYSCALL_DEFINE3() 231 if (!spt) { in SYSCALL_DEFINE3() [all …]
|
A D | mmu_context.c | 123 if (current->mm->context.hash_context->spt) { in hash__init_new_context() 124 mm->context.hash_context->spt = kmalloc(sizeof(struct subpage_prot_table), in hash__init_new_context() 126 if (!mm->context.hash_context->spt) { in hash__init_new_context() 137 kfree(mm->context.hash_context->spt); in hash__init_new_context()
|
A D | hash_utils.c | 1221 struct subpage_prot_table *spt = mm_ctx_subpage_prot(&mm->context); in subpage_protection() local 1225 if (!spt) in subpage_protection() 1228 if (ea >= spt->maxaddr) in subpage_protection() 1232 sbpm = spt->low_prot; in subpage_protection() 1234 sbpm = spt->protptrs[ea >> SBP_L3_SHIFT]; in subpage_protection()
|
/linux/arch/x86/kvm/mmu/ |
A D | mmu_audit.c | 40 u64 *ent = sp->spt; in __mmu_spte_walk() 113 gfn = kvm_mmu_page_get_gfn(sp, sptep - sp->spt); in audit_mappings() 136 gfn = kvm_mmu_page_get_gfn(rev_sp, sptep - rev_sp->spt); in inspect_spte_has_rmap() 145 (long int)(sptep - rev_sp->spt), rev_sp->gfn); in inspect_spte_has_rmap() 183 if (!is_shadow_present_pte(sp->spt[i])) in check_mappings_rmap() 186 inspect_spte_has_rmap(kvm, sp->spt + i); in check_mappings_rmap()
|
A D | tdp_mmu.c | 62 free_page((unsigned long)sp->spt); in tdp_mmu_free_sp() 181 sp->spt = kvm_mmu_memory_cache_alloc(&vcpu->arch.mmu_shadow_page_cache); in alloc_tdp_mmu_page() 182 set_page_private(virt_to_page(sp->spt), (unsigned long)sp); in alloc_tdp_mmu_page() 218 return __pa(root->spt); in kvm_tdp_mmu_get_vcpu_root_hpa() 627 for_each_tdp_pte(_iter, _root->spt, _root->role.level, _start, _end) 727 for_each_tdp_pte_min_level(iter, root->spt, root->role.level, in zap_gfn_range() 1003 child_pt = sp->spt; in kvm_tdp_mmu_map() 1183 for_each_tdp_pte_min_level(iter, root->spt, root->role.level, in wrprot_gfn_range() 1439 for_each_tdp_pte_min_level(iter, root->spt, root->role.level, in write_protect_gfn()
|
A D | paging_tmpl.h | 638 i = (sptep - sp->spt) & ~(PTE_PREFETCH_NUM - 1); in FNAME() 639 spte = sp->spt + i; in FNAME() 979 pte_gpa += (sptep - sp->spt) * sizeof(pt_element_t); in FNAME() 1101 if (!sp->spt[i]) in FNAME() 1110 if (FNAME(prefetch_invalid_gpte)(vcpu, sp, &sp->spt[i], gpte)) { in FNAME() 1120 if (sync_mmio_spte(vcpu, &sp->spt[i], gfn, pte_access)) in FNAME() 1124 drop_spte(vcpu->kvm, &sp->spt[i]); in FNAME() 1129 sptep = &sp->spt[i]; in FNAME()
|
A D | mmu_internal.h | 53 u64 *spt; member
|
A D | mmu.c | 1087 gfn = kvm_mmu_page_get_gfn(sp, spte - sp->spt); in rmap_remove() 1640 kvm_mmu_page_set_gfn(sp, spte - sp->spt, gfn); in rmap_add() 1678 static int is_empty_shadow_page(u64 *spt) in is_empty_shadow_page() argument 1707 MMU_WARN_ON(!is_empty_shadow_page(sp->spt)); in kvm_mmu_free_page() 1710 free_page((unsigned long)sp->spt); in kvm_mmu_free_page() 1781 index = spte - sp->spt; in mark_unsync() 1835 u64 ent = sp->spt[i]; in __mmu_unsync_walk() 2770 gfn = kvm_mmu_page_get_gfn(sp, start - sp->spt); in direct_pte_prefetch_many() 2797 spte = sp->spt + i; in __direct_pte_prefetch() 3358 return __pa(sp->spt); in mmu_alloc_root() [all …]
|
/linux/drivers/block/ |
A D | ataflop.c | 171 unsigned spt; /* sectors per track */ member 779 nsect = UDT->spt; in do_format() 1109 if (addr >= PhysTrackBuffer + SUDT->spt*512) { in fd_readtrack_check() 1487 ReqTrack = block / UDT->spt; in setup_req_params() 1488 ReqSector = block - ReqTrack * UDT->spt + 1; in setup_req_params() 1617 getprm.sect = dtp->spt; in fd_locked_ioctl() 1619 getprm.track = dtp->blocks/dtp->spt/2; in fd_locked_ioctl() 1679 && dtp->spt == setprm.sect in fd_locked_ioctl() 1690 default_params[drive].spt = dtp->spt; in fd_locked_ioctl() 1712 dtp->spt = setprm.sect; in fd_locked_ioctl() [all …]
|
/linux/drivers/scsi/esas2r/ |
A D | esas2r_ioctl.c | 695 spt->sense_length = rq->sense_len; in scsi_passthru_comp_cb() 696 spt->residual_length = in scsi_passthru_comp_cb() 734 spt->req_status = sts; in scsi_passthru_comp_cb() 737 spt->target_id = in scsi_passthru_comp_cb() 1012 rq->target_id = (u16)spt->target_id; in hba_ioctl_callback() 1014 memcpy(rq->vrq->scsi.cdb, spt->cdb, 16); in hba_ioctl_callback() 1016 rq->sense_len = spt->sense_length; in hba_ioctl_callback() 1017 rq->sense_buf = (u8 *)spt->sense_data; in hba_ioctl_callback() 1029 if (spt->flags & ATTO_SPTF_DATA_IN) { in hba_ioctl_callback() 1040 if (spt->flags & ATTO_SPTF_ORDERED_Q) in hba_ioctl_callback() [all …]
|
/linux/kernel/ |
A D | torture.c | 726 int spt; in stutter_wait() local 729 spt = READ_ONCE(stutter_pause_test); in stutter_wait() 730 for (; spt; spt = READ_ONCE(stutter_pause_test)) { in stutter_wait() 735 if (spt == 1) { in stutter_wait() 737 } else if (spt == 2) { in stutter_wait()
|
/linux/Documentation/virt/kvm/ |
A D | mmu.rst | 208 spt: 211 The page pointed to by spt will have its page->private pointing back 213 sptes in spt point either at guest pages, or at lower-level shadow pages. 214 Specifically, if sp1 and sp2 are shadow pages, then sp1->spt[n] may point 215 at __pa(sp2->spt). sp2 will point back at sp1 through parent_pte. 216 The spt array forms a DAG structure with the shadow page as a node, and 228 The reverse mapping for the pte/ptes pointing at this page's spt. If 243 A bitmap indicating which sptes in spt point (directly or indirectly) at
|
/linux/arch/powerpc/include/asm/book3s/64/ |
A D | mmu.h | 186 return ctx->hash_context->spt; in mm_ctx_subpage_prot()
|
A D | mmu-hash.h | 723 struct subpage_prot_table *spt; member
|
/linux/arch/s390/boot/ |
A D | head.S | 311 spt 6f-.LPG0(%r13)
|
/linux/drivers/scsi/qla2xxx/ |
A D | qla_isr.c | 2912 struct t10_pi_tuple *spt; in qla2x00_handle_dif_error() local 2934 spt = page_address(sg_page(sg)) + sg->offset; in qla2x00_handle_dif_error() 2935 spt += j; in qla2x00_handle_dif_error() 2937 spt->app_tag = T10_PI_APP_ESCAPE; in qla2x00_handle_dif_error() 2939 spt->ref_tag = T10_PI_REF_ESCAPE; in qla2x00_handle_dif_error()
|
/linux/arch/s390/tools/ |
A D | opcodes.txt | 199 b208 spt S_RD
|
/linux/drivers/gpu/drm/i915/ |
A D | i915_irq.c | 4358 HPD_FUNCS(spt);
|