/linux/arch/x86/kvm/ |
A D | cpuid.c | 308 if (cpuid->nent) { in kvm_vcpu_ioctl_set_cpuid() 352 if (cpuid->nent) { in kvm_vcpu_ioctl_set_cpuid2() 609 int nent; member 664 ++array->nent; in __do_cpuid_func_emulated() 668 ++array->nent; in __do_cpuid_func_emulated() 675 ++array->nent; in __do_cpuid_func_emulated() 859 --array->nent; in __do_cpuid_func() 1108 .nent = 0, in kvm_dev_ioctl_get_cpuid() 1112 if (cpuid->nent < 1) in kvm_dev_ioctl_get_cpuid() 1121 cpuid->nent)); in kvm_dev_ioctl_get_cpuid() [all …]
|
A D | hyperv.c | 2414 int i, nent = ARRAY_SIZE(cpuid_entries); in kvm_get_hv_cpuid() local 2421 --nent; in kvm_get_hv_cpuid() 2423 if (cpuid->nent < nent) in kvm_get_hv_cpuid() 2426 if (cpuid->nent > nent) in kvm_get_hv_cpuid() 2427 cpuid->nent = nent; in kvm_get_hv_cpuid() 2429 for (i = 0; i < nent; i++) { in kvm_get_hv_cpuid() 2549 nent * sizeof(struct kvm_cpuid_entry2))) in kvm_get_hv_cpuid()
|
/linux/tools/testing/selftests/kvm/x86_64/ |
A D | hyperv_cpuid.c | 52 int nent = 9; in test_hv_cpuid() local 56 nent += 1; /* 0x4000000A */ in test_hv_cpuid() 58 TEST_ASSERT(hv_cpuid_entries->nent == nent, in test_hv_cpuid() 61 nent, evmcs_expected, hv_cpuid_entries->nent); in test_hv_cpuid() 63 for (i = 0; i < hv_cpuid_entries->nent; i++) { in test_hv_cpuid() 115 static struct kvm_cpuid2 cpuid = {.nent = 0}; in test_hv_cpuid_e2big()
|
A D | get_cpuid_test.c | 35 for (i = 0; i < guest_cpuid->nent; i++) { in test_guest_cpuids() 88 for (i = 0; i < cpuid->nent; i++) { in check_cpuid() 114 for (i = 0; i < cpuid1->nent; i++) in compare_cpuids() 117 for (i = 0; i < cpuid2->nent; i++) in compare_cpuids() 147 int size = sizeof(*cpuid) + cpuid->nent * sizeof(cpuid->entries[0]); in vcpu_alloc_cpuid()
|
A D | vmx_set_nested_state_test.c | 252 for (i = 0; i < cpuid->nent; ++i) in disable_vmx() 256 TEST_ASSERT(i != cpuid->nent, "CPUID function 1 not found"); in disable_vmx()
|
/linux/drivers/gpu/drm/nouveau/nvkm/engine/gr/ |
A D | gk20a.c | 44 int nent; in gk20a_gr_av_to_init() local 52 nent = (blob.size / sizeof(struct gk20a_fw_av)); in gk20a_gr_av_to_init() 53 pack = vzalloc((sizeof(*pack) * 2) + (sizeof(*init) * (nent + 1))); in gk20a_gr_av_to_init() 62 for (i = 0; i < nent; i++) { in gk20a_gr_av_to_init() 94 int nent; in gk20a_gr_aiv_to_init() local 102 nent = (blob.size / sizeof(struct gk20a_fw_aiv)); in gk20a_gr_aiv_to_init() 112 for (i = 0; i < nent; i++) { in gk20a_gr_aiv_to_init() 140 int nent; in gk20a_gr_av_to_method() local 148 nent = (blob.size / sizeof(struct gk20a_fw_av)); in gk20a_gr_av_to_method() 151 (sizeof(*init) * (nent + max_classes + 1))); in gk20a_gr_av_to_method() [all …]
|
/linux/drivers/tty/vt/ |
A D | conmakehash.c | 82 int i, nuni, nent; in main() local 272 nent = 0; in main() 275 while ( nent >= unicount[fp0] ) in main() 278 nent = 0; in main() 280 printf("0x%04x", unitable[fp0][nent++]); in main()
|
/linux/tools/testing/selftests/kvm/lib/x86_64/ |
A D | processor.c | 696 int nent = 100; in allocate_kvm_cpuid2() local 707 cpuid->nent = nent; in allocate_kvm_cpuid2() 798 max_ent = cpuid->nent; in vcpu_get_cpuid() 800 for (cpuid->nent = 1; cpuid->nent <= max_ent; cpuid->nent++) { in vcpu_get_cpuid() 837 for (i = 0; i < cpuid->nent; i++) { in kvm_get_supported_cpuid_index() 1345 for (i = 0; i < cpuid->nent; i++) { in set_cpuid() 1393 int i, nent = 0; in vcpu_set_hv_cpuid() local 1400 (cpuid_sys->nent + cpuid_hv->nent) * in vcpu_set_hv_cpuid() 1408 for (i = 0; i < cpuid_sys->nent; i++) { in vcpu_set_hv_cpuid() 1413 nent++; in vcpu_set_hv_cpuid() [all …]
|
/linux/drivers/infiniband/hw/qib/ |
A D | qib_pcie.c | 204 int qib_pcie_params(struct qib_devdata *dd, u32 minw, u32 *nent) in qib_pcie_params() argument 222 maxvec = (nent && *nent) ? *nent : 1; in qib_pcie_params() 232 if (nent) in qib_pcie_params() 233 *nent = !dd->pcidev->msix_enabled ? 0 : nvec; in qib_pcie_params()
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/ |
A D | eq.c | 263 u8 log_eq_size = order_base_2(param->nent + MLX5_NUM_SPARE_EQE); in create_map_eq() 636 .nent = MLX5_NUM_CMD_EQE, in create_async_eqs() 649 .nent = MLX5_NUM_ASYNC_EQE, in create_async_eqs() 659 .nent = /* TODO: sriov max_vf + */ 1, in create_async_eqs() 754 u32 nent = eq_get_size(eq); in mlx5_eq_get_eqe() local 757 eqe = get_eqe(eq, ci & (nent - 1)); in mlx5_eq_get_eqe() 758 eqe = ((eqe->owner & 1) ^ !!(ci & nent)) ? NULL : eqe; in mlx5_eq_get_eqe() 804 int nent; in create_comp_eqs() local 810 nent = MLX5_COMP_EQ_SIZE; in create_comp_eqs() 829 .nent = nent, in create_comp_eqs()
|
/linux/drivers/infiniband/hw/mthca/ |
A D | mthca_eq.c | 184 mthca_write64(MTHCA_EQ_DB_SET_CI | eq->eqn, ci & (eq->nent - 1), in tavor_set_eq_ci() 230 unsigned long off = (entry & (eq->nent - 1)) * MTHCA_EQ_ENTRY_SIZE; in get_eqe() 466 int nent, in mthca_create_eq() argument 479 eq->nent = roundup_pow_of_two(max(nent, 2)); in mthca_create_eq() 480 npages = ALIGN(eq->nent * MTHCA_EQ_ENTRY_SIZE, PAGE_SIZE) / PAGE_SIZE; in mthca_create_eq() 511 for (i = 0; i < eq->nent; ++i) in mthca_create_eq() 535 eq_context->logsize_usrpage = cpu_to_be32((ffs(eq->nent) - 1) << 24); in mthca_create_eq() 560 eq->eqn, eq->nent); in mthca_create_eq() 593 int npages = (eq->nent * MTHCA_EQ_ENTRY_SIZE + PAGE_SIZE - 1) / in mthca_free_eq()
|
A D | mthca_allocator.c | 160 int mthca_array_init(struct mthca_array *array, int nent) in mthca_array_init() argument 162 int npage = (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; in mthca_array_init() 178 void mthca_array_cleanup(struct mthca_array *array, int nent) in mthca_array_cleanup() argument 182 for (i = 0; i < (nent * sizeof (void *) + PAGE_SIZE - 1) / PAGE_SIZE; ++i) in mthca_array_cleanup()
|
A D | mthca_cq.c | 348 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent) in mthca_alloc_cq_buf() argument 353 ret = mthca_buf_alloc(dev, nent * MTHCA_CQ_ENTRY_SIZE, in mthca_alloc_cq_buf() 360 for (i = 0; i < nent; ++i) in mthca_alloc_cq_buf() 768 int mthca_init_cq(struct mthca_dev *dev, int nent, in mthca_init_cq() argument 776 cq->ibcq.cqe = nent - 1; in mthca_init_cq() 814 err = mthca_alloc_cq_buf(dev, &cq->buf, nent); in mthca_init_cq() 828 cq_context->logsize_usrpage = cpu_to_be32((ffs(nent) - 1) << 24); in mthca_init_cq()
|
A D | mthca_dev.h | 421 int mthca_array_init(struct mthca_array *array, int nent); 422 void mthca_array_cleanup(struct mthca_array *array, int nent); 487 int mthca_init_cq(struct mthca_dev *dev, int nent, 498 int mthca_alloc_cq_buf(struct mthca_dev *dev, struct mthca_cq_buf *buf, int nent);
|
A D | mthca_cmd.c | 664 int nent = 0; in mthca_map_cmd() local 694 pages[nent * 2] = cpu_to_be64(virt); in mthca_map_cmd() 698 pages[nent * 2 + 1] = in mthca_map_cmd() 704 if (++nent == MTHCA_MAILBOX_SIZE / 16) { in mthca_map_cmd() 705 err = mthca_cmd(dev, mailbox->dma, nent, 0, op, in mthca_map_cmd() 709 nent = 0; in mthca_map_cmd() 714 if (nent) in mthca_map_cmd() 715 err = mthca_cmd(dev, mailbox->dma, nent, 0, op, in mthca_map_cmd()
|
/linux/fs/nfs/ |
A D | nfs42xattr.c | 73 atomic_long_t nent; member 298 atomic_long_set(&cache->nent, 0); in nfs4_xattr_alloc_cache() 407 atomic_long_set(&cache->nent, 0); in nfs4_xattr_discard_cache() 549 atomic_long_inc(&cache->nent); in nfs4_xattr_hash_add() 578 atomic_long_dec(&cache->nent); in nfs4_xattr_hash_remove() 832 if (atomic_long_read(&cache->nent) > 1) in cache_lru_isolate() 918 atomic_long_dec(&cache->nent); in entry_lru_isolate() 988 atomic_long_set(&cache->nent, 0); in nfs4_xattr_cache_init_once()
|
/linux/drivers/net/ethernet/mellanox/mlx4/ |
A D | eq.c | 110 unsigned long offset = (entry & (eq->nent - 1)) * eqe_size; in get_eqe() 782 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 785 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 807 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 809 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 819 eq->cons_index, eqe->owner, eq->nent, in mlx4_eq_int() 822 !!(eq->cons_index & eq->nent) ? "HW" : "SW"); in mlx4_eq_int() 969 static int mlx4_create_eq(struct mlx4_dev *dev, int nent, in mlx4_create_eq() argument 983 eq->nent = roundup_pow_of_two(max(nent, 2)); in mlx4_create_eq() 987 npages = PAGE_ALIGN(eq->nent * dev->caps.eqe_size) / PAGE_SIZE; in mlx4_create_eq() [all …]
|
A D | cq.c | 342 int mlx4_cq_alloc(struct mlx4_dev *dev, int nent, in mlx4_cq_alloc() argument 382 cpu_to_be32((ilog2(nent) << 24) | in mlx4_cq_alloc() 394 err = mlx4_init_user_cqes(buf_addr, nent, in mlx4_cq_alloc() 399 mlx4_init_kernel_cqes(buf_addr, nent, in mlx4_cq_alloc()
|
/linux/drivers/tty/serial/ |
A D | pch_uart.c | 232 int nent; member 774 for (i = 0; i < priv->nent; i++, sg++) { in pch_dma_tx_complete() 782 priv->nent = 0; in pch_dma_tx_complete() 923 int nent; in dma_handle_tx() local 999 nent = dma_map_sg(port->dev, sg, num, DMA_TO_DEVICE); in dma_handle_tx() 1000 if (!nent) { in dma_handle_tx() 1005 priv->nent = nent; in dma_handle_tx() 1007 for (i = 0; i < nent; i++, sg++) { in dma_handle_tx() 1012 if (i == (nent - 1)) in dma_handle_tx() 1019 priv->sg_tx_p, nent, DMA_MEM_TO_DEV, in dma_handle_tx() [all …]
|
/linux/drivers/iommu/ |
A D | omap-iommu.c | 704 int nent = 1; in iopgtable_clear_entry_core() local 718 nent *= 16; in iopgtable_clear_entry_core() 722 bytes *= nent; in iopgtable_clear_entry_core() 723 memset(iopte, 0, nent * sizeof(*iopte)); in iopgtable_clear_entry_core() 725 flush_iopte_range(obj->dev, pt_dma, pt_offset, nent); in iopgtable_clear_entry_core() 736 nent = 1; /* for the next L1 entry */ in iopgtable_clear_entry_core() 740 nent *= 16; in iopgtable_clear_entry_core() 744 bytes *= nent; in iopgtable_clear_entry_core() 746 memset(iopgd, 0, nent * sizeof(*iopgd)); in iopgtable_clear_entry_core() 747 flush_iopte_range(obj->dev, obj->pd_dma, pd_offset, nent); in iopgtable_clear_entry_core()
|
/linux/lib/ |
A D | scatterlist.c | 604 unsigned int nent, nalloc; in sgl_alloc_order() local 607 nent = round_up(length, PAGE_SIZE << order) >> (PAGE_SHIFT + order); in sgl_alloc_order() 609 if (length > (nent << (PAGE_SHIFT + order))) in sgl_alloc_order() 611 nalloc = nent; in sgl_alloc_order() 639 *nent_p = nent; in sgl_alloc_order()
|
/linux/drivers/infiniband/hw/mlx5/ |
A D | cq.c | 74 static u8 sw_ownership_bit(int n, int nent) in sw_ownership_bit() argument 76 return (n & nent) ? 1 : 0; in sw_ownership_bit() 660 int nent, in alloc_cq_frag_buf() argument 669 nent * cqe_size, in alloc_cq_frag_buf() 678 buf->nent = nent; in alloc_cq_frag_buf() 862 for (i = 0; i < buf->nent; i++) { in init_cq_frag_buf() 1228 (i + 1) & cq->resize_buf->nent); in copy_resize_cqes() 1230 sw_own = sw_ownership_bit(i + 1, cq->resize_buf->nent); in copy_resize_cqes()
|
/linux/include/linux/mlx5/ |
A D | eq.h | 16 int nent; member
|
/linux/arch/x86/include/uapi/asm/ |
A D | kvm.h | 242 __u32 nent; member 264 __u32 nent; member
|
/linux/tools/arch/x86/include/uapi/asm/ |
A D | kvm.h | 242 __u32 nent; member 264 __u32 nent; member
|