/linux/drivers/net/ethernet/mellanox/mlx5/core/steering/ |
A D | dr_table.c | 159 ret = dr_table_init_nic(tbl->dmn, &tbl->rx); in dr_table_init_fdb() 163 ret = dr_table_init_nic(tbl->dmn, &tbl->tx); in dr_table_init_fdb() 185 tbl->rx.nic_dmn = &tbl->dmn->info.rx; in dr_table_init() 186 ret = dr_table_init_nic(tbl->dmn, &tbl->rx); in dr_table_init() 190 tbl->tx.nic_dmn = &tbl->dmn->info.tx; in dr_table_init() 191 ret = dr_table_init_nic(tbl->dmn, &tbl->tx); in dr_table_init() 195 tbl->rx.nic_dmn = &tbl->dmn->info.rx; in dr_table_init() 196 tbl->tx.nic_dmn = &tbl->dmn->info.tx; in dr_table_init() 252 tbl = kzalloc(sizeof(*tbl), GFP_KERNEL); in mlx5dr_table_create() 253 if (!tbl) in mlx5dr_table_create() [all …]
|
/linux/fs/nfs/ |
A D | nfs4session.c | 45 p = &tbl->slots; in nfs4_shrink_slot_table() 53 tbl->max_slots--; in nfs4_shrink_slot_table() 111 slot->table = tbl; in nfs4_new_slot() 125 p = &tbl->slots; in nfs4_find_or_create_slot() 128 *p = nfs4_new_slot(tbl, tbl->max_slots, in nfs4_find_or_create_slot() 132 tbl->max_slots++; in nfs4_find_or_create_slot() 248 __func__, tbl->used_slots[0], tbl->highest_used_slotid, in nfs4_alloc_slot() 250 slotid = find_first_zero_bit(tbl->used_slots, tbl->max_slotid + 1); in nfs4_alloc_slot() 257 __func__, tbl->used_slots[0], tbl->highest_used_slotid, in nfs4_alloc_slot() 279 p = &tbl->slots; in nfs4_reset_slot_table() [all …]
|
A D | nfs4session.h | 83 extern int nfs4_setup_slot_table(struct nfs4_slot_table *tbl, 85 extern void nfs4_shutdown_slot_table(struct nfs4_slot_table *tbl); 86 extern struct nfs4_slot *nfs4_alloc_slot(struct nfs4_slot_table *tbl); 88 extern int nfs4_slot_wait_on_seqid(struct nfs4_slot_table *tbl, 93 extern void nfs4_slot_tbl_drain_complete(struct nfs4_slot_table *tbl); 94 bool nfs41_wake_and_assign_slot(struct nfs4_slot_table *tbl, 96 void nfs41_wake_slot_table(struct nfs4_slot_table *tbl); 100 return !!test_bit(NFS4_SLOT_TBL_DRAINING, &tbl->slot_tbl_state); in nfs4_slot_tbl_draining() 106 return !!test_bit(slotid, tbl->used_slots); in nfs4_test_locked_slot() 115 extern void nfs41_set_target_slotid(struct nfs4_slot_table *tbl, [all …]
|
/linux/arch/powerpc/kernel/ |
A D | iommu.c | 46 *val = bitmap_weight(tbl->it_map, tbl->it_size); in iommu_debugfs_weight_get() 371 tbl->it_ops->flush(tbl); in iommu_alloc() 441 tbl->it_ops->clear(tbl, entry, npages); in __iommu_free() 458 tbl->it_ops->flush(tbl); in iommu_free() 569 tbl->it_ops->flush(tbl); in ppc_iommu_map_sg() 633 tbl->it_ops->flush(tbl); in ppc_iommu_unmap_sg() 645 tbl->it_ops->clear(tbl, tbl->it_offset, tbl->it_size); in iommu_table_clear() 655 tceval = tbl->it_ops->get(tbl, index + tbl->it_offset); in iommu_table_clear() 743 tbl->poolsize = (tbl->it_size * 3 / 4) / tbl->nr_pools; in iommu_init_table() 795 tbl->it_ops->free(tbl); in iommu_table_free() [all …]
|
/linux/drivers/infiniband/hw/vmw_pvrdma/ |
A D | pvrdma_doorbell.c | 56 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_init() local 61 tbl->last = 0; in pvrdma_uar_table_init() 62 tbl->top = 0; in pvrdma_uar_table_init() 78 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_table_cleanup() local 89 tbl = &dev->uar_table.tbl; in pvrdma_uar_alloc() 92 obj = find_next_zero_bit(tbl->table, tbl->max, tbl->last); in pvrdma_uar_alloc() 94 tbl->top = (tbl->top + tbl->max) & tbl->mask; in pvrdma_uar_alloc() 95 obj = find_first_zero_bit(tbl->table, tbl->max); in pvrdma_uar_alloc() 117 struct pvrdma_id_table *tbl = &dev->uar_table.tbl; in pvrdma_uar_free() local 124 tbl->last = min(tbl->last, obj); in pvrdma_uar_free() [all …]
|
/linux/include/linux/ |
A D | rhashtable.h | 289 return unlikely(tbl->nest) ? rht_bucket_nested(tbl, hash) : in rht_bucket() 296 return unlikely(tbl->nest) ? __rht_bucket_nested(tbl, hash) : in rht_bucket_var() 427 tbl, hash) 453 rht_ptr(rht_bucket(tbl, hash), tbl, hash), \ 594 tbl = rht_dereference_rcu(ht->tbl, ht); in __rhashtable_lookup() 614 tbl = rht_dereference_rcu(tbl->future_tbl, ht); in __rhashtable_lookup() 715 tbl = rht_dereference_rcu(ht->tbl, ht); in __rhashtable_insert_fast() 1076 tbl = rht_dereference_rcu(ht->tbl, ht); in __rhashtable_remove_fast() 1085 (tbl = rht_dereference_rcu(tbl->future_tbl, ht))) in __rhashtable_remove_fast() 1210 tbl = rht_dereference_rcu(ht->tbl, ht); in rhashtable_replace_fast() [all …]
|
/linux/tools/perf/util/ |
A D | syscalltbl.c | 73 if (tbl->syscalls.entries == NULL) in syscalltbl__init_native() 92 struct syscalltbl *tbl = malloc(sizeof(*tbl)); in syscalltbl__new() local 93 if (tbl) { in syscalltbl__new() 95 free(tbl); in syscalltbl__new() 99 return tbl; in syscalltbl__new() 104 zfree(&tbl->syscalls.entries); in syscalltbl__delete() 105 free(tbl); in syscalltbl__delete() 149 struct syscalltbl *tbl = zalloc(sizeof(*tbl)); in syscalltbl__new() local 150 if (tbl) in syscalltbl__new() 152 return tbl; in syscalltbl__new() [all …]
|
/linux/arch/powerpc/platforms/powernv/ |
A D | pci-ioda-tce.c | 52 tbl->it_blocksize = 16; in pnv_pci_setup_iommu_table() 55 tbl->it_offset = dma_offset >> tbl->it_page_shift; in pnv_pci_setup_iommu_table() 56 tbl->it_index = 0; in pnv_pci_setup_iommu_table() 58 tbl->it_busno = 0; in pnv_pci_setup_iommu_table() 59 tbl->it_type = TCE_PCI; in pnv_pci_setup_iommu_table() 85 __be64 *tmp = user ? tbl->it_userspace : (__be64 *) tbl->it_base; in pnv_tce() 187 return pnv_tce(tbl, true, index - tbl->it_offset, alloc); in pnv_tce_useraddrptr() 209 __be64 *ptce = pnv_tce(tbl, false, index - tbl->it_offset, false); in pnv_tce_get() 244 tbl->it_level_size : tbl->it_size; in pnv_pci_ioda2_table_free_pages() 246 if (!tbl->it_size) in pnv_pci_ioda2_table_free_pages() [all …]
|
/linux/lib/ |
A D | rhashtable.c | 96 unsigned int size = tbl->size >> tbl->nest; in nested_bucket_table_free() 158 size = sizeof(*tbl) + sizeof(tbl->buckets[0]); in nested_bucket_table_alloc() 161 if (!tbl) in nested_bucket_table_alloc() 218 tbl = rht_dereference_rcu(tbl->future_tbl, ht); in rhashtable_last_table() 416 tbl = rht_dereference(ht->tbl, ht); in rht_deferred_worker() 417 tbl = rhashtable_last_table(ht, tbl); in rht_deferred_worker() 789 struct bucket_table *tbl = iter->walker.tbl; in __rhashtable_walk_find_next() local 1069 RCU_INIT_POINTER(ht->tbl, tbl); in rhashtable_init() 1140 tbl = rht_dereference(ht->tbl, ht); in rhashtable_free_and_destroy() 1179 unsigned int size = tbl->size >> tbl->nest; in __rht_bucket_nested() [all …]
|
/linux/drivers/net/wireless/intel/iwlwifi/dvm/ |
A D | rs.c | 566 if (is_siso(tbl->lq_type) && tbl->is_SGI) { in rate_n_flags_from_tbl() 597 tbl->is_dup = 0; in rs_get_tbl_info_from_mcs() 650 if (!tbl->ant_type || tbl->ant_type > ANT_ABC) in rs_toggle_antenna() 1540 tbl->action++; in rs_move_legacy_other() 1553 tbl->action++; in rs_move_legacy_other() 1723 tbl->action++; in rs_move_siso_to_other() 1890 tbl->action++; in rs_move_mimo2_to_other() 2064 tbl->action++; in rs_move_mimo3_to_other() 2469 (is_mimo2(tbl->lq_type) || is_mimo3(tbl->lq_type))) in rs_rate_scale_perform() 2473 (is_mimo2(tbl->lq_type) || is_mimo3(tbl->lq_type))) { in rs_rate_scale_perform() [all …]
|
/linux/arch/powerpc/include/asm/ |
A D | iommu.h | 40 int (*set)(struct iommu_table *tbl, 51 int (*xchg_no_kill)(struct iommu_table *tbl, 57 void (*tce_kill)(struct iommu_table *tbl, 64 void (*clear)(struct iommu_table *tbl, 68 void (*flush)(struct iommu_table *tbl); 69 void (*free)(struct iommu_table *tbl); 120 ((tbl)->it_ops->useraddrptr((tbl), (entry), false)) 122 ((tbl)->it_ops->useraddrptr((tbl), (entry), true)) 215 struct iommu_table *tbl, 299 (tbl)->it_offset, (tbl)->it_size, \ [all …]
|
/linux/net/netfilter/ipvs/ |
A D | ip_vs_lblc.c | 237 tbl->dead = true; in ip_vs_lblc_flush() 279 tbl->rover = j; in ip_vs_lblc_full_check() 311 if (atomic_read(&tbl->entries) <= tbl->max_size) { in ip_vs_lblc_check_expire() 312 tbl->counter++; in ip_vs_lblc_check_expire() 316 goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3; in ip_vs_lblc_check_expire() 336 tbl->rover = j; in ip_vs_lblc_check_expire() 351 tbl = kmalloc(sizeof(*tbl), GFP_KERNEL); in ip_vs_lblc_init_svc() 352 if (tbl == NULL) in ip_vs_lblc_init_svc() 366 tbl->rover = 0; in ip_vs_lblc_init_svc() 369 tbl->svc = svc; in ip_vs_lblc_init_svc() [all …]
|
A D | ip_vs_lblcr.c | 443 tbl->rover = j; in ip_vs_lblcr_full_check() 475 if (atomic_read(&tbl->entries) <= tbl->max_size) { in ip_vs_lblcr_check_expire() 476 tbl->counter++; in ip_vs_lblcr_check_expire() 480 goal = (atomic_read(&tbl->entries) - tbl->max_size)*4/3; in ip_vs_lblcr_check_expire() 500 tbl->rover = j; in ip_vs_lblcr_check_expire() 514 tbl = kmalloc(sizeof(*tbl), GFP_KERNEL); in ip_vs_lblcr_init_svc() 515 if (tbl == NULL) in ip_vs_lblcr_init_svc() 529 tbl->rover = 0; in ip_vs_lblcr_init_svc() 532 tbl->svc = svc; in ip_vs_lblcr_init_svc() 690 if (!tbl->dead) in ip_vs_lblcr_schedule() [all …]
|
/linux/net/core/ |
A D | neighbour.c | 449 n->tbl = tbl; in neigh_alloc() 775 if (tbl->pconstructor && tbl->pconstructor(n)) { in pneigh_lookup() 942 if (atomic_read(&tbl->entries) < tbl->gc_thresh1) in neigh_periodic_work() 1664 p->tbl = tbl; in neigh_parms_alloc() 1728 list_add(&tbl->parms.list, &tbl->parms_list); in neigh_table_init() 1749 if (!tbl->nht || !tbl->phash_buckets) in neigh_table_init() 3233 struct neigh_table *tbl = state->tbl; in pneigh_get_first() local 3256 struct neigh_table *tbl = state->tbl; in pneigh_get_next() local 3312 state->tbl = tbl; in neigh_seq_start() 3356 struct neigh_table *tbl = state->tbl; in neigh_seq_stop() local [all …]
|
/linux/net/mac80211/ |
A D | mesh_pathtbl.c | 45 struct mesh_table *tbl = tblptr; in mesh_path_rht_free() local 47 mesh_path_free_rcu(tbl, mpath); in mesh_path_rht_free() 54 atomic_set(&tbl->entries, 0); in mesh_table_init() 300 struct mesh_table *tbl; in mesh_path_add_gate() local 315 spin_lock(&tbl->gates_lock); in mesh_path_add_gate() 396 struct mesh_table *tbl; in mesh_path_add() local 438 struct mesh_table *tbl; in mpp_path_add() local 513 mesh_gate_del(tbl, mpath); in mesh_path_free_rcu() 517 atomic_dec(&tbl->entries); in mesh_path_free_rcu() 672 struct mesh_table *tbl; in mesh_path_send_to_gates() local [all …]
|
/linux/drivers/net/wireless/intel/iwlegacy/ |
A D | 4965-rs.c | 517 if (tbl->is_SGI) in il4965_rate_n_flags_from_tbl() 522 if (is_siso(tbl->lq_type) && tbl->is_SGI) { in il4965_rate_n_flags_from_tbl() 554 tbl->is_dup = 0; in il4965_rs_get_tbl_info_from_mcs() 602 if (!tbl->ant_type || tbl->ant_type > ANT_ABC) in il4965_rs_toggle_antenna() 1032 if (is_siso(tbl->lq_type) && (!tbl->is_ht40 || lq_sta->is_dup)) in il4965_rs_set_expected_tpt_table() 1352 tbl->action++; in il4965_rs_move_legacy_other() 1365 tbl->action++; in il4965_rs_move_legacy_other() 1476 tbl->action++; in il4965_rs_move_siso_to_other() 1488 tbl->action++; in il4965_rs_move_siso_to_other() 1601 tbl->action++; in il4965_rs_move_mimo2_to_other() [all …]
|
/linux/drivers/net/wireless/marvell/mwifiex/ |
A D | 11n_rxreorder.c | 121 min((start_win - tbl->start_win), tbl->win_size) : in mwifiex_11n_dispatch_pkt_until_start_win() 182 tbl->rx_reorder_ptr[j] = tbl->rx_reorder_ptr[i + j]; in mwifiex_11n_scan_and_dispatch() 206 if (!tbl) in mwifiex_del_rx_reorder_entry() 229 kfree(tbl); in mwifiex_del_rx_reorder_entry() 248 if (!memcmp(tbl->ta, ta, ETH_ALEN) && tbl->tid == tid) { in mwifiex_11n_get_rx_reorder_tbl() 250 return tbl; in mwifiex_11n_get_rx_reorder_tbl() 354 if (tbl) { in mwifiex_11n_create_rx_reorder_tbl() 570 if (!tbl) { in mwifiex_11n_rx_reorder_pkt() 692 if (!tbl) { in mwifiex_del_ba_tbl() 746 if (tbl) in mwifiex_ret_11n_addba_resp() [all …]
|
/linux/drivers/vfio/ |
A D | vfio_iommu_spapr_tce.c | 210 if (tbl) { in tce_iommu_find_table() 216 *ptbl = tbl; in tce_iommu_find_table() 363 if (!tbl) in tce_iommu_release() 366 tce_iommu_clear(container, tbl, tbl->it_offset, tbl->it_size); in tce_iommu_release() 442 if (tbl->it_indirect_levels && tbl->it_userspace) { in tce_iommu_clear() 695 *start_addr = tbl->it_offset << tbl->it_page_shift; in tce_iommu_create_window() 741 tce_iommu_clear(container, tbl, tbl->it_offset, tbl->it_size); in tce_iommu_remove_window() 1150 if (!tbl) in tce_iommu_release_ownership() 1153 tce_iommu_clear(container, tbl, tbl->it_offset, tbl->it_size); in tce_iommu_release_ownership() 1169 if (!tbl || !tbl->it_map) in tce_iommu_take_ownership() [all …]
|
/linux/drivers/firmware/efi/ |
A D | memattr.c | 24 efi_memory_attributes_table_t *tbl; in efi_memattr_init() local 29 tbl = early_memremap(efi_mem_attr_table, sizeof(*tbl)); in efi_memattr_init() 30 if (!tbl) { in efi_memattr_init() 36 if (tbl->version > 1) { in efi_memattr_init() 38 tbl->version); in efi_memattr_init() 42 tbl_size = sizeof(*tbl) + tbl->num_entries * tbl->desc_size; in efi_memattr_init() 47 early_memunmap(tbl, sizeof(*tbl)); in efi_memattr_init() 134 if (tbl_size <= sizeof(*tbl)) in efi_memattr_apply_permissions() 147 if (!tbl) { in efi_memattr_apply_permissions() 162 valid = entry_is_valid((void *)tbl->entry + i * tbl->desc_size, in efi_memattr_apply_permissions() [all …]
|
/linux/arch/powerpc/kvm/ |
A D | book3s_64_vio_hv.c | 171 u64 *tbl; in kvmppc_rm_tce_put() local 183 tbl = kvmppc_page_address(page); in kvmppc_rm_tce_put() 185 tbl[idx % TCES_PER_PAGE] = tce; in kvmppc_rm_tce_put() 221 struct iommu_table *tbl, in iommu_tce_xchg_no_kill_rm() argument 227 ret = tbl->it_ops->xchg_no_kill(tbl, entry, hpa, direction, true); in iommu_tce_xchg_no_kill_rm() 246 if (tbl->it_ops->tce_kill) in iommu_tce_kill_rm() 247 tbl->it_ops->tce_kill(tbl, entry, pages, true); in iommu_tce_kill_rm() 418 stit->tbl, entry); in kvmppc_rm_h_put_tce() 569 stit->tbl, entry + i, ua, in kvmppc_rm_h_put_tce_indirect() 617 stit->tbl, entry + i); in kvmppc_rm_h_stuff_tce() [all …]
|
A D | book3s_64_vio.c | 53 iommu_tce_table_put(stit->tbl); in kvm_spapr_tce_iommu_table_free() 101 struct iommu_table *tbl = NULL; in kvm_spapr_tce_attach_iommu_group() local 148 if (!tbl) in kvm_spapr_tce_attach_iommu_group() 153 if (tbl != stit->tbl) in kvm_spapr_tce_attach_iommu_group() 158 iommu_tce_table_put(tbl); in kvm_spapr_tce_attach_iommu_group() 173 iommu_tce_table_put(tbl); in kvm_spapr_tce_attach_iommu_group() 177 stit->tbl = tbl; in kvm_spapr_tce_attach_iommu_group() 402 u64 *tbl; in kvmppc_tce_put() local 418 tbl = page_to_virt(page); in kvmppc_tce_put() 588 stit->tbl, entry); in kvmppc_h_put_tce() [all …]
|
/linux/net/netfilter/ |
A D | xt_repldata.h | 24 } *tbl; \ 28 tbl = kzalloc(term_offset + sizeof(*term), GFP_KERNEL); \ 29 if (tbl == NULL) \ 32 strncpy(tbl->repl.name, info->name, sizeof(tbl->repl.name)); \ 34 tbl->repl.valid_hooks = hook_mask; \ 35 tbl->repl.num_entries = nhooks + 1; \ 36 tbl->repl.size = nhooks * sizeof(struct type##_standard) + \ 41 tbl->repl.hook_entry[hooknum] = bytes; \ 42 tbl->repl.underflow[hooknum] = bytes; \ 43 tbl->entries[i++] = (struct type##_standard) \ [all …]
|
/linux/drivers/net/ethernet/mellanox/mlx5/core/en/ |
A D | mod_hdr.c | 44 mutex_init(&tbl->lock); in mlx5e_mod_hdr_tbl_init() 45 hash_init(tbl->hlist); in mlx5e_mod_hdr_tbl_init() 50 mutex_destroy(&tbl->lock); in mlx5e_mod_hdr_tbl_destroy() 72 struct mod_hdr_tbl *tbl, in mlx5e_mod_hdr_attach() argument 89 mutex_lock(&tbl->lock); in mlx5e_mod_hdr_attach() 92 mutex_unlock(&tbl->lock); in mlx5e_mod_hdr_attach() 104 mutex_unlock(&tbl->lock); in mlx5e_mod_hdr_attach() 115 mutex_unlock(&tbl->lock); in mlx5e_mod_hdr_attach() 134 mlx5e_mod_hdr_detach(mdev, tbl, mh); in mlx5e_mod_hdr_attach() 139 struct mod_hdr_tbl *tbl, in mlx5e_mod_hdr_detach() argument [all …]
|
/linux/net/sctp/ |
A D | sysctl.c | 374 struct ctl_table tbl; in proc_sctp_do_hmac_alg() local 383 tbl.data = tmp; in proc_sctp_do_hmac_alg() 387 tbl.maxlen = strlen(tbl.data); in proc_sctp_do_hmac_alg() 421 struct ctl_table tbl; in proc_sctp_do_rto_min() local 428 tbl.data = &new_value; in proc_sctp_do_rto_min() 449 struct ctl_table tbl; in proc_sctp_do_rto_max() local 456 tbl.data = &new_value; in proc_sctp_do_rto_max() 485 struct ctl_table tbl; in proc_sctp_do_auth() local 492 tbl.data = &new_value; in proc_sctp_do_auth() 516 struct ctl_table tbl; in proc_sctp_do_udp_port() local [all …]
|
/linux/include/net/ |
A D | neighbour.h | 75 struct neigh_table *tbl; member 132 #define NEIGH_CACHE_STAT_INC(tbl, field) this_cpu_inc((tbl)->stats->field) argument 136 struct neigh_table *tbl; member 244 return p->tbl->family; in neigh_parms_family() 252 return (char *)n + n->tbl->entry_size; in neighbour_priv() 294 struct neigh_table *tbl, in ___neigh_lookup_noref() argument 321 return ___neigh_lookup_noref(tbl, tbl->key_eq, tbl->hash, pkey, dev); in __neigh_lookup_noref() 355 struct neigh_table *tbl); 392 struct neigh_table *tbl; member 538 n = neigh_create(tbl, pkey, dev); in __neigh_lookup() [all …]
|