Home
last modified time | relevance | path

Searched refs:mw (Results 1 – 25 of 39) sorted by relevance

12

/linux/drivers/infiniband/sw/rxe/
A Drxe_mw.c25 mw->state = (mw->ibmw.type == IB_MW_TYPE_2) ? in rxe_alloc_mw()
34 if (mw->mr) { in rxe_do_dealloc_mw()
42 if (mw->qp) { in rxe_do_dealloc_mw()
50 mw->addr = 0; in rxe_do_dealloc_mw()
152 ((wqe->wr.wr.mw.addr + wqe->wr.wr.mw.length) > in rxe_check_bind_mw()
168 mw->rkey = (mw->rkey & ~0xff) | key; in rxe_do_bind_mw()
169 mw->access = wqe->wr.wr.mw.access; in rxe_do_bind_mw()
171 mw->addr = wqe->wr.wr.mw.addr; in rxe_do_bind_mw()
172 mw->length = wqe->wr.wr.mw.length; in rxe_do_bind_mw()
288 if (!mw) { in rxe_invalidate_mw()
[all …]
A Drxe_verbs.h465 static inline struct rxe_mw *to_rmw(struct ib_mw *mw) in to_rmw() argument
467 return mw ? container_of(mw, struct rxe_mw, ibmw) : NULL; in to_rmw()
480 static inline struct rxe_pd *rxe_mw_pd(struct rxe_mw *mw) in rxe_mw_pd() argument
482 return to_rpd(mw->ibmw.pd); in rxe_mw_pd()
A Drxe_resp.c407 struct rxe_mw *mw = NULL; in check_rkey() local
449 mw = rxe_lookup_mw(qp, access, rkey); in check_rkey()
450 if (!mw) { in check_rkey()
456 mr = mw->mr; in check_rkey()
463 if (mw->access & IB_ZERO_BASED) in check_rkey()
464 qp->resp.offset = mw->addr; in check_rkey()
466 rxe_drop_ref(mw); in check_rkey()
511 if (mw) in check_rkey()
512 rxe_drop_ref(mw); in check_rkey()
/linux/net/netfilter/ipvs/
A Dip_vs_wrr.c63 int mw; /* maximum weight */ member
119 mark->mw = ip_vs_wrr_max_weight(svc) - (mark->di - 1); in ip_vs_wrr_init_svc()
120 mark->cw = mark->mw; in ip_vs_wrr_init_svc()
146 mark->mw = ip_vs_wrr_max_weight(svc) - (mark->di - 1); in ip_vs_wrr_dest_changed()
147 if (mark->cw > mark->mw || !mark->cw) in ip_vs_wrr_dest_changed()
148 mark->cw = mark->mw; in ip_vs_wrr_dest_changed()
172 if (mark->mw == 0) in ip_vs_wrr_schedule()
188 mark->cw = mark->mw; in ip_vs_wrr_schedule()
A Dip_vs_mh.c348 int mw, shift; in ip_vs_mh_shift_weight() local
366 mw = weight / gcd; in ip_vs_mh_shift_weight()
369 shift = fls(mw) - IP_VS_MH_TAB_BITS; in ip_vs_mh_shift_weight()
/linux/drivers/ntb/
A Dntb_transport.c631 if (!mw->virt_addr) in ntb_transport_setup_qp_mw()
796 if (!mw->virt_addr) in ntb_free_mw()
801 mw->alloc_addr, mw->dma_addr); in ntb_free_mw()
802 mw->xlat_size = 0; in ntb_free_mw()
803 mw->buff_size = 0; in ntb_free_mw()
820 mw->alloc_size); in ntb_alloc_mw_buffer()
832 if (mw->alloc_size > mw->buff_size) { in ntb_alloc_mw_buffer()
878 if (mw->buff_size) in ntb_set_mw()
1317 mw->vbase = ioremap_wc(mw->phys_addr, mw->phys_size); in ntb_transport_probe()
1318 if (!mw->vbase) { in ntb_transport_probe()
[all …]
/linux/fs/ocfs2/
A Ddlmglue.c474 struct ocfs2_mask_waiter *mw; in ocfs2_track_lock_wait() local
898 if ((lockres->l_flags & mw->mw_mask) != mw->mw_goal) in lockres_set_flags()
902 mw->mw_status = 0; in lockres_set_flags()
903 complete(&mw->mw_complete); in lockres_set_flags()
1393 ocfs2_init_start_time(mw); in ocfs2_init_mask_waiter()
1401 return mw->mw_status; in ocfs2_wait_for_mask()
1414 mw->mw_mask = mask; in lockres_add_mask_waiter()
1415 mw->mw_goal = goal; in lockres_add_mask_waiter()
1428 if ((lockres->l_flags & mw->mw_mask) != mw->mw_goal) in __lockres_remove_mask_waiter()
1462 ret = mw->mw_status; in ocfs2_wait_for_mask_interruptible()
[all …]
/linux/drivers/net/ethernet/mellanox/mlx4/
A Dmr.c815 struct mlx4_mw *mw) in mlx4_mw_alloc() argument
830 mw->pd = pd; in mlx4_mw_alloc()
831 mw->type = type; in mlx4_mw_alloc()
832 mw->enabled = MLX4_MPT_DISABLED; in mlx4_mw_alloc()
860 if (mw->type == MLX4_MW_TYPE_2) { in mlx4_mw_enable()
867 key_to_hw_index(mw->key) & in mlx4_mw_enable()
873 mw->enabled = MLX4_MPT_EN_HW; in mlx4_mw_enable()
892 if (mw->enabled == MLX4_MPT_EN_HW) { in mlx4_mw_free()
894 key_to_hw_index(mw->key) & in mlx4_mw_free()
899 mw->enabled = MLX4_MPT_EN_SW; in mlx4_mw_free()
[all …]
/linux/drivers/infiniband/hw/hns/
A Dhns_roce_mr.c483 struct hns_roce_mw *mw) in hns_roce_mw_free() argument
488 if (mw->enabled) { in hns_roce_mw_free()
496 key_to_hw_index(mw->rkey)); in hns_roce_mw_free()
500 (int)key_to_hw_index(mw->rkey)); in hns_roce_mw_free()
504 struct hns_roce_mw *mw) in hns_roce_mw_enable() argument
536 mw->enabled = 1; in hns_roce_mw_enable()
568 mw->rkey = hw_index_to_key(id); in hns_roce_alloc_mw()
570 ibmw->rkey = mw->rkey; in hns_roce_alloc_mw()
571 mw->pdn = to_hr_pd(ibmw->pd)->pdn; in hns_roce_alloc_mw()
583 hns_roce_mw_free(hr_dev, mw); in hns_roce_alloc_mw()
[all …]
/linux/drivers/clk/rockchip/
A Dclk.h473 .mux_width = mw, \
494 .mux_width = mw, \
542 #define COMPOSITE_NODIV(_id, cname, pnames, f, mo, ms, mw, mf, \ argument
553 .mux_width = mw, \
571 .mux_width = mw, \
580 mw, mf, ds, dw, df, dt) \ argument
590 .mux_width = mw, \
650 #define COMPOSITE_DDRCLK(_id, cname, pnames, f, mo, ms, mw, \ argument
661 .mux_width = mw, \
802 .mux_width = mw, \
[all …]
/linux/arch/mips/txx9/generic/
A Dmem_tx4927.c46 unsigned int mw = 0; in tx4927_process_sdccr() local
61 mw = 8 >> sdccr_mw; in tx4927_process_sdccr()
64 return rs * cs * mw * bs; in tx4927_process_sdccr()
/linux/drivers/net/ethernet/marvell/octeontx2/af/
A Drvu.c2110 mbox = &mw->mbox; in __rvu_mbox_handler()
2195 mbox = &mw->mbox_up; in __rvu_mbox_up_handler()
2350 if (!mw->mbox_wq) { in rvu_mbox_init()
2357 if (!mw->mbox_wrk) { in rvu_mbox_init()
2364 if (!mw->mbox_wrk_up) { in rvu_mbox_init()
2407 if (mw->mbox_wq) { in rvu_mbox_destroy()
2409 mw->mbox_wq = NULL; in rvu_mbox_destroy()
2435 mbox = &mw->mbox; in rvu_queue_work()
2450 queue_work(mw->mbox_wq, &mw->mbox_wrk[i].work); in rvu_queue_work()
2452 mbox = &mw->mbox_up; in rvu_queue_work()
[all …]
/linux/include/linux/usb/
A Dpd.h249 #define PDO_BATT_MAX_POWER(mw) ((((mw) / 250) & PDO_PWR_MASK) << PDO_BATT_MAX_PWR_SHIFT) argument
374 #define RDO_BATT_OP_PWR(mw) ((((mw) / 250) & RDO_PWR_MASK) << RDO_BATT_OP_PWR_SHIFT) argument
375 #define RDO_BATT_MAX_PWR(mw) ((((mw) / 250) & RDO_PWR_MASK) << RDO_BATT_MAX_PWR_SHIFT) argument
/linux/drivers/infiniband/hw/mlx4/
A Dmr.c617 struct mlx4_ib_mw *mw = to_mmw(ibmw); in mlx4_ib_alloc_mw() local
621 to_mlx4_type(ibmw->type), &mw->mmw); in mlx4_ib_alloc_mw()
625 err = mlx4_mw_enable(dev->dev, &mw->mmw); in mlx4_ib_alloc_mw()
629 ibmw->rkey = mw->mmw.key; in mlx4_ib_alloc_mw()
633 mlx4_mw_free(dev->dev, &mw->mmw); in mlx4_ib_alloc_mw()
639 struct mlx4_ib_mw *mw = to_mmw(ibmw); in mlx4_ib_dealloc_mw() local
641 mlx4_mw_free(to_mdev(ibmw->device)->dev, &mw->mmw); in mlx4_ib_dealloc_mw()
/linux/drivers/mtd/maps/
A Dphysmap-core.c154 map_word mw; in physmap_addr_gpios_read() local
162 mw.x[0] = word; in physmap_addr_gpios_read()
163 return mw; in physmap_addr_gpios_read()
188 static void physmap_addr_gpios_write(struct map_info *map, map_word mw, in physmap_addr_gpios_write() argument
199 word = mw.x[0]; in physmap_addr_gpios_write()
/linux/drivers/pci/endpoint/functions/
A Dpci-epf-ntb.c236 enum pci_epc_interface_type type, u32 mw) in epf_ntb_configure_mw() argument
252 peer_barno = peer_ntb_epc->epf_ntb_bar[mw + NTB_MW_OFFSET]; in epf_ntb_configure_mw()
259 if (mw + NTB_MW_OFFSET == BAR_DB_MW1) in epf_ntb_configure_mw()
262 if (size > ntb->mws_size[mw]) { in epf_ntb_configure_mw()
265 pci_epc_interface_string(type), mw, size, in epf_ntb_configure_mw()
266 ntb->mws_size[mw]); in epf_ntb_configure_mw()
278 pci_epc_interface_string(type), mw); in epf_ntb_configure_mw()
295 enum pci_epc_interface_type type, u32 mw) in epf_ntb_teardown_mw() argument
309 peer_barno = peer_ntb_epc->epf_ntb_bar[mw + NTB_MW_OFFSET]; in epf_ntb_teardown_mw()
314 if (mw + NTB_MW_OFFSET == BAR_DB_MW1) in epf_ntb_teardown_mw()
/linux/drivers/infiniband/hw/bnxt_re/
A Dib_verbs.c448 if (fence->mw) { in bnxt_re_destroy_fence_mr()
450 fence->mw = NULL; in bnxt_re_destroy_fence_mr()
476 struct ib_mw *mw; in bnxt_re_create_fence_mr() local
521 if (IS_ERR(mw)) { in bnxt_re_create_fence_mr()
524 rc = PTR_ERR(mw); in bnxt_re_create_fence_mr()
527 fence->mw = mw; in bnxt_re_create_fence_mr()
3728 mw = kzalloc(sizeof(*mw), GFP_KERNEL); in bnxt_re_alloc_mw()
3729 if (!mw) in bnxt_re_alloc_mw()
3742 mw->ib_mw.rkey = mw->qplib_mw.rkey; in bnxt_re_alloc_mw()
3748 kfree(mw); in bnxt_re_alloc_mw()
[all …]
A Dib_verbs.h53 struct ib_mw *mw; member
206 int bnxt_re_dealloc_mw(struct ib_mw *mw);
/linux/include/dt-bindings/usb/
A Dpd.h44 #define PDO_BATT_MAX_POWER(mw) ((((mw) / 250) & PDO_PWR_MASK) << PDO_BATT_MAX_PWR_SHIFT) argument
/linux/drivers/infiniband/core/
A Duverbs_main.c105 int uverbs_dealloc_mw(struct ib_mw *mw) in uverbs_dealloc_mw() argument
107 struct ib_pd *pd = mw->pd; in uverbs_dealloc_mw()
110 ret = mw->device->ops.dealloc_mw(mw); in uverbs_dealloc_mw()
115 kfree(mw); in uverbs_dealloc_mw()
A Duverbs.h245 int uverbs_dealloc_mw(struct ib_mw *mw);
A Duverbs_cmd.c906 struct ib_mw *mw; in ib_uverbs_alloc_mw() local
929 mw = rdma_zalloc_drv_obj(ib_dev, ib_mw); in ib_uverbs_alloc_mw()
930 if (!mw) { in ib_uverbs_alloc_mw()
935 mw->device = ib_dev; in ib_uverbs_alloc_mw()
936 mw->pd = pd; in ib_uverbs_alloc_mw()
937 mw->uobject = uobj; in ib_uverbs_alloc_mw()
938 mw->type = cmd.mw_type; in ib_uverbs_alloc_mw()
940 ret = pd->device->ops.alloc_mw(mw, &attrs->driver_udata); in ib_uverbs_alloc_mw()
946 uobj->object = mw; in ib_uverbs_alloc_mw()
950 resp.rkey = mw->rkey; in ib_uverbs_alloc_mw()
[all …]
/linux/include/uapi/rdma/
A Drdma_user_rxe.h113 } mw; member
/linux/drivers/infiniband/hw/mlx5/
A Dmr.c2258 struct mlx5_ib_mw *mw = to_mmw(ibmw); in mlx5_ib_alloc_mw() local
2300 err = mlx5_ib_create_mkey(dev, &mw->mmkey, in, inlen); in mlx5_ib_alloc_mw()
2304 mw->mmkey.type = MLX5_MKEY_MW; in mlx5_ib_alloc_mw()
2305 ibmw->rkey = mw->mmkey.key; in mlx5_ib_alloc_mw()
2306 mw->mmkey.ndescs = ndescs; in mlx5_ib_alloc_mw()
2317 err = mlx5r_store_odp_mkey(dev, &mw->mmkey); in mlx5_ib_alloc_mw()
2326 mlx5_core_destroy_mkey(dev->mdev, mw->mmkey.key); in mlx5_ib_alloc_mw()
2332 int mlx5_ib_dealloc_mw(struct ib_mw *mw) in mlx5_ib_dealloc_mw() argument
2334 struct mlx5_ib_dev *dev = to_mdev(mw->device); in mlx5_ib_dealloc_mw()
2335 struct mlx5_ib_mw *mmw = to_mmw(mw); in mlx5_ib_dealloc_mw()
/linux/drivers/net/ethernet/marvell/octeontx2/nic/
A Dotx2_pf.c293 static void otx2_queue_work(struct mbox *mw, struct workqueue_struct *mbox_wq, in otx2_queue_work() argument
306 mbox = &mw->mbox; in otx2_queue_work()
319 mw[i].num_msgs = hdr->num_msgs; in otx2_queue_work()
326 queue_work(mbox_wq, &mw[i].mbox_wrk); in otx2_queue_work()
329 mbox = &mw->mbox_up; in otx2_queue_work()
335 mw[i].up_num_msgs = hdr->num_msgs; in otx2_queue_work()
342 queue_work(mbox_wq, &mw[i].mbox_up_wrk); in otx2_queue_work()

Completed in 103 milliseconds

12