Lines Matching refs:bq
426 struct xdp_bulk_queue *bq; in __cpu_map_entry_alloc() local
442 bq = per_cpu_ptr(rcpu->bulkq, i); in __cpu_map_entry_alloc()
443 bq->obj = rcpu; in __cpu_map_entry_alloc()
689 static void bq_flush_to_queue(struct xdp_bulk_queue *bq) in bq_flush_to_queue() argument
691 struct bpf_cpu_map_entry *rcpu = bq->obj; in bq_flush_to_queue()
697 if (unlikely(!bq->count)) in bq_flush_to_queue()
703 for (i = 0; i < bq->count; i++) { in bq_flush_to_queue()
704 struct xdp_frame *xdpf = bq->q[i]; in bq_flush_to_queue()
714 bq->count = 0; in bq_flush_to_queue()
717 __list_del_clearprev(&bq->flush_node); in bq_flush_to_queue()
729 struct xdp_bulk_queue *bq = this_cpu_ptr(rcpu->bulkq); in bq_enqueue() local
731 if (unlikely(bq->count == CPU_MAP_BULK_SIZE)) in bq_enqueue()
732 bq_flush_to_queue(bq); in bq_enqueue()
743 bq->q[bq->count++] = xdpf; in bq_enqueue()
745 if (!bq->flush_node.prev) in bq_enqueue()
746 list_add(&bq->flush_node, flush_list); in bq_enqueue()
787 struct xdp_bulk_queue *bq, *tmp; in __cpu_map_flush() local
789 list_for_each_entry_safe(bq, tmp, flush_list, flush_node) { in __cpu_map_flush()
790 bq_flush_to_queue(bq); in __cpu_map_flush()
793 wake_up_process(bq->obj->kthread); in __cpu_map_flush()