Home
last modified time | relevance | path

Searched refs:q (Results 1 – 25 of 1341) sorted by relevance

12345678910>>...54

/linux/crypto/
A Dgf128mul.c57 q(0x00), q(0x01), q(0x02), q(0x03), q(0x04), q(0x05), q(0x06), q(0x07),\
58 q(0x08), q(0x09), q(0x0a), q(0x0b), q(0x0c), q(0x0d), q(0x0e), q(0x0f),\
59 q(0x10), q(0x11), q(0x12), q(0x13), q(0x14), q(0x15), q(0x16), q(0x17),\
60 q(0x18), q(0x19), q(0x1a), q(0x1b), q(0x1c), q(0x1d), q(0x1e), q(0x1f),\
61 q(0x20), q(0x21), q(0x22), q(0x23), q(0x24), q(0x25), q(0x26), q(0x27),\
62 q(0x28), q(0x29), q(0x2a), q(0x2b), q(0x2c), q(0x2d), q(0x2e), q(0x2f),\
63 q(0x30), q(0x31), q(0x32), q(0x33), q(0x34), q(0x35), q(0x36), q(0x37),\
64 q(0x38), q(0x39), q(0x3a), q(0x3b), q(0x3c), q(0x3d), q(0x3e), q(0x3f),\
65 q(0x40), q(0x41), q(0x42), q(0x43), q(0x44), q(0x45), q(0x46), q(0x47),\
66 q(0x48), q(0x49), q(0x4a), q(0x4b), q(0x4c), q(0x4d), q(0x4e), q(0x4f),\
[all …]
/linux/Documentation/networking/
A Dtls-offload-layers.svg1q-0.609375 -0.359375 -1.203125 -0.359375q-0.546875 0 -0.96875 0.328125q-0.421875 0.328125 -0.60937…
/linux/drivers/media/v4l2-core/
A Dvideobuf-core.c254 q->ops->buf_release(q, q->bufs[i]); in __videobuf_free()
289 q->ops->buf_release(q, q->bufs[i]); in videobuf_queue_cancel()
812 q->ops->buf_queue(q, q->read_buf); in videobuf_read_zerocopy()
816 CALL(q, sync, q, q->read_buf); in videobuf_read_zerocopy()
825 q->ops->buf_release(q, q->read_buf); in videobuf_read_zerocopy()
919 q->ops->buf_queue(q, q->read_buf); in videobuf_read_one()
930 CALL(q, sync, q, q->read_buf); in videobuf_read_one()
934 q->ops->buf_release(q, q->read_buf); in videobuf_read_one()
949 q->ops->buf_release(q, q->read_buf); in videobuf_read_one()
990 q->ops->buf_queue(q, q->bufs[i]); in __videobuf_read_start()
[all …]
/linux/sound/core/seq/
A Dseq_queue.c76 if (q) { in queue_list_remove()
100 q = kzalloc(sizeof(*q), GFP_KERNEL); in queue_new()
101 if (!q) in queue_new()
113 if (q->tickq == NULL || q->timeq == NULL || q->timer == NULL) { in queue_new()
125 return q; in queue_new()
143 kfree(q); in queue_delete()
183 return q; in snd_seq_queue_alloc()
212 if (q) in queueptr()
346 return (q->owner == client) || (!q->locked && !q->klocked); in check_access()
382 if (! q) in snd_seq_queue_check_access()
[all …]
/linux/drivers/media/common/videobuf2/
A Dvideobuf2-core.c158 ((q)->ops->op ? (q)->ops->op(args) : 0)
180 if (q && q->buf_ops && q->buf_ops->op) \
187 if (q && q->buf_ops && q->buf_ops->op) \
235 q->alloc_devs[plane] ? : q->dev, in __vb2_buf_mem_alloc()
1828 call_void_qop(q, wait_prepare, q); in __vb2_wait_for_done_vb()
1842 call_void_qop(q, wait_finish, q); in __vb2_wait_for_done_vb()
2500 if (q->is_output && q->fileio && q->queued_count < q->num_buffers) in vb2_core_poll()
2710 vb2_core_streamoff(q, q->type); in __vb2_cleanup_fileio()
2938 call_void_qop(q, wait_finish, q); in vb2_thread()
2953 call_void_qop(q, wait_finish, q); in vb2_thread()
[all …]
/linux/sound/core/seq/oss/
A Dseq_oss_readq.c37 q = kzalloc(sizeof(*q), GFP_KERNEL); in snd_seq_oss_readq_new()
42 if (!q->q) { in snd_seq_oss_readq_new()
49 q->head = q->tail = 0; in snd_seq_oss_readq_new()
65 kfree(q->q); in snd_seq_oss_readq_delete()
78 q->head = q->tail = 0; in snd_seq_oss_readq_clear()
151 memcpy(&q->q[q->tail], ev, sizeof(*ev)); in snd_seq_oss_readq_put_event()
152 q->tail = (q->tail + 1) % q->maxlen; in snd_seq_oss_readq_put_event()
173 memcpy(rec, &q->q[q->head], sizeof(*rec)); in snd_seq_oss_readq_pick()
184 (q->qlen > 0 || q->head == q->tail), in snd_seq_oss_readq_wait()
196 q->head = (q->head + 1) % q->maxlen; in snd_seq_oss_readq_free()
[all …]
/linux/net/sched/
A Dsch_choke.c78 return (q->tail - q->head) & q->tab_mask; in choke_len()
97 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes()
98 if (q->head == q->tail) in choke_zap_head_holes()
107 q->tail = (q->tail - 1) & q->tab_mask; in choke_zap_tail_holes()
108 if (q->head == q->tail) in choke_zap_tail_holes()
206 if (q->head == q->tail) in choke_match_random()
270 q->tail = (q->tail + 1) & q->tab_mask; in choke_enqueue()
295 skb = q->tab[q->head]; in choke_dequeue()
312 q->head = (q->head + 1) & q->tab_mask; in choke_reset()
389 q->head = (q->head + 1) & q->tab_mask; in choke_change()
[all …]
A Dsch_netem.c291 return q->loss && q->loss >= get_crandom(&q->loss_cor); in loss_event()
373 rtnl_kfree_skbs(q->t_head, q->t_tail); in tfifo_reset()
451 if (q->duplicate && q->duplicate >= get_crandom(&q->dup_cor)) in netem_enqueue()
470 if (q->latency || q->jitter || q->rate) in netem_enqueue()
494 if (q->corrupt && q->corrupt >= get_crandom(&q->corrupt_cor)) { in netem_enqueue()
536 delay = tabledist(q->latency, q->jitter, in netem_enqueue()
537 &q->delay_cor, q->delay_dist); in netem_enqueue()
645 q->slot.bytes_left = q->slot_config.max_bytes; in get_slot_next()
737 skb = q->qdisc->ops->dequeue(q->qdisc); in netem_dequeue()
748 skb = q->qdisc->ops->dequeue(q->qdisc); in netem_dequeue()
[all …]
A Dsch_sfq.c461 if (++sch->q.qlen <= q->limit) in sfq_enqueue()
615 if (!q->filter_list && q->tail) in sfq_perturbation()
660 q->scaled_quantum = SFQ_ALLOT_SIZE(q->quantum); in sfq_change()
667 q->maxflows = min_t(u32, q->maxflows, q->divisor); in sfq_change()
685 q->limit = min_t(u32, ctl->limit, q->maxdepth * q->maxflows); in sfq_change()
686 q->maxflows = min_t(u32, q->maxflows, q->limit); in sfq_change()
690 while (sch->q.qlen > q->limit) { in sfq_change()
757 q->scaled_quantum = SFQ_ALLOT_SIZE(q->quantum); in sfq_init()
767 q->ht = sfq_alloc(sizeof(q->ht[0]) * q->divisor); in sfq_init()
768 q->slots = sfq_alloc(sizeof(q->slots[0]) * q->maxflows); in sfq_init()
[all …]
A Dsch_sfb.c192 memset(&q->bins, 0, sizeof(q->bins)); in sfb_zero_all_buckets()
202 const struct sfb_bucket *b = &q->bins[q->slot].bins[0][0]; in sfb_compute_qlen()
226 sfb_init_perturbation(q->slot, q); in sfb_swap_slot()
227 q->slot ^= 1; in sfb_swap_slot()
236 if (q->penalty_rate == 0 || q->penalty_burst == 0) in sfb_rate_limit()
242 q->tokens_avail = (age * q->penalty_rate) / HZ; in sfb_rate_limit()
243 if (q->tokens_avail > q->penalty_burst) in sfb_rate_limit()
244 q->tokens_avail = q->penalty_burst; in sfb_rate_limit()
293 if (unlikely(sch->q.qlen >= q->limit)) { in sfb_enqueue()
458 q->slot = 0; in sfb_reset()
[all …]
A Dsch_fq_pie.c159 q->overmemory++; in fq_pie_qdisc_enqueue()
183 sch->q.qlen++; in fq_pie_qdisc_enqueue()
256 sch->q.qlen--; in fq_pie_qdisc_dequeue()
262 if (head == &q->new_flows && !list_empty(&q->old_flows)) in fq_pie_qdisc_dequeue()
307 if (!q->flows_cnt || q->flows_cnt > 65536) { in fq_pie_change()
374 struct fq_pie_sched_data *q = from_timer(q, t, adapt_timer); in fq_pie_timer() local
383 pie_calculate_probability(&q->p_params, &q->flows[idx].vars, in fq_pie_timer()
388 mod_timer(&q->adapt_timer, jiffies + q->p_params.tupdate); in fq_pie_timer()
404 q->sch = sch; in fq_pie_init()
420 err = tcf_block_get(&q->block, &q->filter_list, sch, extack); in fq_pie_init()
[all …]
A Dsch_fq_codel.c180 sch->q.qlen -= i; in fq_codel_drop()
216 memory_limited = q->memory_usage > q->memory_limit; in fq_codel_enqueue()
265 q->backlogs[flow - q->flows] -= qdisc_pkt_len(skb); in dequeue_func()
267 sch->q.qlen--; in dequeue_func()
309 if ((head == &q->new_flows) && !list_empty(&q->old_flows)) in fq_codel_dequeue()
320 if (q->cstats.drop_count && sch->q.qlen) { in fq_codel_dequeue()
349 memset(q->backlogs, 0, q->flows_cnt * sizeof(u32)); in fq_codel_reset()
385 if (q->flows) in fq_codel_change()
440 q->memory_usage > q->memory_limit) { in fq_codel_change()
489 err = tcf_block_get(&q->block, &q->filter_list, sch, extack); in fq_codel_init()
[all …]
A Dsch_fq.c302 root = &q->fq_root[hash_ptr(sk, q->fq_trees_log)]; in fq_classify()
304 if (q->flows >= (2U << q->fq_trees_log) && in fq_classify()
305 q->inactive_flows > q->flows/2) in fq_classify()
359 q->flows++; in fq_classify()
401 sch->q.qlen--; in fq_dequeue_skb()
467 skb->tstamp = q->ktime_cache + q->horizon; in fq_enqueue()
494 sch->q.qlen++; in fq_enqueue()
511 q->unthrottle_latency_ns -= q->unthrottle_latency_ns >> 3; in fq_check_throttled()
588 if ((head == &q->new_flows) && q->old_flows.first) { in fq_dequeue()
754 if (q->fq_root && log == q->fq_trees_log) in fq_resize()
[all …]
A Dsch_red.c77 q->vars.qavg = red_calc_qavg(&q->parms, in red_enqueue()
84 switch (red_action(&q->parms, &q->vars, q->vars.qavg)) { in red_enqueue()
110 if (red_use_harddrop(q) || !red_use_ecn(q)) { in red_enqueue()
198 opt.set.min = q->parms.qth_min >> q->parms.Wlog; in red_offload()
199 opt.set.max = q->parms.qth_max >> q->parms.Wlog; in red_offload()
302 if (!q->qdisc->q.qlen) in __red_change()
327 red_adaptative_algo(&q->parms, &q->vars); in red_adaptative_timer()
417 .qth_min = q->parms.qth_min >> q->parms.Wlog, in red_dump()
418 .qth_max = q->parms.qth_max >> q->parms.Wlog, in red_dump()
464 st.early = q->stats.prob_drop + q->stats.forced_drop; in red_dump_stats()
[all …]
/linux/net/xdp/
A Dxsk_queue.h124 if (q->cached_cons != q->cached_prod) { in xskq_cons_read_addr_unchecked()
194 while (q->cached_cons != q->cached_prod) { in xskq_cons_read_desc()
196 u32 idx = q->cached_cons & q->ring_mask; in xskq_cons_read_desc()
271 if (q->cached_prod == q->cached_cons) in xskq_cons_peek_addr_unchecked()
280 if (q->cached_prod == q->cached_cons) in xskq_cons_peek_desc()
324 u32 free_entries = q->nentries - (q->cached_prod - q->cached_cons); in xskq_prod_nb_free()
331 free_entries = q->nentries - (q->cached_prod - q->cached_cons); in xskq_prod_nb_free()
395 idx = q->cached_prod++ & q->ring_mask; in xskq_prod_reserve_desc()
409 __xskq_prod_submit(q, q->cached_prod); in xskq_prod_submit()
437 return q ? q->invalid_descs : 0; in xskq_nb_invalid_descs()
[all …]
/linux/block/
A Dblk-sysfs.c73 if (!queue_is_mq(q)) in queue_requests_store()
94 if (!q->disk) in queue_ra_show()
106 if (!q->disk) in queue_ra_store()
258 if (q->disk) in queue_max_sectors_store()
411 if (!q->mq_ops || !q->mq_ops->poll) in queue_poll_delay_store()
680 (!q->mq_ops || !q->mq_ops->timeout)) in queue_attr_visible()
750 __elevator_exit(q, q->elevator); in blk_exit_queue()
789 blk_stat_remove_callback(q, q->poll_cb); in blk_release_queue()
883 if (q->elevator) in blk_register_queue()
929 if (WARN_ON(!q)) in blk_unregister_queue()
[all …]
A Dblk-pm.c33 q->dev = dev; in blk_pm_runtime_init()
65 if (!q->dev) in blk_pre_runtime_suspend()
80 blk_set_pm_only(q); in blk_pre_runtime_suspend()
94 blk_mq_unfreeze_queue(q); in blk_pre_runtime_suspend()
102 blk_clear_pm_only(q); in blk_pre_runtime_suspend()
124 if (!q->dev) in blk_post_runtime_suspend()
137 blk_clear_pm_only(q); in blk_post_runtime_suspend()
154 if (!q->dev) in blk_pre_runtime_resume()
178 if (!q->dev) in blk_post_runtime_resume()
211 if (!q->dev) in blk_set_runtime_active()
[all …]
A Delevator.c62 struct request_queue *q = rq->q; in elv_iosched_allow_bio_merge() local
321 if (q->last_merge && elv_bio_merge_ok(q->last_merge, bio)) { in elv_merge()
371 if (q->last_merge && blk_attempt_req_merge(q, q->last_merge, rq)) { in elv_attempt_insert_merge()
519 if (q) { in elv_unregister_queue()
593 if (q->elevator) { in elevator_switch_mq()
598 elevator_exit(q, q->elevator); in elevator_switch_mq()
608 elevator_exit(q, q->elevator); in elevator_switch_mq()
625 (q->tag_set && (q->tag_set->flags & BLK_MQ_F_NO_SCHED))) in elv_support_iosched()
636 if (q->tag_set && q->tag_set->flags & BLK_MQ_F_NO_SCHED_BY_DEFAULT) in elevator_get_default()
801 if (!q->elevator) in elv_iosched_show()
[all …]
/linux/drivers/spi/
A Dspi-fsl-qspi.c346 reg = qspi_readl(q, q->iobase + QUADSPI_FR); in fsl_qspi_irq_handler()
347 qspi_writel(q, reg, q->iobase + QUADSPI_FR); in fsl_qspi_irq_handler()
512 reg = qspi_readl(q, q->iobase + QUADSPI_MCR); in fsl_qspi_invalidate()
514 qspi_writel(q, reg, q->iobase + QUADSPI_MCR); in fsl_qspi_invalidate()
523 qspi_writel(q, reg, q->iobase + QUADSPI_MCR); in fsl_qspi_invalidate()
555 q->ahb_addr + q->selected * q->devtype_data->ahb_buf_size, in fsl_qspi_read_ahb()
665 qspi_writel(q, qspi_readl(q, base + QUADSPI_MCR) | in fsl_qspi_exec_op()
754 qspi_writel(q, qspi_readl(q, base + QUADSPI_FLSHCR) & in fsl_qspi_default_setup()
802 qspi_writel(q, 0xffffffff, q->iobase + QUADSPI_FR); in fsl_qspi_default_setup()
881 q->ahb_addr = devm_ioremap(dev, q->memmap_phy, in fsl_qspi_probe()
[all …]
/linux/drivers/s390/cio/
A Dqdio_main.c148 q->handler(q->irq_ptr->cdev, QDIO_ERROR_GET_BUF_STATE, q->nr, in qdio_do_eqbs()
149 q->first_to_check, count, q->irq_ptr->int_parm); in qdio_do_eqbs()
196 q->handler(q->irq_ptr->cdev, QDIO_ERROR_SET_BUF_STATE, q->nr, in qdio_do_sqbs()
197 q->first_to_check, count, q->irq_ptr->int_parm); in qdio_do_sqbs()
308 return qdio_siga_sync(q, 0, q->mask); in qdio_sync_input_queue()
313 return qdio_siga_sync(q, q->mask, 0); in qdio_sync_output_queue()
359 "%4x cc2 BB1:%1d", SCH_NO(q), q->nr); in qdio_siga_output()
604 DBF_ERROR("%4x cc2 BBC:%1d", SCH_NO(q), q->nr); in qdio_kick_outbound_q()
618 DBF_ERROR("%4x cc2 BB2:%1d", SCH_NO(q), q->nr); in qdio_kick_outbound_q()
671 q->handler(q->irq_ptr->cdev, QDIO_ERROR_ACTIVATE, in qdio_handle_activate_check()
[all …]
/linux/drivers/infiniband/sw/rxe/
A Drxe_queue.h105 prod = q->index; in queue_get_producer()
126 cons = q->index; in queue_get_consumer()
180 prod = q->index; in queue_advance_producer()
182 q->index = prod; in queue_advance_producer()
205 cons = q->index; in queue_advance_consumer()
207 q->index = cons; in queue_advance_consumer()
232 return q->buf->data + (prod << q->log2_elem_size); in queue_producer_addr()
240 return q->buf->data + (cons << q->log2_elem_size); in queue_consumer_addr()
245 return q->buf->data + ((index & q->index_mask) in queue_addr_from_index()
252 return (((u8 *)addr - q->buf->data) >> q->log2_elem_size) in queue_index_from_addr()
[all …]
/linux/drivers/net/wireless/broadcom/b43/
A Dpio.c76 return q; in parse_cookie()
130 q = kzalloc(sizeof(*q), GFP_KERNEL); in b43_setup_pioqueue_tx()
131 if (!q) in b43_setup_pioqueue_tx()
156 return q; in b43_setup_pioqueue_tx()
164 q = kzalloc(sizeof(*q), GFP_KERNEL); in b43_setup_pioqueue_rx()
165 if (!q) in b43_setup_pioqueue_rx()
175 return q; in b43_setup_pioqueue_rx()
195 if (!q) in b43_destroy_pioqueue_tx()
204 if (!q) in b43_destroy_pioqueue_rx()
523 B43_WARN_ON(q->buffer_used > q->buffer_size); in b43_pio_tx()
[all …]
/linux/include/linux/
A Dblkdev.h448 #define blk_queue_pm_only(q) atomic_read(&(q)->pm_only) argument
506 return blk_queue_is_zoned(q) ? q->nr_zones : 0; in blk_queue_nr_zones()
943 if (q && q->limits.logical_block_size) in queue_logical_block_size()
1070 if (q) in bdev_write_same()
1080 if (q) in bdev_write_zeroes_sectors()
1090 if (q) in bdev_zoned_model()
1100 if (q) in bdev_is_zoned()
1110 if (q) in bdev_zone_sectors()
1119 if (q) in bdev_max_open_zones()
1128 if (q) in bdev_max_active_zones()
[all …]
/linux/drivers/net/wireless/mediatek/mt76/
A Ddma.c85 writel(q->desc_dma, &q->regs->desc_base); in mt76_dma_sync_idx()
86 writel(q->ndesc, &q->regs->ring_size); in mt76_dma_sync_idx()
87 q->head = readl(&q->regs->dma_idx); in mt76_dma_sync_idx()
88 q->tail = q->head; in mt76_dma_sync_idx()
128 size = q->ndesc * sizeof(*q->entry); in mt76_dma_alloc_queue()
150 q->entry[q->head].skip_buf0 = true; in mt76_dma_add_buf()
157 q->head = (q->head + 1) % q->ndesc; in mt76_dma_add_buf()
227 writel(q->head, &q->regs->cpu_idx); in mt76_dma_kick_queue()
312 q->tail = (q->tail + 1) % q->ndesc; in mt76_dma_dequeue()
325 if (q->queued + 1 >= q->ndesc - 1) in mt76_dma_tx_queue_skb_raw()
[all …]
/linux/drivers/net/ethernet/chelsio/cxgb3/
A Dsge.c583 q->sdesc[q->pidx] = q->sdesc[idx]; in recycle_rx_buf()
590 if (++q->pidx == q->size) { in recycle_rx_buf()
650 memset(q, 0, sizeof(*q)); in t3_reset_qset()
1325 if (q->pidx >= q->size) { in t3_eth_xmit()
1326 q->pidx -= q->size; in t3_eth_xmit()
1502 write_imm(&q->desc[q->pidx], skb, skb->len, q->gen); in ctrl_xmit()
1535 write_imm(&q->desc[q->pidx], skb, skb->len, q->gen); in restart_ctrlq()
1725 if (q->pidx >= q->size) { in ofld_xmit()
1726 q->pidx -= q->size; in ofld_xmit()
1780 q->pidx -= q->size; in restart_offloadq()
[all …]

Completed in 104 milliseconds

12345678910>>...54