/linux/crypto/ |
A D | gf128mul.c | 57 q(0x00), q(0x01), q(0x02), q(0x03), q(0x04), q(0x05), q(0x06), q(0x07),\ 58 q(0x08), q(0x09), q(0x0a), q(0x0b), q(0x0c), q(0x0d), q(0x0e), q(0x0f),\ 59 q(0x10), q(0x11), q(0x12), q(0x13), q(0x14), q(0x15), q(0x16), q(0x17),\ 60 q(0x18), q(0x19), q(0x1a), q(0x1b), q(0x1c), q(0x1d), q(0x1e), q(0x1f),\ 61 q(0x20), q(0x21), q(0x22), q(0x23), q(0x24), q(0x25), q(0x26), q(0x27),\ 62 q(0x28), q(0x29), q(0x2a), q(0x2b), q(0x2c), q(0x2d), q(0x2e), q(0x2f),\ 63 q(0x30), q(0x31), q(0x32), q(0x33), q(0x34), q(0x35), q(0x36), q(0x37),\ 64 q(0x38), q(0x39), q(0x3a), q(0x3b), q(0x3c), q(0x3d), q(0x3e), q(0x3f),\ 65 q(0x40), q(0x41), q(0x42), q(0x43), q(0x44), q(0x45), q(0x46), q(0x47),\ 66 q(0x48), q(0x49), q(0x4a), q(0x4b), q(0x4c), q(0x4d), q(0x4e), q(0x4f),\ [all …]
|
/linux/Documentation/networking/ |
A D | tls-offload-layers.svg | 1 …q-0.609375 -0.359375 -1.203125 -0.359375q-0.546875 0 -0.96875 0.328125q-0.421875 0.328125 -0.60937…
|
/linux/drivers/media/v4l2-core/ |
A D | videobuf-core.c | 254 q->ops->buf_release(q, q->bufs[i]); in __videobuf_free() 289 q->ops->buf_release(q, q->bufs[i]); in videobuf_queue_cancel() 812 q->ops->buf_queue(q, q->read_buf); in videobuf_read_zerocopy() 816 CALL(q, sync, q, q->read_buf); in videobuf_read_zerocopy() 825 q->ops->buf_release(q, q->read_buf); in videobuf_read_zerocopy() 919 q->ops->buf_queue(q, q->read_buf); in videobuf_read_one() 930 CALL(q, sync, q, q->read_buf); in videobuf_read_one() 934 q->ops->buf_release(q, q->read_buf); in videobuf_read_one() 949 q->ops->buf_release(q, q->read_buf); in videobuf_read_one() 990 q->ops->buf_queue(q, q->bufs[i]); in __videobuf_read_start() [all …]
|
/linux/sound/core/seq/ |
A D | seq_queue.c | 76 if (q) { in queue_list_remove() 100 q = kzalloc(sizeof(*q), GFP_KERNEL); in queue_new() 101 if (!q) in queue_new() 113 if (q->tickq == NULL || q->timeq == NULL || q->timer == NULL) { in queue_new() 125 return q; in queue_new() 143 kfree(q); in queue_delete() 183 return q; in snd_seq_queue_alloc() 212 if (q) in queueptr() 346 return (q->owner == client) || (!q->locked && !q->klocked); in check_access() 382 if (! q) in snd_seq_queue_check_access() [all …]
|
/linux/drivers/media/common/videobuf2/ |
A D | videobuf2-core.c | 158 ((q)->ops->op ? (q)->ops->op(args) : 0) 180 if (q && q->buf_ops && q->buf_ops->op) \ 187 if (q && q->buf_ops && q->buf_ops->op) \ 235 q->alloc_devs[plane] ? : q->dev, in __vb2_buf_mem_alloc() 1828 call_void_qop(q, wait_prepare, q); in __vb2_wait_for_done_vb() 1842 call_void_qop(q, wait_finish, q); in __vb2_wait_for_done_vb() 2500 if (q->is_output && q->fileio && q->queued_count < q->num_buffers) in vb2_core_poll() 2710 vb2_core_streamoff(q, q->type); in __vb2_cleanup_fileio() 2938 call_void_qop(q, wait_finish, q); in vb2_thread() 2953 call_void_qop(q, wait_finish, q); in vb2_thread() [all …]
|
/linux/sound/core/seq/oss/ |
A D | seq_oss_readq.c | 37 q = kzalloc(sizeof(*q), GFP_KERNEL); in snd_seq_oss_readq_new() 42 if (!q->q) { in snd_seq_oss_readq_new() 49 q->head = q->tail = 0; in snd_seq_oss_readq_new() 65 kfree(q->q); in snd_seq_oss_readq_delete() 78 q->head = q->tail = 0; in snd_seq_oss_readq_clear() 151 memcpy(&q->q[q->tail], ev, sizeof(*ev)); in snd_seq_oss_readq_put_event() 152 q->tail = (q->tail + 1) % q->maxlen; in snd_seq_oss_readq_put_event() 173 memcpy(rec, &q->q[q->head], sizeof(*rec)); in snd_seq_oss_readq_pick() 184 (q->qlen > 0 || q->head == q->tail), in snd_seq_oss_readq_wait() 196 q->head = (q->head + 1) % q->maxlen; in snd_seq_oss_readq_free() [all …]
|
/linux/net/sched/ |
A D | sch_choke.c | 78 return (q->tail - q->head) & q->tab_mask; in choke_len() 97 q->head = (q->head + 1) & q->tab_mask; in choke_zap_head_holes() 98 if (q->head == q->tail) in choke_zap_head_holes() 107 q->tail = (q->tail - 1) & q->tab_mask; in choke_zap_tail_holes() 108 if (q->head == q->tail) in choke_zap_tail_holes() 206 if (q->head == q->tail) in choke_match_random() 270 q->tail = (q->tail + 1) & q->tab_mask; in choke_enqueue() 295 skb = q->tab[q->head]; in choke_dequeue() 312 q->head = (q->head + 1) & q->tab_mask; in choke_reset() 389 q->head = (q->head + 1) & q->tab_mask; in choke_change() [all …]
|
A D | sch_netem.c | 291 return q->loss && q->loss >= get_crandom(&q->loss_cor); in loss_event() 373 rtnl_kfree_skbs(q->t_head, q->t_tail); in tfifo_reset() 451 if (q->duplicate && q->duplicate >= get_crandom(&q->dup_cor)) in netem_enqueue() 470 if (q->latency || q->jitter || q->rate) in netem_enqueue() 494 if (q->corrupt && q->corrupt >= get_crandom(&q->corrupt_cor)) { in netem_enqueue() 536 delay = tabledist(q->latency, q->jitter, in netem_enqueue() 537 &q->delay_cor, q->delay_dist); in netem_enqueue() 645 q->slot.bytes_left = q->slot_config.max_bytes; in get_slot_next() 737 skb = q->qdisc->ops->dequeue(q->qdisc); in netem_dequeue() 748 skb = q->qdisc->ops->dequeue(q->qdisc); in netem_dequeue() [all …]
|
A D | sch_sfq.c | 461 if (++sch->q.qlen <= q->limit) in sfq_enqueue() 615 if (!q->filter_list && q->tail) in sfq_perturbation() 660 q->scaled_quantum = SFQ_ALLOT_SIZE(q->quantum); in sfq_change() 667 q->maxflows = min_t(u32, q->maxflows, q->divisor); in sfq_change() 685 q->limit = min_t(u32, ctl->limit, q->maxdepth * q->maxflows); in sfq_change() 686 q->maxflows = min_t(u32, q->maxflows, q->limit); in sfq_change() 690 while (sch->q.qlen > q->limit) { in sfq_change() 757 q->scaled_quantum = SFQ_ALLOT_SIZE(q->quantum); in sfq_init() 767 q->ht = sfq_alloc(sizeof(q->ht[0]) * q->divisor); in sfq_init() 768 q->slots = sfq_alloc(sizeof(q->slots[0]) * q->maxflows); in sfq_init() [all …]
|
A D | sch_sfb.c | 192 memset(&q->bins, 0, sizeof(q->bins)); in sfb_zero_all_buckets() 202 const struct sfb_bucket *b = &q->bins[q->slot].bins[0][0]; in sfb_compute_qlen() 226 sfb_init_perturbation(q->slot, q); in sfb_swap_slot() 227 q->slot ^= 1; in sfb_swap_slot() 236 if (q->penalty_rate == 0 || q->penalty_burst == 0) in sfb_rate_limit() 242 q->tokens_avail = (age * q->penalty_rate) / HZ; in sfb_rate_limit() 243 if (q->tokens_avail > q->penalty_burst) in sfb_rate_limit() 244 q->tokens_avail = q->penalty_burst; in sfb_rate_limit() 293 if (unlikely(sch->q.qlen >= q->limit)) { in sfb_enqueue() 458 q->slot = 0; in sfb_reset() [all …]
|
A D | sch_fq_pie.c | 159 q->overmemory++; in fq_pie_qdisc_enqueue() 183 sch->q.qlen++; in fq_pie_qdisc_enqueue() 256 sch->q.qlen--; in fq_pie_qdisc_dequeue() 262 if (head == &q->new_flows && !list_empty(&q->old_flows)) in fq_pie_qdisc_dequeue() 307 if (!q->flows_cnt || q->flows_cnt > 65536) { in fq_pie_change() 374 struct fq_pie_sched_data *q = from_timer(q, t, adapt_timer); in fq_pie_timer() local 383 pie_calculate_probability(&q->p_params, &q->flows[idx].vars, in fq_pie_timer() 388 mod_timer(&q->adapt_timer, jiffies + q->p_params.tupdate); in fq_pie_timer() 404 q->sch = sch; in fq_pie_init() 420 err = tcf_block_get(&q->block, &q->filter_list, sch, extack); in fq_pie_init() [all …]
|
A D | sch_fq_codel.c | 180 sch->q.qlen -= i; in fq_codel_drop() 216 memory_limited = q->memory_usage > q->memory_limit; in fq_codel_enqueue() 265 q->backlogs[flow - q->flows] -= qdisc_pkt_len(skb); in dequeue_func() 267 sch->q.qlen--; in dequeue_func() 309 if ((head == &q->new_flows) && !list_empty(&q->old_flows)) in fq_codel_dequeue() 320 if (q->cstats.drop_count && sch->q.qlen) { in fq_codel_dequeue() 349 memset(q->backlogs, 0, q->flows_cnt * sizeof(u32)); in fq_codel_reset() 385 if (q->flows) in fq_codel_change() 440 q->memory_usage > q->memory_limit) { in fq_codel_change() 489 err = tcf_block_get(&q->block, &q->filter_list, sch, extack); in fq_codel_init() [all …]
|
A D | sch_fq.c | 302 root = &q->fq_root[hash_ptr(sk, q->fq_trees_log)]; in fq_classify() 304 if (q->flows >= (2U << q->fq_trees_log) && in fq_classify() 305 q->inactive_flows > q->flows/2) in fq_classify() 359 q->flows++; in fq_classify() 401 sch->q.qlen--; in fq_dequeue_skb() 467 skb->tstamp = q->ktime_cache + q->horizon; in fq_enqueue() 494 sch->q.qlen++; in fq_enqueue() 511 q->unthrottle_latency_ns -= q->unthrottle_latency_ns >> 3; in fq_check_throttled() 588 if ((head == &q->new_flows) && q->old_flows.first) { in fq_dequeue() 754 if (q->fq_root && log == q->fq_trees_log) in fq_resize() [all …]
|
A D | sch_red.c | 77 q->vars.qavg = red_calc_qavg(&q->parms, in red_enqueue() 84 switch (red_action(&q->parms, &q->vars, q->vars.qavg)) { in red_enqueue() 110 if (red_use_harddrop(q) || !red_use_ecn(q)) { in red_enqueue() 198 opt.set.min = q->parms.qth_min >> q->parms.Wlog; in red_offload() 199 opt.set.max = q->parms.qth_max >> q->parms.Wlog; in red_offload() 302 if (!q->qdisc->q.qlen) in __red_change() 327 red_adaptative_algo(&q->parms, &q->vars); in red_adaptative_timer() 417 .qth_min = q->parms.qth_min >> q->parms.Wlog, in red_dump() 418 .qth_max = q->parms.qth_max >> q->parms.Wlog, in red_dump() 464 st.early = q->stats.prob_drop + q->stats.forced_drop; in red_dump_stats() [all …]
|
/linux/net/xdp/ |
A D | xsk_queue.h | 124 if (q->cached_cons != q->cached_prod) { in xskq_cons_read_addr_unchecked() 194 while (q->cached_cons != q->cached_prod) { in xskq_cons_read_desc() 196 u32 idx = q->cached_cons & q->ring_mask; in xskq_cons_read_desc() 271 if (q->cached_prod == q->cached_cons) in xskq_cons_peek_addr_unchecked() 280 if (q->cached_prod == q->cached_cons) in xskq_cons_peek_desc() 324 u32 free_entries = q->nentries - (q->cached_prod - q->cached_cons); in xskq_prod_nb_free() 331 free_entries = q->nentries - (q->cached_prod - q->cached_cons); in xskq_prod_nb_free() 395 idx = q->cached_prod++ & q->ring_mask; in xskq_prod_reserve_desc() 409 __xskq_prod_submit(q, q->cached_prod); in xskq_prod_submit() 437 return q ? q->invalid_descs : 0; in xskq_nb_invalid_descs() [all …]
|
/linux/block/ |
A D | blk-sysfs.c | 73 if (!queue_is_mq(q)) in queue_requests_store() 94 if (!q->disk) in queue_ra_show() 106 if (!q->disk) in queue_ra_store() 258 if (q->disk) in queue_max_sectors_store() 411 if (!q->mq_ops || !q->mq_ops->poll) in queue_poll_delay_store() 680 (!q->mq_ops || !q->mq_ops->timeout)) in queue_attr_visible() 750 __elevator_exit(q, q->elevator); in blk_exit_queue() 789 blk_stat_remove_callback(q, q->poll_cb); in blk_release_queue() 883 if (q->elevator) in blk_register_queue() 929 if (WARN_ON(!q)) in blk_unregister_queue() [all …]
|
A D | blk-pm.c | 33 q->dev = dev; in blk_pm_runtime_init() 65 if (!q->dev) in blk_pre_runtime_suspend() 80 blk_set_pm_only(q); in blk_pre_runtime_suspend() 94 blk_mq_unfreeze_queue(q); in blk_pre_runtime_suspend() 102 blk_clear_pm_only(q); in blk_pre_runtime_suspend() 124 if (!q->dev) in blk_post_runtime_suspend() 137 blk_clear_pm_only(q); in blk_post_runtime_suspend() 154 if (!q->dev) in blk_pre_runtime_resume() 178 if (!q->dev) in blk_post_runtime_resume() 211 if (!q->dev) in blk_set_runtime_active() [all …]
|
A D | elevator.c | 62 struct request_queue *q = rq->q; in elv_iosched_allow_bio_merge() local 321 if (q->last_merge && elv_bio_merge_ok(q->last_merge, bio)) { in elv_merge() 371 if (q->last_merge && blk_attempt_req_merge(q, q->last_merge, rq)) { in elv_attempt_insert_merge() 519 if (q) { in elv_unregister_queue() 593 if (q->elevator) { in elevator_switch_mq() 598 elevator_exit(q, q->elevator); in elevator_switch_mq() 608 elevator_exit(q, q->elevator); in elevator_switch_mq() 625 (q->tag_set && (q->tag_set->flags & BLK_MQ_F_NO_SCHED))) in elv_support_iosched() 636 if (q->tag_set && q->tag_set->flags & BLK_MQ_F_NO_SCHED_BY_DEFAULT) in elevator_get_default() 801 if (!q->elevator) in elv_iosched_show() [all …]
|
/linux/drivers/spi/ |
A D | spi-fsl-qspi.c | 346 reg = qspi_readl(q, q->iobase + QUADSPI_FR); in fsl_qspi_irq_handler() 347 qspi_writel(q, reg, q->iobase + QUADSPI_FR); in fsl_qspi_irq_handler() 512 reg = qspi_readl(q, q->iobase + QUADSPI_MCR); in fsl_qspi_invalidate() 514 qspi_writel(q, reg, q->iobase + QUADSPI_MCR); in fsl_qspi_invalidate() 523 qspi_writel(q, reg, q->iobase + QUADSPI_MCR); in fsl_qspi_invalidate() 555 q->ahb_addr + q->selected * q->devtype_data->ahb_buf_size, in fsl_qspi_read_ahb() 665 qspi_writel(q, qspi_readl(q, base + QUADSPI_MCR) | in fsl_qspi_exec_op() 754 qspi_writel(q, qspi_readl(q, base + QUADSPI_FLSHCR) & in fsl_qspi_default_setup() 802 qspi_writel(q, 0xffffffff, q->iobase + QUADSPI_FR); in fsl_qspi_default_setup() 881 q->ahb_addr = devm_ioremap(dev, q->memmap_phy, in fsl_qspi_probe() [all …]
|
/linux/drivers/s390/cio/ |
A D | qdio_main.c | 148 q->handler(q->irq_ptr->cdev, QDIO_ERROR_GET_BUF_STATE, q->nr, in qdio_do_eqbs() 149 q->first_to_check, count, q->irq_ptr->int_parm); in qdio_do_eqbs() 196 q->handler(q->irq_ptr->cdev, QDIO_ERROR_SET_BUF_STATE, q->nr, in qdio_do_sqbs() 197 q->first_to_check, count, q->irq_ptr->int_parm); in qdio_do_sqbs() 308 return qdio_siga_sync(q, 0, q->mask); in qdio_sync_input_queue() 313 return qdio_siga_sync(q, q->mask, 0); in qdio_sync_output_queue() 359 "%4x cc2 BB1:%1d", SCH_NO(q), q->nr); in qdio_siga_output() 604 DBF_ERROR("%4x cc2 BBC:%1d", SCH_NO(q), q->nr); in qdio_kick_outbound_q() 618 DBF_ERROR("%4x cc2 BB2:%1d", SCH_NO(q), q->nr); in qdio_kick_outbound_q() 671 q->handler(q->irq_ptr->cdev, QDIO_ERROR_ACTIVATE, in qdio_handle_activate_check() [all …]
|
/linux/drivers/infiniband/sw/rxe/ |
A D | rxe_queue.h | 105 prod = q->index; in queue_get_producer() 126 cons = q->index; in queue_get_consumer() 180 prod = q->index; in queue_advance_producer() 182 q->index = prod; in queue_advance_producer() 205 cons = q->index; in queue_advance_consumer() 207 q->index = cons; in queue_advance_consumer() 232 return q->buf->data + (prod << q->log2_elem_size); in queue_producer_addr() 240 return q->buf->data + (cons << q->log2_elem_size); in queue_consumer_addr() 245 return q->buf->data + ((index & q->index_mask) in queue_addr_from_index() 252 return (((u8 *)addr - q->buf->data) >> q->log2_elem_size) in queue_index_from_addr() [all …]
|
/linux/drivers/net/wireless/broadcom/b43/ |
A D | pio.c | 76 return q; in parse_cookie() 130 q = kzalloc(sizeof(*q), GFP_KERNEL); in b43_setup_pioqueue_tx() 131 if (!q) in b43_setup_pioqueue_tx() 156 return q; in b43_setup_pioqueue_tx() 164 q = kzalloc(sizeof(*q), GFP_KERNEL); in b43_setup_pioqueue_rx() 165 if (!q) in b43_setup_pioqueue_rx() 175 return q; in b43_setup_pioqueue_rx() 195 if (!q) in b43_destroy_pioqueue_tx() 204 if (!q) in b43_destroy_pioqueue_rx() 523 B43_WARN_ON(q->buffer_used > q->buffer_size); in b43_pio_tx() [all …]
|
/linux/include/linux/ |
A D | blkdev.h | 448 #define blk_queue_pm_only(q) atomic_read(&(q)->pm_only) argument 506 return blk_queue_is_zoned(q) ? q->nr_zones : 0; in blk_queue_nr_zones() 943 if (q && q->limits.logical_block_size) in queue_logical_block_size() 1070 if (q) in bdev_write_same() 1080 if (q) in bdev_write_zeroes_sectors() 1090 if (q) in bdev_zoned_model() 1100 if (q) in bdev_is_zoned() 1110 if (q) in bdev_zone_sectors() 1119 if (q) in bdev_max_open_zones() 1128 if (q) in bdev_max_active_zones() [all …]
|
/linux/drivers/net/wireless/mediatek/mt76/ |
A D | dma.c | 85 writel(q->desc_dma, &q->regs->desc_base); in mt76_dma_sync_idx() 86 writel(q->ndesc, &q->regs->ring_size); in mt76_dma_sync_idx() 87 q->head = readl(&q->regs->dma_idx); in mt76_dma_sync_idx() 88 q->tail = q->head; in mt76_dma_sync_idx() 128 size = q->ndesc * sizeof(*q->entry); in mt76_dma_alloc_queue() 150 q->entry[q->head].skip_buf0 = true; in mt76_dma_add_buf() 157 q->head = (q->head + 1) % q->ndesc; in mt76_dma_add_buf() 227 writel(q->head, &q->regs->cpu_idx); in mt76_dma_kick_queue() 312 q->tail = (q->tail + 1) % q->ndesc; in mt76_dma_dequeue() 325 if (q->queued + 1 >= q->ndesc - 1) in mt76_dma_tx_queue_skb_raw() [all …]
|
/linux/drivers/net/ethernet/chelsio/cxgb3/ |
A D | sge.c | 583 q->sdesc[q->pidx] = q->sdesc[idx]; in recycle_rx_buf() 590 if (++q->pidx == q->size) { in recycle_rx_buf() 650 memset(q, 0, sizeof(*q)); in t3_reset_qset() 1325 if (q->pidx >= q->size) { in t3_eth_xmit() 1326 q->pidx -= q->size; in t3_eth_xmit() 1502 write_imm(&q->desc[q->pidx], skb, skb->len, q->gen); in ctrl_xmit() 1535 write_imm(&q->desc[q->pidx], skb, skb->len, q->gen); in restart_ctrlq() 1725 if (q->pidx >= q->size) { in ofld_xmit() 1726 q->pidx -= q->size; in ofld_xmit() 1780 q->pidx -= q->size; in restart_offloadq() [all …]
|