/linux/drivers/crypto/ccp/ |
A D | ccp-crypto-main.c | 55 struct list_head *backlog; member 102 *backlog = NULL; in ccp_crypto_cmd_complete() 126 *backlog = container_of(req_queue.backlog, in ccp_crypto_cmd_complete() 128 req_queue.backlog = req_queue.backlog->next; in ccp_crypto_cmd_complete() 167 if (backlog) { in ccp_crypto_complete() 168 backlog->ret = -EINPROGRESS; in ccp_crypto_complete() 169 backlog->req->complete(backlog->req, -EINPROGRESS); in ccp_crypto_complete() 199 if (backlog) { in ccp_crypto_complete() 200 backlog->ret = -EINPROGRESS; in ccp_crypto_complete() 201 backlog->req->complete(backlog->req, -EINPROGRESS); in ccp_crypto_complete() [all …]
|
A D | ccp-dev.c | 315 list_add_tail(&cmd->entry, &ccp->backlog); in ccp_enqueue_cmd() 378 struct ccp_cmd *backlog = NULL; in ccp_dequeue_cmd() local 403 if (!list_empty(&ccp->backlog)) { in ccp_dequeue_cmd() 404 backlog = list_first_entry(&ccp->backlog, struct ccp_cmd, in ccp_dequeue_cmd() 406 list_del(&backlog->entry); in ccp_dequeue_cmd() 411 if (backlog) { in ccp_dequeue_cmd() 412 INIT_WORK(&backlog->work, ccp_do_cmd_backlog); in ccp_dequeue_cmd() 413 schedule_work(&backlog->work); in ccp_dequeue_cmd() 488 INIT_LIST_HEAD(&ccp->backlog); in ccp_alloc_struct()
|
/linux/tools/testing/selftests/net/forwarding/ |
A D | sch_red.sh | 163 qdisc_stats_get $swp3 11: .backlog 257 local backlog 265 backlog=$(build_backlog $((2 * limit / 3)) udp) 268 check_err $? "backlog $backlog / $limit Got $pct% marked packets, expected == 0." 275 backlog=$(build_backlog $((3 * limit / 2)) tcp tos=0x01) 331 local backlog 341 backlog=$(build_backlog $((2 * limit / 3)) tcp tos=0x01) 349 backlog=$(build_backlog $((3 * limit / 2)) tcp tos=0x01) 362 local backlog 417 backlog=$(build_backlog $((2 * limit / 3)) tcp tos=0x01) [all …]
|
/linux/tools/perf/ui/tui/ |
A D | helpline.c | 34 static int backlog; in tui_helpline__show() local 37 ret = vscnprintf(ui_helpline__last_msg + backlog, in tui_helpline__show() 38 sizeof(ui_helpline__last_msg) - backlog, format, ap); in tui_helpline__show() 39 backlog += ret; in tui_helpline__show() 43 if (ui_helpline__last_msg[backlog - 1] == '\n') { in tui_helpline__show() 46 backlog = 0; in tui_helpline__show()
|
/linux/tools/perf/ui/gtk/ |
A D | helpline.c | 32 static int backlog; in gtk_helpline_show() local 34 ret = vscnprintf(ui_helpline__current + backlog, in gtk_helpline_show() 35 sizeof(ui_helpline__current) - backlog, fmt, ap); in gtk_helpline_show() 36 backlog += ret; in gtk_helpline_show() 40 if (ptr && (ptr - ui_helpline__current) <= backlog) { in gtk_helpline_show() 43 backlog = 0; in gtk_helpline_show()
|
/linux/net/sched/ |
A D | sch_pie.c | 34 struct pie_vars *vars, u32 backlog, u32 packet_size) in pie_drop_early() argument 54 if (backlog < 2 * mtu) in pie_drop_early() 96 if (!pie_drop_early(sch, &q->params, &q->vars, sch->qstats.backlog, in pie_qdisc_enqueue() 209 struct pie_vars *vars, u32 backlog) in pie_process_dequeue() argument 225 if (backlog == 0) in pie_process_dequeue() 277 if (backlog < QUEUE_THRESHOLD) { in pie_process_dequeue() 301 u32 backlog) in pie_calculate_probability() argument 316 qdelay = (backlog << PIE_SCALE) / vars->avg_dq_rate; in pie_calculate_probability() 327 if (qdelay == 0 && backlog != 0) in pie_calculate_probability() 403 vars->backlog_old = backlog; in pie_calculate_probability() [all …]
|
A D | sch_gred.c | 118 return sch->qstats.backlog; in gred_backlog() 120 return q->backlog; in gred_backlog() 248 q->backlog += qdisc_pkt_len(skb); in gred_enqueue() 276 q->backlog -= qdisc_pkt_len(skb); in gred_dequeue() 279 if (!sch->qstats.backlog) in gred_dequeue() 282 if (!q->backlog) in gred_dequeue() 307 q->backlog = 0; in gred_reset() 346 opt->set.tab[i].backlog = &q->backlog; in gred_offload() 385 table->tab[i]->backlog += hw_stats->stats.qstats[i].backlog; in gred_offload_dump_stats() 390 sch->qstats.backlog += hw_stats->stats.qstats[i].backlog; in gred_offload_dump_stats() [all …]
|
A D | sch_fq_pie.c | 47 u32 backlog; member 163 sel_flow->backlog, skb->len)) { in fq_pie_qdisc_enqueue() 182 sch->qstats.backlog += pkt_len; in fq_pie_qdisc_enqueue() 190 sel_flow->backlog = 0; in fq_pie_qdisc_enqueue() 193 sel_flow->backlog += pkt_len; in fq_pie_qdisc_enqueue() 255 sch->qstats.backlog -= pkt_len; in fq_pie_qdisc_dequeue() 271 flow->backlog -= pkt_len; in fq_pie_qdisc_dequeue() 273 pie_process_dequeue(skb, &q->p_params, &flow->vars, flow->backlog); in fq_pie_qdisc_dequeue() 384 q->flows[idx].backlog); in fq_pie_timer() 526 sch->qstats.backlog = 0; in fq_pie_reset()
|
A D | sch_sfq.c | 109 unsigned int backlog; member 308 slot->backlog -= len; in sfq_drop() 381 slot->backlog); in sfq_enqueue() 432 sch->qstats.backlog -= delta; in sfq_enqueue() 433 slot->backlog -= delta; in sfq_enqueue() 443 slot->backlog += qdisc_pkt_len(skb); in sfq_enqueue() 504 slot->backlog -= qdisc_pkt_len(skb); in sfq_dequeue() 556 slot->backlog = 0; in sfq_rehash() 587 slot->backlog); in sfq_rehash() 588 slot->backlog += qdisc_pkt_len(skb); in sfq_rehash() [all …]
|
A D | sch_skbprio.c | 85 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue() 108 q->qstats[prio].backlog += qdisc_pkt_len(skb); in skbprio_enqueue() 117 q->qstats[lp].backlog -= qdisc_pkt_len(to_drop); in skbprio_enqueue() 152 q->qstats[q->highest_prio].backlog -= qdisc_pkt_len(skb); in skbprio_dequeue() 216 sch->qstats.backlog = 0; in skbprio_reset()
|
/linux/tools/testing/selftests/drivers/net/mlxsw/ |
A D | sch_red_core.sh | 317 qdisc_stats_get $swp3 $(get_qdisc_handle $vlan) .backlog 420 local backlog 428 backlog=$(build_backlog $vlan $((2 * limit / 3)) udp) 431 check_err $? "backlog $backlog / $limit Got $pct% marked packets, expected == 0." 438 backlog=$(build_backlog $vlan $((3 * limit / 2)) tcp tos=0x01) 515 local backlog 525 backlog=$(build_backlog $vlan $((2 * limit / 3)) tcp tos=0x01) 533 backlog=$(build_backlog $vlan $((3 * limit / 2)) tcp tos=0x01) 537 local diff=$((limit - backlog)) 540 check_err $? "backlog $backlog / $limit expected <= 10% distance" [all …]
|
/linux/include/net/ |
A D | fq_impl.h | 24 flow->backlog -= bytes; in __fq_adjust_removal() 25 fq->backlog -= packets; in __fq_adjust_removal() 28 if (flow->backlog) in __fq_adjust_removal() 176 cur_len = cur->backlog; in fq_find_fattest_flow() 185 unsigned int cur_len = tin->default_flow.backlog; in fq_find_fattest_flow() 209 if (!flow->backlog) { in fq_tin_enqueue() 217 flow->backlog += skb->len; in fq_tin_enqueue() 221 fq->backlog++; in fq_tin_enqueue() 231 while (fq->backlog > fq->limit || oom) { in fq_tin_enqueue() 304 WARN_ON_ONCE(flow->backlog); in fq_flow_reset()
|
A D | codel_impl.h | 109 u32 *backlog, in codel_should_drop() argument 127 *backlog <= params->mtu) { in codel_should_drop() 145 u32 *backlog, in codel_dequeue() argument 164 skb_len_func, skb_time_func, backlog, now); in codel_dequeue() 200 backlog, now)) { in codel_dequeue() 225 skb_time_func, backlog, now); in codel_dequeue()
|
A D | sch_generic.h | 881 sch->qstats.backlog -= qdisc_pkt_len(skb); in qdisc_qstats_backlog_dec() 893 sch->qstats.backlog += qdisc_pkt_len(skb); in qdisc_qstats_backlog_inc() 955 __u32 *backlog) in qdisc_qstats_qlen_backlog() argument 961 *backlog = qstats.backlog; in qdisc_qstats_qlen_backlog() 966 __u32 qlen, backlog; in qdisc_tree_flush_backlog() local 968 qdisc_qstats_qlen_backlog(sch, &qlen, &backlog); in qdisc_tree_flush_backlog() 969 qdisc_tree_reduce_backlog(sch, qlen, backlog); in qdisc_tree_flush_backlog() 974 __u32 qlen, backlog; in qdisc_purge_queue() local 978 qdisc_tree_reduce_backlog(sch, qlen, backlog); in qdisc_purge_queue() 1133 sch->qstats.backlog += pkt_len; in qdisc_update_stats_at_enqueue() [all …]
|
A D | pie.h | 127 struct pie_vars *vars, u32 backlog, u32 packet_size); 130 struct pie_vars *vars, u32 backlog); 133 u32 backlog);
|
A D | red.h | 340 unsigned int backlog) in red_calc_qavg_no_idle_time() argument 351 return v->qavg + (backlog - (v->qavg >> p->Wlog)); in red_calc_qavg_no_idle_time() 356 unsigned int backlog) in red_calc_qavg() argument 359 return red_calc_qavg_no_idle_time(p, v, backlog); in red_calc_qavg()
|
/linux/drivers/net/ethernet/mellanox/mlxsw/ |
A D | spectrum_qdisc.c | 79 u64 backlog; member 200 tmp->stats_base.backlog -= mlxsw_sp_qdisc->stats_base.backlog; in mlxsw_sp_qdisc_reduce_parent_backlog() 569 backlog -= stats_base->backlog; in mlxsw_sp_qdisc_update_stats() 575 stats_base->backlog += backlog; in mlxsw_sp_qdisc_update_stats() 751 u64 backlog; in mlxsw_sp_qdisc_leaf_unoffload() local 755 qstats->backlog -= backlog; in mlxsw_sp_qdisc_leaf_unoffload() 1298 u64 backlog; in mlxsw_sp_qdisc_walk_cb_clean_stats() local 1301 backlog = mlxsw_sp_qdisc->stats_base.backlog; in mlxsw_sp_qdisc_walk_cb_clean_stats() 1305 mlxsw_sp_qdisc->stats_base.backlog = backlog; in mlxsw_sp_qdisc_walk_cb_clean_stats() 1422 u64 backlog; in __mlxsw_sp_qdisc_ets_unoffload() local [all …]
|
/linux/net/core/ |
A D | gen_stats.c | 348 qstats->qlen += qcpu->backlog; in gnet_stats_add_queue_cpu() 349 qstats->backlog += qcpu->backlog; in gnet_stats_add_queue_cpu() 364 qstats->backlog += q->backlog; in gnet_stats_add_queue() 399 d->tc_stats.backlog = qstats.backlog; in gnet_stats_copy_queue()
|
/linux/drivers/crypto/qce/ |
A D | core.c | 83 struct crypto_async_request *async_req, *backlog; in qce_handle_queue() local 98 backlog = crypto_get_backlog(&qce->queue); in qce_handle_queue() 108 if (backlog) { in qce_handle_queue() 110 backlog->complete(backlog, -EINPROGRESS); in qce_handle_queue()
|
/linux/net/rxrpc/ |
A D | call_accept.c | 165 struct rxrpc_backlog *b = rx->backlog; in rxrpc_service_prealloc() 171 rx->backlog = b; in rxrpc_service_prealloc() 182 struct rxrpc_backlog *b = rx->backlog; in rxrpc_discard_prealloc() 188 rx->backlog = NULL; in rxrpc_discard_prealloc() 267 struct rxrpc_backlog *b = rx->backlog; in rxrpc_alloc_incoming_call() 450 struct rxrpc_backlog *b = rx->backlog; in rxrpc_user_charge_accept() 482 struct rxrpc_backlog *b = rx->backlog; in rxrpc_kernel_charge_accept()
|
/linux/net/sunrpc/ |
A D | stats.c | 153 ktime_t backlog, execute, now; in rpc_count_iostats_metrics() local 169 backlog = 0; in rpc_count_iostats_metrics() 171 backlog = ktime_sub(req->rq_xtime, task->tk_start); in rpc_count_iostats_metrics() 172 op_metrics->om_queue = ktime_add(op_metrics->om_queue, backlog); in rpc_count_iostats_metrics() 184 trace_rpc_stats_latency(req->rq_task, backlog, req->rq_rtt, execute); in rpc_count_iostats_metrics()
|
/linux/drivers/crypto/rockchip/ |
A D | rk3288_crypto.c | 207 struct crypto_async_request *async_req, *backlog; in rk_crypto_queue_task_cb() local 213 backlog = crypto_get_backlog(&dev->queue); in rk_crypto_queue_task_cb() 223 if (backlog) { in rk_crypto_queue_task_cb() 224 backlog->complete(backlog, -EINPROGRESS); in rk_crypto_queue_task_cb() 225 backlog = NULL; in rk_crypto_queue_task_cb()
|
/linux/drivers/crypto/ |
A D | mxs-dcp.c | 395 struct crypto_async_request *backlog; in dcp_chan_thread_aes() local 404 backlog = crypto_get_backlog(&sdcp->queue[chan]); in dcp_chan_thread_aes() 408 if (!backlog && !arq) { in dcp_chan_thread_aes() 415 if (backlog) in dcp_chan_thread_aes() 416 backlog->complete(backlog, -EINPROGRESS); in dcp_chan_thread_aes() 692 struct crypto_async_request *backlog; in dcp_chan_thread_sha() local 700 backlog = crypto_get_backlog(&sdcp->queue[chan]); in dcp_chan_thread_sha() 704 if (!backlog && !arq) { in dcp_chan_thread_sha() 711 if (backlog) in dcp_chan_thread_sha() 712 backlog->complete(backlog, -EINPROGRESS); in dcp_chan_thread_sha()
|
/linux/crypto/ |
A D | crypto_engine.c | 73 struct crypto_async_request *async_req, *backlog; in crypto_pump_requests() local 118 backlog = crypto_get_backlog(&engine->queue); in crypto_pump_requests() 131 if (backlog) in crypto_pump_requests() 132 backlog->complete(backlog, -EINPROGRESS); in crypto_pump_requests()
|
/linux/drivers/crypto/marvell/cesa/ |
A D | cesa.c | 39 struct crypto_async_request **backlog) in mv_cesa_dequeue_req_locked() argument 43 *backlog = crypto_get_backlog(&engine->queue); in mv_cesa_dequeue_req_locked() 54 struct crypto_async_request *req = NULL, *backlog = NULL; in mv_cesa_rearm_engine() local 60 req = mv_cesa_dequeue_req_locked(engine, &backlog); in mv_cesa_rearm_engine() 68 if (backlog) in mv_cesa_rearm_engine() 69 backlog->complete(backlog, -EINPROGRESS); in mv_cesa_rearm_engine()
|