/linux/net/dccp/ |
A D | qpolicy.c | 17 skb_queue_tail(&sk->sk_write_queue, skb); in qpolicy_simple_push() 23 sk->sk_write_queue.qlen >= dccp_sk(sk)->dccps_tx_qlen; in qpolicy_simple_full() 28 return skb_peek(&sk->sk_write_queue); in qpolicy_simple_top() 40 skb_queue_walk(&sk->sk_write_queue, skb) in qpolicy_prio_best_skb() 50 skb_queue_walk(&sk->sk_write_queue, skb) in qpolicy_prio_worst_skb() 108 skb_unlink(skb, &sk->sk_write_queue); in dccp_qpolicy_drop() 125 skb_unlink(skb, &sk->sk_write_queue); in dccp_qpolicy_pop()
|
A D | output.c | 317 while (*time_budget > 0 && (skb = skb_peek(&sk->sk_write_queue))) { in dccp_flush_write_queue() 343 skb_dequeue(&sk->sk_write_queue); in dccp_flush_write_queue()
|
A D | proto.c | 215 __skb_queue_purge(&sk->sk_write_queue); in dccp_destroy_sock() 285 __skb_queue_purge(&sk->sk_write_queue); in dccp_disconnect() 1037 __skb_queue_purge(&sk->sk_write_queue); in dccp_close()
|
A D | input.c | 181 if (sk->sk_write_queue.qlen > 0 || !(sk->sk_shutdown & SEND_SHUTDOWN)) in dccp_deliver_input_to_ccids()
|
/linux/net/x25/ |
A D | x25_out.c | 100 skb_queue_tail(&sk->sk_write_queue, skbn); in x25_output() 106 skb_queue_tail(&sk->sk_write_queue, skb); in x25_output() 159 if (!skb_peek(&sk->sk_write_queue)) in x25_kick() 177 skb = skb_dequeue(&sk->sk_write_queue); in x25_kick() 181 skb_queue_head(&sk->sk_write_queue, skb); in x25_kick() 200 (skb = skb_dequeue(&sk->sk_write_queue)) != NULL); in x25_kick()
|
A D | x25_subr.c | 38 skb_queue_purge(&sk->sk_write_queue); in x25_clear_queues() 79 skb_queue_head(&sk->sk_write_queue, skb); in x25_requeue_frames() 81 skb_append(skb_prev, skb, &sk->sk_write_queue); in x25_requeue_frames()
|
/linux/net/nfc/ |
A D | rawsock.c | 40 spin_lock_bh(&sk->sk_write_queue.lock); in rawsock_write_queue_purge() 41 __skb_queue_purge(&sk->sk_write_queue); in rawsock_write_queue_purge() 43 spin_unlock_bh(&sk->sk_write_queue.lock); in rawsock_write_queue_purge() 158 spin_lock_bh(&sk->sk_write_queue.lock); in rawsock_data_exchange_complete() 159 if (!skb_queue_empty(&sk->sk_write_queue)) in rawsock_data_exchange_complete() 163 spin_unlock_bh(&sk->sk_write_queue.lock); in rawsock_data_exchange_complete() 191 skb = skb_dequeue(&sk->sk_write_queue); in rawsock_tx_work() 227 spin_lock_bh(&sk->sk_write_queue.lock); in rawsock_sendmsg() 228 __skb_queue_tail(&sk->sk_write_queue, skb); in rawsock_sendmsg() 233 spin_unlock_bh(&sk->sk_write_queue.lock); in rawsock_sendmsg()
|
/linux/net/rose/ |
A D | rose_out.c | 57 if (!skb_peek(&sk->sk_write_queue)) in rose_kick() 73 skb = skb_dequeue(&sk->sk_write_queue); in rose_kick() 77 skb_queue_head(&sk->sk_write_queue, skb); in rose_kick() 96 (skb = skb_dequeue(&sk->sk_write_queue)) != NULL); in rose_kick()
|
A D | rose_subr.c | 34 skb_queue_purge(&sk->sk_write_queue); in rose_clear_queues() 71 skb_queue_head(&sk->sk_write_queue, skb); in rose_requeue_frames() 73 skb_append(skb_prev, skb, &sk->sk_write_queue); in rose_requeue_frames()
|
A D | af_rose.c | 1197 skb_queue_tail(&sk->sk_write_queue, skbn); /* Throw it on the queue */ in rose_sendmsg() 1203 skb_queue_tail(&sk->sk_write_queue, skb); /* Throw it on the queue */ in rose_sendmsg() 1206 skb_queue_tail(&sk->sk_write_queue, skb); /* Shove it onto the queue */ in rose_sendmsg()
|
/linux/net/netrom/ |
A D | nr_out.c | 64 skb_queue_tail(&sk->sk_write_queue, skbn); /* Throw it on the queue */ in nr_output() 69 skb_queue_tail(&sk->sk_write_queue, skb); /* Throw it on the queue */ in nr_output() 134 if (!skb_peek(&sk->sk_write_queue)) in nr_kick() 153 skb = skb_dequeue(&sk->sk_write_queue); in nr_kick() 157 skb_queue_head(&sk->sk_write_queue, skb); in nr_kick() 176 (skb = skb_dequeue(&sk->sk_write_queue)) != NULL); in nr_kick()
|
A D | nr_subr.c | 35 skb_queue_purge(&sk->sk_write_queue); in nr_clear_queues() 74 skb_queue_head(&sk->sk_write_queue, skb); in nr_requeue_frames() 76 skb_append(skb_prev, skb, &sk->sk_write_queue); in nr_requeue_frames()
|
/linux/net/bluetooth/bnep/ |
A D | netdev.c | 104 skb_queue_tail(&sk->sk_write_queue, skb); in bnep_net_set_mc_list() 192 skb_queue_tail(&sk->sk_write_queue, skb); in bnep_net_xmit() 195 if (skb_queue_len(&sk->sk_write_queue) >= BNEP_TX_QUEUE_LEN) { in bnep_net_xmit()
|
/linux/net/llc/ |
A D | llc_conn.c | 180 skb_queue_tail(&sk->sk_write_queue, skb); in llc_conn_send_pdu() 231 skb_queue_tail(&sk->sk_write_queue, skb); in llc_conn_resend_i_pdu_as_cmd() 272 skb_queue_tail(&sk->sk_write_queue, skb); in llc_conn_resend_i_pdu_as_rsp() 331 while ((skb = skb_dequeue(&sk->sk_write_queue)) != NULL) { in llc_conn_send_pdus() 971 skb_queue_len(&sk->sk_write_queue)); in llc_sk_free() 974 skb_queue_purge(&sk->sk_write_queue); in llc_sk_free() 1003 skb_queue_purge(&sk->sk_write_queue); in llc_sk_reset()
|
/linux/net/kcm/ |
A D | kcmsock.c | 591 if (skb_queue_empty(&sk->sk_write_queue)) in kcm_write_msgs() 594 kcm_tx_msg(skb_peek(&sk->sk_write_queue))->sent = 0; in kcm_write_msgs() 596 } else if (skb_queue_empty(&sk->sk_write_queue)) { in kcm_write_msgs() 600 head = skb_peek(&sk->sk_write_queue); in kcm_write_msgs() 700 skb_dequeue(&sk->sk_write_queue); in kcm_write_msgs() 705 } while ((head = skb_peek(&sk->sk_write_queue))); in kcm_write_msgs() 709 WARN_ON(!skb_queue_empty(&sk->sk_write_queue)); in kcm_write_msgs() 853 bool not_busy = skb_queue_empty(&sk->sk_write_queue); in kcm_sendpage() 856 __skb_queue_tail(&sk->sk_write_queue, head); in kcm_sendpage() 1026 __skb_queue_tail(&sk->sk_write_queue, head); in kcm_sendmsg() [all …]
|
A D | kcmproc.c | 121 kcm->sk.sk_write_queue.qlen, in kcm_format_sock() 151 psock->sk->sk_write_queue.qlen, in kcm_format_psock()
|
/linux/include/net/ |
A D | tcp.h | 1808 return skb_peek(&sk->sk_write_queue); in tcp_write_queue_head() 1813 return skb_peek_tail(&sk->sk_write_queue); in tcp_write_queue_tail() 1817 skb_queue_walk_from_safe(&(sk)->sk_write_queue, skb, tmp) 1821 return skb_peek(&sk->sk_write_queue); in tcp_send_head() 1827 return skb_queue_is_last(&sk->sk_write_queue, skb); in tcp_skb_is_last() 1856 __skb_queue_tail(&sk->sk_write_queue, skb); in tcp_add_write_queue_tail() 1859 if (sk->sk_write_queue.next == skb) in tcp_add_write_queue_tail() 1868 __skb_queue_before(&sk->sk_write_queue, skb, new); in tcp_insert_write_queue_before() 1874 __skb_unlink(skb, &sk->sk_write_queue); in tcp_unlink_write_queue()
|
A D | udplite.h | 104 skb_queue_walk(&sk->sk_write_queue, skb) { in udplite_csum_outgoing()
|
A D | udp.h | 132 skb_queue_walk(&sk->sk_write_queue, skb) { in udp_csum_outgoing()
|
/linux/net/ipv4/ |
A D | ip_output.c | 1302 if (skb_queue_empty(&sk->sk_write_queue)) { in ip_append_data() 1310 return __ip_append_data(sk, fl4, &sk->sk_write_queue, &inet->cork.base, in ip_append_data() 1335 if (skb_queue_empty(&sk->sk_write_queue)) in ip_append_page() 1359 skb = skb_peek_tail(&sk->sk_write_queue); in ip_append_page() 1412 __skb_queue_tail(&sk->sk_write_queue, skb); in ip_append_page() 1595 __ip_flush_pending_frames(sk, &sk->sk_write_queue, &inet_sk(sk)->cork.base); in ip_flush_pending_frames() 1708 nskb = skb_peek(&sk->sk_write_queue); in ip_send_unicast_reply()
|
A D | icmp.c | 380 } else if ((skb = skb_peek(&sk->sk_write_queue)) != NULL) { in icmp_push_reply() 388 skb_queue_walk(&sk->sk_write_queue, skb1) { in icmp_push_reply()
|
/linux/net/core/ |
A D | stream.c | 199 WARN_ON(!skb_queue_empty(&sk->sk_write_queue)); in sk_stream_kill_queues()
|
/linux/net/tipc/ |
A D | socket.c | 565 __skb_queue_purge(&sk->sk_write_queue); in __tipc_shutdown() 1275 struct sk_buff_head *txq = &tsk->sk.sk_write_queue; in tipc_sk_push_backlog() 1507 if (unlikely(syn && !tipc_msg_skb_clone(&pkts, &sk->sk_write_queue))) { in __tipc_sendmsg() 1560 struct sk_buff_head *txq = &sk->sk_write_queue; in __tipc_sendstream() 1680 __skb_queue_purge(&sk->sk_write_queue); in tipc_sk_finish_conn() 2234 if (skb_queue_empty(&sk->sk_write_queue)) in tipc_sk_filter_connect() 2253 if (!skb_queue_empty(&sk->sk_write_queue)) in tipc_sk_filter_connect() 2859 tipc_msg_skb_clone(&sk->sk_write_queue, list); in tipc_sk_retry_connect() 3663 skb_queue_len(&sk->sk_write_queue)) || in tipc_sk_fill_sock_diag() 3996 i += tipc_list_dump(&sk->sk_write_queue, false, buf + i); in tipc_sk_dump()
|
/linux/net/ipv6/ |
A D | icmp.c | 282 skb = skb_peek(&sk->sk_write_queue); in icmpv6_push_pending_frames() 290 if (skb_queue_len(&sk->sk_write_queue) == 1) { in icmpv6_push_pending_frames() 300 skb_queue_walk(&sk->sk_write_queue, skb) { in icmpv6_push_pending_frames()
|
A D | raw.c | 553 skb = skb_peek(&sk->sk_write_queue); in rawv6_push_pending_frames() 566 if (skb_queue_len(&sk->sk_write_queue) == 1) { in rawv6_push_pending_frames() 575 skb_queue_walk(&sk->sk_write_queue, skb) { in rawv6_push_pending_frames()
|