1 // SPDX-License-Identifier: GPL-2.0-only
2 /*
3  * Queue between the tx operation and the bh workqueue.
4  *
5  * Copyright (c) 2017-2020, Silicon Laboratories, Inc.
6  * Copyright (c) 2010, ST-Ericsson
7  */
8 #include <linux/sched.h>
9 #include <net/mac80211.h>
10 
11 #include "queue.h"
12 #include "wfx.h"
13 #include "sta.h"
14 #include "data_tx.h"
15 #include "traces.h"
16 
wfx_tx_lock(struct wfx_dev * wdev)17 void wfx_tx_lock(struct wfx_dev *wdev)
18 {
19 	atomic_inc(&wdev->tx_lock);
20 }
21 
wfx_tx_unlock(struct wfx_dev * wdev)22 void wfx_tx_unlock(struct wfx_dev *wdev)
23 {
24 	int tx_lock = atomic_dec_return(&wdev->tx_lock);
25 
26 	WARN(tx_lock < 0, "inconsistent tx_lock value");
27 	if (!tx_lock)
28 		wfx_bh_request_tx(wdev);
29 }
30 
wfx_tx_flush(struct wfx_dev * wdev)31 void wfx_tx_flush(struct wfx_dev *wdev)
32 {
33 	int ret;
34 
35 	/* Do not wait for any reply if chip is frozen */
36 	if (wdev->chip_frozen)
37 		return;
38 
39 	wfx_tx_lock(wdev);
40 	mutex_lock(&wdev->hif_cmd.lock);
41 	ret = wait_event_timeout(wdev->hif.tx_buffers_empty,
42 				 !wdev->hif.tx_buffers_used,
43 				 msecs_to_jiffies(3000));
44 	if (!ret) {
45 		dev_warn(wdev->dev, "cannot flush tx buffers (%d still busy)\n",
46 			 wdev->hif.tx_buffers_used);
47 		wfx_pending_dump_old_frames(wdev, 3000);
48 		/* FIXME: drop pending frames here */
49 		wdev->chip_frozen = true;
50 	}
51 	mutex_unlock(&wdev->hif_cmd.lock);
52 	wfx_tx_unlock(wdev);
53 }
54 
wfx_tx_lock_flush(struct wfx_dev * wdev)55 void wfx_tx_lock_flush(struct wfx_dev *wdev)
56 {
57 	wfx_tx_lock(wdev);
58 	wfx_tx_flush(wdev);
59 }
60 
wfx_tx_queues_init(struct wfx_vif * wvif)61 void wfx_tx_queues_init(struct wfx_vif *wvif)
62 {
63 	/* The device is in charge to respect the details of the QoS parameters.
64 	 * The driver just ensure that it roughtly respect the priorities to
65 	 * avoid any shortage.
66 	 */
67 	const int priorities[IEEE80211_NUM_ACS] = { 1, 2, 64, 128 };
68 	int i;
69 
70 	for (i = 0; i < IEEE80211_NUM_ACS; ++i) {
71 		skb_queue_head_init(&wvif->tx_queue[i].normal);
72 		skb_queue_head_init(&wvif->tx_queue[i].cab);
73 		wvif->tx_queue[i].priority = priorities[i];
74 	}
75 }
76 
wfx_tx_queue_empty(struct wfx_vif * wvif,struct wfx_queue * queue)77 bool wfx_tx_queue_empty(struct wfx_vif *wvif, struct wfx_queue *queue)
78 {
79 	return skb_queue_empty_lockless(&queue->normal) &&
80 	       skb_queue_empty_lockless(&queue->cab);
81 }
82 
wfx_tx_queues_check_empty(struct wfx_vif * wvif)83 void wfx_tx_queues_check_empty(struct wfx_vif *wvif)
84 {
85 	int i;
86 
87 	for (i = 0; i < IEEE80211_NUM_ACS; ++i) {
88 		WARN_ON(atomic_read(&wvif->tx_queue[i].pending_frames));
89 		WARN_ON(!wfx_tx_queue_empty(wvif, &wvif->tx_queue[i]));
90 	}
91 }
92 
__wfx_tx_queue_drop(struct wfx_vif * wvif,struct sk_buff_head * skb_queue,struct sk_buff_head * dropped)93 static void __wfx_tx_queue_drop(struct wfx_vif *wvif,
94 				struct sk_buff_head *skb_queue,
95 				struct sk_buff_head *dropped)
96 {
97 	struct sk_buff *skb, *tmp;
98 
99 	spin_lock_bh(&skb_queue->lock);
100 	skb_queue_walk_safe(skb_queue, skb, tmp) {
101 		__skb_unlink(skb, skb_queue);
102 		skb_queue_head(dropped, skb);
103 	}
104 	spin_unlock_bh(&skb_queue->lock);
105 }
106 
wfx_tx_queue_drop(struct wfx_vif * wvif,struct wfx_queue * queue,struct sk_buff_head * dropped)107 void wfx_tx_queue_drop(struct wfx_vif *wvif, struct wfx_queue *queue,
108 		       struct sk_buff_head *dropped)
109 {
110 	__wfx_tx_queue_drop(wvif, &queue->cab, dropped);
111 	__wfx_tx_queue_drop(wvif, &queue->normal, dropped);
112 	wake_up(&wvif->wdev->tx_dequeue);
113 }
114 
wfx_tx_queues_put(struct wfx_vif * wvif,struct sk_buff * skb)115 void wfx_tx_queues_put(struct wfx_vif *wvif, struct sk_buff *skb)
116 {
117 	struct wfx_queue *queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
118 	struct ieee80211_tx_info *tx_info = IEEE80211_SKB_CB(skb);
119 
120 	if (tx_info->flags & IEEE80211_TX_CTL_SEND_AFTER_DTIM)
121 		skb_queue_tail(&queue->cab, skb);
122 	else
123 		skb_queue_tail(&queue->normal, skb);
124 }
125 
wfx_pending_drop(struct wfx_dev * wdev,struct sk_buff_head * dropped)126 void wfx_pending_drop(struct wfx_dev *wdev, struct sk_buff_head *dropped)
127 {
128 	struct wfx_queue *queue;
129 	struct wfx_vif *wvif;
130 	struct hif_msg *hif;
131 	struct sk_buff *skb;
132 
133 	WARN(!wdev->chip_frozen, "%s should only be used to recover a frozen device",
134 	     __func__);
135 	while ((skb = skb_dequeue(&wdev->tx_pending)) != NULL) {
136 		hif = (struct hif_msg *)skb->data;
137 		wvif = wdev_to_wvif(wdev, hif->interface);
138 		if (wvif) {
139 			queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
140 			WARN_ON(skb_get_queue_mapping(skb) > 3);
141 			WARN_ON(!atomic_read(&queue->pending_frames));
142 			atomic_dec(&queue->pending_frames);
143 		}
144 		skb_queue_head(dropped, skb);
145 	}
146 }
147 
wfx_pending_get(struct wfx_dev * wdev,u32 packet_id)148 struct sk_buff *wfx_pending_get(struct wfx_dev *wdev, u32 packet_id)
149 {
150 	struct wfx_queue *queue;
151 	struct hif_req_tx *req;
152 	struct wfx_vif *wvif;
153 	struct hif_msg *hif;
154 	struct sk_buff *skb;
155 
156 	spin_lock_bh(&wdev->tx_pending.lock);
157 	skb_queue_walk(&wdev->tx_pending, skb) {
158 		hif = (struct hif_msg *)skb->data;
159 		req = (struct hif_req_tx *)hif->body;
160 		if (req->packet_id != packet_id)
161 			continue;
162 		spin_unlock_bh(&wdev->tx_pending.lock);
163 		wvif = wdev_to_wvif(wdev, hif->interface);
164 		if (wvif) {
165 			queue = &wvif->tx_queue[skb_get_queue_mapping(skb)];
166 			WARN_ON(skb_get_queue_mapping(skb) > 3);
167 			WARN_ON(!atomic_read(&queue->pending_frames));
168 			atomic_dec(&queue->pending_frames);
169 		}
170 		skb_unlink(skb, &wdev->tx_pending);
171 		return skb;
172 	}
173 	spin_unlock_bh(&wdev->tx_pending.lock);
174 	WARN(1, "cannot find packet in pending queue");
175 	return NULL;
176 }
177 
wfx_pending_dump_old_frames(struct wfx_dev * wdev,unsigned int limit_ms)178 void wfx_pending_dump_old_frames(struct wfx_dev *wdev, unsigned int limit_ms)
179 {
180 	ktime_t now = ktime_get();
181 	struct wfx_tx_priv *tx_priv;
182 	struct hif_req_tx *req;
183 	struct sk_buff *skb;
184 	bool first = true;
185 
186 	spin_lock_bh(&wdev->tx_pending.lock);
187 	skb_queue_walk(&wdev->tx_pending, skb) {
188 		tx_priv = wfx_skb_tx_priv(skb);
189 		req = wfx_skb_txreq(skb);
190 		if (ktime_after(now, ktime_add_ms(tx_priv->xmit_timestamp,
191 						  limit_ms))) {
192 			if (first) {
193 				dev_info(wdev->dev, "frames stuck in firmware since %dms or more:\n",
194 					 limit_ms);
195 				first = false;
196 			}
197 			dev_info(wdev->dev, "   id %08x sent %lldms ago\n",
198 				 req->packet_id,
199 				 ktime_ms_delta(now, tx_priv->xmit_timestamp));
200 		}
201 	}
202 	spin_unlock_bh(&wdev->tx_pending.lock);
203 }
204 
wfx_pending_get_pkt_us_delay(struct wfx_dev * wdev,struct sk_buff * skb)205 unsigned int wfx_pending_get_pkt_us_delay(struct wfx_dev *wdev,
206 					  struct sk_buff *skb)
207 {
208 	ktime_t now = ktime_get();
209 	struct wfx_tx_priv *tx_priv = wfx_skb_tx_priv(skb);
210 
211 	return ktime_us_delta(now, tx_priv->xmit_timestamp);
212 }
213 
wfx_tx_queues_has_cab(struct wfx_vif * wvif)214 bool wfx_tx_queues_has_cab(struct wfx_vif *wvif)
215 {
216 	int i;
217 
218 	if (wvif->vif->type != NL80211_IFTYPE_AP)
219 		return false;
220 	for (i = 0; i < IEEE80211_NUM_ACS; ++i)
221 		/* Note: since only AP can have mcast frames in queue and only
222 		 * one vif can be AP, all queued frames has same interface id
223 		 */
224 		if (!skb_queue_empty_lockless(&wvif->tx_queue[i].cab))
225 			return true;
226 	return false;
227 }
228 
wfx_tx_queue_get_weight(struct wfx_queue * queue)229 static int wfx_tx_queue_get_weight(struct wfx_queue *queue)
230 {
231 	return atomic_read(&queue->pending_frames) * queue->priority;
232 }
233 
wfx_tx_queues_get_skb(struct wfx_dev * wdev)234 static struct sk_buff *wfx_tx_queues_get_skb(struct wfx_dev *wdev)
235 {
236 	struct wfx_queue *queues[IEEE80211_NUM_ACS * ARRAY_SIZE(wdev->vif)];
237 	int i, j, num_queues = 0;
238 	struct wfx_vif *wvif;
239 	struct hif_msg *hif;
240 	struct sk_buff *skb;
241 
242 	/* sort the queues */
243 	wvif = NULL;
244 	while ((wvif = wvif_iterate(wdev, wvif)) != NULL) {
245 		for (i = 0; i < IEEE80211_NUM_ACS; i++) {
246 			WARN_ON(num_queues >= ARRAY_SIZE(queues));
247 			queues[num_queues] = &wvif->tx_queue[i];
248 			for (j = num_queues; j > 0; j--)
249 				if (wfx_tx_queue_get_weight(queues[j]) <
250 				    wfx_tx_queue_get_weight(queues[j - 1]))
251 					swap(queues[j - 1], queues[j]);
252 			num_queues++;
253 		}
254 	}
255 
256 	wvif = NULL;
257 	while ((wvif = wvif_iterate(wdev, wvif)) != NULL) {
258 		if (!wvif->after_dtim_tx_allowed)
259 			continue;
260 		for (i = 0; i < num_queues; i++) {
261 			skb = skb_dequeue(&queues[i]->cab);
262 			if (!skb)
263 				continue;
264 			/* Note: since only AP can have mcast frames in queue
265 			 * and only one vif can be AP, all queued frames has
266 			 * same interface id
267 			 */
268 			hif = (struct hif_msg *)skb->data;
269 			WARN_ON(hif->interface != wvif->id);
270 			WARN_ON(queues[i] !=
271 				&wvif->tx_queue[skb_get_queue_mapping(skb)]);
272 			atomic_inc(&queues[i]->pending_frames);
273 			trace_queues_stats(wdev, queues[i]);
274 			return skb;
275 		}
276 		/* No more multicast to sent */
277 		wvif->after_dtim_tx_allowed = false;
278 		schedule_work(&wvif->update_tim_work);
279 	}
280 
281 	for (i = 0; i < num_queues; i++) {
282 		skb = skb_dequeue(&queues[i]->normal);
283 		if (skb) {
284 			atomic_inc(&queues[i]->pending_frames);
285 			trace_queues_stats(wdev, queues[i]);
286 			return skb;
287 		}
288 	}
289 	return NULL;
290 }
291 
wfx_tx_queues_get(struct wfx_dev * wdev)292 struct hif_msg *wfx_tx_queues_get(struct wfx_dev *wdev)
293 {
294 	struct wfx_tx_priv *tx_priv;
295 	struct sk_buff *skb;
296 
297 	if (atomic_read(&wdev->tx_lock))
298 		return NULL;
299 	skb = wfx_tx_queues_get_skb(wdev);
300 	if (!skb)
301 		return NULL;
302 	skb_queue_tail(&wdev->tx_pending, skb);
303 	wake_up(&wdev->tx_dequeue);
304 	tx_priv = wfx_skb_tx_priv(skb);
305 	tx_priv->xmit_timestamp = ktime_get();
306 	return (struct hif_msg *)skb->data;
307 }
308