wlcore: improved Tx scheduling algorithm

Prioritize EDCA by choosing the AC before anything else. Use the
fast/slow link bitmap in FW to improve the scheduling algorithm for
the multi-link scenario.

Set packet thresholds to determine if a given link is high or low
priority according to its speed. A slow link will be given high priority
if the amount of packets queued for it in the FW is lower than the
slow-threshold. Similarly, a fast link will be given high priority if
the number of its packets queued in FW is smaller than the high-threshold.

The improved algorithm:
1. Choose optimal AC according to FW utilization
2. Traversing the VIFs in a round-robin fashion, try to choose a high
   priority link. Links are traversed in a round-robin fashion inside a
   VIF.
3. If no high priority links are found, choose the first non-empty
   (low priority) link found in the round robin.

Signed-off-by: Arik Nemtsov <arik@wizery.com>
Signed-off-by: Luciano Coelho <coelho@ti.com>
This commit is contained in:
Arik Nemtsov 2012-11-27 08:45:00 +02:00 committed by Luciano Coelho
parent 9ebcb23215
commit 0e81047996
6 changed files with 112 additions and 37 deletions

View File

@ -210,6 +210,8 @@ static struct wlcore_conf wl12xx_conf = {
.tmpl_short_retry_limit = 10, .tmpl_short_retry_limit = 10,
.tmpl_long_retry_limit = 10, .tmpl_long_retry_limit = 10,
.tx_watchdog_timeout = 5000, .tx_watchdog_timeout = 5000,
.slow_link_thold = 3,
.fast_link_thold = 10,
}, },
.conn = { .conn = {
.wake_up_event = CONF_WAKE_UP_EVENT_DTIM, .wake_up_event = CONF_WAKE_UP_EVENT_DTIM,

View File

@ -337,6 +337,8 @@ static struct wlcore_conf wl18xx_conf = {
.tmpl_short_retry_limit = 10, .tmpl_short_retry_limit = 10,
.tmpl_long_retry_limit = 10, .tmpl_long_retry_limit = 10,
.tx_watchdog_timeout = 5000, .tx_watchdog_timeout = 5000,
.slow_link_thold = 3,
.fast_link_thold = 30,
}, },
.conn = { .conn = {
.wake_up_event = CONF_WAKE_UP_EVENT_DTIM, .wake_up_event = CONF_WAKE_UP_EVENT_DTIM,

View File

@ -677,6 +677,18 @@ struct conf_tx_settings {
/* Time in ms for Tx watchdog timer to expire */ /* Time in ms for Tx watchdog timer to expire */
u32 tx_watchdog_timeout; u32 tx_watchdog_timeout;
/*
* when a slow link has this much packets pending, it becomes a low
* priority link, scheduling-wise
*/
u8 slow_link_thold;
/*
* when a fast link has this much packets pending, it becomes a low
* priority link, scheduling-wise
*/
u8 fast_link_thold;
} __packed; } __packed;
enum { enum {
@ -1281,7 +1293,7 @@ struct conf_recovery_settings {
* version, the two LSB are the lower driver's private conf * version, the two LSB are the lower driver's private conf
* version. * version.
*/ */
#define WLCORE_CONF_VERSION (0x0004 << 16) #define WLCORE_CONF_VERSION (0x0005 << 16)
#define WLCORE_CONF_MASK 0xffff0000 #define WLCORE_CONF_MASK 0xffff0000
#define WLCORE_CONF_SIZE (sizeof(struct wlcore_conf_header) + \ #define WLCORE_CONF_SIZE (sizeof(struct wlcore_conf_header) + \
sizeof(struct wlcore_conf)) sizeof(struct wlcore_conf))

View File

@ -362,8 +362,6 @@ static void wl12xx_irq_update_links_status(struct wl1271 *wl,
u32 cur_fw_ps_map; u32 cur_fw_ps_map;
u8 hlid; u8 hlid;
/* TODO: also use link_fast_bitmap here */
cur_fw_ps_map = le32_to_cpu(status->link_ps_bitmap); cur_fw_ps_map = le32_to_cpu(status->link_ps_bitmap);
if (wl->ap_fw_ps_map != cur_fw_ps_map) { if (wl->ap_fw_ps_map != cur_fw_ps_map) {
wl1271_debug(DEBUG_PSM, wl1271_debug(DEBUG_PSM,
@ -479,6 +477,8 @@ static int wlcore_fw_status(struct wl1271 *wl,
wl->time_offset = (timespec_to_ns(&ts) >> 10) - wl->time_offset = (timespec_to_ns(&ts) >> 10) -
(s64)le32_to_cpu(status_2->fw_localtime); (s64)le32_to_cpu(status_2->fw_localtime);
wl->fw_fast_lnk_map = le32_to_cpu(status_2->link_fast_bitmap);
return 0; return 0;
} }

View File

@ -467,8 +467,7 @@ void wl1271_handle_tx_low_watermark(struct wl1271 *wl)
} }
} }
static struct sk_buff_head *wl1271_select_queue(struct wl1271 *wl, static int wlcore_select_ac(struct wl1271 *wl)
struct sk_buff_head *queues)
{ {
int i, q = -1, ac; int i, q = -1, ac;
u32 min_pkts = 0xffffffff; u32 min_pkts = 0xffffffff;
@ -482,33 +481,24 @@ static struct sk_buff_head *wl1271_select_queue(struct wl1271 *wl,
*/ */
for (i = 0; i < NUM_TX_QUEUES; i++) { for (i = 0; i < NUM_TX_QUEUES; i++) {
ac = wl1271_tx_get_queue(i); ac = wl1271_tx_get_queue(i);
if (!skb_queue_empty(&queues[ac]) && if (wl->tx_queue_count[ac] &&
(wl->tx_allocated_pkts[ac] < min_pkts)) { wl->tx_allocated_pkts[ac] < min_pkts) {
q = ac; q = ac;
min_pkts = wl->tx_allocated_pkts[q]; min_pkts = wl->tx_allocated_pkts[q];
} }
} }
if (q == -1) return q;
return NULL;
return &queues[q];
} }
static struct sk_buff *wl12xx_lnk_skb_dequeue(struct wl1271 *wl, static struct sk_buff *wlcore_lnk_dequeue(struct wl1271 *wl,
struct wl1271_link *lnk) struct wl1271_link *lnk, u8 q)
{ {
struct sk_buff *skb; struct sk_buff *skb;
unsigned long flags; unsigned long flags;
struct sk_buff_head *queue;
queue = wl1271_select_queue(wl, lnk->tx_queue); skb = skb_dequeue(&lnk->tx_queue[q]);
if (!queue)
return NULL;
skb = skb_dequeue(queue);
if (skb) { if (skb) {
int q = wl1271_tx_get_queue(skb_get_queue_mapping(skb));
spin_lock_irqsave(&wl->wl_lock, flags); spin_lock_irqsave(&wl->wl_lock, flags);
WARN_ON_ONCE(wl->tx_queue_count[q] <= 0); WARN_ON_ONCE(wl->tx_queue_count[q] <= 0);
wl->tx_queue_count[q]--; wl->tx_queue_count[q]--;
@ -522,9 +512,41 @@ static struct sk_buff *wl12xx_lnk_skb_dequeue(struct wl1271 *wl,
return skb; return skb;
} }
static struct sk_buff *wl12xx_vif_skb_dequeue(struct wl1271 *wl, static bool wlcore_lnk_high_prio(struct wl1271 *wl, u8 hlid,
struct wl12xx_vif *wlvif, struct wl1271_link *lnk)
u8 *hlid) {
u8 thold;
if (test_bit(hlid, (unsigned long *)&wl->fw_fast_lnk_map))
thold = wl->conf.tx.fast_link_thold;
else
thold = wl->conf.tx.slow_link_thold;
return lnk->allocated_pkts < thold;
}
static struct sk_buff *wlcore_lnk_dequeue_high_prio(struct wl1271 *wl,
u8 hlid, u8 ac,
u8 *low_prio_hlid)
{
struct wl1271_link *lnk = &wl->links[hlid];
if (!wlcore_lnk_high_prio(wl, hlid, lnk)) {
if (*low_prio_hlid == WL12XX_INVALID_LINK_ID &&
!skb_queue_empty(&lnk->tx_queue[ac]))
/* we found the first non-empty low priority queue */
*low_prio_hlid = hlid;
return NULL;
}
return wlcore_lnk_dequeue(wl, lnk, ac);
}
static struct sk_buff *wlcore_vif_dequeue_high_prio(struct wl1271 *wl,
struct wl12xx_vif *wlvif,
u8 ac, u8 *hlid,
u8 *low_prio_hlid)
{ {
struct sk_buff *skb = NULL; struct sk_buff *skb = NULL;
int i, h, start_hlid; int i, h, start_hlid;
@ -540,7 +562,8 @@ static struct sk_buff *wl12xx_vif_skb_dequeue(struct wl1271 *wl,
if (!test_bit(h, wlvif->links_map)) if (!test_bit(h, wlvif->links_map))
continue; continue;
skb = wl12xx_lnk_skb_dequeue(wl, &wl->links[h]); skb = wlcore_lnk_dequeue_high_prio(wl, h, ac,
low_prio_hlid);
if (!skb) if (!skb)
continue; continue;
@ -560,42 +583,74 @@ static struct sk_buff *wl1271_skb_dequeue(struct wl1271 *wl, u8 *hlid)
unsigned long flags; unsigned long flags;
struct wl12xx_vif *wlvif = wl->last_wlvif; struct wl12xx_vif *wlvif = wl->last_wlvif;
struct sk_buff *skb = NULL; struct sk_buff *skb = NULL;
int ac;
u8 low_prio_hlid = WL12XX_INVALID_LINK_ID;
ac = wlcore_select_ac(wl);
if (ac < 0)
goto out;
/* continue from last wlvif (round robin) */ /* continue from last wlvif (round robin) */
if (wlvif) { if (wlvif) {
wl12xx_for_each_wlvif_continue(wl, wlvif) { wl12xx_for_each_wlvif_continue(wl, wlvif) {
skb = wl12xx_vif_skb_dequeue(wl, wlvif, hlid); if (!wlvif->tx_queue_count[ac])
if (skb) { continue;
wl->last_wlvif = wlvif;
break; skb = wlcore_vif_dequeue_high_prio(wl, wlvif, ac, hlid,
} &low_prio_hlid);
if (!skb)
continue;
wl->last_wlvif = wlvif;
break;
} }
} }
/* dequeue from the system HLID before the restarting wlvif list */ /* dequeue from the system HLID before the restarting wlvif list */
if (!skb) { if (!skb) {
skb = wl12xx_lnk_skb_dequeue(wl, &wl->links[wl->system_hlid]); skb = wlcore_lnk_dequeue_high_prio(wl, wl->system_hlid,
*hlid = wl->system_hlid; ac, &low_prio_hlid);
if (skb) {
*hlid = wl->system_hlid;
wl->last_wlvif = NULL;
}
} }
/* do a new pass over the wlvif list */ /* Do a new pass over the wlvif list. But no need to continue
* after last_wlvif. The previous pass should have found it. */
if (!skb) { if (!skb) {
wl12xx_for_each_wlvif(wl, wlvif) { wl12xx_for_each_wlvif(wl, wlvif) {
skb = wl12xx_vif_skb_dequeue(wl, wlvif, hlid); if (!wlvif->tx_queue_count[ac])
goto next;
skb = wlcore_vif_dequeue_high_prio(wl, wlvif, ac, hlid,
&low_prio_hlid);
if (skb) { if (skb) {
wl->last_wlvif = wlvif; wl->last_wlvif = wlvif;
break; break;
} }
/* next:
* No need to continue after last_wlvif. The previous
* pass should have found it.
*/
if (wlvif == wl->last_wlvif) if (wlvif == wl->last_wlvif)
break; break;
} }
} }
/* no high priority skbs found - but maybe a low priority one? */
if (!skb && low_prio_hlid != WL12XX_INVALID_LINK_ID) {
struct wl1271_link *lnk = &wl->links[low_prio_hlid];
skb = wlcore_lnk_dequeue(wl, lnk, ac);
WARN_ON(!skb); /* we checked this before */
*hlid = low_prio_hlid;
/* ensure proper round robin in the vif/link levels */
wl->last_wlvif = lnk->wlvif;
if (lnk->wlvif)
lnk->wlvif->last_tx_hlid = low_prio_hlid;
}
if (!skb && if (!skb &&
test_and_clear_bit(WL1271_FLAG_DUMMY_PACKET_PENDING, &wl->flags)) { test_and_clear_bit(WL1271_FLAG_DUMMY_PACKET_PENDING, &wl->flags)) {
int q; int q;
@ -609,6 +664,7 @@ static struct sk_buff *wl1271_skb_dequeue(struct wl1271 *wl, u8 *hlid)
spin_unlock_irqrestore(&wl->wl_lock, flags); spin_unlock_irqrestore(&wl->wl_lock, flags);
} }
out:
return skb; return skb;
} }

View File

@ -361,6 +361,9 @@ struct wl1271 {
*/ */
struct wl1271_link links[WL12XX_MAX_LINKS]; struct wl1271_link links[WL12XX_MAX_LINKS];
/* Fast/slow links bitmap according to FW */
u32 fw_fast_lnk_map;
/* AP-mode - a bitmap of links currently in PS mode according to FW */ /* AP-mode - a bitmap of links currently in PS mode according to FW */
u32 ap_fw_ps_map; u32 ap_fw_ps_map;