forked from luck/tmp_suning_uos_patched
dba0491f00
Because we can miss AP wakeup (beacon) while scanning other channels, it's better go into wakeup state and inform the AP of that upon returning to the operating channel, rather than staying asleep and waiting for the next TIM indicating traffic for us. This saves precious time, especially when we only have 200ms inter- scan period for monitoring the active connection. Signed-off-by: Loic Poulain <loic.poulain@linaro.org> Link: https://lore.kernel.org/r/1593420923-26668-1-git-send-email-loic.poulain@linaro.org [rewrite commit message a bit] Signed-off-by: Johannes Berg <johannes.berg@intel.com>
993 lines
26 KiB
C
993 lines
26 KiB
C
// SPDX-License-Identifier: GPL-2.0-only
|
|
/*
|
|
* Off-channel operation helpers
|
|
*
|
|
* Copyright 2003, Jouni Malinen <jkmaline@cc.hut.fi>
|
|
* Copyright 2004, Instant802 Networks, Inc.
|
|
* Copyright 2005, Devicescape Software, Inc.
|
|
* Copyright 2006-2007 Jiri Benc <jbenc@suse.cz>
|
|
* Copyright 2007, Michael Wu <flamingice@sourmilk.net>
|
|
* Copyright 2009 Johannes Berg <johannes@sipsolutions.net>
|
|
* Copyright (C) 2019 Intel Corporation
|
|
*/
|
|
#include <linux/export.h>
|
|
#include <net/mac80211.h>
|
|
#include "ieee80211_i.h"
|
|
#include "driver-ops.h"
|
|
|
|
/*
|
|
* Tell our hardware to disable PS.
|
|
* Optionally inform AP that we will go to sleep so that it will buffer
|
|
* the frames while we are doing off-channel work. This is optional
|
|
* because we *may* be doing work on-operating channel, and want our
|
|
* hardware unconditionally awake, but still let the AP send us normal frames.
|
|
*/
|
|
static void ieee80211_offchannel_ps_enable(struct ieee80211_sub_if_data *sdata)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_if_managed *ifmgd = &sdata->u.mgd;
|
|
bool offchannel_ps_enabled = false;
|
|
|
|
/* FIXME: what to do when local->pspolling is true? */
|
|
|
|
del_timer_sync(&local->dynamic_ps_timer);
|
|
del_timer_sync(&ifmgd->bcn_mon_timer);
|
|
del_timer_sync(&ifmgd->conn_mon_timer);
|
|
|
|
cancel_work_sync(&local->dynamic_ps_enable_work);
|
|
|
|
if (local->hw.conf.flags & IEEE80211_CONF_PS) {
|
|
offchannel_ps_enabled = true;
|
|
local->hw.conf.flags &= ~IEEE80211_CONF_PS;
|
|
ieee80211_hw_config(local, IEEE80211_CONF_CHANGE_PS);
|
|
}
|
|
|
|
if (!offchannel_ps_enabled ||
|
|
!ieee80211_hw_check(&local->hw, PS_NULLFUNC_STACK))
|
|
/*
|
|
* If power save was enabled, no need to send a nullfunc
|
|
* frame because AP knows that we are sleeping. But if the
|
|
* hardware is creating the nullfunc frame for power save
|
|
* status (ie. IEEE80211_HW_PS_NULLFUNC_STACK is not
|
|
* enabled) and power save was enabled, the firmware just
|
|
* sent a null frame with power save disabled. So we need
|
|
* to send a new nullfunc frame to inform the AP that we
|
|
* are again sleeping.
|
|
*/
|
|
ieee80211_send_nullfunc(local, sdata, true);
|
|
}
|
|
|
|
/* inform AP that we are awake again */
|
|
static void ieee80211_offchannel_ps_disable(struct ieee80211_sub_if_data *sdata)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
|
|
if (!local->ps_sdata)
|
|
ieee80211_send_nullfunc(local, sdata, false);
|
|
else if (local->hw.conf.dynamic_ps_timeout > 0) {
|
|
/*
|
|
* the dynamic_ps_timer had been running before leaving the
|
|
* operating channel, restart the timer now and send a nullfunc
|
|
* frame to inform the AP that we are awake so that AP sends
|
|
* the buffered packets (if any).
|
|
*/
|
|
ieee80211_send_nullfunc(local, sdata, false);
|
|
mod_timer(&local->dynamic_ps_timer, jiffies +
|
|
msecs_to_jiffies(local->hw.conf.dynamic_ps_timeout));
|
|
}
|
|
|
|
ieee80211_sta_reset_beacon_monitor(sdata);
|
|
ieee80211_sta_reset_conn_monitor(sdata);
|
|
}
|
|
|
|
void ieee80211_offchannel_stop_vifs(struct ieee80211_local *local)
|
|
{
|
|
struct ieee80211_sub_if_data *sdata;
|
|
|
|
if (WARN_ON(local->use_chanctx))
|
|
return;
|
|
|
|
/*
|
|
* notify the AP about us leaving the channel and stop all
|
|
* STA interfaces.
|
|
*/
|
|
|
|
/*
|
|
* Stop queues and transmit all frames queued by the driver
|
|
* before sending nullfunc to enable powersave at the AP.
|
|
*/
|
|
ieee80211_stop_queues_by_reason(&local->hw, IEEE80211_MAX_QUEUE_MAP,
|
|
IEEE80211_QUEUE_STOP_REASON_OFFCHANNEL,
|
|
false);
|
|
ieee80211_flush_queues(local, NULL, false);
|
|
|
|
mutex_lock(&local->iflist_mtx);
|
|
list_for_each_entry(sdata, &local->interfaces, list) {
|
|
if (!ieee80211_sdata_running(sdata))
|
|
continue;
|
|
|
|
if (sdata->vif.type == NL80211_IFTYPE_P2P_DEVICE ||
|
|
sdata->vif.type == NL80211_IFTYPE_NAN)
|
|
continue;
|
|
|
|
if (sdata->vif.type != NL80211_IFTYPE_MONITOR)
|
|
set_bit(SDATA_STATE_OFFCHANNEL, &sdata->state);
|
|
|
|
/* Check to see if we should disable beaconing. */
|
|
if (sdata->vif.bss_conf.enable_beacon) {
|
|
set_bit(SDATA_STATE_OFFCHANNEL_BEACON_STOPPED,
|
|
&sdata->state);
|
|
sdata->vif.bss_conf.enable_beacon = false;
|
|
ieee80211_bss_info_change_notify(
|
|
sdata, BSS_CHANGED_BEACON_ENABLED);
|
|
}
|
|
|
|
if (sdata->vif.type == NL80211_IFTYPE_STATION &&
|
|
sdata->u.mgd.associated)
|
|
ieee80211_offchannel_ps_enable(sdata);
|
|
}
|
|
mutex_unlock(&local->iflist_mtx);
|
|
}
|
|
|
|
void ieee80211_offchannel_return(struct ieee80211_local *local)
|
|
{
|
|
struct ieee80211_sub_if_data *sdata;
|
|
|
|
if (WARN_ON(local->use_chanctx))
|
|
return;
|
|
|
|
mutex_lock(&local->iflist_mtx);
|
|
list_for_each_entry(sdata, &local->interfaces, list) {
|
|
if (sdata->vif.type == NL80211_IFTYPE_P2P_DEVICE)
|
|
continue;
|
|
|
|
if (sdata->vif.type != NL80211_IFTYPE_MONITOR)
|
|
clear_bit(SDATA_STATE_OFFCHANNEL, &sdata->state);
|
|
|
|
if (!ieee80211_sdata_running(sdata))
|
|
continue;
|
|
|
|
/* Tell AP we're back */
|
|
if (sdata->vif.type == NL80211_IFTYPE_STATION &&
|
|
sdata->u.mgd.associated)
|
|
ieee80211_offchannel_ps_disable(sdata);
|
|
|
|
if (test_and_clear_bit(SDATA_STATE_OFFCHANNEL_BEACON_STOPPED,
|
|
&sdata->state)) {
|
|
sdata->vif.bss_conf.enable_beacon = true;
|
|
ieee80211_bss_info_change_notify(
|
|
sdata, BSS_CHANGED_BEACON_ENABLED);
|
|
}
|
|
}
|
|
mutex_unlock(&local->iflist_mtx);
|
|
|
|
ieee80211_wake_queues_by_reason(&local->hw, IEEE80211_MAX_QUEUE_MAP,
|
|
IEEE80211_QUEUE_STOP_REASON_OFFCHANNEL,
|
|
false);
|
|
}
|
|
|
|
static void ieee80211_roc_notify_destroy(struct ieee80211_roc_work *roc)
|
|
{
|
|
/* was never transmitted */
|
|
if (roc->frame) {
|
|
cfg80211_mgmt_tx_status(&roc->sdata->wdev, roc->mgmt_tx_cookie,
|
|
roc->frame->data, roc->frame->len,
|
|
false, GFP_KERNEL);
|
|
ieee80211_free_txskb(&roc->sdata->local->hw, roc->frame);
|
|
}
|
|
|
|
if (!roc->mgmt_tx_cookie)
|
|
cfg80211_remain_on_channel_expired(&roc->sdata->wdev,
|
|
roc->cookie, roc->chan,
|
|
GFP_KERNEL);
|
|
else
|
|
cfg80211_tx_mgmt_expired(&roc->sdata->wdev,
|
|
roc->mgmt_tx_cookie,
|
|
roc->chan, GFP_KERNEL);
|
|
|
|
list_del(&roc->list);
|
|
kfree(roc);
|
|
}
|
|
|
|
static unsigned long ieee80211_end_finished_rocs(struct ieee80211_local *local,
|
|
unsigned long now)
|
|
{
|
|
struct ieee80211_roc_work *roc, *tmp;
|
|
long remaining_dur_min = LONG_MAX;
|
|
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
list_for_each_entry_safe(roc, tmp, &local->roc_list, list) {
|
|
long remaining;
|
|
|
|
if (!roc->started)
|
|
break;
|
|
|
|
remaining = roc->start_time +
|
|
msecs_to_jiffies(roc->duration) -
|
|
now;
|
|
|
|
/* In case of HW ROC, it is possible that the HW finished the
|
|
* ROC session before the actual requested time. In such a case
|
|
* end the ROC session (disregarding the remaining time).
|
|
*/
|
|
if (roc->abort || roc->hw_begun || remaining <= 0)
|
|
ieee80211_roc_notify_destroy(roc);
|
|
else
|
|
remaining_dur_min = min(remaining_dur_min, remaining);
|
|
}
|
|
|
|
return remaining_dur_min;
|
|
}
|
|
|
|
static bool ieee80211_recalc_sw_work(struct ieee80211_local *local,
|
|
unsigned long now)
|
|
{
|
|
long dur = ieee80211_end_finished_rocs(local, now);
|
|
|
|
if (dur == LONG_MAX)
|
|
return false;
|
|
|
|
mod_delayed_work(local->workqueue, &local->roc_work, dur);
|
|
return true;
|
|
}
|
|
|
|
static void ieee80211_handle_roc_started(struct ieee80211_roc_work *roc,
|
|
unsigned long start_time)
|
|
{
|
|
if (WARN_ON(roc->notified))
|
|
return;
|
|
|
|
roc->start_time = start_time;
|
|
roc->started = true;
|
|
|
|
if (roc->mgmt_tx_cookie) {
|
|
if (!WARN_ON(!roc->frame)) {
|
|
ieee80211_tx_skb_tid_band(roc->sdata, roc->frame, 7,
|
|
roc->chan->band);
|
|
roc->frame = NULL;
|
|
}
|
|
} else {
|
|
cfg80211_ready_on_channel(&roc->sdata->wdev, roc->cookie,
|
|
roc->chan, roc->req_duration,
|
|
GFP_KERNEL);
|
|
}
|
|
|
|
roc->notified = true;
|
|
}
|
|
|
|
static void ieee80211_hw_roc_start(struct work_struct *work)
|
|
{
|
|
struct ieee80211_local *local =
|
|
container_of(work, struct ieee80211_local, hw_roc_start);
|
|
struct ieee80211_roc_work *roc;
|
|
|
|
mutex_lock(&local->mtx);
|
|
|
|
list_for_each_entry(roc, &local->roc_list, list) {
|
|
if (!roc->started)
|
|
break;
|
|
|
|
roc->hw_begun = true;
|
|
ieee80211_handle_roc_started(roc, local->hw_roc_start_time);
|
|
}
|
|
|
|
mutex_unlock(&local->mtx);
|
|
}
|
|
|
|
void ieee80211_ready_on_channel(struct ieee80211_hw *hw)
|
|
{
|
|
struct ieee80211_local *local = hw_to_local(hw);
|
|
|
|
local->hw_roc_start_time = jiffies;
|
|
|
|
trace_api_ready_on_channel(local);
|
|
|
|
ieee80211_queue_work(hw, &local->hw_roc_start);
|
|
}
|
|
EXPORT_SYMBOL_GPL(ieee80211_ready_on_channel);
|
|
|
|
static void _ieee80211_start_next_roc(struct ieee80211_local *local)
|
|
{
|
|
struct ieee80211_roc_work *roc, *tmp;
|
|
enum ieee80211_roc_type type;
|
|
u32 min_dur, max_dur;
|
|
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
if (WARN_ON(list_empty(&local->roc_list)))
|
|
return;
|
|
|
|
roc = list_first_entry(&local->roc_list, struct ieee80211_roc_work,
|
|
list);
|
|
|
|
if (WARN_ON(roc->started))
|
|
return;
|
|
|
|
min_dur = roc->duration;
|
|
max_dur = roc->duration;
|
|
type = roc->type;
|
|
|
|
list_for_each_entry(tmp, &local->roc_list, list) {
|
|
if (tmp == roc)
|
|
continue;
|
|
if (tmp->sdata != roc->sdata || tmp->chan != roc->chan)
|
|
break;
|
|
max_dur = max(tmp->duration, max_dur);
|
|
min_dur = min(tmp->duration, min_dur);
|
|
type = max(tmp->type, type);
|
|
}
|
|
|
|
if (local->ops->remain_on_channel) {
|
|
int ret = drv_remain_on_channel(local, roc->sdata, roc->chan,
|
|
max_dur, type);
|
|
|
|
if (ret) {
|
|
wiphy_warn(local->hw.wiphy,
|
|
"failed to start next HW ROC (%d)\n", ret);
|
|
/*
|
|
* queue the work struct again to avoid recursion
|
|
* when multiple failures occur
|
|
*/
|
|
list_for_each_entry(tmp, &local->roc_list, list) {
|
|
if (tmp->sdata != roc->sdata ||
|
|
tmp->chan != roc->chan)
|
|
break;
|
|
tmp->started = true;
|
|
tmp->abort = true;
|
|
}
|
|
ieee80211_queue_work(&local->hw, &local->hw_roc_done);
|
|
return;
|
|
}
|
|
|
|
/* we'll notify about the start once the HW calls back */
|
|
list_for_each_entry(tmp, &local->roc_list, list) {
|
|
if (tmp->sdata != roc->sdata || tmp->chan != roc->chan)
|
|
break;
|
|
tmp->started = true;
|
|
}
|
|
} else {
|
|
/* If actually operating on the desired channel (with at least
|
|
* 20 MHz channel width) don't stop all the operations but still
|
|
* treat it as though the ROC operation started properly, so
|
|
* other ROC operations won't interfere with this one.
|
|
*/
|
|
roc->on_channel = roc->chan == local->_oper_chandef.chan &&
|
|
local->_oper_chandef.width != NL80211_CHAN_WIDTH_5 &&
|
|
local->_oper_chandef.width != NL80211_CHAN_WIDTH_10;
|
|
|
|
/* start this ROC */
|
|
ieee80211_recalc_idle(local);
|
|
|
|
if (!roc->on_channel) {
|
|
ieee80211_offchannel_stop_vifs(local);
|
|
|
|
local->tmp_channel = roc->chan;
|
|
ieee80211_hw_config(local, 0);
|
|
}
|
|
|
|
ieee80211_queue_delayed_work(&local->hw, &local->roc_work,
|
|
msecs_to_jiffies(min_dur));
|
|
|
|
/* tell userspace or send frame(s) */
|
|
list_for_each_entry(tmp, &local->roc_list, list) {
|
|
if (tmp->sdata != roc->sdata || tmp->chan != roc->chan)
|
|
break;
|
|
|
|
tmp->on_channel = roc->on_channel;
|
|
ieee80211_handle_roc_started(tmp, jiffies);
|
|
}
|
|
}
|
|
}
|
|
|
|
void ieee80211_start_next_roc(struct ieee80211_local *local)
|
|
{
|
|
struct ieee80211_roc_work *roc;
|
|
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
if (list_empty(&local->roc_list)) {
|
|
ieee80211_run_deferred_scan(local);
|
|
return;
|
|
}
|
|
|
|
/* defer roc if driver is not started (i.e. during reconfig) */
|
|
if (local->in_reconfig)
|
|
return;
|
|
|
|
roc = list_first_entry(&local->roc_list, struct ieee80211_roc_work,
|
|
list);
|
|
|
|
if (WARN_ON_ONCE(roc->started))
|
|
return;
|
|
|
|
if (local->ops->remain_on_channel) {
|
|
_ieee80211_start_next_roc(local);
|
|
} else {
|
|
/* delay it a bit */
|
|
ieee80211_queue_delayed_work(&local->hw, &local->roc_work,
|
|
round_jiffies_relative(HZ/2));
|
|
}
|
|
}
|
|
|
|
static void __ieee80211_roc_work(struct ieee80211_local *local)
|
|
{
|
|
struct ieee80211_roc_work *roc;
|
|
bool on_channel;
|
|
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
if (WARN_ON(local->ops->remain_on_channel))
|
|
return;
|
|
|
|
roc = list_first_entry_or_null(&local->roc_list,
|
|
struct ieee80211_roc_work, list);
|
|
if (!roc)
|
|
return;
|
|
|
|
if (!roc->started) {
|
|
WARN_ON(local->use_chanctx);
|
|
_ieee80211_start_next_roc(local);
|
|
} else {
|
|
on_channel = roc->on_channel;
|
|
if (ieee80211_recalc_sw_work(local, jiffies))
|
|
return;
|
|
|
|
/* careful - roc pointer became invalid during recalc */
|
|
|
|
if (!on_channel) {
|
|
ieee80211_flush_queues(local, NULL, false);
|
|
|
|
local->tmp_channel = NULL;
|
|
ieee80211_hw_config(local, 0);
|
|
|
|
ieee80211_offchannel_return(local);
|
|
}
|
|
|
|
ieee80211_recalc_idle(local);
|
|
ieee80211_start_next_roc(local);
|
|
}
|
|
}
|
|
|
|
static void ieee80211_roc_work(struct work_struct *work)
|
|
{
|
|
struct ieee80211_local *local =
|
|
container_of(work, struct ieee80211_local, roc_work.work);
|
|
|
|
mutex_lock(&local->mtx);
|
|
__ieee80211_roc_work(local);
|
|
mutex_unlock(&local->mtx);
|
|
}
|
|
|
|
static void ieee80211_hw_roc_done(struct work_struct *work)
|
|
{
|
|
struct ieee80211_local *local =
|
|
container_of(work, struct ieee80211_local, hw_roc_done);
|
|
|
|
mutex_lock(&local->mtx);
|
|
|
|
ieee80211_end_finished_rocs(local, jiffies);
|
|
|
|
/* if there's another roc, start it now */
|
|
ieee80211_start_next_roc(local);
|
|
|
|
mutex_unlock(&local->mtx);
|
|
}
|
|
|
|
void ieee80211_remain_on_channel_expired(struct ieee80211_hw *hw)
|
|
{
|
|
struct ieee80211_local *local = hw_to_local(hw);
|
|
|
|
trace_api_remain_on_channel_expired(local);
|
|
|
|
ieee80211_queue_work(hw, &local->hw_roc_done);
|
|
}
|
|
EXPORT_SYMBOL_GPL(ieee80211_remain_on_channel_expired);
|
|
|
|
static bool
|
|
ieee80211_coalesce_hw_started_roc(struct ieee80211_local *local,
|
|
struct ieee80211_roc_work *new_roc,
|
|
struct ieee80211_roc_work *cur_roc)
|
|
{
|
|
unsigned long now = jiffies;
|
|
unsigned long remaining;
|
|
|
|
if (WARN_ON(!cur_roc->started))
|
|
return false;
|
|
|
|
/* if it was scheduled in the hardware, but not started yet,
|
|
* we can only combine if the older one had a longer duration
|
|
*/
|
|
if (!cur_roc->hw_begun && new_roc->duration > cur_roc->duration)
|
|
return false;
|
|
|
|
remaining = cur_roc->start_time +
|
|
msecs_to_jiffies(cur_roc->duration) -
|
|
now;
|
|
|
|
/* if it doesn't fit entirely, schedule a new one */
|
|
if (new_roc->duration > jiffies_to_msecs(remaining))
|
|
return false;
|
|
|
|
/* add just after the current one so we combine their finish later */
|
|
list_add(&new_roc->list, &cur_roc->list);
|
|
|
|
/* if the existing one has already begun then let this one also
|
|
* begin, otherwise they'll both be marked properly by the work
|
|
* struct that runs once the driver notifies us of the beginning
|
|
*/
|
|
if (cur_roc->hw_begun) {
|
|
new_roc->hw_begun = true;
|
|
ieee80211_handle_roc_started(new_roc, now);
|
|
}
|
|
|
|
return true;
|
|
}
|
|
|
|
static int ieee80211_start_roc_work(struct ieee80211_local *local,
|
|
struct ieee80211_sub_if_data *sdata,
|
|
struct ieee80211_channel *channel,
|
|
unsigned int duration, u64 *cookie,
|
|
struct sk_buff *txskb,
|
|
enum ieee80211_roc_type type)
|
|
{
|
|
struct ieee80211_roc_work *roc, *tmp;
|
|
bool queued = false, combine_started = true;
|
|
int ret;
|
|
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
if (channel->freq_offset)
|
|
/* this may work, but is untested */
|
|
return -EOPNOTSUPP;
|
|
|
|
if (local->use_chanctx && !local->ops->remain_on_channel)
|
|
return -EOPNOTSUPP;
|
|
|
|
roc = kzalloc(sizeof(*roc), GFP_KERNEL);
|
|
if (!roc)
|
|
return -ENOMEM;
|
|
|
|
/*
|
|
* If the duration is zero, then the driver
|
|
* wouldn't actually do anything. Set it to
|
|
* 10 for now.
|
|
*
|
|
* TODO: cancel the off-channel operation
|
|
* when we get the SKB's TX status and
|
|
* the wait time was zero before.
|
|
*/
|
|
if (!duration)
|
|
duration = 10;
|
|
|
|
roc->chan = channel;
|
|
roc->duration = duration;
|
|
roc->req_duration = duration;
|
|
roc->frame = txskb;
|
|
roc->type = type;
|
|
roc->sdata = sdata;
|
|
|
|
/*
|
|
* cookie is either the roc cookie (for normal roc)
|
|
* or the SKB (for mgmt TX)
|
|
*/
|
|
if (!txskb) {
|
|
roc->cookie = ieee80211_mgmt_tx_cookie(local);
|
|
*cookie = roc->cookie;
|
|
} else {
|
|
roc->mgmt_tx_cookie = *cookie;
|
|
}
|
|
|
|
/* if there's no need to queue, handle it immediately */
|
|
if (list_empty(&local->roc_list) &&
|
|
!local->scanning && !ieee80211_is_radar_required(local)) {
|
|
/* if not HW assist, just queue & schedule work */
|
|
if (!local->ops->remain_on_channel) {
|
|
list_add_tail(&roc->list, &local->roc_list);
|
|
ieee80211_queue_delayed_work(&local->hw,
|
|
&local->roc_work, 0);
|
|
} else {
|
|
/* otherwise actually kick it off here
|
|
* (for error handling)
|
|
*/
|
|
ret = drv_remain_on_channel(local, sdata, channel,
|
|
duration, type);
|
|
if (ret) {
|
|
kfree(roc);
|
|
return ret;
|
|
}
|
|
roc->started = true;
|
|
list_add_tail(&roc->list, &local->roc_list);
|
|
}
|
|
|
|
return 0;
|
|
}
|
|
|
|
/* otherwise handle queueing */
|
|
|
|
list_for_each_entry(tmp, &local->roc_list, list) {
|
|
if (tmp->chan != channel || tmp->sdata != sdata)
|
|
continue;
|
|
|
|
/*
|
|
* Extend this ROC if possible: If it hasn't started, add
|
|
* just after the new one to combine.
|
|
*/
|
|
if (!tmp->started) {
|
|
list_add(&roc->list, &tmp->list);
|
|
queued = true;
|
|
break;
|
|
}
|
|
|
|
if (!combine_started)
|
|
continue;
|
|
|
|
if (!local->ops->remain_on_channel) {
|
|
/* If there's no hardware remain-on-channel, and
|
|
* doing so won't push us over the maximum r-o-c
|
|
* we allow, then we can just add the new one to
|
|
* the list and mark it as having started now.
|
|
* If it would push over the limit, don't try to
|
|
* combine with other started ones (that haven't
|
|
* been running as long) but potentially sort it
|
|
* with others that had the same fate.
|
|
*/
|
|
unsigned long now = jiffies;
|
|
u32 elapsed = jiffies_to_msecs(now - tmp->start_time);
|
|
struct wiphy *wiphy = local->hw.wiphy;
|
|
u32 max_roc = wiphy->max_remain_on_channel_duration;
|
|
|
|
if (elapsed + roc->duration > max_roc) {
|
|
combine_started = false;
|
|
continue;
|
|
}
|
|
|
|
list_add(&roc->list, &tmp->list);
|
|
queued = true;
|
|
roc->on_channel = tmp->on_channel;
|
|
ieee80211_handle_roc_started(roc, now);
|
|
ieee80211_recalc_sw_work(local, now);
|
|
break;
|
|
}
|
|
|
|
queued = ieee80211_coalesce_hw_started_roc(local, roc, tmp);
|
|
if (queued)
|
|
break;
|
|
/* if it wasn't queued, perhaps it can be combined with
|
|
* another that also couldn't get combined previously,
|
|
* but no need to check for already started ones, since
|
|
* that can't work.
|
|
*/
|
|
combine_started = false;
|
|
}
|
|
|
|
if (!queued)
|
|
list_add_tail(&roc->list, &local->roc_list);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int ieee80211_remain_on_channel(struct wiphy *wiphy, struct wireless_dev *wdev,
|
|
struct ieee80211_channel *chan,
|
|
unsigned int duration, u64 *cookie)
|
|
{
|
|
struct ieee80211_sub_if_data *sdata = IEEE80211_WDEV_TO_SUB_IF(wdev);
|
|
struct ieee80211_local *local = sdata->local;
|
|
int ret;
|
|
|
|
mutex_lock(&local->mtx);
|
|
ret = ieee80211_start_roc_work(local, sdata, chan,
|
|
duration, cookie, NULL,
|
|
IEEE80211_ROC_TYPE_NORMAL);
|
|
mutex_unlock(&local->mtx);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static int ieee80211_cancel_roc(struct ieee80211_local *local,
|
|
u64 cookie, bool mgmt_tx)
|
|
{
|
|
struct ieee80211_roc_work *roc, *tmp, *found = NULL;
|
|
int ret;
|
|
|
|
if (!cookie)
|
|
return -ENOENT;
|
|
|
|
flush_work(&local->hw_roc_start);
|
|
|
|
mutex_lock(&local->mtx);
|
|
list_for_each_entry_safe(roc, tmp, &local->roc_list, list) {
|
|
if (!mgmt_tx && roc->cookie != cookie)
|
|
continue;
|
|
else if (mgmt_tx && roc->mgmt_tx_cookie != cookie)
|
|
continue;
|
|
|
|
found = roc;
|
|
break;
|
|
}
|
|
|
|
if (!found) {
|
|
mutex_unlock(&local->mtx);
|
|
return -ENOENT;
|
|
}
|
|
|
|
if (!found->started) {
|
|
ieee80211_roc_notify_destroy(found);
|
|
goto out_unlock;
|
|
}
|
|
|
|
if (local->ops->remain_on_channel) {
|
|
ret = drv_cancel_remain_on_channel(local, roc->sdata);
|
|
if (WARN_ON_ONCE(ret)) {
|
|
mutex_unlock(&local->mtx);
|
|
return ret;
|
|
}
|
|
|
|
/* TODO:
|
|
* if multiple items were combined here then we really shouldn't
|
|
* cancel them all - we should wait for as much time as needed
|
|
* for the longest remaining one, and only then cancel ...
|
|
*/
|
|
list_for_each_entry_safe(roc, tmp, &local->roc_list, list) {
|
|
if (!roc->started)
|
|
break;
|
|
if (roc == found)
|
|
found = NULL;
|
|
ieee80211_roc_notify_destroy(roc);
|
|
}
|
|
|
|
/* that really must not happen - it was started */
|
|
WARN_ON(found);
|
|
|
|
ieee80211_start_next_roc(local);
|
|
} else {
|
|
/* go through work struct to return to the operating channel */
|
|
found->abort = true;
|
|
mod_delayed_work(local->workqueue, &local->roc_work, 0);
|
|
}
|
|
|
|
out_unlock:
|
|
mutex_unlock(&local->mtx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int ieee80211_cancel_remain_on_channel(struct wiphy *wiphy,
|
|
struct wireless_dev *wdev, u64 cookie)
|
|
{
|
|
struct ieee80211_sub_if_data *sdata = IEEE80211_WDEV_TO_SUB_IF(wdev);
|
|
struct ieee80211_local *local = sdata->local;
|
|
|
|
return ieee80211_cancel_roc(local, cookie, false);
|
|
}
|
|
|
|
int ieee80211_mgmt_tx(struct wiphy *wiphy, struct wireless_dev *wdev,
|
|
struct cfg80211_mgmt_tx_params *params, u64 *cookie)
|
|
{
|
|
struct ieee80211_sub_if_data *sdata = IEEE80211_WDEV_TO_SUB_IF(wdev);
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct sk_buff *skb;
|
|
struct sta_info *sta;
|
|
const struct ieee80211_mgmt *mgmt = (void *)params->buf;
|
|
bool need_offchan = false;
|
|
u32 flags;
|
|
int ret;
|
|
u8 *data;
|
|
|
|
if (params->dont_wait_for_ack)
|
|
flags = IEEE80211_TX_CTL_NO_ACK;
|
|
else
|
|
flags = IEEE80211_TX_INTFL_NL80211_FRAME_TX |
|
|
IEEE80211_TX_CTL_REQ_TX_STATUS;
|
|
|
|
if (params->no_cck)
|
|
flags |= IEEE80211_TX_CTL_NO_CCK_RATE;
|
|
|
|
switch (sdata->vif.type) {
|
|
case NL80211_IFTYPE_ADHOC:
|
|
if (!sdata->vif.bss_conf.ibss_joined)
|
|
need_offchan = true;
|
|
#ifdef CONFIG_MAC80211_MESH
|
|
fallthrough;
|
|
case NL80211_IFTYPE_MESH_POINT:
|
|
if (ieee80211_vif_is_mesh(&sdata->vif) &&
|
|
!sdata->u.mesh.mesh_id_len)
|
|
need_offchan = true;
|
|
#endif
|
|
fallthrough;
|
|
case NL80211_IFTYPE_AP:
|
|
case NL80211_IFTYPE_AP_VLAN:
|
|
case NL80211_IFTYPE_P2P_GO:
|
|
if (sdata->vif.type != NL80211_IFTYPE_ADHOC &&
|
|
!ieee80211_vif_is_mesh(&sdata->vif) &&
|
|
!rcu_access_pointer(sdata->bss->beacon))
|
|
need_offchan = true;
|
|
if (!ieee80211_is_action(mgmt->frame_control) ||
|
|
mgmt->u.action.category == WLAN_CATEGORY_PUBLIC ||
|
|
mgmt->u.action.category == WLAN_CATEGORY_SELF_PROTECTED ||
|
|
mgmt->u.action.category == WLAN_CATEGORY_SPECTRUM_MGMT)
|
|
break;
|
|
rcu_read_lock();
|
|
sta = sta_info_get_bss(sdata, mgmt->da);
|
|
rcu_read_unlock();
|
|
if (!sta)
|
|
return -ENOLINK;
|
|
break;
|
|
case NL80211_IFTYPE_STATION:
|
|
case NL80211_IFTYPE_P2P_CLIENT:
|
|
sdata_lock(sdata);
|
|
if (!sdata->u.mgd.associated ||
|
|
(params->offchan && params->wait &&
|
|
local->ops->remain_on_channel &&
|
|
memcmp(sdata->u.mgd.associated->bssid,
|
|
mgmt->bssid, ETH_ALEN)))
|
|
need_offchan = true;
|
|
sdata_unlock(sdata);
|
|
break;
|
|
case NL80211_IFTYPE_P2P_DEVICE:
|
|
need_offchan = true;
|
|
break;
|
|
case NL80211_IFTYPE_NAN:
|
|
default:
|
|
return -EOPNOTSUPP;
|
|
}
|
|
|
|
/* configurations requiring offchan cannot work if no channel has been
|
|
* specified
|
|
*/
|
|
if (need_offchan && !params->chan)
|
|
return -EINVAL;
|
|
|
|
mutex_lock(&local->mtx);
|
|
|
|
/* Check if the operating channel is the requested channel */
|
|
if (!need_offchan) {
|
|
struct ieee80211_chanctx_conf *chanctx_conf;
|
|
|
|
rcu_read_lock();
|
|
chanctx_conf = rcu_dereference(sdata->vif.chanctx_conf);
|
|
|
|
if (chanctx_conf) {
|
|
need_offchan = params->chan &&
|
|
(params->chan !=
|
|
chanctx_conf->def.chan);
|
|
} else if (!params->chan) {
|
|
ret = -EINVAL;
|
|
rcu_read_unlock();
|
|
goto out_unlock;
|
|
} else {
|
|
need_offchan = true;
|
|
}
|
|
rcu_read_unlock();
|
|
}
|
|
|
|
if (need_offchan && !params->offchan) {
|
|
ret = -EBUSY;
|
|
goto out_unlock;
|
|
}
|
|
|
|
skb = dev_alloc_skb(local->hw.extra_tx_headroom + params->len);
|
|
if (!skb) {
|
|
ret = -ENOMEM;
|
|
goto out_unlock;
|
|
}
|
|
skb_reserve(skb, local->hw.extra_tx_headroom);
|
|
|
|
data = skb_put_data(skb, params->buf, params->len);
|
|
|
|
/* Update CSA counters */
|
|
if (sdata->vif.csa_active &&
|
|
(sdata->vif.type == NL80211_IFTYPE_AP ||
|
|
sdata->vif.type == NL80211_IFTYPE_MESH_POINT ||
|
|
sdata->vif.type == NL80211_IFTYPE_ADHOC) &&
|
|
params->n_csa_offsets) {
|
|
int i;
|
|
struct beacon_data *beacon = NULL;
|
|
|
|
rcu_read_lock();
|
|
|
|
if (sdata->vif.type == NL80211_IFTYPE_AP)
|
|
beacon = rcu_dereference(sdata->u.ap.beacon);
|
|
else if (sdata->vif.type == NL80211_IFTYPE_ADHOC)
|
|
beacon = rcu_dereference(sdata->u.ibss.presp);
|
|
else if (ieee80211_vif_is_mesh(&sdata->vif))
|
|
beacon = rcu_dereference(sdata->u.mesh.beacon);
|
|
|
|
if (beacon)
|
|
for (i = 0; i < params->n_csa_offsets; i++)
|
|
data[params->csa_offsets[i]] =
|
|
beacon->cntdwn_current_counter;
|
|
|
|
rcu_read_unlock();
|
|
}
|
|
|
|
IEEE80211_SKB_CB(skb)->flags = flags;
|
|
|
|
skb->dev = sdata->dev;
|
|
|
|
if (!params->dont_wait_for_ack) {
|
|
/* make a copy to preserve the frame contents
|
|
* in case of encryption.
|
|
*/
|
|
ret = ieee80211_attach_ack_skb(local, skb, cookie, GFP_KERNEL);
|
|
if (ret) {
|
|
kfree_skb(skb);
|
|
goto out_unlock;
|
|
}
|
|
} else {
|
|
/* Assign a dummy non-zero cookie, it's not sent to
|
|
* userspace in this case but we rely on its value
|
|
* internally in the need_offchan case to distinguish
|
|
* mgmt-tx from remain-on-channel.
|
|
*/
|
|
*cookie = 0xffffffff;
|
|
}
|
|
|
|
if (!need_offchan) {
|
|
ieee80211_tx_skb(sdata, skb);
|
|
ret = 0;
|
|
goto out_unlock;
|
|
}
|
|
|
|
IEEE80211_SKB_CB(skb)->flags |= IEEE80211_TX_CTL_TX_OFFCHAN |
|
|
IEEE80211_TX_INTFL_OFFCHAN_TX_OK;
|
|
if (ieee80211_hw_check(&local->hw, QUEUE_CONTROL))
|
|
IEEE80211_SKB_CB(skb)->hw_queue =
|
|
local->hw.offchannel_tx_hw_queue;
|
|
|
|
/* This will handle all kinds of coalescing and immediate TX */
|
|
ret = ieee80211_start_roc_work(local, sdata, params->chan,
|
|
params->wait, cookie, skb,
|
|
IEEE80211_ROC_TYPE_MGMT_TX);
|
|
if (ret)
|
|
ieee80211_free_txskb(&local->hw, skb);
|
|
out_unlock:
|
|
mutex_unlock(&local->mtx);
|
|
return ret;
|
|
}
|
|
|
|
int ieee80211_mgmt_tx_cancel_wait(struct wiphy *wiphy,
|
|
struct wireless_dev *wdev, u64 cookie)
|
|
{
|
|
struct ieee80211_local *local = wiphy_priv(wiphy);
|
|
|
|
return ieee80211_cancel_roc(local, cookie, true);
|
|
}
|
|
|
|
void ieee80211_roc_setup(struct ieee80211_local *local)
|
|
{
|
|
INIT_WORK(&local->hw_roc_start, ieee80211_hw_roc_start);
|
|
INIT_WORK(&local->hw_roc_done, ieee80211_hw_roc_done);
|
|
INIT_DELAYED_WORK(&local->roc_work, ieee80211_roc_work);
|
|
INIT_LIST_HEAD(&local->roc_list);
|
|
}
|
|
|
|
void ieee80211_roc_purge(struct ieee80211_local *local,
|
|
struct ieee80211_sub_if_data *sdata)
|
|
{
|
|
struct ieee80211_roc_work *roc, *tmp;
|
|
bool work_to_do = false;
|
|
|
|
mutex_lock(&local->mtx);
|
|
list_for_each_entry_safe(roc, tmp, &local->roc_list, list) {
|
|
if (sdata && roc->sdata != sdata)
|
|
continue;
|
|
|
|
if (roc->started) {
|
|
if (local->ops->remain_on_channel) {
|
|
/* can race, so ignore return value */
|
|
drv_cancel_remain_on_channel(local, sdata);
|
|
ieee80211_roc_notify_destroy(roc);
|
|
} else {
|
|
roc->abort = true;
|
|
work_to_do = true;
|
|
}
|
|
} else {
|
|
ieee80211_roc_notify_destroy(roc);
|
|
}
|
|
}
|
|
if (work_to_do)
|
|
__ieee80211_roc_work(local);
|
|
mutex_unlock(&local->mtx);
|
|
}
|