mirror of
https://github.com/torvalds/linux
synced 2024-11-05 18:23:50 +00:00
5d52ee8110
With single-channel drivers, we need to be able to change a running chanctx if we want to use chanctx reservation. Not all drivers may be able to do this, so add a flag that indicates support for it. Changing a running chanctx can also be used as an optimization in multi-channel drivers when the context needs to be reserved for future usage. Introduce IEEE80211_CHANCTX_RESERVED chanctx mode to mark a channel as reserved so nobody else can use it (since we know it's going to change). In the future, we may allow several vifs to use the same reservation as long as they plan to use the chanctx on the same future channel. Signed-off-by: Luciano Coelho <luciano.coelho@intel.com> Signed-off-by: Johannes Berg <johannes.berg@intel.com>
939 lines
24 KiB
C
939 lines
24 KiB
C
/*
|
|
* mac80211 - channel management
|
|
*/
|
|
|
|
#include <linux/nl80211.h>
|
|
#include <linux/export.h>
|
|
#include <linux/rtnetlink.h>
|
|
#include <net/cfg80211.h>
|
|
#include "ieee80211_i.h"
|
|
#include "driver-ops.h"
|
|
|
|
static enum nl80211_chan_width ieee80211_get_sta_bw(struct ieee80211_sta *sta)
|
|
{
|
|
switch (sta->bandwidth) {
|
|
case IEEE80211_STA_RX_BW_20:
|
|
if (sta->ht_cap.ht_supported)
|
|
return NL80211_CHAN_WIDTH_20;
|
|
else
|
|
return NL80211_CHAN_WIDTH_20_NOHT;
|
|
case IEEE80211_STA_RX_BW_40:
|
|
return NL80211_CHAN_WIDTH_40;
|
|
case IEEE80211_STA_RX_BW_80:
|
|
return NL80211_CHAN_WIDTH_80;
|
|
case IEEE80211_STA_RX_BW_160:
|
|
/*
|
|
* This applied for both 160 and 80+80. since we use
|
|
* the returned value to consider degradation of
|
|
* ctx->conf.min_def, we have to make sure to take
|
|
* the bigger one (NL80211_CHAN_WIDTH_160).
|
|
* Otherwise we might try degrading even when not
|
|
* needed, as the max required sta_bw returned (80+80)
|
|
* might be smaller than the configured bw (160).
|
|
*/
|
|
return NL80211_CHAN_WIDTH_160;
|
|
default:
|
|
WARN_ON(1);
|
|
return NL80211_CHAN_WIDTH_20;
|
|
}
|
|
}
|
|
|
|
static enum nl80211_chan_width
|
|
ieee80211_get_max_required_bw(struct ieee80211_sub_if_data *sdata)
|
|
{
|
|
enum nl80211_chan_width max_bw = NL80211_CHAN_WIDTH_20_NOHT;
|
|
struct sta_info *sta;
|
|
|
|
rcu_read_lock();
|
|
list_for_each_entry_rcu(sta, &sdata->local->sta_list, list) {
|
|
if (sdata != sta->sdata &&
|
|
!(sta->sdata->bss && sta->sdata->bss == sdata->bss))
|
|
continue;
|
|
|
|
if (!sta->uploaded)
|
|
continue;
|
|
|
|
max_bw = max(max_bw, ieee80211_get_sta_bw(&sta->sta));
|
|
}
|
|
rcu_read_unlock();
|
|
|
|
return max_bw;
|
|
}
|
|
|
|
static enum nl80211_chan_width
|
|
ieee80211_get_chanctx_max_required_bw(struct ieee80211_local *local,
|
|
struct ieee80211_chanctx_conf *conf)
|
|
{
|
|
struct ieee80211_sub_if_data *sdata;
|
|
enum nl80211_chan_width max_bw = NL80211_CHAN_WIDTH_20_NOHT;
|
|
|
|
rcu_read_lock();
|
|
list_for_each_entry_rcu(sdata, &local->interfaces, list) {
|
|
struct ieee80211_vif *vif = &sdata->vif;
|
|
enum nl80211_chan_width width = NL80211_CHAN_WIDTH_20_NOHT;
|
|
|
|
if (!ieee80211_sdata_running(sdata))
|
|
continue;
|
|
|
|
if (rcu_access_pointer(sdata->vif.chanctx_conf) != conf)
|
|
continue;
|
|
|
|
switch (vif->type) {
|
|
case NL80211_IFTYPE_AP:
|
|
case NL80211_IFTYPE_AP_VLAN:
|
|
width = ieee80211_get_max_required_bw(sdata);
|
|
break;
|
|
case NL80211_IFTYPE_P2P_DEVICE:
|
|
continue;
|
|
case NL80211_IFTYPE_STATION:
|
|
case NL80211_IFTYPE_ADHOC:
|
|
case NL80211_IFTYPE_WDS:
|
|
case NL80211_IFTYPE_MESH_POINT:
|
|
width = vif->bss_conf.chandef.width;
|
|
break;
|
|
case NL80211_IFTYPE_UNSPECIFIED:
|
|
case NUM_NL80211_IFTYPES:
|
|
case NL80211_IFTYPE_MONITOR:
|
|
case NL80211_IFTYPE_P2P_CLIENT:
|
|
case NL80211_IFTYPE_P2P_GO:
|
|
WARN_ON_ONCE(1);
|
|
}
|
|
max_bw = max(max_bw, width);
|
|
}
|
|
|
|
/* use the configured bandwidth in case of monitor interface */
|
|
sdata = rcu_dereference(local->monitor_sdata);
|
|
if (sdata && rcu_access_pointer(sdata->vif.chanctx_conf) == conf)
|
|
max_bw = max(max_bw, conf->def.width);
|
|
|
|
rcu_read_unlock();
|
|
|
|
return max_bw;
|
|
}
|
|
|
|
/*
|
|
* recalc the min required chan width of the channel context, which is
|
|
* the max of min required widths of all the interfaces bound to this
|
|
* channel context.
|
|
*/
|
|
void ieee80211_recalc_chanctx_min_def(struct ieee80211_local *local,
|
|
struct ieee80211_chanctx *ctx)
|
|
{
|
|
enum nl80211_chan_width max_bw;
|
|
struct cfg80211_chan_def min_def;
|
|
|
|
lockdep_assert_held(&local->chanctx_mtx);
|
|
|
|
/* don't optimize 5MHz, 10MHz, and radar_enabled confs */
|
|
if (ctx->conf.def.width == NL80211_CHAN_WIDTH_5 ||
|
|
ctx->conf.def.width == NL80211_CHAN_WIDTH_10 ||
|
|
ctx->conf.radar_enabled) {
|
|
ctx->conf.min_def = ctx->conf.def;
|
|
return;
|
|
}
|
|
|
|
max_bw = ieee80211_get_chanctx_max_required_bw(local, &ctx->conf);
|
|
|
|
/* downgrade chandef up to max_bw */
|
|
min_def = ctx->conf.def;
|
|
while (min_def.width > max_bw)
|
|
ieee80211_chandef_downgrade(&min_def);
|
|
|
|
if (cfg80211_chandef_identical(&ctx->conf.min_def, &min_def))
|
|
return;
|
|
|
|
ctx->conf.min_def = min_def;
|
|
if (!ctx->driver_present)
|
|
return;
|
|
|
|
drv_change_chanctx(local, ctx, IEEE80211_CHANCTX_CHANGE_MIN_WIDTH);
|
|
}
|
|
|
|
static void ieee80211_change_chanctx(struct ieee80211_local *local,
|
|
struct ieee80211_chanctx *ctx,
|
|
const struct cfg80211_chan_def *chandef)
|
|
{
|
|
if (cfg80211_chandef_identical(&ctx->conf.def, chandef))
|
|
return;
|
|
|
|
WARN_ON(!cfg80211_chandef_compatible(&ctx->conf.def, chandef));
|
|
|
|
ctx->conf.def = *chandef;
|
|
drv_change_chanctx(local, ctx, IEEE80211_CHANCTX_CHANGE_WIDTH);
|
|
ieee80211_recalc_chanctx_min_def(local, ctx);
|
|
|
|
if (!local->use_chanctx) {
|
|
local->_oper_chandef = *chandef;
|
|
ieee80211_hw_config(local, 0);
|
|
}
|
|
}
|
|
|
|
static bool ieee80211_chanctx_is_reserved(struct ieee80211_local *local,
|
|
struct ieee80211_chanctx *ctx)
|
|
{
|
|
struct ieee80211_sub_if_data *sdata;
|
|
bool ret = false;
|
|
|
|
lockdep_assert_held(&local->chanctx_mtx);
|
|
rcu_read_lock();
|
|
list_for_each_entry_rcu(sdata, &local->interfaces, list) {
|
|
if (!ieee80211_sdata_running(sdata))
|
|
continue;
|
|
if (sdata->reserved_chanctx == ctx) {
|
|
ret = true;
|
|
break;
|
|
}
|
|
}
|
|
|
|
rcu_read_unlock();
|
|
return ret;
|
|
}
|
|
|
|
static struct ieee80211_chanctx *
|
|
ieee80211_find_chanctx(struct ieee80211_local *local,
|
|
const struct cfg80211_chan_def *chandef,
|
|
enum ieee80211_chanctx_mode mode)
|
|
{
|
|
struct ieee80211_chanctx *ctx;
|
|
|
|
lockdep_assert_held(&local->chanctx_mtx);
|
|
|
|
if (mode == IEEE80211_CHANCTX_EXCLUSIVE)
|
|
return NULL;
|
|
|
|
list_for_each_entry(ctx, &local->chanctx_list, list) {
|
|
const struct cfg80211_chan_def *compat;
|
|
|
|
/* We don't support chanctx reservation for multiple
|
|
* vifs yet, so don't allow reserved chanctxs to be
|
|
* reused.
|
|
*/
|
|
if ((ctx->mode == IEEE80211_CHANCTX_EXCLUSIVE) ||
|
|
ieee80211_chanctx_is_reserved(local, ctx))
|
|
continue;
|
|
|
|
compat = cfg80211_chandef_compatible(&ctx->conf.def, chandef);
|
|
if (!compat)
|
|
continue;
|
|
|
|
ieee80211_change_chanctx(local, ctx, compat);
|
|
|
|
return ctx;
|
|
}
|
|
|
|
return NULL;
|
|
}
|
|
|
|
static bool ieee80211_is_radar_required(struct ieee80211_local *local)
|
|
{
|
|
struct ieee80211_sub_if_data *sdata;
|
|
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
rcu_read_lock();
|
|
list_for_each_entry_rcu(sdata, &local->interfaces, list) {
|
|
if (sdata->radar_required) {
|
|
rcu_read_unlock();
|
|
return true;
|
|
}
|
|
}
|
|
rcu_read_unlock();
|
|
|
|
return false;
|
|
}
|
|
|
|
static struct ieee80211_chanctx *
|
|
ieee80211_new_chanctx(struct ieee80211_local *local,
|
|
const struct cfg80211_chan_def *chandef,
|
|
enum ieee80211_chanctx_mode mode)
|
|
{
|
|
struct ieee80211_chanctx *ctx;
|
|
u32 changed;
|
|
int err;
|
|
|
|
lockdep_assert_held(&local->chanctx_mtx);
|
|
|
|
ctx = kzalloc(sizeof(*ctx) + local->hw.chanctx_data_size, GFP_KERNEL);
|
|
if (!ctx)
|
|
return ERR_PTR(-ENOMEM);
|
|
|
|
ctx->conf.def = *chandef;
|
|
ctx->conf.rx_chains_static = 1;
|
|
ctx->conf.rx_chains_dynamic = 1;
|
|
ctx->mode = mode;
|
|
ctx->conf.radar_enabled = ieee80211_is_radar_required(local);
|
|
ieee80211_recalc_chanctx_min_def(local, ctx);
|
|
if (!local->use_chanctx)
|
|
local->hw.conf.radar_enabled = ctx->conf.radar_enabled;
|
|
|
|
/* we hold the mutex to prevent idle from changing */
|
|
lockdep_assert_held(&local->mtx);
|
|
/* turn idle off *before* setting channel -- some drivers need that */
|
|
changed = ieee80211_idle_off(local);
|
|
if (changed)
|
|
ieee80211_hw_config(local, changed);
|
|
|
|
if (!local->use_chanctx) {
|
|
local->_oper_chandef = *chandef;
|
|
ieee80211_hw_config(local, 0);
|
|
} else {
|
|
err = drv_add_chanctx(local, ctx);
|
|
if (err) {
|
|
kfree(ctx);
|
|
ieee80211_recalc_idle(local);
|
|
return ERR_PTR(err);
|
|
}
|
|
}
|
|
|
|
/* and keep the mutex held until the new chanctx is on the list */
|
|
list_add_rcu(&ctx->list, &local->chanctx_list);
|
|
|
|
return ctx;
|
|
}
|
|
|
|
static void ieee80211_free_chanctx(struct ieee80211_local *local,
|
|
struct ieee80211_chanctx *ctx)
|
|
{
|
|
bool check_single_channel = false;
|
|
lockdep_assert_held(&local->chanctx_mtx);
|
|
|
|
WARN_ON_ONCE(ctx->refcount != 0);
|
|
|
|
if (!local->use_chanctx) {
|
|
struct cfg80211_chan_def *chandef = &local->_oper_chandef;
|
|
chandef->width = NL80211_CHAN_WIDTH_20_NOHT;
|
|
chandef->center_freq1 = chandef->chan->center_freq;
|
|
chandef->center_freq2 = 0;
|
|
|
|
/* NOTE: Disabling radar is only valid here for
|
|
* single channel context. To be sure, check it ...
|
|
*/
|
|
if (local->hw.conf.radar_enabled)
|
|
check_single_channel = true;
|
|
local->hw.conf.radar_enabled = false;
|
|
|
|
ieee80211_hw_config(local, 0);
|
|
} else {
|
|
drv_remove_chanctx(local, ctx);
|
|
}
|
|
|
|
list_del_rcu(&ctx->list);
|
|
kfree_rcu(ctx, rcu_head);
|
|
|
|
/* throw a warning if this wasn't the only channel context. */
|
|
WARN_ON(check_single_channel && !list_empty(&local->chanctx_list));
|
|
|
|
ieee80211_recalc_idle(local);
|
|
}
|
|
|
|
static void ieee80211_recalc_chanctx_chantype(struct ieee80211_local *local,
|
|
struct ieee80211_chanctx *ctx)
|
|
{
|
|
struct ieee80211_chanctx_conf *conf = &ctx->conf;
|
|
struct ieee80211_sub_if_data *sdata;
|
|
const struct cfg80211_chan_def *compat = NULL;
|
|
|
|
lockdep_assert_held(&local->chanctx_mtx);
|
|
|
|
rcu_read_lock();
|
|
list_for_each_entry_rcu(sdata, &local->interfaces, list) {
|
|
|
|
if (!ieee80211_sdata_running(sdata))
|
|
continue;
|
|
if (rcu_access_pointer(sdata->vif.chanctx_conf) != conf)
|
|
continue;
|
|
|
|
if (!compat)
|
|
compat = &sdata->vif.bss_conf.chandef;
|
|
|
|
compat = cfg80211_chandef_compatible(
|
|
&sdata->vif.bss_conf.chandef, compat);
|
|
if (!compat)
|
|
break;
|
|
}
|
|
rcu_read_unlock();
|
|
|
|
if (WARN_ON_ONCE(!compat))
|
|
return;
|
|
|
|
ieee80211_change_chanctx(local, ctx, compat);
|
|
}
|
|
|
|
static void ieee80211_recalc_radar_chanctx(struct ieee80211_local *local,
|
|
struct ieee80211_chanctx *chanctx)
|
|
{
|
|
bool radar_enabled;
|
|
|
|
lockdep_assert_held(&local->chanctx_mtx);
|
|
/* for setting local->radar_detect_enabled */
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
radar_enabled = ieee80211_is_radar_required(local);
|
|
|
|
if (radar_enabled == chanctx->conf.radar_enabled)
|
|
return;
|
|
|
|
chanctx->conf.radar_enabled = radar_enabled;
|
|
local->radar_detect_enabled = chanctx->conf.radar_enabled;
|
|
|
|
if (!local->use_chanctx) {
|
|
local->hw.conf.radar_enabled = chanctx->conf.radar_enabled;
|
|
ieee80211_hw_config(local, IEEE80211_CONF_CHANGE_CHANNEL);
|
|
}
|
|
|
|
drv_change_chanctx(local, chanctx, IEEE80211_CHANCTX_CHANGE_RADAR);
|
|
}
|
|
|
|
static int ieee80211_assign_vif_chanctx(struct ieee80211_sub_if_data *sdata,
|
|
struct ieee80211_chanctx *new_ctx)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_chanctx_conf *conf;
|
|
struct ieee80211_chanctx *curr_ctx = NULL;
|
|
int ret = 0;
|
|
|
|
conf = rcu_dereference_protected(sdata->vif.chanctx_conf,
|
|
lockdep_is_held(&local->chanctx_mtx));
|
|
|
|
if (conf) {
|
|
curr_ctx = container_of(conf, struct ieee80211_chanctx, conf);
|
|
|
|
curr_ctx->refcount--;
|
|
drv_unassign_vif_chanctx(local, sdata, curr_ctx);
|
|
conf = NULL;
|
|
}
|
|
|
|
if (new_ctx) {
|
|
ret = drv_assign_vif_chanctx(local, sdata, new_ctx);
|
|
if (ret)
|
|
goto out;
|
|
|
|
new_ctx->refcount++;
|
|
conf = &new_ctx->conf;
|
|
}
|
|
|
|
out:
|
|
rcu_assign_pointer(sdata->vif.chanctx_conf, conf);
|
|
|
|
sdata->vif.bss_conf.idle = !conf;
|
|
|
|
if (curr_ctx && curr_ctx->refcount > 0) {
|
|
ieee80211_recalc_chanctx_chantype(local, curr_ctx);
|
|
ieee80211_recalc_smps_chanctx(local, curr_ctx);
|
|
ieee80211_recalc_radar_chanctx(local, curr_ctx);
|
|
ieee80211_recalc_chanctx_min_def(local, curr_ctx);
|
|
}
|
|
|
|
if (new_ctx && new_ctx->refcount > 0) {
|
|
ieee80211_recalc_txpower(sdata);
|
|
ieee80211_recalc_chanctx_min_def(local, new_ctx);
|
|
}
|
|
|
|
if (sdata->vif.type != NL80211_IFTYPE_P2P_DEVICE &&
|
|
sdata->vif.type != NL80211_IFTYPE_MONITOR)
|
|
ieee80211_bss_info_change_notify(sdata,
|
|
BSS_CHANGED_IDLE);
|
|
|
|
return ret;
|
|
}
|
|
|
|
static void __ieee80211_vif_release_channel(struct ieee80211_sub_if_data *sdata)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_chanctx_conf *conf;
|
|
struct ieee80211_chanctx *ctx;
|
|
|
|
lockdep_assert_held(&local->chanctx_mtx);
|
|
|
|
conf = rcu_dereference_protected(sdata->vif.chanctx_conf,
|
|
lockdep_is_held(&local->chanctx_mtx));
|
|
if (!conf)
|
|
return;
|
|
|
|
ctx = container_of(conf, struct ieee80211_chanctx, conf);
|
|
|
|
if (sdata->reserved_chanctx)
|
|
ieee80211_vif_unreserve_chanctx(sdata);
|
|
|
|
ieee80211_assign_vif_chanctx(sdata, NULL);
|
|
if (ctx->refcount == 0)
|
|
ieee80211_free_chanctx(local, ctx);
|
|
}
|
|
|
|
void ieee80211_recalc_smps_chanctx(struct ieee80211_local *local,
|
|
struct ieee80211_chanctx *chanctx)
|
|
{
|
|
struct ieee80211_sub_if_data *sdata;
|
|
u8 rx_chains_static, rx_chains_dynamic;
|
|
|
|
lockdep_assert_held(&local->chanctx_mtx);
|
|
|
|
rx_chains_static = 1;
|
|
rx_chains_dynamic = 1;
|
|
|
|
rcu_read_lock();
|
|
list_for_each_entry_rcu(sdata, &local->interfaces, list) {
|
|
u8 needed_static, needed_dynamic;
|
|
|
|
if (!ieee80211_sdata_running(sdata))
|
|
continue;
|
|
|
|
if (rcu_access_pointer(sdata->vif.chanctx_conf) !=
|
|
&chanctx->conf)
|
|
continue;
|
|
|
|
switch (sdata->vif.type) {
|
|
case NL80211_IFTYPE_P2P_DEVICE:
|
|
continue;
|
|
case NL80211_IFTYPE_STATION:
|
|
if (!sdata->u.mgd.associated)
|
|
continue;
|
|
break;
|
|
case NL80211_IFTYPE_AP_VLAN:
|
|
continue;
|
|
case NL80211_IFTYPE_AP:
|
|
case NL80211_IFTYPE_ADHOC:
|
|
case NL80211_IFTYPE_WDS:
|
|
case NL80211_IFTYPE_MESH_POINT:
|
|
break;
|
|
default:
|
|
WARN_ON_ONCE(1);
|
|
}
|
|
|
|
switch (sdata->smps_mode) {
|
|
default:
|
|
WARN_ONCE(1, "Invalid SMPS mode %d\n",
|
|
sdata->smps_mode);
|
|
/* fall through */
|
|
case IEEE80211_SMPS_OFF:
|
|
needed_static = sdata->needed_rx_chains;
|
|
needed_dynamic = sdata->needed_rx_chains;
|
|
break;
|
|
case IEEE80211_SMPS_DYNAMIC:
|
|
needed_static = 1;
|
|
needed_dynamic = sdata->needed_rx_chains;
|
|
break;
|
|
case IEEE80211_SMPS_STATIC:
|
|
needed_static = 1;
|
|
needed_dynamic = 1;
|
|
break;
|
|
}
|
|
|
|
rx_chains_static = max(rx_chains_static, needed_static);
|
|
rx_chains_dynamic = max(rx_chains_dynamic, needed_dynamic);
|
|
}
|
|
rcu_read_unlock();
|
|
|
|
if (!local->use_chanctx) {
|
|
if (rx_chains_static > 1)
|
|
local->smps_mode = IEEE80211_SMPS_OFF;
|
|
else if (rx_chains_dynamic > 1)
|
|
local->smps_mode = IEEE80211_SMPS_DYNAMIC;
|
|
else
|
|
local->smps_mode = IEEE80211_SMPS_STATIC;
|
|
ieee80211_hw_config(local, 0);
|
|
}
|
|
|
|
if (rx_chains_static == chanctx->conf.rx_chains_static &&
|
|
rx_chains_dynamic == chanctx->conf.rx_chains_dynamic)
|
|
return;
|
|
|
|
chanctx->conf.rx_chains_static = rx_chains_static;
|
|
chanctx->conf.rx_chains_dynamic = rx_chains_dynamic;
|
|
drv_change_chanctx(local, chanctx, IEEE80211_CHANCTX_CHANGE_RX_CHAINS);
|
|
}
|
|
|
|
int ieee80211_vif_use_channel(struct ieee80211_sub_if_data *sdata,
|
|
const struct cfg80211_chan_def *chandef,
|
|
enum ieee80211_chanctx_mode mode)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_chanctx *ctx;
|
|
u8 radar_detect_width = 0;
|
|
int ret;
|
|
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
WARN_ON(sdata->dev && netif_carrier_ok(sdata->dev));
|
|
|
|
mutex_lock(&local->chanctx_mtx);
|
|
|
|
ret = cfg80211_chandef_dfs_required(local->hw.wiphy,
|
|
chandef,
|
|
sdata->wdev.iftype);
|
|
if (ret < 0)
|
|
goto out;
|
|
if (ret > 0)
|
|
radar_detect_width = BIT(chandef->width);
|
|
|
|
sdata->radar_required = ret;
|
|
|
|
ret = ieee80211_check_combinations(sdata, chandef, mode,
|
|
radar_detect_width);
|
|
if (ret < 0)
|
|
goto out;
|
|
|
|
__ieee80211_vif_release_channel(sdata);
|
|
|
|
ctx = ieee80211_find_chanctx(local, chandef, mode);
|
|
if (!ctx)
|
|
ctx = ieee80211_new_chanctx(local, chandef, mode);
|
|
if (IS_ERR(ctx)) {
|
|
ret = PTR_ERR(ctx);
|
|
goto out;
|
|
}
|
|
|
|
sdata->vif.bss_conf.chandef = *chandef;
|
|
|
|
ret = ieee80211_assign_vif_chanctx(sdata, ctx);
|
|
if (ret) {
|
|
/* if assign fails refcount stays the same */
|
|
if (ctx->refcount == 0)
|
|
ieee80211_free_chanctx(local, ctx);
|
|
goto out;
|
|
}
|
|
|
|
ieee80211_recalc_smps_chanctx(local, ctx);
|
|
ieee80211_recalc_radar_chanctx(local, ctx);
|
|
out:
|
|
mutex_unlock(&local->chanctx_mtx);
|
|
return ret;
|
|
}
|
|
|
|
static int __ieee80211_vif_change_channel(struct ieee80211_sub_if_data *sdata,
|
|
struct ieee80211_chanctx *ctx,
|
|
u32 *changed)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
const struct cfg80211_chan_def *chandef = &sdata->csa_chandef;
|
|
u32 chanctx_changed = 0;
|
|
|
|
if (!cfg80211_chandef_usable(sdata->local->hw.wiphy, chandef,
|
|
IEEE80211_CHAN_DISABLED))
|
|
return -EINVAL;
|
|
|
|
if (ctx->refcount != 1)
|
|
return -EINVAL;
|
|
|
|
if (sdata->vif.bss_conf.chandef.width != chandef->width) {
|
|
chanctx_changed = IEEE80211_CHANCTX_CHANGE_WIDTH;
|
|
*changed |= BSS_CHANGED_BANDWIDTH;
|
|
}
|
|
|
|
sdata->vif.bss_conf.chandef = *chandef;
|
|
ctx->conf.def = *chandef;
|
|
|
|
chanctx_changed |= IEEE80211_CHANCTX_CHANGE_CHANNEL;
|
|
drv_change_chanctx(local, ctx, chanctx_changed);
|
|
|
|
ieee80211_recalc_chanctx_chantype(local, ctx);
|
|
ieee80211_recalc_smps_chanctx(local, ctx);
|
|
ieee80211_recalc_radar_chanctx(local, ctx);
|
|
ieee80211_recalc_chanctx_min_def(local, ctx);
|
|
|
|
return 0;
|
|
}
|
|
|
|
int ieee80211_vif_change_channel(struct ieee80211_sub_if_data *sdata,
|
|
u32 *changed)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_chanctx_conf *conf;
|
|
struct ieee80211_chanctx *ctx;
|
|
int ret;
|
|
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
/* should never be called if not performing a channel switch. */
|
|
if (WARN_ON(!sdata->vif.csa_active))
|
|
return -EINVAL;
|
|
|
|
mutex_lock(&local->chanctx_mtx);
|
|
conf = rcu_dereference_protected(sdata->vif.chanctx_conf,
|
|
lockdep_is_held(&local->chanctx_mtx));
|
|
if (!conf) {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
ctx = container_of(conf, struct ieee80211_chanctx, conf);
|
|
|
|
ret = __ieee80211_vif_change_channel(sdata, ctx, changed);
|
|
out:
|
|
mutex_unlock(&local->chanctx_mtx);
|
|
return ret;
|
|
}
|
|
|
|
static void
|
|
__ieee80211_vif_copy_chanctx_to_vlans(struct ieee80211_sub_if_data *sdata,
|
|
bool clear)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_sub_if_data *vlan;
|
|
struct ieee80211_chanctx_conf *conf;
|
|
|
|
if (WARN_ON(sdata->vif.type != NL80211_IFTYPE_AP))
|
|
return;
|
|
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
/* Check that conf exists, even when clearing this function
|
|
* must be called with the AP's channel context still there
|
|
* as it would otherwise cause VLANs to have an invalid
|
|
* channel context pointer for a while, possibly pointing
|
|
* to a channel context that has already been freed.
|
|
*/
|
|
conf = rcu_dereference_protected(sdata->vif.chanctx_conf,
|
|
lockdep_is_held(&local->chanctx_mtx));
|
|
WARN_ON(!conf);
|
|
|
|
if (clear)
|
|
conf = NULL;
|
|
|
|
list_for_each_entry(vlan, &sdata->u.ap.vlans, u.vlan.list)
|
|
rcu_assign_pointer(vlan->vif.chanctx_conf, conf);
|
|
}
|
|
|
|
void ieee80211_vif_copy_chanctx_to_vlans(struct ieee80211_sub_if_data *sdata,
|
|
bool clear)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
|
|
mutex_lock(&local->chanctx_mtx);
|
|
|
|
__ieee80211_vif_copy_chanctx_to_vlans(sdata, clear);
|
|
|
|
mutex_unlock(&local->chanctx_mtx);
|
|
}
|
|
|
|
int ieee80211_vif_unreserve_chanctx(struct ieee80211_sub_if_data *sdata)
|
|
{
|
|
lockdep_assert_held(&sdata->local->chanctx_mtx);
|
|
|
|
if (WARN_ON(!sdata->reserved_chanctx))
|
|
return -EINVAL;
|
|
|
|
if (--sdata->reserved_chanctx->refcount == 0)
|
|
ieee80211_free_chanctx(sdata->local, sdata->reserved_chanctx);
|
|
|
|
sdata->reserved_chanctx = NULL;
|
|
|
|
return 0;
|
|
}
|
|
|
|
int ieee80211_vif_reserve_chanctx(struct ieee80211_sub_if_data *sdata,
|
|
const struct cfg80211_chan_def *chandef,
|
|
enum ieee80211_chanctx_mode mode)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_chanctx_conf *conf;
|
|
struct ieee80211_chanctx *new_ctx, *curr_ctx;
|
|
int ret = 0;
|
|
|
|
mutex_lock(&local->chanctx_mtx);
|
|
|
|
conf = rcu_dereference_protected(sdata->vif.chanctx_conf,
|
|
lockdep_is_held(&local->chanctx_mtx));
|
|
if (!conf) {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
curr_ctx = container_of(conf, struct ieee80211_chanctx, conf);
|
|
|
|
/* try to find another context with the chandef we want */
|
|
new_ctx = ieee80211_find_chanctx(local, chandef, mode);
|
|
if (!new_ctx) {
|
|
if (curr_ctx->refcount == 1 &&
|
|
(local->hw.flags & IEEE80211_HW_CHANGE_RUNNING_CHANCTX)) {
|
|
/* if we're the only users of the chanctx and
|
|
* the driver supports changing a running
|
|
* context, reserve our current context
|
|
*/
|
|
new_ctx = curr_ctx;
|
|
} else {
|
|
/* create a new context and reserve it */
|
|
new_ctx = ieee80211_new_chanctx(local, chandef, mode);
|
|
if (IS_ERR(new_ctx)) {
|
|
ret = PTR_ERR(new_ctx);
|
|
goto out;
|
|
}
|
|
}
|
|
}
|
|
|
|
new_ctx->refcount++;
|
|
sdata->reserved_chanctx = new_ctx;
|
|
sdata->reserved_chandef = *chandef;
|
|
out:
|
|
mutex_unlock(&local->chanctx_mtx);
|
|
return ret;
|
|
}
|
|
|
|
int ieee80211_vif_use_reserved_context(struct ieee80211_sub_if_data *sdata,
|
|
u32 *changed)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_chanctx *ctx;
|
|
struct ieee80211_chanctx *old_ctx;
|
|
struct ieee80211_chanctx_conf *conf;
|
|
int ret;
|
|
u32 tmp_changed = *changed;
|
|
|
|
/* TODO: need to recheck if the chandef is usable etc.? */
|
|
|
|
lockdep_assert_held(&local->mtx);
|
|
|
|
mutex_lock(&local->chanctx_mtx);
|
|
|
|
ctx = sdata->reserved_chanctx;
|
|
if (WARN_ON(!ctx)) {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
conf = rcu_dereference_protected(sdata->vif.chanctx_conf,
|
|
lockdep_is_held(&local->chanctx_mtx));
|
|
if (!conf) {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
old_ctx = container_of(conf, struct ieee80211_chanctx, conf);
|
|
|
|
if (sdata->vif.bss_conf.chandef.width != sdata->reserved_chandef.width)
|
|
tmp_changed |= BSS_CHANGED_BANDWIDTH;
|
|
|
|
sdata->vif.bss_conf.chandef = sdata->reserved_chandef;
|
|
|
|
/* unref our reservation */
|
|
ctx->refcount--;
|
|
sdata->reserved_chanctx = NULL;
|
|
|
|
if (old_ctx == ctx) {
|
|
/* This is our own context, just change it */
|
|
ret = __ieee80211_vif_change_channel(sdata, old_ctx,
|
|
&tmp_changed);
|
|
if (ret)
|
|
goto out;
|
|
} else {
|
|
ret = ieee80211_assign_vif_chanctx(sdata, ctx);
|
|
if (old_ctx->refcount == 0)
|
|
ieee80211_free_chanctx(local, old_ctx);
|
|
if (ret) {
|
|
/* if assign fails refcount stays the same */
|
|
if (ctx->refcount == 0)
|
|
ieee80211_free_chanctx(local, ctx);
|
|
goto out;
|
|
}
|
|
|
|
if (sdata->vif.type == NL80211_IFTYPE_AP)
|
|
__ieee80211_vif_copy_chanctx_to_vlans(sdata, false);
|
|
}
|
|
|
|
*changed = tmp_changed;
|
|
|
|
ieee80211_recalc_chanctx_chantype(local, ctx);
|
|
ieee80211_recalc_smps_chanctx(local, ctx);
|
|
ieee80211_recalc_radar_chanctx(local, ctx);
|
|
ieee80211_recalc_chanctx_min_def(local, ctx);
|
|
out:
|
|
mutex_unlock(&local->chanctx_mtx);
|
|
return ret;
|
|
}
|
|
|
|
int ieee80211_vif_change_bandwidth(struct ieee80211_sub_if_data *sdata,
|
|
const struct cfg80211_chan_def *chandef,
|
|
u32 *changed)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_chanctx_conf *conf;
|
|
struct ieee80211_chanctx *ctx;
|
|
int ret;
|
|
|
|
if (!cfg80211_chandef_usable(sdata->local->hw.wiphy, chandef,
|
|
IEEE80211_CHAN_DISABLED))
|
|
return -EINVAL;
|
|
|
|
mutex_lock(&local->chanctx_mtx);
|
|
if (cfg80211_chandef_identical(chandef, &sdata->vif.bss_conf.chandef)) {
|
|
ret = 0;
|
|
goto out;
|
|
}
|
|
|
|
if (chandef->width == NL80211_CHAN_WIDTH_20_NOHT ||
|
|
sdata->vif.bss_conf.chandef.width == NL80211_CHAN_WIDTH_20_NOHT) {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
conf = rcu_dereference_protected(sdata->vif.chanctx_conf,
|
|
lockdep_is_held(&local->chanctx_mtx));
|
|
if (!conf) {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
ctx = container_of(conf, struct ieee80211_chanctx, conf);
|
|
if (!cfg80211_chandef_compatible(&conf->def, chandef)) {
|
|
ret = -EINVAL;
|
|
goto out;
|
|
}
|
|
|
|
sdata->vif.bss_conf.chandef = *chandef;
|
|
|
|
ieee80211_recalc_chanctx_chantype(local, ctx);
|
|
|
|
*changed |= BSS_CHANGED_BANDWIDTH;
|
|
ret = 0;
|
|
out:
|
|
mutex_unlock(&local->chanctx_mtx);
|
|
return ret;
|
|
}
|
|
|
|
void ieee80211_vif_release_channel(struct ieee80211_sub_if_data *sdata)
|
|
{
|
|
WARN_ON(sdata->dev && netif_carrier_ok(sdata->dev));
|
|
|
|
lockdep_assert_held(&sdata->local->mtx);
|
|
|
|
mutex_lock(&sdata->local->chanctx_mtx);
|
|
__ieee80211_vif_release_channel(sdata);
|
|
mutex_unlock(&sdata->local->chanctx_mtx);
|
|
}
|
|
|
|
void ieee80211_vif_vlan_copy_chanctx(struct ieee80211_sub_if_data *sdata)
|
|
{
|
|
struct ieee80211_local *local = sdata->local;
|
|
struct ieee80211_sub_if_data *ap;
|
|
struct ieee80211_chanctx_conf *conf;
|
|
|
|
if (WARN_ON(sdata->vif.type != NL80211_IFTYPE_AP_VLAN || !sdata->bss))
|
|
return;
|
|
|
|
ap = container_of(sdata->bss, struct ieee80211_sub_if_data, u.ap);
|
|
|
|
mutex_lock(&local->chanctx_mtx);
|
|
|
|
conf = rcu_dereference_protected(ap->vif.chanctx_conf,
|
|
lockdep_is_held(&local->chanctx_mtx));
|
|
rcu_assign_pointer(sdata->vif.chanctx_conf, conf);
|
|
mutex_unlock(&local->chanctx_mtx);
|
|
}
|
|
|
|
void ieee80211_iter_chan_contexts_atomic(
|
|
struct ieee80211_hw *hw,
|
|
void (*iter)(struct ieee80211_hw *hw,
|
|
struct ieee80211_chanctx_conf *chanctx_conf,
|
|
void *data),
|
|
void *iter_data)
|
|
{
|
|
struct ieee80211_local *local = hw_to_local(hw);
|
|
struct ieee80211_chanctx *ctx;
|
|
|
|
rcu_read_lock();
|
|
list_for_each_entry_rcu(ctx, &local->chanctx_list, list)
|
|
if (ctx->driver_present)
|
|
iter(hw, &ctx->conf, iter_data);
|
|
rcu_read_unlock();
|
|
}
|
|
EXPORT_SYMBOL_GPL(ieee80211_iter_chan_contexts_atomic);
|