void (*sta_ps)(struct mt76_dev *dev, struct ieee80211_sta *sta,
bool ps);
+ s8 (*get_tx_txpwr_adj)(struct mt76_dev *dev, s8 txpwr,
+ s8 max_txpwr_adj);
};
struct mt76_channel_state {
}
spin_unlock_bh(&dev->mt76.lock);
- txpwr_adj = mt76x2_tx_get_txpwr_adj(dev, dev->mt76.txpower_conf,
+ txpwr_adj = mt76x2_tx_get_txpwr_adj(&dev->mt76, dev->mt76.txpower_conf,
max_txpwr_adj);
txwi->ctl2 = FIELD_PREP(MT_TX_PWR_ADJ, txpwr_adj);
void mt76x2_update_channel(struct mt76_dev *mdev);
-s8 mt76x2_tx_get_txpwr_adj(struct mt76x2_dev *dev, s8 txpwr, s8 max_txpwr_adj);
+s8 mt76x2_tx_get_txpwr_adj(struct mt76_dev *mdev, s8 txpwr, s8 max_txpwr_adj);
void mt76x2_tx_set_txpwr_auto(struct mt76x2_dev *dev, s8 txpwr);
.rx_skb = mt76x2_queue_rx_skb,
.rx_poll_complete = mt76x2_rx_poll_complete,
.sta_ps = mt76x2_sta_ps,
+ .get_tx_txpwr_adj = mt76x2_tx_get_txpwr_adj,
};
struct mt76x2_dev *dev;
struct mt76_dev *mdev;
}
EXPORT_SYMBOL_GPL(mt76x2_tx);
-s8 mt76x2_tx_get_txpwr_adj(struct mt76x2_dev *dev, s8 txpwr, s8 max_txpwr_adj)
+s8 mt76x2_tx_get_txpwr_adj(struct mt76_dev *mdev, s8 txpwr, s8 max_txpwr_adj)
{
+ struct mt76x2_dev *dev = container_of(mdev, struct mt76x2_dev, mt76);
+
txpwr = min_t(s8, txpwr, dev->mt76.txpower_conf);
txpwr -= (dev->target_power + dev->target_power_delta[0]);
txpwr = min_t(s8, txpwr, max_txpwr_adj);
{
s8 txpwr_adj;
- txpwr_adj = mt76x2_tx_get_txpwr_adj(dev, txpwr,
+ txpwr_adj = mt76x2_tx_get_txpwr_adj(&dev->mt76, txpwr,
dev->mt76.rate_power.ofdm[4]);
mt76_rmw_field(dev, MT_PROT_AUTO_TX_CFG,
MT_PROT_AUTO_TX_CFG_PROT_PADJ, txpwr_adj);