2 * Copyright (C) 2016 Felix Fietkau <nbd@nbd.name>
3 * Copyright (C) 2018 Stanislaw Gruszka <stf_xl@wp.pl>
5 * Permission to use, copy, modify, and/or distribute this software for any
6 * purpose with or without fee is hereby granted, provided that the above
7 * copyright notice and this permission notice appear in all copies.
9 * THE SOFTWARE IS PROVIDED "AS IS" AND THE AUTHOR DISCLAIMS ALL WARRANTIES
10 * WITH REGARD TO THIS SOFTWARE INCLUDING ALL IMPLIED WARRANTIES OF
11 * MERCHANTABILITY AND FITNESS. IN NO EVENT SHALL THE AUTHOR BE LIABLE FOR
12 * ANY SPECIAL, DIRECT, INDIRECT, OR CONSEQUENTIAL DAMAGES OR ANY DAMAGES
13 * WHATSOEVER RESULTING FROM LOSS OF USE, DATA OR PROFITS, WHETHER IN AN
14 * ACTION OF CONTRACT, NEGLIGENCE OR OTHER TORTIOUS ACTION, ARISING OUT OF
15 * OR IN CONNECTION WITH THE USE OR PERFORMANCE OF THIS SOFTWARE.
18 #ifndef __MT76X02_UTIL_H
19 #define __MT76X02_UTIL_H
21 #include <linux/kfifo.h>
24 #include "mt76x02_regs.h"
25 #include "mt76x02_mac.h"
26 #include "mt76x02_dfs.h"
27 #include "mt76x02_dma.h"
29 struct mt76x02_mac_stats {
37 #define MT_MAX_CHAINS 2
38 struct mt76x02_rx_freq_cal {
39 s8 high_gain[MT_MAX_CHAINS];
40 s8 rssi_offset[MT_MAX_CHAINS];
47 struct mt76x02_calibration {
48 struct mt76x02_rx_freq_cal rx;
50 u8 agc_gain_init[MT_MAX_CHAINS];
51 u8 agc_gain_cur[MT_MAX_CHAINS];
63 bool tssi_comp_pending;
65 bool channel_cal_done;
69 struct mt76_dev mt76; /* must be first */
71 struct mac_address macaddr_list[8];
73 struct mutex phy_mutex;
76 DECLARE_KFIFO_PTR(txstatus_fifo, struct mt76x02_tx_status);
78 struct sk_buff *rx_head;
80 struct tasklet_struct tx_tasklet;
81 struct tasklet_struct pre_tbtt_tasklet;
82 struct delayed_work cal_work;
83 struct delayed_work mac_work;
85 struct mt76x02_mac_stats stats;
86 atomic_t avg_ampdu_len;
89 struct sk_buff *beacons[8];
96 struct mt76x02_calibration cal;
99 s8 target_power_delta[2];
107 struct mt76x02_dfs_pattern_detector dfs_pd;
110 extern struct ieee80211_rate mt76x02_rates[12];
112 void mt76x02_configure_filter(struct ieee80211_hw *hw,
113 unsigned int changed_flags,
114 unsigned int *total_flags, u64 multicast);
115 int mt76x02_sta_add(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
116 struct ieee80211_sta *sta);
117 int mt76x02_sta_remove(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
118 struct ieee80211_sta *sta);
120 void mt76x02_vif_init(struct mt76x02_dev *dev, struct ieee80211_vif *vif,
122 int mt76x02_add_interface(struct ieee80211_hw *hw,
123 struct ieee80211_vif *vif);
124 void mt76x02_remove_interface(struct ieee80211_hw *hw,
125 struct ieee80211_vif *vif);
127 int mt76x02_ampdu_action(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
128 struct ieee80211_ampdu_params *params);
129 int mt76x02_set_key(struct ieee80211_hw *hw, enum set_key_cmd cmd,
130 struct ieee80211_vif *vif, struct ieee80211_sta *sta,
131 struct ieee80211_key_conf *key);
132 int mt76x02_conf_tx(struct ieee80211_hw *hw, struct ieee80211_vif *vif,
133 u16 queue, const struct ieee80211_tx_queue_params *params);
134 void mt76x02_sta_rate_tbl_update(struct ieee80211_hw *hw,
135 struct ieee80211_vif *vif,
136 struct ieee80211_sta *sta);
137 s8 mt76x02_tx_get_max_txpwr_adj(struct mt76x02_dev *dev,
138 const struct ieee80211_tx_rate *rate);
139 s8 mt76x02_tx_get_txpwr_adj(struct mt76x02_dev *dev, s8 txpwr,
141 void mt76x02_tx_set_txpwr_auto(struct mt76x02_dev *dev, s8 txpwr);
142 int mt76x02_insert_hdr_pad(struct sk_buff *skb);
143 void mt76x02_remove_hdr_pad(struct sk_buff *skb, int len);
144 void mt76x02_tx_complete(struct mt76_dev *dev, struct sk_buff *skb);
145 bool mt76x02_tx_status_data(struct mt76_dev *mdev, u8 *update);
146 void mt76x02_queue_rx_skb(struct mt76_dev *mdev, enum mt76_rxq_id q,
147 struct sk_buff *skb);
148 void mt76x02_rx_poll_complete(struct mt76_dev *mdev, enum mt76_rxq_id q);
149 irqreturn_t mt76x02_irq_handler(int irq, void *dev_instance);
150 void mt76x02_tx(struct ieee80211_hw *hw, struct ieee80211_tx_control *control,
151 struct sk_buff *skb);
152 int mt76x02_tx_prepare_skb(struct mt76_dev *mdev, void *txwi,
153 struct sk_buff *skb, struct mt76_queue *q,
154 struct mt76_wcid *wcid, struct ieee80211_sta *sta,
157 extern const u16 mt76x02_beacon_offsets[16];
158 void mt76x02_set_beacon_offsets(struct mt76x02_dev *dev);
159 void mt76x02_set_irq_mask(struct mt76x02_dev *dev, u32 clear, u32 set);
160 void mt76x02_mac_start(struct mt76x02_dev *dev);
162 static inline bool is_mt76x2(struct mt76x02_dev *dev)
164 return mt76_chip(&dev->mt76) == 0x7612 ||
165 mt76_chip(&dev->mt76) == 0x7662 ||
166 mt76_chip(&dev->mt76) == 0x7602;
169 static inline void mt76x02_irq_enable(struct mt76x02_dev *dev, u32 mask)
171 mt76x02_set_irq_mask(dev, 0, mask);
174 static inline void mt76x02_irq_disable(struct mt76x02_dev *dev, u32 mask)
176 mt76x02_set_irq_mask(dev, mask, 0);
180 mt76x02_wait_for_txrx_idle(struct mt76_dev *dev)
182 return __mt76_poll_msec(dev, MT_MAC_STATUS,
183 MT_MAC_STATUS_TX | MT_MAC_STATUS_RX,
187 static inline struct mt76x02_sta *
188 mt76x02_rx_get_sta(struct mt76_dev *dev, u8 idx)
190 struct mt76_wcid *wcid;
192 if (idx >= ARRAY_SIZE(dev->wcid))
195 wcid = rcu_dereference(dev->wcid[idx]);
199 return container_of(wcid, struct mt76x02_sta, wcid);
202 static inline struct mt76_wcid *
203 mt76x02_rx_get_sta_wcid(struct mt76x02_sta *sta, bool unicast)
211 return &sta->vif->group_wcid;