1 // SPDX-License-Identifier: GPL-2.0
2 /* Copyright(c) 2013 - 2018 Intel Corporation. */
4 #include <linux/vmalloc.h>
9 /* The stat_string is expected to be a format string formatted using
10 * vsnprintf by fm10k_add_stat_strings. Every member of a stats array
11 * should use the same format specifiers as they will be formatted
12 * using the same variadic arguments.
14 char stat_string[ETH_GSTRING_LEN];
19 #define FM10K_STAT_FIELDS(_type, _name, _stat) { \
20 .stat_string = _name, \
21 .sizeof_stat = FIELD_SIZEOF(_type, _stat), \
22 .stat_offset = offsetof(_type, _stat) \
25 /* netdevice statistics */
26 #define FM10K_NETDEV_STAT(_net_stat) \
27 FM10K_STAT_FIELDS(struct net_device_stats, __stringify(_net_stat), \
30 static const struct fm10k_stats fm10k_gstrings_net_stats[] = {
31 FM10K_NETDEV_STAT(tx_packets),
32 FM10K_NETDEV_STAT(tx_bytes),
33 FM10K_NETDEV_STAT(tx_errors),
34 FM10K_NETDEV_STAT(rx_packets),
35 FM10K_NETDEV_STAT(rx_bytes),
36 FM10K_NETDEV_STAT(rx_errors),
37 FM10K_NETDEV_STAT(rx_dropped),
39 /* detailed Rx errors */
40 FM10K_NETDEV_STAT(rx_length_errors),
41 FM10K_NETDEV_STAT(rx_crc_errors),
42 FM10K_NETDEV_STAT(rx_fifo_errors),
45 #define FM10K_NETDEV_STATS_LEN ARRAY_SIZE(fm10k_gstrings_net_stats)
47 /* General interface statistics */
48 #define FM10K_STAT(_name, _stat) \
49 FM10K_STAT_FIELDS(struct fm10k_intfc, _name, _stat)
51 static const struct fm10k_stats fm10k_gstrings_global_stats[] = {
52 FM10K_STAT("tx_restart_queue", restart_queue),
53 FM10K_STAT("tx_busy", tx_busy),
54 FM10K_STAT("tx_csum_errors", tx_csum_errors),
55 FM10K_STAT("rx_alloc_failed", alloc_failed),
56 FM10K_STAT("rx_csum_errors", rx_csum_errors),
58 FM10K_STAT("tx_packets_nic", tx_packets_nic),
59 FM10K_STAT("tx_bytes_nic", tx_bytes_nic),
60 FM10K_STAT("rx_packets_nic", rx_packets_nic),
61 FM10K_STAT("rx_bytes_nic", rx_bytes_nic),
62 FM10K_STAT("rx_drops_nic", rx_drops_nic),
63 FM10K_STAT("rx_overrun_pf", rx_overrun_pf),
64 FM10K_STAT("rx_overrun_vf", rx_overrun_vf),
66 FM10K_STAT("swapi_status", hw.swapi.status),
67 FM10K_STAT("mac_rules_used", hw.swapi.mac.used),
68 FM10K_STAT("mac_rules_avail", hw.swapi.mac.avail),
70 FM10K_STAT("reset_while_pending", hw.mac.reset_while_pending),
72 FM10K_STAT("tx_hang_count", tx_timeout_count),
75 static const struct fm10k_stats fm10k_gstrings_pf_stats[] = {
76 FM10K_STAT("timeout", stats.timeout.count),
77 FM10K_STAT("ur", stats.ur.count),
78 FM10K_STAT("ca", stats.ca.count),
79 FM10K_STAT("um", stats.um.count),
80 FM10K_STAT("xec", stats.xec.count),
81 FM10K_STAT("vlan_drop", stats.vlan_drop.count),
82 FM10K_STAT("loopback_drop", stats.loopback_drop.count),
83 FM10K_STAT("nodesc_drop", stats.nodesc_drop.count),
86 /* mailbox statistics */
87 #define FM10K_MBX_STAT(_name, _stat) \
88 FM10K_STAT_FIELDS(struct fm10k_mbx_info, _name, _stat)
90 static const struct fm10k_stats fm10k_gstrings_mbx_stats[] = {
91 FM10K_MBX_STAT("mbx_tx_busy", tx_busy),
92 FM10K_MBX_STAT("mbx_tx_dropped", tx_dropped),
93 FM10K_MBX_STAT("mbx_tx_messages", tx_messages),
94 FM10K_MBX_STAT("mbx_tx_dwords", tx_dwords),
95 FM10K_MBX_STAT("mbx_tx_mbmem_pulled", tx_mbmem_pulled),
96 FM10K_MBX_STAT("mbx_rx_messages", rx_messages),
97 FM10K_MBX_STAT("mbx_rx_dwords", rx_dwords),
98 FM10K_MBX_STAT("mbx_rx_parse_err", rx_parse_err),
99 FM10K_MBX_STAT("mbx_rx_mbmem_pushed", rx_mbmem_pushed),
102 /* per-queue ring statistics */
103 #define FM10K_QUEUE_STAT(_name, _stat) \
104 FM10K_STAT_FIELDS(struct fm10k_ring, _name, _stat)
106 static const struct fm10k_stats fm10k_gstrings_queue_stats[] = {
107 FM10K_QUEUE_STAT("%s_queue_%u_packets", stats.packets),
108 FM10K_QUEUE_STAT("%s_queue_%u_bytes", stats.bytes),
111 #define FM10K_GLOBAL_STATS_LEN ARRAY_SIZE(fm10k_gstrings_global_stats)
112 #define FM10K_PF_STATS_LEN ARRAY_SIZE(fm10k_gstrings_pf_stats)
113 #define FM10K_MBX_STATS_LEN ARRAY_SIZE(fm10k_gstrings_mbx_stats)
114 #define FM10K_QUEUE_STATS_LEN ARRAY_SIZE(fm10k_gstrings_queue_stats)
116 #define FM10K_STATIC_STATS_LEN (FM10K_GLOBAL_STATS_LEN + \
117 FM10K_NETDEV_STATS_LEN + \
120 static const char fm10k_gstrings_test[][ETH_GSTRING_LEN] = {
121 "Mailbox test (on/offline)"
124 #define FM10K_TEST_LEN (sizeof(fm10k_gstrings_test) / ETH_GSTRING_LEN)
126 enum fm10k_self_test_types {
128 FM10K_TEST_MAX = FM10K_TEST_LEN
135 static const char fm10k_prv_flags[FM10K_PRV_FLAG_LEN][ETH_GSTRING_LEN] = {
138 static void fm10k_add_stat_strings(u8 **p, const struct fm10k_stats stats[],
139 const unsigned int size, ...)
143 for (i = 0; i < size; i++) {
146 va_start(args, size);
147 vsnprintf(*p, ETH_GSTRING_LEN, stats[i].stat_string, args);
148 *p += ETH_GSTRING_LEN;
153 static void fm10k_get_stat_strings(struct net_device *dev, u8 *data)
155 struct fm10k_intfc *interface = netdev_priv(dev);
158 fm10k_add_stat_strings(&data, fm10k_gstrings_net_stats,
159 FM10K_NETDEV_STATS_LEN);
161 fm10k_add_stat_strings(&data, fm10k_gstrings_global_stats,
162 FM10K_GLOBAL_STATS_LEN);
164 fm10k_add_stat_strings(&data, fm10k_gstrings_mbx_stats,
165 FM10K_MBX_STATS_LEN);
167 if (interface->hw.mac.type != fm10k_mac_vf)
168 fm10k_add_stat_strings(&data, fm10k_gstrings_pf_stats,
171 for (i = 0; i < interface->hw.mac.max_queues; i++) {
172 fm10k_add_stat_strings(&data, fm10k_gstrings_queue_stats,
173 FM10K_QUEUE_STATS_LEN,
176 fm10k_add_stat_strings(&data, fm10k_gstrings_queue_stats,
177 FM10K_QUEUE_STATS_LEN,
182 static void fm10k_get_strings(struct net_device *dev,
183 u32 stringset, u8 *data)
187 memcpy(data, fm10k_gstrings_test,
188 FM10K_TEST_LEN * ETH_GSTRING_LEN);
191 fm10k_get_stat_strings(dev, data);
193 case ETH_SS_PRIV_FLAGS:
194 memcpy(data, fm10k_prv_flags,
195 FM10K_PRV_FLAG_LEN * ETH_GSTRING_LEN);
200 static int fm10k_get_sset_count(struct net_device *dev, int sset)
202 struct fm10k_intfc *interface = netdev_priv(dev);
203 struct fm10k_hw *hw = &interface->hw;
204 int stats_len = FM10K_STATIC_STATS_LEN;
208 return FM10K_TEST_LEN;
210 stats_len += hw->mac.max_queues * 2 * FM10K_QUEUE_STATS_LEN;
212 if (hw->mac.type != fm10k_mac_vf)
213 stats_len += FM10K_PF_STATS_LEN;
216 case ETH_SS_PRIV_FLAGS:
217 return FM10K_PRV_FLAG_LEN;
223 static void fm10k_add_ethtool_stats(u64 **data, void *pointer,
224 const struct fm10k_stats stats[],
225 const unsigned int size)
231 /* memory is not zero allocated so we have to clear it */
232 for (i = 0; i < size; i++)
237 for (i = 0; i < size; i++) {
238 p = (char *)pointer + stats[i].stat_offset;
240 switch (stats[i].sizeof_stat) {
242 *((*data)++) = *(u64 *)p;
245 *((*data)++) = *(u32 *)p;
248 *((*data)++) = *(u16 *)p;
251 *((*data)++) = *(u8 *)p;
259 static void fm10k_get_ethtool_stats(struct net_device *netdev,
260 struct ethtool_stats __always_unused *stats,
263 struct fm10k_intfc *interface = netdev_priv(netdev);
264 struct net_device_stats *net_stats = &netdev->stats;
267 fm10k_update_stats(interface);
269 fm10k_add_ethtool_stats(&data, net_stats, fm10k_gstrings_net_stats,
270 FM10K_NETDEV_STATS_LEN);
272 fm10k_add_ethtool_stats(&data, interface, fm10k_gstrings_global_stats,
273 FM10K_GLOBAL_STATS_LEN);
275 fm10k_add_ethtool_stats(&data, &interface->hw.mbx,
276 fm10k_gstrings_mbx_stats,
277 FM10K_MBX_STATS_LEN);
279 if (interface->hw.mac.type != fm10k_mac_vf) {
280 fm10k_add_ethtool_stats(&data, interface,
281 fm10k_gstrings_pf_stats,
285 for (i = 0; i < interface->hw.mac.max_queues; i++) {
286 struct fm10k_ring *ring;
288 ring = interface->tx_ring[i];
289 fm10k_add_ethtool_stats(&data, ring,
290 fm10k_gstrings_queue_stats,
291 FM10K_QUEUE_STATS_LEN);
293 ring = interface->rx_ring[i];
294 fm10k_add_ethtool_stats(&data, ring,
295 fm10k_gstrings_queue_stats,
296 FM10K_QUEUE_STATS_LEN);
300 /* If function below adds more registers this define needs to be updated */
301 #define FM10K_REGS_LEN_Q 29
303 static void fm10k_get_reg_q(struct fm10k_hw *hw, u32 *buff, int i)
307 buff[idx++] = fm10k_read_reg(hw, FM10K_RDBAL(i));
308 buff[idx++] = fm10k_read_reg(hw, FM10K_RDBAH(i));
309 buff[idx++] = fm10k_read_reg(hw, FM10K_RDLEN(i));
310 buff[idx++] = fm10k_read_reg(hw, FM10K_TPH_RXCTRL(i));
311 buff[idx++] = fm10k_read_reg(hw, FM10K_RDH(i));
312 buff[idx++] = fm10k_read_reg(hw, FM10K_RDT(i));
313 buff[idx++] = fm10k_read_reg(hw, FM10K_RXQCTL(i));
314 buff[idx++] = fm10k_read_reg(hw, FM10K_RXDCTL(i));
315 buff[idx++] = fm10k_read_reg(hw, FM10K_RXINT(i));
316 buff[idx++] = fm10k_read_reg(hw, FM10K_SRRCTL(i));
317 buff[idx++] = fm10k_read_reg(hw, FM10K_QPRC(i));
318 buff[idx++] = fm10k_read_reg(hw, FM10K_QPRDC(i));
319 buff[idx++] = fm10k_read_reg(hw, FM10K_QBRC_L(i));
320 buff[idx++] = fm10k_read_reg(hw, FM10K_QBRC_H(i));
321 buff[idx++] = fm10k_read_reg(hw, FM10K_TDBAL(i));
322 buff[idx++] = fm10k_read_reg(hw, FM10K_TDBAH(i));
323 buff[idx++] = fm10k_read_reg(hw, FM10K_TDLEN(i));
324 buff[idx++] = fm10k_read_reg(hw, FM10K_TPH_TXCTRL(i));
325 buff[idx++] = fm10k_read_reg(hw, FM10K_TDH(i));
326 buff[idx++] = fm10k_read_reg(hw, FM10K_TDT(i));
327 buff[idx++] = fm10k_read_reg(hw, FM10K_TXDCTL(i));
328 buff[idx++] = fm10k_read_reg(hw, FM10K_TXQCTL(i));
329 buff[idx++] = fm10k_read_reg(hw, FM10K_TXINT(i));
330 buff[idx++] = fm10k_read_reg(hw, FM10K_QPTC(i));
331 buff[idx++] = fm10k_read_reg(hw, FM10K_QBTC_L(i));
332 buff[idx++] = fm10k_read_reg(hw, FM10K_QBTC_H(i));
333 buff[idx++] = fm10k_read_reg(hw, FM10K_TQDLOC(i));
334 buff[idx++] = fm10k_read_reg(hw, FM10K_TX_SGLORT(i));
335 buff[idx++] = fm10k_read_reg(hw, FM10K_PFVTCTL(i));
337 BUG_ON(idx != FM10K_REGS_LEN_Q);
340 /* If function above adds more registers this define needs to be updated */
341 #define FM10K_REGS_LEN_VSI 43
343 static void fm10k_get_reg_vsi(struct fm10k_hw *hw, u32 *buff, int i)
347 buff[idx++] = fm10k_read_reg(hw, FM10K_MRQC(i));
348 for (j = 0; j < 10; j++)
349 buff[idx++] = fm10k_read_reg(hw, FM10K_RSSRK(i, j));
350 for (j = 0; j < 32; j++)
351 buff[idx++] = fm10k_read_reg(hw, FM10K_RETA(i, j));
353 BUG_ON(idx != FM10K_REGS_LEN_VSI);
356 static void fm10k_get_regs(struct net_device *netdev,
357 struct ethtool_regs *regs, void *p)
359 struct fm10k_intfc *interface = netdev_priv(netdev);
360 struct fm10k_hw *hw = &interface->hw;
364 regs->version = BIT(24) | (hw->revision_id << 16) | hw->device_id;
366 switch (hw->mac.type) {
368 /* General PF Registers */
369 *(buff++) = fm10k_read_reg(hw, FM10K_CTRL);
370 *(buff++) = fm10k_read_reg(hw, FM10K_CTRL_EXT);
371 *(buff++) = fm10k_read_reg(hw, FM10K_GCR);
372 *(buff++) = fm10k_read_reg(hw, FM10K_GCR_EXT);
374 for (i = 0; i < 8; i++) {
375 *(buff++) = fm10k_read_reg(hw, FM10K_DGLORTMAP(i));
376 *(buff++) = fm10k_read_reg(hw, FM10K_DGLORTDEC(i));
379 for (i = 0; i < 65; i++) {
380 fm10k_get_reg_vsi(hw, buff, i);
381 buff += FM10K_REGS_LEN_VSI;
384 *(buff++) = fm10k_read_reg(hw, FM10K_DMA_CTRL);
385 *(buff++) = fm10k_read_reg(hw, FM10K_DMA_CTRL2);
387 for (i = 0; i < FM10K_MAX_QUEUES_PF; i++) {
388 fm10k_get_reg_q(hw, buff, i);
389 buff += FM10K_REGS_LEN_Q;
392 *(buff++) = fm10k_read_reg(hw, FM10K_TPH_CTRL);
394 for (i = 0; i < 8; i++)
395 *(buff++) = fm10k_read_reg(hw, FM10K_INT_MAP(i));
397 /* Interrupt Throttling Registers */
398 for (i = 0; i < 130; i++)
399 *(buff++) = fm10k_read_reg(hw, FM10K_ITR(i));
403 /* General VF registers */
404 *(buff++) = fm10k_read_reg(hw, FM10K_VFCTRL);
405 *(buff++) = fm10k_read_reg(hw, FM10K_VFINT_MAP);
406 *(buff++) = fm10k_read_reg(hw, FM10K_VFSYSTIME);
408 /* Interrupt Throttling Registers */
409 for (i = 0; i < 8; i++)
410 *(buff++) = fm10k_read_reg(hw, FM10K_VFITR(i));
412 fm10k_get_reg_vsi(hw, buff, 0);
413 buff += FM10K_REGS_LEN_VSI;
415 for (i = 0; i < FM10K_MAX_QUEUES_POOL; i++) {
416 if (i < hw->mac.max_queues)
417 fm10k_get_reg_q(hw, buff, i);
419 memset(buff, 0, sizeof(u32) * FM10K_REGS_LEN_Q);
420 buff += FM10K_REGS_LEN_Q;
429 /* If function above adds more registers these define need to be updated */
430 #define FM10K_REGS_LEN_PF \
431 (162 + (65 * FM10K_REGS_LEN_VSI) + (FM10K_MAX_QUEUES_PF * FM10K_REGS_LEN_Q))
432 #define FM10K_REGS_LEN_VF \
433 (11 + FM10K_REGS_LEN_VSI + (FM10K_MAX_QUEUES_POOL * FM10K_REGS_LEN_Q))
435 static int fm10k_get_regs_len(struct net_device *netdev)
437 struct fm10k_intfc *interface = netdev_priv(netdev);
438 struct fm10k_hw *hw = &interface->hw;
440 switch (hw->mac.type) {
442 return FM10K_REGS_LEN_PF * sizeof(u32);
444 return FM10K_REGS_LEN_VF * sizeof(u32);
450 static void fm10k_get_drvinfo(struct net_device *dev,
451 struct ethtool_drvinfo *info)
453 struct fm10k_intfc *interface = netdev_priv(dev);
455 strncpy(info->driver, fm10k_driver_name,
456 sizeof(info->driver) - 1);
457 strncpy(info->version, fm10k_driver_version,
458 sizeof(info->version) - 1);
459 strncpy(info->bus_info, pci_name(interface->pdev),
460 sizeof(info->bus_info) - 1);
463 static void fm10k_get_pauseparam(struct net_device *dev,
464 struct ethtool_pauseparam *pause)
466 struct fm10k_intfc *interface = netdev_priv(dev);
468 /* record fixed values for autoneg and tx pause */
472 pause->rx_pause = interface->rx_pause ? 1 : 0;
475 static int fm10k_set_pauseparam(struct net_device *dev,
476 struct ethtool_pauseparam *pause)
478 struct fm10k_intfc *interface = netdev_priv(dev);
479 struct fm10k_hw *hw = &interface->hw;
481 if (pause->autoneg || !pause->tx_pause)
484 /* we can only support pause on the PF to avoid head-of-line blocking */
485 if (hw->mac.type == fm10k_mac_pf)
486 interface->rx_pause = pause->rx_pause ? ~0 : 0;
487 else if (pause->rx_pause)
490 if (netif_running(dev))
491 fm10k_update_rx_drop_en(interface);
496 static u32 fm10k_get_msglevel(struct net_device *netdev)
498 struct fm10k_intfc *interface = netdev_priv(netdev);
500 return interface->msg_enable;
503 static void fm10k_set_msglevel(struct net_device *netdev, u32 data)
505 struct fm10k_intfc *interface = netdev_priv(netdev);
507 interface->msg_enable = data;
510 static void fm10k_get_ringparam(struct net_device *netdev,
511 struct ethtool_ringparam *ring)
513 struct fm10k_intfc *interface = netdev_priv(netdev);
515 ring->rx_max_pending = FM10K_MAX_RXD;
516 ring->tx_max_pending = FM10K_MAX_TXD;
517 ring->rx_mini_max_pending = 0;
518 ring->rx_jumbo_max_pending = 0;
519 ring->rx_pending = interface->rx_ring_count;
520 ring->tx_pending = interface->tx_ring_count;
521 ring->rx_mini_pending = 0;
522 ring->rx_jumbo_pending = 0;
525 static int fm10k_set_ringparam(struct net_device *netdev,
526 struct ethtool_ringparam *ring)
528 struct fm10k_intfc *interface = netdev_priv(netdev);
529 struct fm10k_ring *temp_ring;
531 u32 new_rx_count, new_tx_count;
533 if ((ring->rx_mini_pending) || (ring->rx_jumbo_pending))
536 new_tx_count = clamp_t(u32, ring->tx_pending,
537 FM10K_MIN_TXD, FM10K_MAX_TXD);
538 new_tx_count = ALIGN(new_tx_count, FM10K_REQ_TX_DESCRIPTOR_MULTIPLE);
540 new_rx_count = clamp_t(u32, ring->rx_pending,
541 FM10K_MIN_RXD, FM10K_MAX_RXD);
542 new_rx_count = ALIGN(new_rx_count, FM10K_REQ_RX_DESCRIPTOR_MULTIPLE);
544 if ((new_tx_count == interface->tx_ring_count) &&
545 (new_rx_count == interface->rx_ring_count)) {
550 while (test_and_set_bit(__FM10K_RESETTING, interface->state))
551 usleep_range(1000, 2000);
553 if (!netif_running(interface->netdev)) {
554 for (i = 0; i < interface->num_tx_queues; i++)
555 interface->tx_ring[i]->count = new_tx_count;
556 for (i = 0; i < interface->num_rx_queues; i++)
557 interface->rx_ring[i]->count = new_rx_count;
558 interface->tx_ring_count = new_tx_count;
559 interface->rx_ring_count = new_rx_count;
563 /* allocate temporary buffer to store rings in */
564 i = max_t(int, interface->num_tx_queues, interface->num_rx_queues);
565 temp_ring = vmalloc(i * sizeof(struct fm10k_ring));
572 fm10k_down(interface);
574 /* Setup new Tx resources and free the old Tx resources in that order.
575 * We can then assign the new resources to the rings via a memcpy.
576 * The advantage to this approach is that we are guaranteed to still
577 * have resources even in the case of an allocation failure.
579 if (new_tx_count != interface->tx_ring_count) {
580 for (i = 0; i < interface->num_tx_queues; i++) {
581 memcpy(&temp_ring[i], interface->tx_ring[i],
582 sizeof(struct fm10k_ring));
584 temp_ring[i].count = new_tx_count;
585 err = fm10k_setup_tx_resources(&temp_ring[i]);
589 fm10k_free_tx_resources(&temp_ring[i]);
595 for (i = 0; i < interface->num_tx_queues; i++) {
596 fm10k_free_tx_resources(interface->tx_ring[i]);
598 memcpy(interface->tx_ring[i], &temp_ring[i],
599 sizeof(struct fm10k_ring));
602 interface->tx_ring_count = new_tx_count;
605 /* Repeat the process for the Rx rings if needed */
606 if (new_rx_count != interface->rx_ring_count) {
607 for (i = 0; i < interface->num_rx_queues; i++) {
608 memcpy(&temp_ring[i], interface->rx_ring[i],
609 sizeof(struct fm10k_ring));
611 temp_ring[i].count = new_rx_count;
612 err = fm10k_setup_rx_resources(&temp_ring[i]);
616 fm10k_free_rx_resources(&temp_ring[i]);
622 for (i = 0; i < interface->num_rx_queues; i++) {
623 fm10k_free_rx_resources(interface->rx_ring[i]);
625 memcpy(interface->rx_ring[i], &temp_ring[i],
626 sizeof(struct fm10k_ring));
629 interface->rx_ring_count = new_rx_count;
636 clear_bit(__FM10K_RESETTING, interface->state);
640 static int fm10k_get_coalesce(struct net_device *dev,
641 struct ethtool_coalesce *ec)
643 struct fm10k_intfc *interface = netdev_priv(dev);
645 ec->use_adaptive_tx_coalesce = ITR_IS_ADAPTIVE(interface->tx_itr);
646 ec->tx_coalesce_usecs = interface->tx_itr & ~FM10K_ITR_ADAPTIVE;
648 ec->use_adaptive_rx_coalesce = ITR_IS_ADAPTIVE(interface->rx_itr);
649 ec->rx_coalesce_usecs = interface->rx_itr & ~FM10K_ITR_ADAPTIVE;
654 static int fm10k_set_coalesce(struct net_device *dev,
655 struct ethtool_coalesce *ec)
657 struct fm10k_intfc *interface = netdev_priv(dev);
658 struct fm10k_q_vector *qv;
663 if ((ec->rx_coalesce_usecs > FM10K_ITR_MAX) ||
664 (ec->tx_coalesce_usecs > FM10K_ITR_MAX))
667 /* record settings */
668 tx_itr = ec->tx_coalesce_usecs;
669 rx_itr = ec->rx_coalesce_usecs;
671 /* set initial values for adaptive ITR */
672 if (ec->use_adaptive_tx_coalesce)
673 tx_itr = FM10K_ITR_ADAPTIVE | FM10K_TX_ITR_DEFAULT;
675 if (ec->use_adaptive_rx_coalesce)
676 rx_itr = FM10K_ITR_ADAPTIVE | FM10K_RX_ITR_DEFAULT;
678 /* update interface */
679 interface->tx_itr = tx_itr;
680 interface->rx_itr = rx_itr;
682 /* update q_vectors */
683 for (i = 0; i < interface->num_q_vectors; i++) {
684 qv = interface->q_vector[i];
692 static int fm10k_get_rss_hash_opts(struct fm10k_intfc *interface,
693 struct ethtool_rxnfc *cmd)
697 /* Report default options for RSS on fm10k */
698 switch (cmd->flow_type) {
701 cmd->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3;
704 if (test_bit(FM10K_FLAG_RSS_FIELD_IPV4_UDP,
706 cmd->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3;
718 cmd->data |= RXH_IP_SRC | RXH_IP_DST;
721 if (test_bit(FM10K_FLAG_RSS_FIELD_IPV6_UDP,
723 cmd->data |= RXH_L4_B_0_1 | RXH_L4_B_2_3;
724 cmd->data |= RXH_IP_SRC | RXH_IP_DST;
733 static int fm10k_get_rxnfc(struct net_device *dev, struct ethtool_rxnfc *cmd,
734 u32 __always_unused *rule_locs)
736 struct fm10k_intfc *interface = netdev_priv(dev);
737 int ret = -EOPNOTSUPP;
740 case ETHTOOL_GRXRINGS:
741 cmd->data = interface->num_rx_queues;
745 ret = fm10k_get_rss_hash_opts(interface, cmd);
754 static int fm10k_set_rss_hash_opt(struct fm10k_intfc *interface,
755 struct ethtool_rxnfc *nfc)
757 int rss_ipv4_udp = test_bit(FM10K_FLAG_RSS_FIELD_IPV4_UDP,
759 int rss_ipv6_udp = test_bit(FM10K_FLAG_RSS_FIELD_IPV6_UDP,
762 /* RSS does not support anything other than hashing
763 * to queues on src and dst IPs and ports
765 if (nfc->data & ~(RXH_IP_SRC | RXH_IP_DST |
766 RXH_L4_B_0_1 | RXH_L4_B_2_3))
769 switch (nfc->flow_type) {
772 if (!(nfc->data & RXH_IP_SRC) ||
773 !(nfc->data & RXH_IP_DST) ||
774 !(nfc->data & RXH_L4_B_0_1) ||
775 !(nfc->data & RXH_L4_B_2_3))
779 if (!(nfc->data & RXH_IP_SRC) ||
780 !(nfc->data & RXH_IP_DST))
782 switch (nfc->data & (RXH_L4_B_0_1 | RXH_L4_B_2_3)) {
784 clear_bit(FM10K_FLAG_RSS_FIELD_IPV4_UDP,
787 case (RXH_L4_B_0_1 | RXH_L4_B_2_3):
788 set_bit(FM10K_FLAG_RSS_FIELD_IPV4_UDP,
796 if (!(nfc->data & RXH_IP_SRC) ||
797 !(nfc->data & RXH_IP_DST))
799 switch (nfc->data & (RXH_L4_B_0_1 | RXH_L4_B_2_3)) {
801 clear_bit(FM10K_FLAG_RSS_FIELD_IPV6_UDP,
804 case (RXH_L4_B_0_1 | RXH_L4_B_2_3):
805 set_bit(FM10K_FLAG_RSS_FIELD_IPV6_UDP,
820 if (!(nfc->data & RXH_IP_SRC) ||
821 !(nfc->data & RXH_IP_DST) ||
822 (nfc->data & RXH_L4_B_0_1) ||
823 (nfc->data & RXH_L4_B_2_3))
830 /* If something changed we need to update the MRQC register. Note that
831 * test_bit() is guaranteed to return strictly 0 or 1, so testing for
834 if ((rss_ipv4_udp != test_bit(FM10K_FLAG_RSS_FIELD_IPV4_UDP,
835 interface->flags)) ||
836 (rss_ipv6_udp != test_bit(FM10K_FLAG_RSS_FIELD_IPV6_UDP,
837 interface->flags))) {
838 struct fm10k_hw *hw = &interface->hw;
842 /* Perform hash on these packet types */
843 mrqc = FM10K_MRQC_IPV4 |
844 FM10K_MRQC_TCP_IPV4 |
848 if (test_bit(FM10K_FLAG_RSS_FIELD_IPV4_UDP,
850 mrqc |= FM10K_MRQC_UDP_IPV4;
853 if (test_bit(FM10K_FLAG_RSS_FIELD_IPV6_UDP,
855 mrqc |= FM10K_MRQC_UDP_IPV6;
859 /* If we enable UDP RSS display a warning that this may cause
860 * fragmented UDP packets to arrive out of order.
863 netif_warn(interface, drv, interface->netdev,
864 "enabling UDP RSS: fragmented packets may arrive out of order to the stack above\n");
866 fm10k_write_reg(hw, FM10K_MRQC(0), mrqc);
872 static int fm10k_set_rxnfc(struct net_device *dev, struct ethtool_rxnfc *cmd)
874 struct fm10k_intfc *interface = netdev_priv(dev);
875 int ret = -EOPNOTSUPP;
879 ret = fm10k_set_rss_hash_opt(interface, cmd);
888 static int fm10k_mbx_test(struct fm10k_intfc *interface, u64 *data)
890 struct fm10k_hw *hw = &interface->hw;
891 struct fm10k_mbx_info *mbx = &hw->mbx;
892 u32 attr_flag, test_msg[6];
893 unsigned long timeout;
896 /* For now this is a VF only feature */
897 if (hw->mac.type != fm10k_mac_vf)
900 /* loop through both nested and unnested attribute types */
901 for (attr_flag = BIT(FM10K_TEST_MSG_UNSET);
902 attr_flag < BIT(2 * FM10K_TEST_MSG_NESTED);
903 attr_flag += attr_flag) {
904 /* generate message to be tested */
905 fm10k_tlv_msg_test_create(test_msg, attr_flag);
907 fm10k_mbx_lock(interface);
908 mbx->test_result = FM10K_NOT_IMPLEMENTED;
909 err = mbx->ops.enqueue_tx(hw, mbx, test_msg);
910 fm10k_mbx_unlock(interface);
912 /* wait up to 1 second for response */
913 timeout = jiffies + HZ;
918 usleep_range(500, 1000);
920 fm10k_mbx_lock(interface);
921 mbx->ops.process(hw, mbx);
922 fm10k_mbx_unlock(interface);
924 err = mbx->test_result;
927 } while (time_is_after_jiffies(timeout));
929 /* reporting errors */
935 *data = err < 0 ? (attr_flag) : (err > 0);
939 static void fm10k_self_test(struct net_device *dev,
940 struct ethtool_test *eth_test, u64 *data)
942 struct fm10k_intfc *interface = netdev_priv(dev);
943 struct fm10k_hw *hw = &interface->hw;
945 memset(data, 0, sizeof(*data) * FM10K_TEST_LEN);
947 if (FM10K_REMOVED(hw->hw_addr)) {
948 netif_err(interface, drv, dev,
949 "Interface removed - test blocked\n");
950 eth_test->flags |= ETH_TEST_FL_FAILED;
954 if (fm10k_mbx_test(interface, &data[FM10K_TEST_MBX]))
955 eth_test->flags |= ETH_TEST_FL_FAILED;
958 static u32 fm10k_get_priv_flags(struct net_device *netdev)
963 static int fm10k_set_priv_flags(struct net_device *netdev, u32 priv_flags)
965 if (priv_flags >= BIT(FM10K_PRV_FLAG_LEN))
971 static u32 fm10k_get_reta_size(struct net_device __always_unused *netdev)
973 return FM10K_RETA_SIZE * FM10K_RETA_ENTRIES_PER_REG;
976 void fm10k_write_reta(struct fm10k_intfc *interface, const u32 *indir)
978 u16 rss_i = interface->ring_feature[RING_F_RSS].indices;
979 struct fm10k_hw *hw = &interface->hw;
983 /* record entries to reta table */
984 for (i = 0; i < FM10K_RETA_SIZE; i++) {
987 /* generate a new table if we weren't given one */
988 for (j = 0; j < 4; j++) {
990 n = indir[4 * i + j];
992 n = ethtool_rxfh_indir_default(4 * i + j,
1003 if (interface->reta[i] == reta)
1006 interface->reta[i] = reta;
1007 fm10k_write_reg(hw, FM10K_RETA(0, i), reta);
1011 static int fm10k_get_reta(struct net_device *netdev, u32 *indir)
1013 struct fm10k_intfc *interface = netdev_priv(netdev);
1019 for (i = 0; i < FM10K_RETA_SIZE; i++, indir += 4) {
1020 u32 reta = interface->reta[i];
1022 indir[0] = (reta << 24) >> 24;
1023 indir[1] = (reta << 16) >> 24;
1024 indir[2] = (reta << 8) >> 24;
1025 indir[3] = (reta) >> 24;
1031 static int fm10k_set_reta(struct net_device *netdev, const u32 *indir)
1033 struct fm10k_intfc *interface = netdev_priv(netdev);
1040 /* Verify user input. */
1041 rss_i = interface->ring_feature[RING_F_RSS].indices;
1042 for (i = fm10k_get_reta_size(netdev); i--;) {
1043 if (indir[i] < rss_i)
1048 fm10k_write_reta(interface, indir);
1053 static u32 fm10k_get_rssrk_size(struct net_device __always_unused *netdev)
1055 return FM10K_RSSRK_SIZE * FM10K_RSSRK_ENTRIES_PER_REG;
1058 static int fm10k_get_rssh(struct net_device *netdev, u32 *indir, u8 *key,
1061 struct fm10k_intfc *interface = netdev_priv(netdev);
1065 *hfunc = ETH_RSS_HASH_TOP;
1067 err = fm10k_get_reta(netdev, indir);
1071 for (i = 0; i < FM10K_RSSRK_SIZE; i++, key += 4)
1072 *(__le32 *)key = cpu_to_le32(interface->rssrk[i]);
1077 static int fm10k_set_rssh(struct net_device *netdev, const u32 *indir,
1078 const u8 *key, const u8 hfunc)
1080 struct fm10k_intfc *interface = netdev_priv(netdev);
1081 struct fm10k_hw *hw = &interface->hw;
1084 /* We do not allow change in unsupported parameters */
1085 if (hfunc != ETH_RSS_HASH_NO_CHANGE && hfunc != ETH_RSS_HASH_TOP)
1088 err = fm10k_set_reta(netdev, indir);
1092 for (i = 0; i < FM10K_RSSRK_SIZE; i++, key += 4) {
1093 u32 rssrk = le32_to_cpu(*(__le32 *)key);
1095 if (interface->rssrk[i] == rssrk)
1098 interface->rssrk[i] = rssrk;
1099 fm10k_write_reg(hw, FM10K_RSSRK(0, i), rssrk);
1105 static unsigned int fm10k_max_channels(struct net_device *dev)
1107 struct fm10k_intfc *interface = netdev_priv(dev);
1108 unsigned int max_combined = interface->hw.mac.max_queues;
1109 u8 tcs = netdev_get_num_tc(dev);
1111 /* For QoS report channels per traffic class */
1113 max_combined = BIT((fls(max_combined / tcs) - 1));
1115 return max_combined;
1118 static void fm10k_get_channels(struct net_device *dev,
1119 struct ethtool_channels *ch)
1121 struct fm10k_intfc *interface = netdev_priv(dev);
1122 struct fm10k_hw *hw = &interface->hw;
1124 /* report maximum channels */
1125 ch->max_combined = fm10k_max_channels(dev);
1127 /* report info for other vector */
1128 ch->max_other = NON_Q_VECTORS(hw);
1129 ch->other_count = ch->max_other;
1131 /* record RSS queues */
1132 ch->combined_count = interface->ring_feature[RING_F_RSS].indices;
1135 static int fm10k_set_channels(struct net_device *dev,
1136 struct ethtool_channels *ch)
1138 struct fm10k_intfc *interface = netdev_priv(dev);
1139 unsigned int count = ch->combined_count;
1140 struct fm10k_hw *hw = &interface->hw;
1142 /* verify they are not requesting separate vectors */
1143 if (!count || ch->rx_count || ch->tx_count)
1146 /* verify other_count has not changed */
1147 if (ch->other_count != NON_Q_VECTORS(hw))
1150 /* verify the number of channels does not exceed hardware limits */
1151 if (count > fm10k_max_channels(dev))
1154 interface->ring_feature[RING_F_RSS].limit = count;
1156 /* use setup TC to update any traffic class queue mapping */
1157 return fm10k_setup_tc(dev, netdev_get_num_tc(dev));
1160 static const struct ethtool_ops fm10k_ethtool_ops = {
1161 .get_strings = fm10k_get_strings,
1162 .get_sset_count = fm10k_get_sset_count,
1163 .get_ethtool_stats = fm10k_get_ethtool_stats,
1164 .get_drvinfo = fm10k_get_drvinfo,
1165 .get_link = ethtool_op_get_link,
1166 .get_pauseparam = fm10k_get_pauseparam,
1167 .set_pauseparam = fm10k_set_pauseparam,
1168 .get_msglevel = fm10k_get_msglevel,
1169 .set_msglevel = fm10k_set_msglevel,
1170 .get_ringparam = fm10k_get_ringparam,
1171 .set_ringparam = fm10k_set_ringparam,
1172 .get_coalesce = fm10k_get_coalesce,
1173 .set_coalesce = fm10k_set_coalesce,
1174 .get_rxnfc = fm10k_get_rxnfc,
1175 .set_rxnfc = fm10k_set_rxnfc,
1176 .get_regs = fm10k_get_regs,
1177 .get_regs_len = fm10k_get_regs_len,
1178 .self_test = fm10k_self_test,
1179 .get_priv_flags = fm10k_get_priv_flags,
1180 .set_priv_flags = fm10k_set_priv_flags,
1181 .get_rxfh_indir_size = fm10k_get_reta_size,
1182 .get_rxfh_key_size = fm10k_get_rssrk_size,
1183 .get_rxfh = fm10k_get_rssh,
1184 .set_rxfh = fm10k_set_rssh,
1185 .get_channels = fm10k_get_channels,
1186 .set_channels = fm10k_set_channels,
1187 .get_ts_info = ethtool_op_get_ts_info,
1190 void fm10k_set_ethtool_ops(struct net_device *dev)
1192 dev->ethtool_ops = &fm10k_ethtool_ops;