]> asedeno.scripts.mit.edu Git - linux.git/blob - drivers/net/ethernet/hisilicon/hns3/hns3pf/hclge_debugfs.c
Linux 5.6-rc7
[linux.git] / drivers / net / ethernet / hisilicon / hns3 / hns3pf / hclge_debugfs.c
1 // SPDX-License-Identifier: GPL-2.0+
2 /* Copyright (c) 2018-2019 Hisilicon Limited. */
3
4 #include <linux/device.h>
5
6 #include "hclge_debugfs.h"
7 #include "hclge_main.h"
8 #include "hclge_tm.h"
9 #include "hnae3.h"
10
11 static struct hclge_dbg_reg_type_info hclge_dbg_reg_info[] = {
12         { .reg_type = "bios common",
13           .dfx_msg = &hclge_dbg_bios_common_reg[0],
14           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_bios_common_reg),
15                        .offset = HCLGE_DBG_DFX_BIOS_OFFSET,
16                        .cmd = HCLGE_OPC_DFX_BIOS_COMMON_REG } },
17         { .reg_type = "ssu",
18           .dfx_msg = &hclge_dbg_ssu_reg_0[0],
19           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_ssu_reg_0),
20                        .offset = HCLGE_DBG_DFX_SSU_0_OFFSET,
21                        .cmd = HCLGE_OPC_DFX_SSU_REG_0 } },
22         { .reg_type = "ssu",
23           .dfx_msg = &hclge_dbg_ssu_reg_1[0],
24           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_ssu_reg_1),
25                        .offset = HCLGE_DBG_DFX_SSU_1_OFFSET,
26                        .cmd = HCLGE_OPC_DFX_SSU_REG_1 } },
27         { .reg_type = "ssu",
28           .dfx_msg = &hclge_dbg_ssu_reg_2[0],
29           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_ssu_reg_2),
30                        .offset = HCLGE_DBG_DFX_SSU_2_OFFSET,
31                        .cmd = HCLGE_OPC_DFX_SSU_REG_2 } },
32         { .reg_type = "igu egu",
33           .dfx_msg = &hclge_dbg_igu_egu_reg[0],
34           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_igu_egu_reg),
35                        .offset = HCLGE_DBG_DFX_IGU_OFFSET,
36                        .cmd = HCLGE_OPC_DFX_IGU_EGU_REG } },
37         { .reg_type = "rpu",
38           .dfx_msg = &hclge_dbg_rpu_reg_0[0],
39           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_rpu_reg_0),
40                        .offset = HCLGE_DBG_DFX_RPU_0_OFFSET,
41                        .cmd = HCLGE_OPC_DFX_RPU_REG_0 } },
42         { .reg_type = "rpu",
43           .dfx_msg = &hclge_dbg_rpu_reg_1[0],
44           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_rpu_reg_1),
45                        .offset = HCLGE_DBG_DFX_RPU_1_OFFSET,
46                        .cmd = HCLGE_OPC_DFX_RPU_REG_1 } },
47         { .reg_type = "ncsi",
48           .dfx_msg = &hclge_dbg_ncsi_reg[0],
49           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_ncsi_reg),
50                        .offset = HCLGE_DBG_DFX_NCSI_OFFSET,
51                        .cmd = HCLGE_OPC_DFX_NCSI_REG } },
52         { .reg_type = "rtc",
53           .dfx_msg = &hclge_dbg_rtc_reg[0],
54           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_rtc_reg),
55                        .offset = HCLGE_DBG_DFX_RTC_OFFSET,
56                        .cmd = HCLGE_OPC_DFX_RTC_REG } },
57         { .reg_type = "ppp",
58           .dfx_msg = &hclge_dbg_ppp_reg[0],
59           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_ppp_reg),
60                        .offset = HCLGE_DBG_DFX_PPP_OFFSET,
61                        .cmd = HCLGE_OPC_DFX_PPP_REG } },
62         { .reg_type = "rcb",
63           .dfx_msg = &hclge_dbg_rcb_reg[0],
64           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_rcb_reg),
65                        .offset = HCLGE_DBG_DFX_RCB_OFFSET,
66                        .cmd = HCLGE_OPC_DFX_RCB_REG } },
67         { .reg_type = "tqp",
68           .dfx_msg = &hclge_dbg_tqp_reg[0],
69           .reg_msg = { .msg_num = ARRAY_SIZE(hclge_dbg_tqp_reg),
70                        .offset = HCLGE_DBG_DFX_TQP_OFFSET,
71                        .cmd = HCLGE_OPC_DFX_TQP_REG } },
72 };
73
74 static int hclge_dbg_get_dfx_bd_num(struct hclge_dev *hdev, int offset)
75 {
76         struct hclge_desc desc[HCLGE_GET_DFX_REG_TYPE_CNT];
77         int entries_per_desc;
78         int index;
79         int ret;
80
81         ret = hclge_query_bd_num_cmd_send(hdev, desc);
82         if (ret) {
83                 dev_err(&hdev->pdev->dev,
84                         "get dfx bdnum fail, ret = %d\n", ret);
85                 return ret;
86         }
87
88         entries_per_desc = ARRAY_SIZE(desc[0].data);
89         index = offset % entries_per_desc;
90         return (int)desc[offset / entries_per_desc].data[index];
91 }
92
93 static int hclge_dbg_cmd_send(struct hclge_dev *hdev,
94                               struct hclge_desc *desc_src,
95                               int index, int bd_num,
96                               enum hclge_opcode_type cmd)
97 {
98         struct hclge_desc *desc = desc_src;
99         int ret, i;
100
101         hclge_cmd_setup_basic_desc(desc, cmd, true);
102         desc->data[0] = cpu_to_le32(index);
103
104         for (i = 1; i < bd_num; i++) {
105                 desc->flag |= cpu_to_le16(HCLGE_CMD_FLAG_NEXT);
106                 desc++;
107                 hclge_cmd_setup_basic_desc(desc, cmd, true);
108         }
109
110         ret = hclge_cmd_send(&hdev->hw, desc_src, bd_num);
111         if (ret)
112                 dev_err(&hdev->pdev->dev,
113                         "cmd(0x%x) send fail, ret = %d\n", cmd, ret);
114         return ret;
115 }
116
117 static void hclge_dbg_dump_reg_common(struct hclge_dev *hdev,
118                                       struct hclge_dbg_reg_type_info *reg_info,
119                                       const char *cmd_buf)
120 {
121 #define IDX_OFFSET      1
122
123         const char *s = &cmd_buf[strlen(reg_info->reg_type) + IDX_OFFSET];
124         struct hclge_dbg_dfx_message *dfx_message = reg_info->dfx_msg;
125         struct hclge_dbg_reg_common_msg *reg_msg = &reg_info->reg_msg;
126         struct hclge_desc *desc_src;
127         struct hclge_desc *desc;
128         int entries_per_desc;
129         int bd_num, buf_len;
130         int index = 0;
131         int min_num;
132         int ret, i;
133
134         if (*s) {
135                 ret = kstrtouint(s, 0, &index);
136                 index = (ret != 0) ? 0 : index;
137         }
138
139         bd_num = hclge_dbg_get_dfx_bd_num(hdev, reg_msg->offset);
140         if (bd_num <= 0) {
141                 dev_err(&hdev->pdev->dev, "get cmd(%d) bd num(%d) failed\n",
142                         reg_msg->offset, bd_num);
143                 return;
144         }
145
146         buf_len = sizeof(struct hclge_desc) * bd_num;
147         desc_src = kzalloc(buf_len, GFP_KERNEL);
148         if (!desc_src) {
149                 dev_err(&hdev->pdev->dev, "call kzalloc failed\n");
150                 return;
151         }
152
153         desc = desc_src;
154         ret = hclge_dbg_cmd_send(hdev, desc, index, bd_num, reg_msg->cmd);
155         if (ret) {
156                 kfree(desc_src);
157                 return;
158         }
159
160         entries_per_desc = ARRAY_SIZE(desc->data);
161         min_num = min_t(int, bd_num * entries_per_desc, reg_msg->msg_num);
162
163         desc = desc_src;
164         for (i = 0; i < min_num; i++) {
165                 if (i > 0 && (i % entries_per_desc) == 0)
166                         desc++;
167                 if (dfx_message->flag)
168                         dev_info(&hdev->pdev->dev, "%s: 0x%x\n",
169                                  dfx_message->message,
170                                  le32_to_cpu(desc->data[i % entries_per_desc]));
171
172                 dfx_message++;
173         }
174
175         kfree(desc_src);
176 }
177
178 static void hclge_dbg_dump_dcb(struct hclge_dev *hdev, const char *cmd_buf)
179 {
180         struct device *dev = &hdev->pdev->dev;
181         struct hclge_dbg_bitmap_cmd *bitmap;
182         int rq_id, pri_id, qset_id;
183         int port_id, nq_id, pg_id;
184         struct hclge_desc desc[2];
185
186         int cnt, ret;
187
188         cnt = sscanf(cmd_buf, "%i %i %i %i %i %i",
189                      &port_id, &pri_id, &pg_id, &rq_id, &nq_id, &qset_id);
190         if (cnt != 6) {
191                 dev_err(&hdev->pdev->dev,
192                         "dump dcb: bad command parameter, cnt=%d\n", cnt);
193                 return;
194         }
195
196         ret = hclge_dbg_cmd_send(hdev, desc, qset_id, 1,
197                                  HCLGE_OPC_QSET_DFX_STS);
198         if (ret)
199                 return;
200
201         bitmap = (struct hclge_dbg_bitmap_cmd *)&desc[0].data[1];
202         dev_info(dev, "roce_qset_mask: 0x%x\n", bitmap->bit0);
203         dev_info(dev, "nic_qs_mask: 0x%x\n", bitmap->bit1);
204         dev_info(dev, "qs_shaping_pass: 0x%x\n", bitmap->bit2);
205         dev_info(dev, "qs_bp_sts: 0x%x\n", bitmap->bit3);
206
207         ret = hclge_dbg_cmd_send(hdev, desc, pri_id, 1, HCLGE_OPC_PRI_DFX_STS);
208         if (ret)
209                 return;
210
211         bitmap = (struct hclge_dbg_bitmap_cmd *)&desc[0].data[1];
212         dev_info(dev, "pri_mask: 0x%x\n", bitmap->bit0);
213         dev_info(dev, "pri_cshaping_pass: 0x%x\n", bitmap->bit1);
214         dev_info(dev, "pri_pshaping_pass: 0x%x\n", bitmap->bit2);
215
216         ret = hclge_dbg_cmd_send(hdev, desc, pg_id, 1, HCLGE_OPC_PG_DFX_STS);
217         if (ret)
218                 return;
219
220         bitmap = (struct hclge_dbg_bitmap_cmd *)&desc[0].data[1];
221         dev_info(dev, "pg_mask: 0x%x\n", bitmap->bit0);
222         dev_info(dev, "pg_cshaping_pass: 0x%x\n", bitmap->bit1);
223         dev_info(dev, "pg_pshaping_pass: 0x%x\n", bitmap->bit2);
224
225         ret = hclge_dbg_cmd_send(hdev, desc, port_id, 1,
226                                  HCLGE_OPC_PORT_DFX_STS);
227         if (ret)
228                 return;
229
230         bitmap = (struct hclge_dbg_bitmap_cmd *)&desc[0].data[1];
231         dev_info(dev, "port_mask: 0x%x\n", bitmap->bit0);
232         dev_info(dev, "port_shaping_pass: 0x%x\n", bitmap->bit1);
233
234         ret = hclge_dbg_cmd_send(hdev, desc, nq_id, 1, HCLGE_OPC_SCH_NQ_CNT);
235         if (ret)
236                 return;
237
238         dev_info(dev, "sch_nq_cnt: 0x%x\n", le32_to_cpu(desc[0].data[1]));
239
240         ret = hclge_dbg_cmd_send(hdev, desc, nq_id, 1, HCLGE_OPC_SCH_RQ_CNT);
241         if (ret)
242                 return;
243
244         dev_info(dev, "sch_rq_cnt: 0x%x\n", le32_to_cpu(desc[0].data[1]));
245
246         ret = hclge_dbg_cmd_send(hdev, desc, 0, 2, HCLGE_OPC_TM_INTERNAL_STS);
247         if (ret)
248                 return;
249
250         dev_info(dev, "pri_bp: 0x%x\n", le32_to_cpu(desc[0].data[1]));
251         dev_info(dev, "fifo_dfx_info: 0x%x\n", le32_to_cpu(desc[0].data[2]));
252         dev_info(dev, "sch_roce_fifo_afull_gap: 0x%x\n",
253                  le32_to_cpu(desc[0].data[3]));
254         dev_info(dev, "tx_private_waterline: 0x%x\n",
255                  le32_to_cpu(desc[0].data[4]));
256         dev_info(dev, "tm_bypass_en: 0x%x\n", le32_to_cpu(desc[0].data[5]));
257         dev_info(dev, "SSU_TM_BYPASS_EN: 0x%x\n", le32_to_cpu(desc[1].data[0]));
258         dev_info(dev, "SSU_RESERVE_CFG: 0x%x\n", le32_to_cpu(desc[1].data[1]));
259
260         ret = hclge_dbg_cmd_send(hdev, desc, port_id, 1,
261                                  HCLGE_OPC_TM_INTERNAL_CNT);
262         if (ret)
263                 return;
264
265         dev_info(dev, "SCH_NIC_NUM: 0x%x\n", le32_to_cpu(desc[0].data[1]));
266         dev_info(dev, "SCH_ROCE_NUM: 0x%x\n", le32_to_cpu(desc[0].data[2]));
267
268         ret = hclge_dbg_cmd_send(hdev, desc, port_id, 1,
269                                  HCLGE_OPC_TM_INTERNAL_STS_1);
270         if (ret)
271                 return;
272
273         dev_info(dev, "TC_MAP_SEL: 0x%x\n", le32_to_cpu(desc[0].data[1]));
274         dev_info(dev, "IGU_PFC_PRI_EN: 0x%x\n", le32_to_cpu(desc[0].data[2]));
275         dev_info(dev, "MAC_PFC_PRI_EN: 0x%x\n", le32_to_cpu(desc[0].data[3]));
276         dev_info(dev, "IGU_PRI_MAP_TC_CFG: 0x%x\n",
277                  le32_to_cpu(desc[0].data[4]));
278         dev_info(dev, "IGU_TX_PRI_MAP_TC_CFG: 0x%x\n",
279                  le32_to_cpu(desc[0].data[5]));
280 }
281
282 static void hclge_dbg_dump_reg_cmd(struct hclge_dev *hdev, const char *cmd_buf)
283 {
284         struct hclge_dbg_reg_type_info *reg_info;
285         bool has_dump = false;
286         int i;
287
288         for (i = 0; i < ARRAY_SIZE(hclge_dbg_reg_info); i++) {
289                 reg_info = &hclge_dbg_reg_info[i];
290                 if (!strncmp(cmd_buf, reg_info->reg_type,
291                              strlen(reg_info->reg_type))) {
292                         hclge_dbg_dump_reg_common(hdev, reg_info, cmd_buf);
293                         has_dump = true;
294                 }
295         }
296
297         if (strncmp(cmd_buf, "dcb", 3) == 0) {
298                 hclge_dbg_dump_dcb(hdev, &cmd_buf[sizeof("dcb")]);
299                 has_dump = true;
300         }
301
302         if (!has_dump) {
303                 dev_info(&hdev->pdev->dev, "unknown command\n");
304                 return;
305         }
306 }
307
308 static void hclge_title_idx_print(struct hclge_dev *hdev, bool flag, int index,
309                                   char *title_buf, char *true_buf,
310                                   char *false_buf)
311 {
312         if (flag)
313                 dev_info(&hdev->pdev->dev, "%s(%d): %s\n", title_buf, index,
314                          true_buf);
315         else
316                 dev_info(&hdev->pdev->dev, "%s(%d): %s\n", title_buf, index,
317                          false_buf);
318 }
319
320 static void hclge_dbg_dump_tc(struct hclge_dev *hdev)
321 {
322         struct hclge_ets_tc_weight_cmd *ets_weight;
323         struct hclge_desc desc;
324         int i, ret;
325
326         if (!hnae3_dev_dcb_supported(hdev)) {
327                 dev_info(&hdev->pdev->dev,
328                          "Only DCB-supported dev supports tc\n");
329                 return;
330         }
331
332         hclge_cmd_setup_basic_desc(&desc, HCLGE_OPC_ETS_TC_WEIGHT, true);
333
334         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
335         if (ret) {
336                 dev_err(&hdev->pdev->dev, "dump tc fail, ret = %d\n", ret);
337                 return;
338         }
339
340         ets_weight = (struct hclge_ets_tc_weight_cmd *)desc.data;
341
342         dev_info(&hdev->pdev->dev, "dump tc\n");
343         dev_info(&hdev->pdev->dev, "weight_offset: %u\n",
344                  ets_weight->weight_offset);
345
346         for (i = 0; i < HNAE3_MAX_TC; i++)
347                 hclge_title_idx_print(hdev, ets_weight->tc_weight[i], i,
348                                       "tc", "no sp mode", "sp mode");
349 }
350
351 static void hclge_dbg_dump_tm_pg(struct hclge_dev *hdev)
352 {
353         struct hclge_port_shapping_cmd *port_shap_cfg_cmd;
354         struct hclge_bp_to_qs_map_cmd *bp_to_qs_map_cmd;
355         struct hclge_pg_shapping_cmd *pg_shap_cfg_cmd;
356         enum hclge_opcode_type cmd;
357         struct hclge_desc desc;
358         int ret;
359
360         cmd = HCLGE_OPC_TM_PG_C_SHAPPING;
361         hclge_cmd_setup_basic_desc(&desc, cmd, true);
362         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
363         if (ret)
364                 goto err_tm_pg_cmd_send;
365
366         pg_shap_cfg_cmd = (struct hclge_pg_shapping_cmd *)desc.data;
367         dev_info(&hdev->pdev->dev, "PG_C pg_id: %u\n", pg_shap_cfg_cmd->pg_id);
368         dev_info(&hdev->pdev->dev, "PG_C pg_shapping: 0x%x\n",
369                  le32_to_cpu(pg_shap_cfg_cmd->pg_shapping_para));
370
371         cmd = HCLGE_OPC_TM_PG_P_SHAPPING;
372         hclge_cmd_setup_basic_desc(&desc, cmd, true);
373         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
374         if (ret)
375                 goto err_tm_pg_cmd_send;
376
377         pg_shap_cfg_cmd = (struct hclge_pg_shapping_cmd *)desc.data;
378         dev_info(&hdev->pdev->dev, "PG_P pg_id: %u\n", pg_shap_cfg_cmd->pg_id);
379         dev_info(&hdev->pdev->dev, "PG_P pg_shapping: 0x%x\n",
380                  le32_to_cpu(pg_shap_cfg_cmd->pg_shapping_para));
381
382         cmd = HCLGE_OPC_TM_PORT_SHAPPING;
383         hclge_cmd_setup_basic_desc(&desc, cmd, true);
384         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
385         if (ret)
386                 goto err_tm_pg_cmd_send;
387
388         port_shap_cfg_cmd = (struct hclge_port_shapping_cmd *)desc.data;
389         dev_info(&hdev->pdev->dev, "PORT port_shapping: 0x%x\n",
390                  le32_to_cpu(port_shap_cfg_cmd->port_shapping_para));
391
392         cmd = HCLGE_OPC_TM_PG_SCH_MODE_CFG;
393         hclge_cmd_setup_basic_desc(&desc, cmd, true);
394         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
395         if (ret)
396                 goto err_tm_pg_cmd_send;
397
398         dev_info(&hdev->pdev->dev, "PG_SCH pg_id: %u\n",
399                  le32_to_cpu(desc.data[0]));
400
401         cmd = HCLGE_OPC_TM_PRI_SCH_MODE_CFG;
402         hclge_cmd_setup_basic_desc(&desc, cmd, true);
403         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
404         if (ret)
405                 goto err_tm_pg_cmd_send;
406
407         dev_info(&hdev->pdev->dev, "PRI_SCH pri_id: %u\n",
408                  le32_to_cpu(desc.data[0]));
409
410         cmd = HCLGE_OPC_TM_QS_SCH_MODE_CFG;
411         hclge_cmd_setup_basic_desc(&desc, cmd, true);
412         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
413         if (ret)
414                 goto err_tm_pg_cmd_send;
415
416         dev_info(&hdev->pdev->dev, "QS_SCH qs_id: %u\n",
417                  le32_to_cpu(desc.data[0]));
418
419         if (!hnae3_dev_dcb_supported(hdev)) {
420                 dev_info(&hdev->pdev->dev,
421                          "Only DCB-supported dev supports tm mapping\n");
422                 return;
423         }
424
425         cmd = HCLGE_OPC_TM_BP_TO_QSET_MAPPING;
426         hclge_cmd_setup_basic_desc(&desc, cmd, true);
427         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
428         if (ret)
429                 goto err_tm_pg_cmd_send;
430
431         bp_to_qs_map_cmd = (struct hclge_bp_to_qs_map_cmd *)desc.data;
432         dev_info(&hdev->pdev->dev, "BP_TO_QSET tc_id: %u\n",
433                  bp_to_qs_map_cmd->tc_id);
434         dev_info(&hdev->pdev->dev, "BP_TO_QSET qs_group_id: 0x%x\n",
435                  bp_to_qs_map_cmd->qs_group_id);
436         dev_info(&hdev->pdev->dev, "BP_TO_QSET qs_bit_map: 0x%x\n",
437                  le32_to_cpu(bp_to_qs_map_cmd->qs_bit_map));
438         return;
439
440 err_tm_pg_cmd_send:
441         dev_err(&hdev->pdev->dev, "dump tm_pg fail(0x%x), ret = %d\n",
442                 cmd, ret);
443 }
444
445 static void hclge_dbg_dump_tm(struct hclge_dev *hdev)
446 {
447         struct hclge_priority_weight_cmd *priority_weight;
448         struct hclge_pg_to_pri_link_cmd *pg_to_pri_map;
449         struct hclge_qs_to_pri_link_cmd *qs_to_pri_map;
450         struct hclge_nq_to_qs_link_cmd *nq_to_qs_map;
451         struct hclge_pri_shapping_cmd *shap_cfg_cmd;
452         struct hclge_pg_weight_cmd *pg_weight;
453         struct hclge_qs_weight_cmd *qs_weight;
454         enum hclge_opcode_type cmd;
455         struct hclge_desc desc;
456         int ret;
457
458         cmd = HCLGE_OPC_TM_PG_TO_PRI_LINK;
459         hclge_cmd_setup_basic_desc(&desc, cmd, true);
460         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
461         if (ret)
462                 goto err_tm_cmd_send;
463
464         pg_to_pri_map = (struct hclge_pg_to_pri_link_cmd *)desc.data;
465         dev_info(&hdev->pdev->dev, "dump tm\n");
466         dev_info(&hdev->pdev->dev, "PG_TO_PRI gp_id: %u\n",
467                  pg_to_pri_map->pg_id);
468         dev_info(&hdev->pdev->dev, "PG_TO_PRI map: 0x%x\n",
469                  pg_to_pri_map->pri_bit_map);
470
471         cmd = HCLGE_OPC_TM_QS_TO_PRI_LINK;
472         hclge_cmd_setup_basic_desc(&desc, cmd, true);
473         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
474         if (ret)
475                 goto err_tm_cmd_send;
476
477         qs_to_pri_map = (struct hclge_qs_to_pri_link_cmd *)desc.data;
478         dev_info(&hdev->pdev->dev, "QS_TO_PRI qs_id: %u\n",
479                  le16_to_cpu(qs_to_pri_map->qs_id));
480         dev_info(&hdev->pdev->dev, "QS_TO_PRI priority: %u\n",
481                  qs_to_pri_map->priority);
482         dev_info(&hdev->pdev->dev, "QS_TO_PRI link_vld: %u\n",
483                  qs_to_pri_map->link_vld);
484
485         cmd = HCLGE_OPC_TM_NQ_TO_QS_LINK;
486         hclge_cmd_setup_basic_desc(&desc, cmd, true);
487         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
488         if (ret)
489                 goto err_tm_cmd_send;
490
491         nq_to_qs_map = (struct hclge_nq_to_qs_link_cmd *)desc.data;
492         dev_info(&hdev->pdev->dev, "NQ_TO_QS nq_id: %u\n",
493                  le16_to_cpu(nq_to_qs_map->nq_id));
494         dev_info(&hdev->pdev->dev, "NQ_TO_QS qset_id: 0x%x\n",
495                  le16_to_cpu(nq_to_qs_map->qset_id));
496
497         cmd = HCLGE_OPC_TM_PG_WEIGHT;
498         hclge_cmd_setup_basic_desc(&desc, cmd, true);
499         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
500         if (ret)
501                 goto err_tm_cmd_send;
502
503         pg_weight = (struct hclge_pg_weight_cmd *)desc.data;
504         dev_info(&hdev->pdev->dev, "PG pg_id: %u\n", pg_weight->pg_id);
505         dev_info(&hdev->pdev->dev, "PG dwrr: %u\n", pg_weight->dwrr);
506
507         cmd = HCLGE_OPC_TM_QS_WEIGHT;
508         hclge_cmd_setup_basic_desc(&desc, cmd, true);
509         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
510         if (ret)
511                 goto err_tm_cmd_send;
512
513         qs_weight = (struct hclge_qs_weight_cmd *)desc.data;
514         dev_info(&hdev->pdev->dev, "QS qs_id: %u\n",
515                  le16_to_cpu(qs_weight->qs_id));
516         dev_info(&hdev->pdev->dev, "QS dwrr: %u\n", qs_weight->dwrr);
517
518         cmd = HCLGE_OPC_TM_PRI_WEIGHT;
519         hclge_cmd_setup_basic_desc(&desc, cmd, true);
520         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
521         if (ret)
522                 goto err_tm_cmd_send;
523
524         priority_weight = (struct hclge_priority_weight_cmd *)desc.data;
525         dev_info(&hdev->pdev->dev, "PRI pri_id: %u\n", priority_weight->pri_id);
526         dev_info(&hdev->pdev->dev, "PRI dwrr: %u\n", priority_weight->dwrr);
527
528         cmd = HCLGE_OPC_TM_PRI_C_SHAPPING;
529         hclge_cmd_setup_basic_desc(&desc, cmd, true);
530         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
531         if (ret)
532                 goto err_tm_cmd_send;
533
534         shap_cfg_cmd = (struct hclge_pri_shapping_cmd *)desc.data;
535         dev_info(&hdev->pdev->dev, "PRI_C pri_id: %u\n", shap_cfg_cmd->pri_id);
536         dev_info(&hdev->pdev->dev, "PRI_C pri_shapping: 0x%x\n",
537                  le32_to_cpu(shap_cfg_cmd->pri_shapping_para));
538
539         cmd = HCLGE_OPC_TM_PRI_P_SHAPPING;
540         hclge_cmd_setup_basic_desc(&desc, cmd, true);
541         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
542         if (ret)
543                 goto err_tm_cmd_send;
544
545         shap_cfg_cmd = (struct hclge_pri_shapping_cmd *)desc.data;
546         dev_info(&hdev->pdev->dev, "PRI_P pri_id: %u\n", shap_cfg_cmd->pri_id);
547         dev_info(&hdev->pdev->dev, "PRI_P pri_shapping: 0x%x\n",
548                  le32_to_cpu(shap_cfg_cmd->pri_shapping_para));
549
550         hclge_dbg_dump_tm_pg(hdev);
551
552         return;
553
554 err_tm_cmd_send:
555         dev_err(&hdev->pdev->dev, "dump tm fail(0x%x), ret = %d\n",
556                 cmd, ret);
557 }
558
559 static void hclge_dbg_dump_tm_map(struct hclge_dev *hdev,
560                                   const char *cmd_buf)
561 {
562         struct hclge_bp_to_qs_map_cmd *bp_to_qs_map_cmd;
563         struct hclge_nq_to_qs_link_cmd *nq_to_qs_map;
564         struct hclge_qs_to_pri_link_cmd *map;
565         struct hclge_tqp_tx_queue_tc_cmd *tc;
566         enum hclge_opcode_type cmd;
567         struct hclge_desc desc;
568         int queue_id, group_id;
569         u32 qset_maping[32];
570         int tc_id, qset_id;
571         int pri_id, ret;
572         u32 i;
573
574         ret = kstrtouint(cmd_buf, 0, &queue_id);
575         queue_id = (ret != 0) ? 0 : queue_id;
576
577         cmd = HCLGE_OPC_TM_NQ_TO_QS_LINK;
578         nq_to_qs_map = (struct hclge_nq_to_qs_link_cmd *)desc.data;
579         hclge_cmd_setup_basic_desc(&desc, cmd, true);
580         nq_to_qs_map->nq_id = cpu_to_le16(queue_id);
581         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
582         if (ret)
583                 goto err_tm_map_cmd_send;
584         qset_id = nq_to_qs_map->qset_id & 0x3FF;
585
586         cmd = HCLGE_OPC_TM_QS_TO_PRI_LINK;
587         map = (struct hclge_qs_to_pri_link_cmd *)desc.data;
588         hclge_cmd_setup_basic_desc(&desc, cmd, true);
589         map->qs_id = cpu_to_le16(qset_id);
590         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
591         if (ret)
592                 goto err_tm_map_cmd_send;
593         pri_id = map->priority;
594
595         cmd = HCLGE_OPC_TQP_TX_QUEUE_TC;
596         tc = (struct hclge_tqp_tx_queue_tc_cmd *)desc.data;
597         hclge_cmd_setup_basic_desc(&desc, cmd, true);
598         tc->queue_id = cpu_to_le16(queue_id);
599         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
600         if (ret)
601                 goto err_tm_map_cmd_send;
602         tc_id = tc->tc_id & 0x7;
603
604         dev_info(&hdev->pdev->dev, "queue_id | qset_id | pri_id | tc_id\n");
605         dev_info(&hdev->pdev->dev, "%04d     | %04d    | %02d     | %02d\n",
606                  queue_id, qset_id, pri_id, tc_id);
607
608         if (!hnae3_dev_dcb_supported(hdev)) {
609                 dev_info(&hdev->pdev->dev,
610                          "Only DCB-supported dev supports tm mapping\n");
611                 return;
612         }
613
614         cmd = HCLGE_OPC_TM_BP_TO_QSET_MAPPING;
615         bp_to_qs_map_cmd = (struct hclge_bp_to_qs_map_cmd *)desc.data;
616         for (group_id = 0; group_id < 32; group_id++) {
617                 hclge_cmd_setup_basic_desc(&desc, cmd, true);
618                 bp_to_qs_map_cmd->tc_id = tc_id;
619                 bp_to_qs_map_cmd->qs_group_id = group_id;
620                 ret = hclge_cmd_send(&hdev->hw, &desc, 1);
621                 if (ret)
622                         goto err_tm_map_cmd_send;
623
624                 qset_maping[group_id] = bp_to_qs_map_cmd->qs_bit_map;
625         }
626
627         dev_info(&hdev->pdev->dev, "index | tm bp qset maping:\n");
628
629         i = 0;
630         for (group_id = 0; group_id < 4; group_id++) {
631                 dev_info(&hdev->pdev->dev,
632                          "%04d  | %08x:%08x:%08x:%08x:%08x:%08x:%08x:%08x\n",
633                          group_id * 256, qset_maping[(u32)(i + 7)],
634                          qset_maping[(u32)(i + 6)], qset_maping[(u32)(i + 5)],
635                          qset_maping[(u32)(i + 4)], qset_maping[(u32)(i + 3)],
636                          qset_maping[(u32)(i + 2)], qset_maping[(u32)(i + 1)],
637                          qset_maping[i]);
638                 i += 8;
639         }
640
641         return;
642
643 err_tm_map_cmd_send:
644         dev_err(&hdev->pdev->dev, "dump tqp map fail(0x%x), ret = %d\n",
645                 cmd, ret);
646 }
647
648 static void hclge_dbg_dump_qos_pause_cfg(struct hclge_dev *hdev)
649 {
650         struct hclge_cfg_pause_param_cmd *pause_param;
651         struct hclge_desc desc;
652         int ret;
653
654         hclge_cmd_setup_basic_desc(&desc, HCLGE_OPC_CFG_MAC_PARA, true);
655
656         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
657         if (ret) {
658                 dev_err(&hdev->pdev->dev, "dump checksum fail, ret = %d\n",
659                         ret);
660                 return;
661         }
662
663         pause_param = (struct hclge_cfg_pause_param_cmd *)desc.data;
664         dev_info(&hdev->pdev->dev, "dump qos pause cfg\n");
665         dev_info(&hdev->pdev->dev, "pause_trans_gap: 0x%x\n",
666                  pause_param->pause_trans_gap);
667         dev_info(&hdev->pdev->dev, "pause_trans_time: 0x%x\n",
668                  le16_to_cpu(pause_param->pause_trans_time));
669 }
670
671 static void hclge_dbg_dump_qos_pri_map(struct hclge_dev *hdev)
672 {
673         struct hclge_qos_pri_map_cmd *pri_map;
674         struct hclge_desc desc;
675         int ret;
676
677         hclge_cmd_setup_basic_desc(&desc, HCLGE_OPC_PRI_TO_TC_MAPPING, true);
678
679         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
680         if (ret) {
681                 dev_err(&hdev->pdev->dev,
682                         "dump qos pri map fail, ret = %d\n", ret);
683                 return;
684         }
685
686         pri_map = (struct hclge_qos_pri_map_cmd *)desc.data;
687         dev_info(&hdev->pdev->dev, "dump qos pri map\n");
688         dev_info(&hdev->pdev->dev, "vlan_to_pri: 0x%x\n", pri_map->vlan_pri);
689         dev_info(&hdev->pdev->dev, "pri_0_to_tc: 0x%x\n", pri_map->pri0_tc);
690         dev_info(&hdev->pdev->dev, "pri_1_to_tc: 0x%x\n", pri_map->pri1_tc);
691         dev_info(&hdev->pdev->dev, "pri_2_to_tc: 0x%x\n", pri_map->pri2_tc);
692         dev_info(&hdev->pdev->dev, "pri_3_to_tc: 0x%x\n", pri_map->pri3_tc);
693         dev_info(&hdev->pdev->dev, "pri_4_to_tc: 0x%x\n", pri_map->pri4_tc);
694         dev_info(&hdev->pdev->dev, "pri_5_to_tc: 0x%x\n", pri_map->pri5_tc);
695         dev_info(&hdev->pdev->dev, "pri_6_to_tc: 0x%x\n", pri_map->pri6_tc);
696         dev_info(&hdev->pdev->dev, "pri_7_to_tc: 0x%x\n", pri_map->pri7_tc);
697 }
698
699 static void hclge_dbg_dump_qos_buf_cfg(struct hclge_dev *hdev)
700 {
701         struct hclge_tx_buff_alloc_cmd *tx_buf_cmd;
702         struct hclge_rx_priv_buff_cmd *rx_buf_cmd;
703         struct hclge_rx_priv_wl_buf *rx_priv_wl;
704         struct hclge_rx_com_wl *rx_packet_cnt;
705         struct hclge_rx_com_thrd *rx_com_thrd;
706         struct hclge_rx_com_wl *rx_com_wl;
707         enum hclge_opcode_type cmd;
708         struct hclge_desc desc[2];
709         int i, ret;
710
711         cmd = HCLGE_OPC_TX_BUFF_ALLOC;
712         hclge_cmd_setup_basic_desc(desc, cmd, true);
713         ret = hclge_cmd_send(&hdev->hw, desc, 1);
714         if (ret)
715                 goto err_qos_cmd_send;
716
717         dev_info(&hdev->pdev->dev, "dump qos buf cfg\n");
718
719         tx_buf_cmd = (struct hclge_tx_buff_alloc_cmd *)desc[0].data;
720         for (i = 0; i < HCLGE_MAX_TC_NUM; i++)
721                 dev_info(&hdev->pdev->dev, "tx_packet_buf_tc_%d: 0x%x\n", i,
722                          le16_to_cpu(tx_buf_cmd->tx_pkt_buff[i]));
723
724         cmd = HCLGE_OPC_RX_PRIV_BUFF_ALLOC;
725         hclge_cmd_setup_basic_desc(desc, cmd, true);
726         ret = hclge_cmd_send(&hdev->hw, desc, 1);
727         if (ret)
728                 goto err_qos_cmd_send;
729
730         dev_info(&hdev->pdev->dev, "\n");
731         rx_buf_cmd = (struct hclge_rx_priv_buff_cmd *)desc[0].data;
732         for (i = 0; i < HCLGE_MAX_TC_NUM; i++)
733                 dev_info(&hdev->pdev->dev, "rx_packet_buf_tc_%d: 0x%x\n", i,
734                          le16_to_cpu(rx_buf_cmd->buf_num[i]));
735
736         dev_info(&hdev->pdev->dev, "rx_share_buf: 0x%x\n",
737                  le16_to_cpu(rx_buf_cmd->shared_buf));
738
739         cmd = HCLGE_OPC_RX_COM_WL_ALLOC;
740         hclge_cmd_setup_basic_desc(desc, cmd, true);
741         ret = hclge_cmd_send(&hdev->hw, desc, 1);
742         if (ret)
743                 goto err_qos_cmd_send;
744
745         rx_com_wl = (struct hclge_rx_com_wl *)desc[0].data;
746         dev_info(&hdev->pdev->dev, "\n");
747         dev_info(&hdev->pdev->dev, "rx_com_wl: high: 0x%x, low: 0x%x\n",
748                  le16_to_cpu(rx_com_wl->com_wl.high),
749                  le16_to_cpu(rx_com_wl->com_wl.low));
750
751         cmd = HCLGE_OPC_RX_GBL_PKT_CNT;
752         hclge_cmd_setup_basic_desc(desc, cmd, true);
753         ret = hclge_cmd_send(&hdev->hw, desc, 1);
754         if (ret)
755                 goto err_qos_cmd_send;
756
757         rx_packet_cnt = (struct hclge_rx_com_wl *)desc[0].data;
758         dev_info(&hdev->pdev->dev,
759                  "rx_global_packet_cnt: high: 0x%x, low: 0x%x\n",
760                  le16_to_cpu(rx_packet_cnt->com_wl.high),
761                  le16_to_cpu(rx_packet_cnt->com_wl.low));
762         dev_info(&hdev->pdev->dev, "\n");
763
764         if (!hnae3_dev_dcb_supported(hdev)) {
765                 dev_info(&hdev->pdev->dev,
766                          "Only DCB-supported dev supports rx priv wl\n");
767                 return;
768         }
769         cmd = HCLGE_OPC_RX_PRIV_WL_ALLOC;
770         hclge_cmd_setup_basic_desc(&desc[0], cmd, true);
771         desc[0].flag |= cpu_to_le16(HCLGE_CMD_FLAG_NEXT);
772         hclge_cmd_setup_basic_desc(&desc[1], cmd, true);
773         ret = hclge_cmd_send(&hdev->hw, desc, 2);
774         if (ret)
775                 goto err_qos_cmd_send;
776
777         rx_priv_wl = (struct hclge_rx_priv_wl_buf *)desc[0].data;
778         for (i = 0; i < HCLGE_TC_NUM_ONE_DESC; i++)
779                 dev_info(&hdev->pdev->dev,
780                          "rx_priv_wl_tc_%d: high: 0x%x, low: 0x%x\n", i,
781                          le16_to_cpu(rx_priv_wl->tc_wl[i].high),
782                          le16_to_cpu(rx_priv_wl->tc_wl[i].low));
783
784         rx_priv_wl = (struct hclge_rx_priv_wl_buf *)desc[1].data;
785         for (i = 0; i < HCLGE_TC_NUM_ONE_DESC; i++)
786                 dev_info(&hdev->pdev->dev,
787                          "rx_priv_wl_tc_%d: high: 0x%x, low: 0x%x\n",
788                          i + HCLGE_TC_NUM_ONE_DESC,
789                          le16_to_cpu(rx_priv_wl->tc_wl[i].high),
790                          le16_to_cpu(rx_priv_wl->tc_wl[i].low));
791
792         cmd = HCLGE_OPC_RX_COM_THRD_ALLOC;
793         hclge_cmd_setup_basic_desc(&desc[0], cmd, true);
794         desc[0].flag |= cpu_to_le16(HCLGE_CMD_FLAG_NEXT);
795         hclge_cmd_setup_basic_desc(&desc[1], cmd, true);
796         ret = hclge_cmd_send(&hdev->hw, desc, 2);
797         if (ret)
798                 goto err_qos_cmd_send;
799
800         dev_info(&hdev->pdev->dev, "\n");
801         rx_com_thrd = (struct hclge_rx_com_thrd *)desc[0].data;
802         for (i = 0; i < HCLGE_TC_NUM_ONE_DESC; i++)
803                 dev_info(&hdev->pdev->dev,
804                          "rx_com_thrd_tc_%d: high: 0x%x, low: 0x%x\n", i,
805                          le16_to_cpu(rx_com_thrd->com_thrd[i].high),
806                          le16_to_cpu(rx_com_thrd->com_thrd[i].low));
807
808         rx_com_thrd = (struct hclge_rx_com_thrd *)desc[1].data;
809         for (i = 0; i < HCLGE_TC_NUM_ONE_DESC; i++)
810                 dev_info(&hdev->pdev->dev,
811                          "rx_com_thrd_tc_%d: high: 0x%x, low: 0x%x\n",
812                          i + HCLGE_TC_NUM_ONE_DESC,
813                          le16_to_cpu(rx_com_thrd->com_thrd[i].high),
814                          le16_to_cpu(rx_com_thrd->com_thrd[i].low));
815         return;
816
817 err_qos_cmd_send:
818         dev_err(&hdev->pdev->dev,
819                 "dump qos buf cfg fail(0x%x), ret = %d\n", cmd, ret);
820 }
821
822 static void hclge_dbg_dump_mng_table(struct hclge_dev *hdev)
823 {
824         struct hclge_mac_ethertype_idx_rd_cmd *req0;
825         char printf_buf[HCLGE_DBG_BUF_LEN];
826         struct hclge_desc desc;
827         int ret, i;
828
829         dev_info(&hdev->pdev->dev, "mng tab:\n");
830         memset(printf_buf, 0, HCLGE_DBG_BUF_LEN);
831         strncat(printf_buf,
832                 "entry|mac_addr         |mask|ether|mask|vlan|mask",
833                 HCLGE_DBG_BUF_LEN - 1);
834         strncat(printf_buf + strlen(printf_buf),
835                 "|i_map|i_dir|e_type|pf_id|vf_id|q_id|drop\n",
836                 HCLGE_DBG_BUF_LEN - strlen(printf_buf) - 1);
837
838         dev_info(&hdev->pdev->dev, "%s", printf_buf);
839
840         for (i = 0; i < HCLGE_DBG_MNG_TBL_MAX; i++) {
841                 hclge_cmd_setup_basic_desc(&desc, HCLGE_MAC_ETHERTYPE_IDX_RD,
842                                            true);
843                 req0 = (struct hclge_mac_ethertype_idx_rd_cmd *)&desc.data;
844                 req0->index = cpu_to_le16(i);
845
846                 ret = hclge_cmd_send(&hdev->hw, &desc, 1);
847                 if (ret) {
848                         dev_err(&hdev->pdev->dev,
849                                 "call hclge_cmd_send fail, ret = %d\n", ret);
850                         return;
851                 }
852
853                 if (!req0->resp_code)
854                         continue;
855
856                 memset(printf_buf, 0, HCLGE_DBG_BUF_LEN);
857                 snprintf(printf_buf, HCLGE_DBG_BUF_LEN,
858                          "%02u   |%02x:%02x:%02x:%02x:%02x:%02x|",
859                          le16_to_cpu(req0->index),
860                          req0->mac_addr[0], req0->mac_addr[1],
861                          req0->mac_addr[2], req0->mac_addr[3],
862                          req0->mac_addr[4], req0->mac_addr[5]);
863
864                 snprintf(printf_buf + strlen(printf_buf),
865                          HCLGE_DBG_BUF_LEN - strlen(printf_buf),
866                          "%x   |%04x |%x   |%04x|%x   |%02x   |%02x   |",
867                          !!(req0->flags & HCLGE_DBG_MNG_MAC_MASK_B),
868                          req0->ethter_type,
869                          !!(req0->flags & HCLGE_DBG_MNG_ETHER_MASK_B),
870                          req0->vlan_tag & HCLGE_DBG_MNG_VLAN_TAG,
871                          !!(req0->flags & HCLGE_DBG_MNG_VLAN_MASK_B),
872                          req0->i_port_bitmap, req0->i_port_direction);
873
874                 snprintf(printf_buf + strlen(printf_buf),
875                          HCLGE_DBG_BUF_LEN - strlen(printf_buf),
876                          "%d     |%d    |%02d   |%04d|%x\n",
877                          !!(req0->egress_port & HCLGE_DBG_MNG_E_TYPE_B),
878                          req0->egress_port & HCLGE_DBG_MNG_PF_ID,
879                          (req0->egress_port >> 3) & HCLGE_DBG_MNG_VF_ID,
880                          req0->egress_queue,
881                          !!(req0->egress_port & HCLGE_DBG_MNG_DROP_B));
882
883                 dev_info(&hdev->pdev->dev, "%s", printf_buf);
884         }
885 }
886
887 static int hclge_dbg_fd_tcam_read(struct hclge_dev *hdev, u8 stage,
888                                   bool sel_x, u32 loc)
889 {
890         struct hclge_fd_tcam_config_1_cmd *req1;
891         struct hclge_fd_tcam_config_2_cmd *req2;
892         struct hclge_fd_tcam_config_3_cmd *req3;
893         struct hclge_desc desc[3];
894         int ret, i;
895         u32 *req;
896
897         hclge_cmd_setup_basic_desc(&desc[0], HCLGE_OPC_FD_TCAM_OP, true);
898         desc[0].flag |= cpu_to_le16(HCLGE_CMD_FLAG_NEXT);
899         hclge_cmd_setup_basic_desc(&desc[1], HCLGE_OPC_FD_TCAM_OP, true);
900         desc[1].flag |= cpu_to_le16(HCLGE_CMD_FLAG_NEXT);
901         hclge_cmd_setup_basic_desc(&desc[2], HCLGE_OPC_FD_TCAM_OP, true);
902
903         req1 = (struct hclge_fd_tcam_config_1_cmd *)desc[0].data;
904         req2 = (struct hclge_fd_tcam_config_2_cmd *)desc[1].data;
905         req3 = (struct hclge_fd_tcam_config_3_cmd *)desc[2].data;
906
907         req1->stage  = stage;
908         req1->xy_sel = sel_x ? 1 : 0;
909         req1->index  = cpu_to_le32(loc);
910
911         ret = hclge_cmd_send(&hdev->hw, desc, 3);
912         if (ret)
913                 return ret;
914
915         dev_info(&hdev->pdev->dev, " read result tcam key %s(%u):\n",
916                  sel_x ? "x" : "y", loc);
917
918         /* tcam_data0 ~ tcam_data1 */
919         req = (u32 *)req1->tcam_data;
920         for (i = 0; i < 2; i++)
921                 dev_info(&hdev->pdev->dev, "%08x\n", *req++);
922
923         /* tcam_data2 ~ tcam_data7 */
924         req = (u32 *)req2->tcam_data;
925         for (i = 0; i < 6; i++)
926                 dev_info(&hdev->pdev->dev, "%08x\n", *req++);
927
928         /* tcam_data8 ~ tcam_data12 */
929         req = (u32 *)req3->tcam_data;
930         for (i = 0; i < 5; i++)
931                 dev_info(&hdev->pdev->dev, "%08x\n", *req++);
932
933         return ret;
934 }
935
936 static int hclge_dbg_get_rules_location(struct hclge_dev *hdev, u16 *rule_locs)
937 {
938         struct hclge_fd_rule *rule;
939         struct hlist_node *node;
940         int cnt = 0;
941
942         spin_lock_bh(&hdev->fd_rule_lock);
943         hlist_for_each_entry_safe(rule, node, &hdev->fd_rule_list, rule_node) {
944                 rule_locs[cnt] = rule->location;
945                 cnt++;
946         }
947         spin_unlock_bh(&hdev->fd_rule_lock);
948
949         if (cnt != hdev->hclge_fd_rule_num)
950                 return -EINVAL;
951
952         return cnt;
953 }
954
955 static void hclge_dbg_fd_tcam(struct hclge_dev *hdev)
956 {
957         int i, ret, rule_cnt;
958         u16 *rule_locs;
959
960         if (!hnae3_dev_fd_supported(hdev)) {
961                 dev_err(&hdev->pdev->dev,
962                         "Only FD-supported dev supports dump fd tcam\n");
963                 return;
964         }
965
966         if (!hdev->hclge_fd_rule_num ||
967             !hdev->fd_cfg.rule_num[HCLGE_FD_STAGE_1])
968                 return;
969
970         rule_locs = kcalloc(hdev->fd_cfg.rule_num[HCLGE_FD_STAGE_1],
971                             sizeof(u16), GFP_KERNEL);
972         if (!rule_locs)
973                 return;
974
975         rule_cnt = hclge_dbg_get_rules_location(hdev, rule_locs);
976         if (rule_cnt <= 0) {
977                 dev_err(&hdev->pdev->dev,
978                         "failed to get rule number, ret = %d\n", rule_cnt);
979                 kfree(rule_locs);
980                 return;
981         }
982
983         for (i = 0; i < rule_cnt; i++) {
984                 ret = hclge_dbg_fd_tcam_read(hdev, 0, true, rule_locs[i]);
985                 if (ret) {
986                         dev_err(&hdev->pdev->dev,
987                                 "failed to get fd tcam key x, ret = %d\n", ret);
988                         kfree(rule_locs);
989                         return;
990                 }
991
992                 ret = hclge_dbg_fd_tcam_read(hdev, 0, false, rule_locs[i]);
993                 if (ret) {
994                         dev_err(&hdev->pdev->dev,
995                                 "failed to get fd tcam key y, ret = %d\n", ret);
996                         kfree(rule_locs);
997                         return;
998                 }
999         }
1000
1001         kfree(rule_locs);
1002 }
1003
1004 void hclge_dbg_dump_rst_info(struct hclge_dev *hdev)
1005 {
1006         dev_info(&hdev->pdev->dev, "PF reset count: %u\n",
1007                  hdev->rst_stats.pf_rst_cnt);
1008         dev_info(&hdev->pdev->dev, "FLR reset count: %u\n",
1009                  hdev->rst_stats.flr_rst_cnt);
1010         dev_info(&hdev->pdev->dev, "GLOBAL reset count: %u\n",
1011                  hdev->rst_stats.global_rst_cnt);
1012         dev_info(&hdev->pdev->dev, "IMP reset count: %u\n",
1013                  hdev->rst_stats.imp_rst_cnt);
1014         dev_info(&hdev->pdev->dev, "reset done count: %u\n",
1015                  hdev->rst_stats.reset_done_cnt);
1016         dev_info(&hdev->pdev->dev, "HW reset done count: %u\n",
1017                  hdev->rst_stats.hw_reset_done_cnt);
1018         dev_info(&hdev->pdev->dev, "reset count: %u\n",
1019                  hdev->rst_stats.reset_cnt);
1020         dev_info(&hdev->pdev->dev, "reset fail count: %u\n",
1021                  hdev->rst_stats.reset_fail_cnt);
1022         dev_info(&hdev->pdev->dev, "vector0 interrupt enable status: 0x%x\n",
1023                  hclge_read_dev(&hdev->hw, HCLGE_MISC_VECTOR_REG_BASE));
1024         dev_info(&hdev->pdev->dev, "reset interrupt source: 0x%x\n",
1025                  hclge_read_dev(&hdev->hw, HCLGE_MISC_RESET_STS_REG));
1026         dev_info(&hdev->pdev->dev, "reset interrupt status: 0x%x\n",
1027                  hclge_read_dev(&hdev->hw, HCLGE_MISC_VECTOR_INT_STS));
1028         dev_info(&hdev->pdev->dev, "hardware reset status: 0x%x\n",
1029                  hclge_read_dev(&hdev->hw, HCLGE_GLOBAL_RESET_REG));
1030         dev_info(&hdev->pdev->dev, "handshake status: 0x%x\n",
1031                  hclge_read_dev(&hdev->hw, HCLGE_NIC_CSQ_DEPTH_REG));
1032         dev_info(&hdev->pdev->dev, "function reset status: 0x%x\n",
1033                  hclge_read_dev(&hdev->hw, HCLGE_FUN_RST_ING));
1034         dev_info(&hdev->pdev->dev, "hdev state: 0x%lx\n", hdev->state);
1035 }
1036
1037 static void hclge_dbg_dump_serv_info(struct hclge_dev *hdev)
1038 {
1039         dev_info(&hdev->pdev->dev, "last_serv_processed: %lu\n",
1040                  hdev->last_serv_processed);
1041         dev_info(&hdev->pdev->dev, "last_serv_cnt: %lu\n",
1042                  hdev->serv_processed_cnt);
1043 }
1044
1045 static void hclge_dbg_get_m7_stats_info(struct hclge_dev *hdev)
1046 {
1047         struct hclge_desc *desc_src, *desc_tmp;
1048         struct hclge_get_m7_bd_cmd *req;
1049         struct hclge_desc desc;
1050         u32 bd_num, buf_len;
1051         int ret, i;
1052
1053         hclge_cmd_setup_basic_desc(&desc, HCLGE_OPC_M7_STATS_BD, true);
1054
1055         req = (struct hclge_get_m7_bd_cmd *)desc.data;
1056         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
1057         if (ret) {
1058                 dev_err(&hdev->pdev->dev,
1059                         "get firmware statistics bd number failed, ret = %d\n",
1060                         ret);
1061                 return;
1062         }
1063
1064         bd_num = le32_to_cpu(req->bd_num);
1065
1066         buf_len  = sizeof(struct hclge_desc) * bd_num;
1067         desc_src = kzalloc(buf_len, GFP_KERNEL);
1068         if (!desc_src) {
1069                 dev_err(&hdev->pdev->dev,
1070                         "allocate desc for get_m7_stats failed\n");
1071                 return;
1072         }
1073
1074         desc_tmp = desc_src;
1075         ret  = hclge_dbg_cmd_send(hdev, desc_tmp, 0, bd_num,
1076                                   HCLGE_OPC_M7_STATS_INFO);
1077         if (ret) {
1078                 kfree(desc_src);
1079                 dev_err(&hdev->pdev->dev,
1080                         "get firmware statistics failed, ret = %d\n", ret);
1081                 return;
1082         }
1083
1084         for (i = 0; i < bd_num; i++) {
1085                 dev_info(&hdev->pdev->dev, "0x%08x  0x%08x  0x%08x\n",
1086                          le32_to_cpu(desc_tmp->data[0]),
1087                          le32_to_cpu(desc_tmp->data[1]),
1088                          le32_to_cpu(desc_tmp->data[2]));
1089                 dev_info(&hdev->pdev->dev, "0x%08x  0x%08x  0x%08x\n",
1090                          le32_to_cpu(desc_tmp->data[3]),
1091                          le32_to_cpu(desc_tmp->data[4]),
1092                          le32_to_cpu(desc_tmp->data[5]));
1093
1094                 desc_tmp++;
1095         }
1096
1097         kfree(desc_src);
1098 }
1099
1100 #define HCLGE_CMD_NCL_CONFIG_BD_NUM     5
1101
1102 static void hclge_ncl_config_data_print(struct hclge_dev *hdev,
1103                                         struct hclge_desc *desc, int *offset,
1104                                         int *length)
1105 {
1106 #define HCLGE_CMD_DATA_NUM              6
1107
1108         int i;
1109         int j;
1110
1111         for (i = 0; i < HCLGE_CMD_NCL_CONFIG_BD_NUM; i++) {
1112                 for (j = 0; j < HCLGE_CMD_DATA_NUM; j++) {
1113                         if (i == 0 && j == 0)
1114                                 continue;
1115
1116                         dev_info(&hdev->pdev->dev, "0x%04x | 0x%08x\n",
1117                                  *offset,
1118                                  le32_to_cpu(desc[i].data[j]));
1119                         *offset += sizeof(u32);
1120                         *length -= sizeof(u32);
1121                         if (*length <= 0)
1122                                 return;
1123                 }
1124         }
1125 }
1126
1127 /* hclge_dbg_dump_ncl_config: print specified range of NCL_CONFIG file
1128  * @hdev: pointer to struct hclge_dev
1129  * @cmd_buf: string that contains offset and length
1130  */
1131 static void hclge_dbg_dump_ncl_config(struct hclge_dev *hdev,
1132                                       const char *cmd_buf)
1133 {
1134 #define HCLGE_MAX_NCL_CONFIG_OFFSET     4096
1135 #define HCLGE_MAX_NCL_CONFIG_LENGTH     (20 + 24 * 4)
1136
1137         struct hclge_desc desc[HCLGE_CMD_NCL_CONFIG_BD_NUM];
1138         int bd_num = HCLGE_CMD_NCL_CONFIG_BD_NUM;
1139         int offset;
1140         int length;
1141         int data0;
1142         int ret;
1143
1144         ret = sscanf(cmd_buf, "%x %x", &offset, &length);
1145         if (ret != 2 || offset >= HCLGE_MAX_NCL_CONFIG_OFFSET ||
1146             length > HCLGE_MAX_NCL_CONFIG_OFFSET - offset) {
1147                 dev_err(&hdev->pdev->dev, "Invalid offset or length.\n");
1148                 return;
1149         }
1150         if (offset < 0 || length <= 0) {
1151                 dev_err(&hdev->pdev->dev, "Non-positive offset or length.\n");
1152                 return;
1153         }
1154
1155         dev_info(&hdev->pdev->dev, "offset |    data\n");
1156
1157         while (length > 0) {
1158                 data0 = offset;
1159                 if (length >= HCLGE_MAX_NCL_CONFIG_LENGTH)
1160                         data0 |= HCLGE_MAX_NCL_CONFIG_LENGTH << 16;
1161                 else
1162                         data0 |= length << 16;
1163                 ret = hclge_dbg_cmd_send(hdev, desc, data0, bd_num,
1164                                          HCLGE_OPC_QUERY_NCL_CONFIG);
1165                 if (ret)
1166                         return;
1167
1168                 hclge_ncl_config_data_print(hdev, desc, &offset, &length);
1169         }
1170 }
1171
1172 /* hclge_dbg_dump_mac_tnl_status: print message about mac tnl interrupt
1173  * @hdev: pointer to struct hclge_dev
1174  */
1175 static void hclge_dbg_dump_mac_tnl_status(struct hclge_dev *hdev)
1176 {
1177 #define HCLGE_BILLION_NANO_SECONDS 1000000000
1178
1179         struct hclge_mac_tnl_stats stats;
1180         unsigned long rem_nsec;
1181
1182         dev_info(&hdev->pdev->dev, "Recently generated mac tnl interruption:\n");
1183
1184         while (kfifo_get(&hdev->mac_tnl_log, &stats)) {
1185                 rem_nsec = do_div(stats.time, HCLGE_BILLION_NANO_SECONDS);
1186                 dev_info(&hdev->pdev->dev, "[%07lu.%03lu] status = 0x%x\n",
1187                          (unsigned long)stats.time, rem_nsec / 1000,
1188                          stats.status);
1189         }
1190 }
1191
1192 static void hclge_dbg_dump_qs_shaper_single(struct hclge_dev *hdev, u16 qsid)
1193 {
1194         struct hclge_qs_shapping_cmd *shap_cfg_cmd;
1195         u8 ir_u, ir_b, ir_s, bs_b, bs_s;
1196         struct hclge_desc desc;
1197         u32 shapping_para;
1198         int ret;
1199
1200         hclge_cmd_setup_basic_desc(&desc, HCLGE_OPC_QCN_SHAPPING_CFG, true);
1201
1202         shap_cfg_cmd = (struct hclge_qs_shapping_cmd *)desc.data;
1203         shap_cfg_cmd->qs_id = cpu_to_le16(qsid);
1204
1205         ret = hclge_cmd_send(&hdev->hw, &desc, 1);
1206         if (ret) {
1207                 dev_err(&hdev->pdev->dev,
1208                         "qs%u failed to get tx_rate, ret=%d\n",
1209                         qsid, ret);
1210                 return;
1211         }
1212
1213         shapping_para = le32_to_cpu(shap_cfg_cmd->qs_shapping_para);
1214         ir_b = hclge_tm_get_field(shapping_para, IR_B);
1215         ir_u = hclge_tm_get_field(shapping_para, IR_U);
1216         ir_s = hclge_tm_get_field(shapping_para, IR_S);
1217         bs_b = hclge_tm_get_field(shapping_para, BS_B);
1218         bs_s = hclge_tm_get_field(shapping_para, BS_S);
1219
1220         dev_info(&hdev->pdev->dev,
1221                  "qs%u ir_b:%u, ir_u:%u, ir_s:%u, bs_b:%u, bs_s:%u\n",
1222                  qsid, ir_b, ir_u, ir_s, bs_b, bs_s);
1223 }
1224
1225 static void hclge_dbg_dump_qs_shaper_all(struct hclge_dev *hdev)
1226 {
1227         struct hnae3_knic_private_info *kinfo;
1228         struct hclge_vport *vport;
1229         int vport_id, i;
1230
1231         for (vport_id = 0; vport_id <= pci_num_vf(hdev->pdev); vport_id++) {
1232                 vport = &hdev->vport[vport_id];
1233                 kinfo = &vport->nic.kinfo;
1234
1235                 dev_info(&hdev->pdev->dev, "qs cfg of vport%d:\n", vport_id);
1236
1237                 for (i = 0; i < kinfo->num_tc; i++) {
1238                         u16 qsid = vport->qs_offset + i;
1239
1240                         hclge_dbg_dump_qs_shaper_single(hdev, qsid);
1241                 }
1242         }
1243 }
1244
1245 static void hclge_dbg_dump_qs_shaper(struct hclge_dev *hdev,
1246                                      const char *cmd_buf)
1247 {
1248 #define HCLGE_MAX_QSET_NUM 1024
1249
1250         u16 qsid;
1251         int ret;
1252
1253         ret = kstrtou16(cmd_buf, 0, &qsid);
1254         if (ret) {
1255                 hclge_dbg_dump_qs_shaper_all(hdev);
1256                 return;
1257         }
1258
1259         if (qsid >= HCLGE_MAX_QSET_NUM) {
1260                 dev_err(&hdev->pdev->dev, "qsid(%u) out of range[0-1023]\n",
1261                         qsid);
1262                 return;
1263         }
1264
1265         hclge_dbg_dump_qs_shaper_single(hdev, qsid);
1266 }
1267
1268 int hclge_dbg_run_cmd(struct hnae3_handle *handle, const char *cmd_buf)
1269 {
1270 #define DUMP_REG        "dump reg"
1271 #define DUMP_TM_MAP     "dump tm map"
1272
1273         struct hclge_vport *vport = hclge_get_vport(handle);
1274         struct hclge_dev *hdev = vport->back;
1275
1276         if (strncmp(cmd_buf, "dump fd tcam", 12) == 0) {
1277                 hclge_dbg_fd_tcam(hdev);
1278         } else if (strncmp(cmd_buf, "dump tc", 7) == 0) {
1279                 hclge_dbg_dump_tc(hdev);
1280         } else if (strncmp(cmd_buf, DUMP_TM_MAP, strlen(DUMP_TM_MAP)) == 0) {
1281                 hclge_dbg_dump_tm_map(hdev, &cmd_buf[sizeof(DUMP_TM_MAP)]);
1282         } else if (strncmp(cmd_buf, "dump tm", 7) == 0) {
1283                 hclge_dbg_dump_tm(hdev);
1284         } else if (strncmp(cmd_buf, "dump qos pause cfg", 18) == 0) {
1285                 hclge_dbg_dump_qos_pause_cfg(hdev);
1286         } else if (strncmp(cmd_buf, "dump qos pri map", 16) == 0) {
1287                 hclge_dbg_dump_qos_pri_map(hdev);
1288         } else if (strncmp(cmd_buf, "dump qos buf cfg", 16) == 0) {
1289                 hclge_dbg_dump_qos_buf_cfg(hdev);
1290         } else if (strncmp(cmd_buf, "dump mng tbl", 12) == 0) {
1291                 hclge_dbg_dump_mng_table(hdev);
1292         } else if (strncmp(cmd_buf, DUMP_REG, strlen(DUMP_REG)) == 0) {
1293                 hclge_dbg_dump_reg_cmd(hdev, &cmd_buf[sizeof(DUMP_REG)]);
1294         } else if (strncmp(cmd_buf, "dump reset info", 15) == 0) {
1295                 hclge_dbg_dump_rst_info(hdev);
1296         } else if (strncmp(cmd_buf, "dump serv info", 14) == 0) {
1297                 hclge_dbg_dump_serv_info(hdev);
1298         } else if (strncmp(cmd_buf, "dump m7 info", 12) == 0) {
1299                 hclge_dbg_get_m7_stats_info(hdev);
1300         } else if (strncmp(cmd_buf, "dump ncl_config", 15) == 0) {
1301                 hclge_dbg_dump_ncl_config(hdev,
1302                                           &cmd_buf[sizeof("dump ncl_config")]);
1303         } else if (strncmp(cmd_buf, "dump mac tnl status", 19) == 0) {
1304                 hclge_dbg_dump_mac_tnl_status(hdev);
1305         } else if (strncmp(cmd_buf, "dump qs shaper", 14) == 0) {
1306                 hclge_dbg_dump_qs_shaper(hdev,
1307                                          &cmd_buf[sizeof("dump qs shaper")]);
1308         } else {
1309                 dev_info(&hdev->pdev->dev, "unknown command\n");
1310                 return -EINVAL;
1311         }
1312
1313         return 0;
1314 }