| 1 | /* |
| 2 | * Copyright (c) 2014-2015 Hisilicon Limited. |
| 3 | * |
| 4 | * This program is free software; you can redistribute it and/or modify |
| 5 | * it under the terms of the GNU General Public License as published by |
| 6 | * the Free Software Foundation; either version 2 of the License, or |
| 7 | * (at your option) any later version. |
| 8 | */ |
| 9 | |
| 10 | #include <linux/etherdevice.h> |
| 11 | #include <linux/netdevice.h> |
| 12 | #include <linux/spinlock.h> |
| 13 | |
| 14 | #include "hnae.h" |
| 15 | #include "hns_dsaf_mac.h" |
| 16 | #include "hns_dsaf_main.h" |
| 17 | #include "hns_dsaf_ppe.h" |
| 18 | #include "hns_dsaf_rcb.h" |
| 19 | |
| 20 | #define AE_NAME_PORT_ID_IDX 6 |
| 21 | #define ETH_STATIC_REG 1 |
| 22 | #define ETH_DUMP_REG 5 |
| 23 | #define ETH_GSTRING_LEN 32 |
| 24 | |
| 25 | static struct hns_mac_cb *hns_get_mac_cb(struct hnae_handle *handle) |
| 26 | { |
| 27 | struct hnae_vf_cb *vf_cb = hns_ae_get_vf_cb(handle); |
| 28 | |
| 29 | return vf_cb->mac_cb; |
| 30 | } |
| 31 | |
| 32 | /** |
| 33 | * hns_ae_map_eport_to_dport - translate enet port id to dsaf port id |
| 34 | * @port_id: enet port id |
| 35 | *: debug port 0-1, service port 2 -7 (dsaf mode only 2) |
| 36 | * return: dsaf port id |
| 37 | *: service ports 0 - 5, debug port 6-7 |
| 38 | **/ |
| 39 | static int hns_ae_map_eport_to_dport(u32 port_id) |
| 40 | { |
| 41 | int port_index; |
| 42 | |
| 43 | if (port_id < DSAF_DEBUG_NW_NUM) |
| 44 | port_index = port_id + DSAF_SERVICE_PORT_NUM_PER_DSAF; |
| 45 | else |
| 46 | port_index = port_id - DSAF_DEBUG_NW_NUM; |
| 47 | |
| 48 | return port_index; |
| 49 | } |
| 50 | |
| 51 | static struct dsaf_device *hns_ae_get_dsaf_dev(struct hnae_ae_dev *dev) |
| 52 | { |
| 53 | return container_of(dev, struct dsaf_device, ae_dev); |
| 54 | } |
| 55 | |
| 56 | static struct hns_ppe_cb *hns_get_ppe_cb(struct hnae_handle *handle) |
| 57 | { |
| 58 | int ppe_index; |
| 59 | int ppe_common_index; |
| 60 | struct ppe_common_cb *ppe_comm; |
| 61 | struct hnae_vf_cb *vf_cb = hns_ae_get_vf_cb(handle); |
| 62 | |
| 63 | if (vf_cb->port_index < DSAF_SERVICE_PORT_NUM_PER_DSAF) { |
| 64 | ppe_index = vf_cb->port_index; |
| 65 | ppe_common_index = 0; |
| 66 | } else { |
| 67 | ppe_index = 0; |
| 68 | ppe_common_index = |
| 69 | vf_cb->port_index - DSAF_SERVICE_PORT_NUM_PER_DSAF + 1; |
| 70 | } |
| 71 | ppe_comm = vf_cb->dsaf_dev->ppe_common[ppe_common_index]; |
| 72 | return &ppe_comm->ppe_cb[ppe_index]; |
| 73 | } |
| 74 | |
| 75 | static int hns_ae_get_q_num_per_vf( |
| 76 | struct dsaf_device *dsaf_dev, int port) |
| 77 | { |
| 78 | int common_idx = hns_dsaf_get_comm_idx_by_port(port); |
| 79 | |
| 80 | return dsaf_dev->rcb_common[common_idx]->max_q_per_vf; |
| 81 | } |
| 82 | |
| 83 | static int hns_ae_get_vf_num_per_port( |
| 84 | struct dsaf_device *dsaf_dev, int port) |
| 85 | { |
| 86 | int common_idx = hns_dsaf_get_comm_idx_by_port(port); |
| 87 | |
| 88 | return dsaf_dev->rcb_common[common_idx]->max_vfn; |
| 89 | } |
| 90 | |
| 91 | static struct ring_pair_cb *hns_ae_get_base_ring_pair( |
| 92 | struct dsaf_device *dsaf_dev, int port) |
| 93 | { |
| 94 | int common_idx = hns_dsaf_get_comm_idx_by_port(port); |
| 95 | struct rcb_common_cb *rcb_comm = dsaf_dev->rcb_common[common_idx]; |
| 96 | int q_num = rcb_comm->max_q_per_vf; |
| 97 | int vf_num = rcb_comm->max_vfn; |
| 98 | |
| 99 | if (common_idx == HNS_DSAF_COMM_SERVICE_NW_IDX) |
| 100 | return &rcb_comm->ring_pair_cb[port * q_num * vf_num]; |
| 101 | else |
| 102 | return &rcb_comm->ring_pair_cb[0]; |
| 103 | } |
| 104 | |
| 105 | static struct ring_pair_cb *hns_ae_get_ring_pair(struct hnae_queue *q) |
| 106 | { |
| 107 | return container_of(q, struct ring_pair_cb, q); |
| 108 | } |
| 109 | |
| 110 | struct hnae_handle *hns_ae_get_handle(struct hnae_ae_dev *dev, |
| 111 | u32 port_id) |
| 112 | { |
| 113 | int port_idx; |
| 114 | int vfnum_per_port; |
| 115 | int qnum_per_vf; |
| 116 | int i; |
| 117 | struct dsaf_device *dsaf_dev; |
| 118 | struct hnae_handle *ae_handle; |
| 119 | struct ring_pair_cb *ring_pair_cb; |
| 120 | struct hnae_vf_cb *vf_cb; |
| 121 | |
| 122 | dsaf_dev = hns_ae_get_dsaf_dev(dev); |
| 123 | port_idx = hns_ae_map_eport_to_dport(port_id); |
| 124 | |
| 125 | ring_pair_cb = hns_ae_get_base_ring_pair(dsaf_dev, port_idx); |
| 126 | vfnum_per_port = hns_ae_get_vf_num_per_port(dsaf_dev, port_idx); |
| 127 | qnum_per_vf = hns_ae_get_q_num_per_vf(dsaf_dev, port_idx); |
| 128 | |
| 129 | vf_cb = kzalloc(sizeof(*vf_cb) + |
| 130 | qnum_per_vf * sizeof(struct hnae_queue *), GFP_KERNEL); |
| 131 | if (unlikely(!vf_cb)) { |
| 132 | dev_err(dsaf_dev->dev, "malloc vf_cb fail!\n"); |
| 133 | ae_handle = ERR_PTR(-ENOMEM); |
| 134 | goto handle_err; |
| 135 | } |
| 136 | ae_handle = &vf_cb->ae_handle; |
| 137 | /* ae_handle Init */ |
| 138 | ae_handle->owner_dev = dsaf_dev->dev; |
| 139 | ae_handle->dev = dev; |
| 140 | ae_handle->q_num = qnum_per_vf; |
| 141 | |
| 142 | /* find ring pair, and set vf id*/ |
| 143 | for (ae_handle->vf_id = 0; |
| 144 | ae_handle->vf_id < vfnum_per_port; ae_handle->vf_id++) { |
| 145 | if (!ring_pair_cb->used_by_vf) |
| 146 | break; |
| 147 | ring_pair_cb += qnum_per_vf; |
| 148 | } |
| 149 | if (ae_handle->vf_id >= vfnum_per_port) { |
| 150 | dev_err(dsaf_dev->dev, "malloc queue fail!\n"); |
| 151 | ae_handle = ERR_PTR(-EINVAL); |
| 152 | goto vf_id_err; |
| 153 | } |
| 154 | |
| 155 | ae_handle->qs = (struct hnae_queue **)(&ae_handle->qs + 1); |
| 156 | for (i = 0; i < qnum_per_vf; i++) { |
| 157 | ae_handle->qs[i] = &ring_pair_cb->q; |
| 158 | ae_handle->qs[i]->rx_ring.q = ae_handle->qs[i]; |
| 159 | ae_handle->qs[i]->tx_ring.q = ae_handle->qs[i]; |
| 160 | |
| 161 | ring_pair_cb->used_by_vf = 1; |
| 162 | if (port_idx < DSAF_SERVICE_PORT_NUM_PER_DSAF) |
| 163 | ring_pair_cb->port_id_in_dsa = port_idx; |
| 164 | else |
| 165 | ring_pair_cb->port_id_in_dsa = 0; |
| 166 | |
| 167 | ring_pair_cb++; |
| 168 | } |
| 169 | |
| 170 | vf_cb->dsaf_dev = dsaf_dev; |
| 171 | vf_cb->port_index = port_idx; |
| 172 | vf_cb->mac_cb = &dsaf_dev->mac_cb[port_idx]; |
| 173 | |
| 174 | ae_handle->phy_if = vf_cb->mac_cb->phy_if; |
| 175 | ae_handle->phy_node = vf_cb->mac_cb->phy_node; |
| 176 | ae_handle->if_support = vf_cb->mac_cb->if_support; |
| 177 | ae_handle->port_type = vf_cb->mac_cb->mac_type; |
| 178 | ae_handle->dport_id = port_idx; |
| 179 | |
| 180 | return ae_handle; |
| 181 | vf_id_err: |
| 182 | kfree(vf_cb); |
| 183 | handle_err: |
| 184 | return ae_handle; |
| 185 | } |
| 186 | |
| 187 | static void hns_ae_put_handle(struct hnae_handle *handle) |
| 188 | { |
| 189 | struct hnae_vf_cb *vf_cb = hns_ae_get_vf_cb(handle); |
| 190 | int i; |
| 191 | |
| 192 | vf_cb->mac_cb = NULL; |
| 193 | |
| 194 | kfree(vf_cb); |
| 195 | |
| 196 | for (i = 0; i < handle->q_num; i++) |
| 197 | hns_ae_get_ring_pair(handle->qs[i])->used_by_vf = 0; |
| 198 | } |
| 199 | |
| 200 | static void hns_ae_ring_enable_all(struct hnae_handle *handle, int val) |
| 201 | { |
| 202 | int q_num = handle->q_num; |
| 203 | int i; |
| 204 | |
| 205 | for (i = 0; i < q_num; i++) |
| 206 | hns_rcb_ring_enable_hw(handle->qs[i], val); |
| 207 | } |
| 208 | |
| 209 | static void hns_ae_init_queue(struct hnae_queue *q) |
| 210 | { |
| 211 | struct ring_pair_cb *ring = |
| 212 | container_of(q, struct ring_pair_cb, q); |
| 213 | |
| 214 | hns_rcb_init_hw(ring); |
| 215 | } |
| 216 | |
| 217 | static void hns_ae_fini_queue(struct hnae_queue *q) |
| 218 | { |
| 219 | struct hnae_vf_cb *vf_cb = hns_ae_get_vf_cb(q->handle); |
| 220 | |
| 221 | if (vf_cb->mac_cb->mac_type == HNAE_PORT_SERVICE) |
| 222 | hns_rcb_reset_ring_hw(q); |
| 223 | } |
| 224 | |
| 225 | static int hns_ae_set_mac_address(struct hnae_handle *handle, void *p) |
| 226 | { |
| 227 | int ret; |
| 228 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 229 | |
| 230 | if (!p || !is_valid_ether_addr((const u8 *)p)) { |
| 231 | dev_err(handle->owner_dev, "is not valid ether addr !\n"); |
| 232 | return -EADDRNOTAVAIL; |
| 233 | } |
| 234 | |
| 235 | ret = hns_mac_change_vf_addr(mac_cb, handle->vf_id, p); |
| 236 | if (ret != 0) { |
| 237 | dev_err(handle->owner_dev, |
| 238 | "set_mac_address fail, ret=%d!\n", ret); |
| 239 | return ret; |
| 240 | } |
| 241 | |
| 242 | return 0; |
| 243 | } |
| 244 | |
| 245 | static int hns_ae_set_multicast_one(struct hnae_handle *handle, void *addr) |
| 246 | { |
| 247 | int ret; |
| 248 | char *mac_addr = (char *)addr; |
| 249 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 250 | |
| 251 | assert(mac_cb); |
| 252 | |
| 253 | if (mac_cb->mac_type != HNAE_PORT_SERVICE) |
| 254 | return 0; |
| 255 | |
| 256 | ret = hns_mac_set_multi(mac_cb, mac_cb->mac_id, mac_addr, true); |
| 257 | if (ret) { |
| 258 | dev_err(handle->owner_dev, |
| 259 | "mac add mul_mac:%pM port%d fail, ret = %#x!\n", |
| 260 | mac_addr, mac_cb->mac_id, ret); |
| 261 | return ret; |
| 262 | } |
| 263 | |
| 264 | ret = hns_mac_set_multi(mac_cb, DSAF_BASE_INNER_PORT_NUM, |
| 265 | mac_addr, true); |
| 266 | if (ret) |
| 267 | dev_err(handle->owner_dev, |
| 268 | "mac add mul_mac:%pM port%d fail, ret = %#x!\n", |
| 269 | mac_addr, DSAF_BASE_INNER_PORT_NUM, ret); |
| 270 | |
| 271 | return ret; |
| 272 | } |
| 273 | |
| 274 | static int hns_ae_set_mtu(struct hnae_handle *handle, int new_mtu) |
| 275 | { |
| 276 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 277 | |
| 278 | return hns_mac_set_mtu(mac_cb, new_mtu); |
| 279 | } |
| 280 | |
| 281 | static void hns_ae_set_tso_stats(struct hnae_handle *handle, int enable) |
| 282 | { |
| 283 | struct hns_ppe_cb *ppe_cb = hns_get_ppe_cb(handle); |
| 284 | |
| 285 | hns_ppe_set_tso_enable(ppe_cb, enable); |
| 286 | } |
| 287 | |
| 288 | static int hns_ae_start(struct hnae_handle *handle) |
| 289 | { |
| 290 | int ret; |
| 291 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 292 | |
| 293 | ret = hns_mac_vm_config_bc_en(mac_cb, 0, true); |
| 294 | if (ret) |
| 295 | return ret; |
| 296 | |
| 297 | hns_ae_ring_enable_all(handle, 1); |
| 298 | msleep(100); |
| 299 | |
| 300 | hns_mac_start(mac_cb); |
| 301 | |
| 302 | return 0; |
| 303 | } |
| 304 | |
| 305 | void hns_ae_stop(struct hnae_handle *handle) |
| 306 | { |
| 307 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 308 | |
| 309 | /* just clean tx fbd, neednot rx fbd*/ |
| 310 | hns_rcb_wait_fbd_clean(handle->qs, handle->q_num, RCB_INT_FLAG_TX); |
| 311 | |
| 312 | msleep(20); |
| 313 | |
| 314 | hns_mac_stop(mac_cb); |
| 315 | |
| 316 | usleep_range(10000, 20000); |
| 317 | |
| 318 | hns_ae_ring_enable_all(handle, 0); |
| 319 | |
| 320 | (void)hns_mac_vm_config_bc_en(mac_cb, 0, false); |
| 321 | } |
| 322 | |
| 323 | static void hns_ae_reset(struct hnae_handle *handle) |
| 324 | { |
| 325 | struct hnae_vf_cb *vf_cb = hns_ae_get_vf_cb(handle); |
| 326 | |
| 327 | if (vf_cb->mac_cb->mac_type == HNAE_PORT_DEBUG) { |
| 328 | u8 ppe_common_index = |
| 329 | vf_cb->port_index - DSAF_SERVICE_PORT_NUM_PER_DSAF + 1; |
| 330 | |
| 331 | hns_mac_reset(vf_cb->mac_cb); |
| 332 | hns_ppe_reset_common(vf_cb->dsaf_dev, ppe_common_index); |
| 333 | } |
| 334 | } |
| 335 | |
| 336 | void hns_ae_toggle_ring_irq(struct hnae_ring *ring, u32 mask) |
| 337 | { |
| 338 | u32 flag; |
| 339 | |
| 340 | if (is_tx_ring(ring)) |
| 341 | flag = RCB_INT_FLAG_TX; |
| 342 | else |
| 343 | flag = RCB_INT_FLAG_RX; |
| 344 | |
| 345 | hns_rcb_int_ctrl_hw(ring->q, flag, mask); |
| 346 | } |
| 347 | |
| 348 | static void hns_aev2_toggle_ring_irq(struct hnae_ring *ring, u32 mask) |
| 349 | { |
| 350 | u32 flag; |
| 351 | |
| 352 | if (is_tx_ring(ring)) |
| 353 | flag = RCB_INT_FLAG_TX; |
| 354 | else |
| 355 | flag = RCB_INT_FLAG_RX; |
| 356 | |
| 357 | hns_rcbv2_int_ctrl_hw(ring->q, flag, mask); |
| 358 | } |
| 359 | |
| 360 | static void hns_ae_toggle_queue_status(struct hnae_queue *queue, u32 val) |
| 361 | { |
| 362 | struct dsaf_device *dsaf_dev = hns_ae_get_dsaf_dev(queue->dev); |
| 363 | |
| 364 | if (AE_IS_VER1(dsaf_dev->dsaf_ver)) |
| 365 | hns_rcb_int_clr_hw(queue, RCB_INT_FLAG_TX | RCB_INT_FLAG_RX); |
| 366 | else |
| 367 | hns_rcbv2_int_clr_hw(queue, RCB_INT_FLAG_TX | RCB_INT_FLAG_RX); |
| 368 | |
| 369 | hns_rcb_start(queue, val); |
| 370 | } |
| 371 | |
| 372 | static int hns_ae_get_link_status(struct hnae_handle *handle) |
| 373 | { |
| 374 | u32 link_status; |
| 375 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 376 | |
| 377 | hns_mac_get_link_status(mac_cb, &link_status); |
| 378 | |
| 379 | return !!link_status; |
| 380 | } |
| 381 | |
| 382 | static int hns_ae_get_mac_info(struct hnae_handle *handle, |
| 383 | u8 *auto_neg, u16 *speed, u8 *duplex) |
| 384 | { |
| 385 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 386 | |
| 387 | return hns_mac_get_port_info(mac_cb, auto_neg, speed, duplex); |
| 388 | } |
| 389 | |
| 390 | static void hns_ae_adjust_link(struct hnae_handle *handle, int speed, |
| 391 | int duplex) |
| 392 | { |
| 393 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 394 | |
| 395 | hns_mac_adjust_link(mac_cb, speed, duplex); |
| 396 | } |
| 397 | |
| 398 | static void hns_ae_get_ring_bdnum_limit(struct hnae_queue *queue, |
| 399 | u32 *uplimit) |
| 400 | { |
| 401 | *uplimit = HNS_RCB_RING_MAX_PENDING_BD; |
| 402 | } |
| 403 | |
| 404 | static void hns_ae_get_pauseparam(struct hnae_handle *handle, |
| 405 | u32 *auto_neg, u32 *rx_en, u32 *tx_en) |
| 406 | { |
| 407 | assert(handle); |
| 408 | |
| 409 | hns_mac_get_autoneg(hns_get_mac_cb(handle), auto_neg); |
| 410 | |
| 411 | hns_mac_get_pauseparam(hns_get_mac_cb(handle), rx_en, tx_en); |
| 412 | } |
| 413 | |
| 414 | static int hns_ae_set_autoneg(struct hnae_handle *handle, u8 enable) |
| 415 | { |
| 416 | assert(handle); |
| 417 | |
| 418 | return hns_mac_set_autoneg(hns_get_mac_cb(handle), enable); |
| 419 | } |
| 420 | |
| 421 | static void hns_ae_set_promisc_mode(struct hnae_handle *handle, u32 en) |
| 422 | { |
| 423 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 424 | |
| 425 | hns_dsaf_set_promisc_mode(hns_ae_get_dsaf_dev(handle->dev), en); |
| 426 | hns_mac_set_promisc(mac_cb, (u8)!!en); |
| 427 | } |
| 428 | |
| 429 | static int hns_ae_get_autoneg(struct hnae_handle *handle) |
| 430 | { |
| 431 | u32 auto_neg; |
| 432 | |
| 433 | assert(handle); |
| 434 | |
| 435 | hns_mac_get_autoneg(hns_get_mac_cb(handle), &auto_neg); |
| 436 | |
| 437 | return auto_neg; |
| 438 | } |
| 439 | |
| 440 | static int hns_ae_set_pauseparam(struct hnae_handle *handle, |
| 441 | u32 autoneg, u32 rx_en, u32 tx_en) |
| 442 | { |
| 443 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 444 | int ret; |
| 445 | |
| 446 | ret = hns_mac_set_autoneg(mac_cb, autoneg); |
| 447 | if (ret) |
| 448 | return ret; |
| 449 | |
| 450 | return hns_mac_set_pauseparam(mac_cb, rx_en, tx_en); |
| 451 | } |
| 452 | |
| 453 | static void hns_ae_get_coalesce_usecs(struct hnae_handle *handle, |
| 454 | u32 *tx_usecs, u32 *rx_usecs) |
| 455 | { |
| 456 | int port; |
| 457 | |
| 458 | port = hns_ae_map_eport_to_dport(handle->eport_id); |
| 459 | |
| 460 | *tx_usecs = hns_rcb_get_coalesce_usecs( |
| 461 | hns_ae_get_dsaf_dev(handle->dev), |
| 462 | hns_dsaf_get_comm_idx_by_port(port)); |
| 463 | *rx_usecs = hns_rcb_get_coalesce_usecs( |
| 464 | hns_ae_get_dsaf_dev(handle->dev), |
| 465 | hns_dsaf_get_comm_idx_by_port(port)); |
| 466 | } |
| 467 | |
| 468 | static void hns_ae_get_rx_max_coalesced_frames(struct hnae_handle *handle, |
| 469 | u32 *tx_frames, u32 *rx_frames) |
| 470 | { |
| 471 | int port; |
| 472 | |
| 473 | assert(handle); |
| 474 | |
| 475 | port = hns_ae_map_eport_to_dport(handle->eport_id); |
| 476 | |
| 477 | *tx_frames = hns_rcb_get_coalesced_frames( |
| 478 | hns_ae_get_dsaf_dev(handle->dev), port); |
| 479 | *rx_frames = hns_rcb_get_coalesced_frames( |
| 480 | hns_ae_get_dsaf_dev(handle->dev), port); |
| 481 | } |
| 482 | |
| 483 | static void hns_ae_set_coalesce_usecs(struct hnae_handle *handle, |
| 484 | u32 timeout) |
| 485 | { |
| 486 | int port; |
| 487 | |
| 488 | assert(handle); |
| 489 | |
| 490 | port = hns_ae_map_eport_to_dport(handle->eport_id); |
| 491 | |
| 492 | hns_rcb_set_coalesce_usecs(hns_ae_get_dsaf_dev(handle->dev), |
| 493 | port, timeout); |
| 494 | } |
| 495 | |
| 496 | static int hns_ae_set_coalesce_frames(struct hnae_handle *handle, |
| 497 | u32 coalesce_frames) |
| 498 | { |
| 499 | int port; |
| 500 | int ret; |
| 501 | |
| 502 | assert(handle); |
| 503 | |
| 504 | port = hns_ae_map_eport_to_dport(handle->eport_id); |
| 505 | |
| 506 | ret = hns_rcb_set_coalesced_frames(hns_ae_get_dsaf_dev(handle->dev), |
| 507 | port, coalesce_frames); |
| 508 | return ret; |
| 509 | } |
| 510 | |
| 511 | void hns_ae_update_stats(struct hnae_handle *handle, |
| 512 | struct net_device_stats *net_stats) |
| 513 | { |
| 514 | int port; |
| 515 | int idx; |
| 516 | struct dsaf_device *dsaf_dev; |
| 517 | struct hns_mac_cb *mac_cb; |
| 518 | struct hns_ppe_cb *ppe_cb; |
| 519 | struct hnae_queue *queue; |
| 520 | struct hnae_vf_cb *vf_cb = hns_ae_get_vf_cb(handle); |
| 521 | u64 tx_bytes = 0, rx_bytes = 0, tx_packets = 0, rx_packets = 0; |
| 522 | u64 rx_errors = 0, tx_errors = 0, tx_dropped = 0; |
| 523 | u64 rx_missed_errors = 0; |
| 524 | |
| 525 | dsaf_dev = hns_ae_get_dsaf_dev(handle->dev); |
| 526 | if (!dsaf_dev) |
| 527 | return; |
| 528 | port = vf_cb->port_index; |
| 529 | ppe_cb = hns_get_ppe_cb(handle); |
| 530 | mac_cb = hns_get_mac_cb(handle); |
| 531 | |
| 532 | for (idx = 0; idx < handle->q_num; idx++) { |
| 533 | queue = handle->qs[idx]; |
| 534 | hns_rcb_update_stats(queue); |
| 535 | |
| 536 | tx_bytes += queue->tx_ring.stats.tx_bytes; |
| 537 | tx_packets += queue->tx_ring.stats.tx_pkts; |
| 538 | rx_bytes += queue->rx_ring.stats.rx_bytes; |
| 539 | rx_packets += queue->rx_ring.stats.rx_pkts; |
| 540 | |
| 541 | rx_errors += queue->rx_ring.stats.err_pkt_len |
| 542 | + queue->rx_ring.stats.l2_err |
| 543 | + queue->rx_ring.stats.l3l4_csum_err; |
| 544 | } |
| 545 | |
| 546 | hns_ppe_update_stats(ppe_cb); |
| 547 | rx_missed_errors = ppe_cb->hw_stats.rx_drop_no_buf; |
| 548 | tx_errors += ppe_cb->hw_stats.tx_err_checksum |
| 549 | + ppe_cb->hw_stats.tx_err_fifo_empty; |
| 550 | |
| 551 | if (mac_cb->mac_type == HNAE_PORT_SERVICE) { |
| 552 | hns_dsaf_update_stats(dsaf_dev, port); |
| 553 | /* for port upline direction, i.e., rx. */ |
| 554 | rx_missed_errors += dsaf_dev->hw_stats[port].bp_drop; |
| 555 | rx_missed_errors += dsaf_dev->hw_stats[port].pad_drop; |
| 556 | rx_missed_errors += dsaf_dev->hw_stats[port].crc_false; |
| 557 | |
| 558 | /* for port downline direction, i.e., tx. */ |
| 559 | port = port + DSAF_PPE_INODE_BASE; |
| 560 | hns_dsaf_update_stats(dsaf_dev, port); |
| 561 | tx_dropped += dsaf_dev->hw_stats[port].bp_drop; |
| 562 | tx_dropped += dsaf_dev->hw_stats[port].pad_drop; |
| 563 | tx_dropped += dsaf_dev->hw_stats[port].crc_false; |
| 564 | tx_dropped += dsaf_dev->hw_stats[port].rslt_drop; |
| 565 | tx_dropped += dsaf_dev->hw_stats[port].vlan_drop; |
| 566 | tx_dropped += dsaf_dev->hw_stats[port].stp_drop; |
| 567 | } |
| 568 | |
| 569 | hns_mac_update_stats(mac_cb); |
| 570 | rx_errors += mac_cb->hw_stats.rx_fifo_overrun_err; |
| 571 | |
| 572 | tx_errors += mac_cb->hw_stats.tx_bad_pkts |
| 573 | + mac_cb->hw_stats.tx_fragment_err |
| 574 | + mac_cb->hw_stats.tx_jabber_err |
| 575 | + mac_cb->hw_stats.tx_underrun_err |
| 576 | + mac_cb->hw_stats.tx_crc_err; |
| 577 | |
| 578 | net_stats->tx_bytes = tx_bytes; |
| 579 | net_stats->tx_packets = tx_packets; |
| 580 | net_stats->rx_bytes = rx_bytes; |
| 581 | net_stats->rx_dropped = 0; |
| 582 | net_stats->rx_packets = rx_packets; |
| 583 | net_stats->rx_errors = rx_errors; |
| 584 | net_stats->tx_errors = tx_errors; |
| 585 | net_stats->tx_dropped = tx_dropped; |
| 586 | net_stats->rx_missed_errors = rx_missed_errors; |
| 587 | net_stats->rx_crc_errors = mac_cb->hw_stats.rx_fcs_err; |
| 588 | net_stats->rx_frame_errors = mac_cb->hw_stats.rx_align_err; |
| 589 | net_stats->rx_fifo_errors = mac_cb->hw_stats.rx_fifo_overrun_err; |
| 590 | net_stats->rx_length_errors = mac_cb->hw_stats.rx_len_err; |
| 591 | net_stats->multicast = mac_cb->hw_stats.rx_mc_pkts; |
| 592 | } |
| 593 | |
| 594 | void hns_ae_get_stats(struct hnae_handle *handle, u64 *data) |
| 595 | { |
| 596 | int idx; |
| 597 | struct hns_mac_cb *mac_cb; |
| 598 | struct hns_ppe_cb *ppe_cb; |
| 599 | u64 *p = data; |
| 600 | struct hnae_vf_cb *vf_cb; |
| 601 | |
| 602 | if (!handle || !data) { |
| 603 | pr_err("hns_ae_get_stats NULL handle or data pointer!\n"); |
| 604 | return; |
| 605 | } |
| 606 | |
| 607 | vf_cb = hns_ae_get_vf_cb(handle); |
| 608 | mac_cb = hns_get_mac_cb(handle); |
| 609 | ppe_cb = hns_get_ppe_cb(handle); |
| 610 | |
| 611 | for (idx = 0; idx < handle->q_num; idx++) { |
| 612 | hns_rcb_get_stats(handle->qs[idx], p); |
| 613 | p += hns_rcb_get_ring_sset_count((int)ETH_SS_STATS); |
| 614 | } |
| 615 | |
| 616 | hns_ppe_get_stats(ppe_cb, p); |
| 617 | p += hns_ppe_get_sset_count((int)ETH_SS_STATS); |
| 618 | |
| 619 | hns_mac_get_stats(mac_cb, p); |
| 620 | p += hns_mac_get_sset_count(mac_cb, (int)ETH_SS_STATS); |
| 621 | |
| 622 | if (mac_cb->mac_type == HNAE_PORT_SERVICE) |
| 623 | hns_dsaf_get_stats(vf_cb->dsaf_dev, p, vf_cb->port_index); |
| 624 | } |
| 625 | |
| 626 | void hns_ae_get_strings(struct hnae_handle *handle, |
| 627 | u32 stringset, u8 *data) |
| 628 | { |
| 629 | int port; |
| 630 | int idx; |
| 631 | struct hns_mac_cb *mac_cb; |
| 632 | struct hns_ppe_cb *ppe_cb; |
| 633 | u8 *p = data; |
| 634 | struct hnae_vf_cb *vf_cb; |
| 635 | |
| 636 | assert(handle); |
| 637 | |
| 638 | vf_cb = hns_ae_get_vf_cb(handle); |
| 639 | port = vf_cb->port_index; |
| 640 | mac_cb = hns_get_mac_cb(handle); |
| 641 | ppe_cb = hns_get_ppe_cb(handle); |
| 642 | |
| 643 | for (idx = 0; idx < handle->q_num; idx++) { |
| 644 | hns_rcb_get_strings(stringset, p, idx); |
| 645 | p += ETH_GSTRING_LEN * hns_rcb_get_ring_sset_count(stringset); |
| 646 | } |
| 647 | |
| 648 | hns_ppe_get_strings(ppe_cb, stringset, p); |
| 649 | p += ETH_GSTRING_LEN * hns_ppe_get_sset_count(stringset); |
| 650 | |
| 651 | hns_mac_get_strings(mac_cb, stringset, p); |
| 652 | p += ETH_GSTRING_LEN * hns_mac_get_sset_count(mac_cb, stringset); |
| 653 | |
| 654 | if (mac_cb->mac_type == HNAE_PORT_SERVICE) |
| 655 | hns_dsaf_get_strings(stringset, p, port); |
| 656 | } |
| 657 | |
| 658 | int hns_ae_get_sset_count(struct hnae_handle *handle, int stringset) |
| 659 | { |
| 660 | u32 sset_count = 0; |
| 661 | struct hns_mac_cb *mac_cb; |
| 662 | |
| 663 | assert(handle); |
| 664 | |
| 665 | mac_cb = hns_get_mac_cb(handle); |
| 666 | |
| 667 | sset_count += hns_rcb_get_ring_sset_count(stringset) * handle->q_num; |
| 668 | sset_count += hns_ppe_get_sset_count(stringset); |
| 669 | sset_count += hns_mac_get_sset_count(mac_cb, stringset); |
| 670 | |
| 671 | if (mac_cb->mac_type == HNAE_PORT_SERVICE) |
| 672 | sset_count += hns_dsaf_get_sset_count(stringset); |
| 673 | |
| 674 | return sset_count; |
| 675 | } |
| 676 | |
| 677 | static int hns_ae_config_loopback(struct hnae_handle *handle, |
| 678 | enum hnae_loop loop, int en) |
| 679 | { |
| 680 | int ret; |
| 681 | struct hnae_vf_cb *vf_cb = hns_ae_get_vf_cb(handle); |
| 682 | struct hns_mac_cb *mac_cb = hns_get_mac_cb(handle); |
| 683 | |
| 684 | switch (loop) { |
| 685 | case MAC_INTERNALLOOP_PHY: |
| 686 | ret = 0; |
| 687 | break; |
| 688 | case MAC_INTERNALLOOP_SERDES: |
| 689 | ret = hns_mac_config_sds_loopback(vf_cb->mac_cb, en); |
| 690 | break; |
| 691 | case MAC_INTERNALLOOP_MAC: |
| 692 | ret = hns_mac_config_mac_loopback(vf_cb->mac_cb, loop, en); |
| 693 | break; |
| 694 | default: |
| 695 | ret = -EINVAL; |
| 696 | } |
| 697 | |
| 698 | if (!ret) |
| 699 | hns_dsaf_set_inner_lb(mac_cb->dsaf_dev, mac_cb->mac_id, en); |
| 700 | |
| 701 | return ret; |
| 702 | } |
| 703 | |
| 704 | void hns_ae_update_led_status(struct hnae_handle *handle) |
| 705 | { |
| 706 | struct hns_mac_cb *mac_cb; |
| 707 | |
| 708 | assert(handle); |
| 709 | mac_cb = hns_get_mac_cb(handle); |
| 710 | if (!mac_cb->cpld_vaddr) |
| 711 | return; |
| 712 | hns_set_led_opt(mac_cb); |
| 713 | } |
| 714 | |
| 715 | int hns_ae_cpld_set_led_id(struct hnae_handle *handle, |
| 716 | enum hnae_led_state status) |
| 717 | { |
| 718 | struct hns_mac_cb *mac_cb; |
| 719 | |
| 720 | assert(handle); |
| 721 | |
| 722 | mac_cb = hns_get_mac_cb(handle); |
| 723 | |
| 724 | return hns_cpld_led_set_id(mac_cb, status); |
| 725 | } |
| 726 | |
| 727 | void hns_ae_get_regs(struct hnae_handle *handle, void *data) |
| 728 | { |
| 729 | u32 *p = data; |
| 730 | u32 rcb_com_idx; |
| 731 | int i; |
| 732 | struct hnae_vf_cb *vf_cb = hns_ae_get_vf_cb(handle); |
| 733 | struct hns_ppe_cb *ppe_cb = hns_get_ppe_cb(handle); |
| 734 | |
| 735 | hns_ppe_get_regs(ppe_cb, p); |
| 736 | p += hns_ppe_get_regs_count(); |
| 737 | |
| 738 | rcb_com_idx = hns_dsaf_get_comm_idx_by_port(vf_cb->port_index); |
| 739 | hns_rcb_get_common_regs(vf_cb->dsaf_dev->rcb_common[rcb_com_idx], p); |
| 740 | p += hns_rcb_get_common_regs_count(); |
| 741 | |
| 742 | for (i = 0; i < handle->q_num; i++) { |
| 743 | hns_rcb_get_ring_regs(handle->qs[i], p); |
| 744 | p += hns_rcb_get_ring_regs_count(); |
| 745 | } |
| 746 | |
| 747 | hns_mac_get_regs(vf_cb->mac_cb, p); |
| 748 | p += hns_mac_get_regs_count(vf_cb->mac_cb); |
| 749 | |
| 750 | if (vf_cb->mac_cb->mac_type == HNAE_PORT_SERVICE) |
| 751 | hns_dsaf_get_regs(vf_cb->dsaf_dev, vf_cb->port_index, p); |
| 752 | } |
| 753 | |
| 754 | int hns_ae_get_regs_len(struct hnae_handle *handle) |
| 755 | { |
| 756 | u32 total_num; |
| 757 | struct hnae_vf_cb *vf_cb = hns_ae_get_vf_cb(handle); |
| 758 | |
| 759 | total_num = hns_ppe_get_regs_count(); |
| 760 | total_num += hns_rcb_get_common_regs_count(); |
| 761 | total_num += hns_rcb_get_ring_regs_count() * handle->q_num; |
| 762 | total_num += hns_mac_get_regs_count(vf_cb->mac_cb); |
| 763 | |
| 764 | if (vf_cb->mac_cb->mac_type == HNAE_PORT_SERVICE) |
| 765 | total_num += hns_dsaf_get_regs_count(); |
| 766 | |
| 767 | return total_num; |
| 768 | } |
| 769 | |
| 770 | static u32 hns_ae_get_rss_key_size(struct hnae_handle *handle) |
| 771 | { |
| 772 | return HNS_PPEV2_RSS_KEY_SIZE; |
| 773 | } |
| 774 | |
| 775 | static u32 hns_ae_get_rss_indir_size(struct hnae_handle *handle) |
| 776 | { |
| 777 | return HNS_PPEV2_RSS_IND_TBL_SIZE; |
| 778 | } |
| 779 | |
| 780 | static int hns_ae_get_rss(struct hnae_handle *handle, u32 *indir, u8 *key, |
| 781 | u8 *hfunc) |
| 782 | { |
| 783 | struct hns_ppe_cb *ppe_cb = hns_get_ppe_cb(handle); |
| 784 | |
| 785 | /* currently we support only one type of hash function i.e. Toep hash */ |
| 786 | if (hfunc) |
| 787 | *hfunc = ETH_RSS_HASH_TOP; |
| 788 | |
| 789 | /* get the RSS Key required by the user */ |
| 790 | if (key) |
| 791 | memcpy(key, ppe_cb->rss_key, HNS_PPEV2_RSS_KEY_SIZE); |
| 792 | |
| 793 | /* update the current hash->queue mappings from the shadow RSS table */ |
| 794 | memcpy(indir, ppe_cb->rss_indir_table, HNS_PPEV2_RSS_IND_TBL_SIZE); |
| 795 | |
| 796 | return 0; |
| 797 | } |
| 798 | |
| 799 | static int hns_ae_set_rss(struct hnae_handle *handle, const u32 *indir, |
| 800 | const u8 *key, const u8 hfunc) |
| 801 | { |
| 802 | struct hns_ppe_cb *ppe_cb = hns_get_ppe_cb(handle); |
| 803 | |
| 804 | /* set the RSS Hash Key if specififed by the user */ |
| 805 | if (key) |
| 806 | hns_ppe_set_rss_key(ppe_cb, (u32 *)key); |
| 807 | |
| 808 | /* update the shadow RSS table with user specified qids */ |
| 809 | memcpy(ppe_cb->rss_indir_table, indir, HNS_PPEV2_RSS_IND_TBL_SIZE); |
| 810 | |
| 811 | /* now update the hardware */ |
| 812 | hns_ppe_set_indir_table(ppe_cb, ppe_cb->rss_indir_table); |
| 813 | |
| 814 | return 0; |
| 815 | } |
| 816 | |
| 817 | static struct hnae_ae_ops hns_dsaf_ops = { |
| 818 | .get_handle = hns_ae_get_handle, |
| 819 | .put_handle = hns_ae_put_handle, |
| 820 | .init_queue = hns_ae_init_queue, |
| 821 | .fini_queue = hns_ae_fini_queue, |
| 822 | .start = hns_ae_start, |
| 823 | .stop = hns_ae_stop, |
| 824 | .reset = hns_ae_reset, |
| 825 | .toggle_ring_irq = hns_ae_toggle_ring_irq, |
| 826 | .toggle_queue_status = hns_ae_toggle_queue_status, |
| 827 | .get_status = hns_ae_get_link_status, |
| 828 | .get_info = hns_ae_get_mac_info, |
| 829 | .adjust_link = hns_ae_adjust_link, |
| 830 | .set_loopback = hns_ae_config_loopback, |
| 831 | .get_ring_bdnum_limit = hns_ae_get_ring_bdnum_limit, |
| 832 | .get_pauseparam = hns_ae_get_pauseparam, |
| 833 | .set_autoneg = hns_ae_set_autoneg, |
| 834 | .get_autoneg = hns_ae_get_autoneg, |
| 835 | .set_pauseparam = hns_ae_set_pauseparam, |
| 836 | .get_coalesce_usecs = hns_ae_get_coalesce_usecs, |
| 837 | .get_rx_max_coalesced_frames = hns_ae_get_rx_max_coalesced_frames, |
| 838 | .set_coalesce_usecs = hns_ae_set_coalesce_usecs, |
| 839 | .set_coalesce_frames = hns_ae_set_coalesce_frames, |
| 840 | .set_promisc_mode = hns_ae_set_promisc_mode, |
| 841 | .set_mac_addr = hns_ae_set_mac_address, |
| 842 | .set_mc_addr = hns_ae_set_multicast_one, |
| 843 | .set_mtu = hns_ae_set_mtu, |
| 844 | .update_stats = hns_ae_update_stats, |
| 845 | .set_tso_stats = hns_ae_set_tso_stats, |
| 846 | .get_stats = hns_ae_get_stats, |
| 847 | .get_strings = hns_ae_get_strings, |
| 848 | .get_sset_count = hns_ae_get_sset_count, |
| 849 | .update_led_status = hns_ae_update_led_status, |
| 850 | .set_led_id = hns_ae_cpld_set_led_id, |
| 851 | .get_regs = hns_ae_get_regs, |
| 852 | .get_regs_len = hns_ae_get_regs_len, |
| 853 | .get_rss_key_size = hns_ae_get_rss_key_size, |
| 854 | .get_rss_indir_size = hns_ae_get_rss_indir_size, |
| 855 | .get_rss = hns_ae_get_rss, |
| 856 | .set_rss = hns_ae_set_rss |
| 857 | }; |
| 858 | |
| 859 | int hns_dsaf_ae_init(struct dsaf_device *dsaf_dev) |
| 860 | { |
| 861 | struct hnae_ae_dev *ae_dev = &dsaf_dev->ae_dev; |
| 862 | static atomic_t id = ATOMIC_INIT(-1); |
| 863 | |
| 864 | switch (dsaf_dev->dsaf_ver) { |
| 865 | case AE_VERSION_1: |
| 866 | hns_dsaf_ops.toggle_ring_irq = hns_ae_toggle_ring_irq; |
| 867 | break; |
| 868 | case AE_VERSION_2: |
| 869 | hns_dsaf_ops.toggle_ring_irq = hns_aev2_toggle_ring_irq; |
| 870 | break; |
| 871 | default: |
| 872 | break; |
| 873 | } |
| 874 | |
| 875 | snprintf(ae_dev->name, AE_NAME_SIZE, "%s%d", DSAF_DEVICE_NAME, |
| 876 | (int)atomic_inc_return(&id)); |
| 877 | ae_dev->ops = &hns_dsaf_ops; |
| 878 | ae_dev->dev = dsaf_dev->dev; |
| 879 | |
| 880 | return hnae_ae_register(ae_dev, THIS_MODULE); |
| 881 | } |
| 882 | |
| 883 | void hns_dsaf_ae_uninit(struct dsaf_device *dsaf_dev) |
| 884 | { |
| 885 | hnae_ae_unregister(&dsaf_dev->ae_dev); |
| 886 | } |