mcdi.c revision 59cfc479b2c8ba344c8497d5c913b6cba2ce3755
1/**************************************************************************** 2 * Driver for Solarflare Solarstorm network controllers and boards 3 * Copyright 2008-2011 Solarflare Communications Inc. 4 * 5 * This program is free software; you can redistribute it and/or modify it 6 * under the terms of the GNU General Public License version 2 as published 7 * by the Free Software Foundation, incorporated herein by reference. 8 */ 9 10#include <linux/delay.h> 11#include "net_driver.h" 12#include "nic.h" 13#include "io.h" 14#include "regs.h" 15#include "mcdi_pcol.h" 16#include "phy.h" 17 18/************************************************************************** 19 * 20 * Management-Controller-to-Driver Interface 21 * 22 ************************************************************************** 23 */ 24 25#define MCDI_RPC_TIMEOUT (10 * HZ) 26 27#define MCDI_PDU(efx) \ 28 (efx_port_num(efx) ? MC_SMEM_P1_PDU_OFST : MC_SMEM_P0_PDU_OFST) 29#define MCDI_DOORBELL(efx) \ 30 (efx_port_num(efx) ? MC_SMEM_P1_DOORBELL_OFST : MC_SMEM_P0_DOORBELL_OFST) 31#define MCDI_STATUS(efx) \ 32 (efx_port_num(efx) ? MC_SMEM_P1_STATUS_OFST : MC_SMEM_P0_STATUS_OFST) 33 34/* A reboot/assertion causes the MCDI status word to be set after the 35 * command word is set or a REBOOT event is sent. If we notice a reboot 36 * via these mechanisms then wait 10ms for the status word to be set. */ 37#define MCDI_STATUS_DELAY_US 100 38#define MCDI_STATUS_DELAY_COUNT 100 39#define MCDI_STATUS_SLEEP_MS \ 40 (MCDI_STATUS_DELAY_US * MCDI_STATUS_DELAY_COUNT / 1000) 41 42#define SEQ_MASK \ 43 EFX_MASK32(EFX_WIDTH(MCDI_HEADER_SEQ)) 44 45static inline struct efx_mcdi_iface *efx_mcdi(struct efx_nic *efx) 46{ 47 struct siena_nic_data *nic_data; 48 EFX_BUG_ON_PARANOID(efx_nic_rev(efx) < EFX_REV_SIENA_A0); 49 nic_data = efx->nic_data; 50 return &nic_data->mcdi; 51} 52 53void efx_mcdi_init(struct efx_nic *efx) 54{ 55 struct efx_mcdi_iface *mcdi; 56 57 if (efx_nic_rev(efx) < EFX_REV_SIENA_A0) 58 return; 59 60 mcdi = efx_mcdi(efx); 61 init_waitqueue_head(&mcdi->wq); 62 spin_lock_init(&mcdi->iface_lock); 63 atomic_set(&mcdi->state, MCDI_STATE_QUIESCENT); 64 mcdi->mode = MCDI_MODE_POLL; 65 66 (void) efx_mcdi_poll_reboot(efx); 67} 68 69static void efx_mcdi_copyin(struct efx_nic *efx, unsigned cmd, 70 const u8 *inbuf, size_t inlen) 71{ 72 struct efx_mcdi_iface *mcdi = efx_mcdi(efx); 73 unsigned pdu = FR_CZ_MC_TREG_SMEM + MCDI_PDU(efx); 74 unsigned doorbell = FR_CZ_MC_TREG_SMEM + MCDI_DOORBELL(efx); 75 unsigned int i; 76 efx_dword_t hdr; 77 u32 xflags, seqno; 78 79 BUG_ON(atomic_read(&mcdi->state) == MCDI_STATE_QUIESCENT); 80 BUG_ON(inlen & 3 || inlen >= MC_SMEM_PDU_LEN); 81 82 seqno = mcdi->seqno & SEQ_MASK; 83 xflags = 0; 84 if (mcdi->mode == MCDI_MODE_EVENTS) 85 xflags |= MCDI_HEADER_XFLAGS_EVREQ; 86 87 EFX_POPULATE_DWORD_6(hdr, 88 MCDI_HEADER_RESPONSE, 0, 89 MCDI_HEADER_RESYNC, 1, 90 MCDI_HEADER_CODE, cmd, 91 MCDI_HEADER_DATALEN, inlen, 92 MCDI_HEADER_SEQ, seqno, 93 MCDI_HEADER_XFLAGS, xflags); 94 95 efx_writed(efx, &hdr, pdu); 96 97 for (i = 0; i < inlen; i += 4) 98 _efx_writed(efx, *((__le32 *)(inbuf + i)), pdu + 4 + i); 99 100 /* Ensure the payload is written out before the header */ 101 wmb(); 102 103 /* ring the doorbell with a distinctive value */ 104 _efx_writed(efx, (__force __le32) 0x45789abc, doorbell); 105} 106 107static void efx_mcdi_copyout(struct efx_nic *efx, u8 *outbuf, size_t outlen) 108{ 109 struct efx_mcdi_iface *mcdi = efx_mcdi(efx); 110 unsigned int pdu = FR_CZ_MC_TREG_SMEM + MCDI_PDU(efx); 111 int i; 112 113 BUG_ON(atomic_read(&mcdi->state) == MCDI_STATE_QUIESCENT); 114 BUG_ON(outlen & 3 || outlen >= MC_SMEM_PDU_LEN); 115 116 for (i = 0; i < outlen; i += 4) 117 *((__le32 *)(outbuf + i)) = _efx_readd(efx, pdu + 4 + i); 118} 119 120static int efx_mcdi_poll(struct efx_nic *efx) 121{ 122 struct efx_mcdi_iface *mcdi = efx_mcdi(efx); 123 unsigned long time, finish; 124 unsigned int respseq, respcmd, error; 125 unsigned int pdu = FR_CZ_MC_TREG_SMEM + MCDI_PDU(efx); 126 unsigned int rc, spins; 127 efx_dword_t reg; 128 129 /* Check for a reboot atomically with respect to efx_mcdi_copyout() */ 130 rc = -efx_mcdi_poll_reboot(efx); 131 if (rc) 132 goto out; 133 134 /* Poll for completion. Poll quickly (once a us) for the 1st jiffy, 135 * because generally mcdi responses are fast. After that, back off 136 * and poll once a jiffy (approximately) 137 */ 138 spins = TICK_USEC; 139 finish = jiffies + MCDI_RPC_TIMEOUT; 140 141 while (1) { 142 if (spins != 0) { 143 --spins; 144 udelay(1); 145 } else { 146 schedule_timeout_uninterruptible(1); 147 } 148 149 time = jiffies; 150 151 rmb(); 152 efx_readd(efx, ®, pdu); 153 154 /* All 1's indicates that shared memory is in reset (and is 155 * not a valid header). Wait for it to come out reset before 156 * completing the command */ 157 if (EFX_DWORD_FIELD(reg, EFX_DWORD_0) != 0xffffffff && 158 EFX_DWORD_FIELD(reg, MCDI_HEADER_RESPONSE)) 159 break; 160 161 if (time_after(time, finish)) 162 return -ETIMEDOUT; 163 } 164 165 mcdi->resplen = EFX_DWORD_FIELD(reg, MCDI_HEADER_DATALEN); 166 respseq = EFX_DWORD_FIELD(reg, MCDI_HEADER_SEQ); 167 respcmd = EFX_DWORD_FIELD(reg, MCDI_HEADER_CODE); 168 error = EFX_DWORD_FIELD(reg, MCDI_HEADER_ERROR); 169 170 if (error && mcdi->resplen == 0) { 171 netif_err(efx, hw, efx->net_dev, "MC rebooted\n"); 172 rc = EIO; 173 } else if ((respseq ^ mcdi->seqno) & SEQ_MASK) { 174 netif_err(efx, hw, efx->net_dev, 175 "MC response mismatch tx seq 0x%x rx seq 0x%x\n", 176 respseq, mcdi->seqno); 177 rc = EIO; 178 } else if (error) { 179 efx_readd(efx, ®, pdu + 4); 180 switch (EFX_DWORD_FIELD(reg, EFX_DWORD_0)) { 181#define TRANSLATE_ERROR(name) \ 182 case MC_CMD_ERR_ ## name: \ 183 rc = name; \ 184 break 185 TRANSLATE_ERROR(ENOENT); 186 TRANSLATE_ERROR(EINTR); 187 TRANSLATE_ERROR(EACCES); 188 TRANSLATE_ERROR(EBUSY); 189 TRANSLATE_ERROR(EINVAL); 190 TRANSLATE_ERROR(EDEADLK); 191 TRANSLATE_ERROR(ENOSYS); 192 TRANSLATE_ERROR(ETIME); 193#undef TRANSLATE_ERROR 194 default: 195 rc = EIO; 196 break; 197 } 198 } else 199 rc = 0; 200 201out: 202 mcdi->resprc = rc; 203 if (rc) 204 mcdi->resplen = 0; 205 206 /* Return rc=0 like wait_event_timeout() */ 207 return 0; 208} 209 210/* Test and clear MC-rebooted flag for this port/function; reset 211 * software state as necessary. 212 */ 213int efx_mcdi_poll_reboot(struct efx_nic *efx) 214{ 215 unsigned int addr = FR_CZ_MC_TREG_SMEM + MCDI_STATUS(efx); 216 efx_dword_t reg; 217 uint32_t value; 218 219 if (efx_nic_rev(efx) < EFX_REV_SIENA_A0) 220 return false; 221 222 efx_readd(efx, ®, addr); 223 value = EFX_DWORD_FIELD(reg, EFX_DWORD_0); 224 225 if (value == 0) 226 return 0; 227 228 /* MAC statistics have been cleared on the NIC; clear our copy 229 * so that efx_update_diff_stat() can continue to work. 230 */ 231 memset(&efx->mac_stats, 0, sizeof(efx->mac_stats)); 232 233 EFX_ZERO_DWORD(reg); 234 efx_writed(efx, ®, addr); 235 236 if (value == MC_STATUS_DWORD_ASSERT) 237 return -EINTR; 238 else 239 return -EIO; 240} 241 242static void efx_mcdi_acquire(struct efx_mcdi_iface *mcdi) 243{ 244 /* Wait until the interface becomes QUIESCENT and we win the race 245 * to mark it RUNNING. */ 246 wait_event(mcdi->wq, 247 atomic_cmpxchg(&mcdi->state, 248 MCDI_STATE_QUIESCENT, 249 MCDI_STATE_RUNNING) 250 == MCDI_STATE_QUIESCENT); 251} 252 253static int efx_mcdi_await_completion(struct efx_nic *efx) 254{ 255 struct efx_mcdi_iface *mcdi = efx_mcdi(efx); 256 257 if (wait_event_timeout( 258 mcdi->wq, 259 atomic_read(&mcdi->state) == MCDI_STATE_COMPLETED, 260 MCDI_RPC_TIMEOUT) == 0) 261 return -ETIMEDOUT; 262 263 /* Check if efx_mcdi_set_mode() switched us back to polled completions. 264 * In which case, poll for completions directly. If efx_mcdi_ev_cpl() 265 * completed the request first, then we'll just end up completing the 266 * request again, which is safe. 267 * 268 * We need an smp_rmb() to synchronise with efx_mcdi_mode_poll(), which 269 * wait_event_timeout() implicitly provides. 270 */ 271 if (mcdi->mode == MCDI_MODE_POLL) 272 return efx_mcdi_poll(efx); 273 274 return 0; 275} 276 277static bool efx_mcdi_complete(struct efx_mcdi_iface *mcdi) 278{ 279 /* If the interface is RUNNING, then move to COMPLETED and wake any 280 * waiters. If the interface isn't in RUNNING then we've received a 281 * duplicate completion after we've already transitioned back to 282 * QUIESCENT. [A subsequent invocation would increment seqno, so would 283 * have failed the seqno check]. 284 */ 285 if (atomic_cmpxchg(&mcdi->state, 286 MCDI_STATE_RUNNING, 287 MCDI_STATE_COMPLETED) == MCDI_STATE_RUNNING) { 288 wake_up(&mcdi->wq); 289 return true; 290 } 291 292 return false; 293} 294 295static void efx_mcdi_release(struct efx_mcdi_iface *mcdi) 296{ 297 atomic_set(&mcdi->state, MCDI_STATE_QUIESCENT); 298 wake_up(&mcdi->wq); 299} 300 301static void efx_mcdi_ev_cpl(struct efx_nic *efx, unsigned int seqno, 302 unsigned int datalen, unsigned int errno) 303{ 304 struct efx_mcdi_iface *mcdi = efx_mcdi(efx); 305 bool wake = false; 306 307 spin_lock(&mcdi->iface_lock); 308 309 if ((seqno ^ mcdi->seqno) & SEQ_MASK) { 310 if (mcdi->credits) 311 /* The request has been cancelled */ 312 --mcdi->credits; 313 else 314 netif_err(efx, hw, efx->net_dev, 315 "MC response mismatch tx seq 0x%x rx " 316 "seq 0x%x\n", seqno, mcdi->seqno); 317 } else { 318 mcdi->resprc = errno; 319 mcdi->resplen = datalen; 320 321 wake = true; 322 } 323 324 spin_unlock(&mcdi->iface_lock); 325 326 if (wake) 327 efx_mcdi_complete(mcdi); 328} 329 330int efx_mcdi_rpc(struct efx_nic *efx, unsigned cmd, 331 const u8 *inbuf, size_t inlen, u8 *outbuf, size_t outlen, 332 size_t *outlen_actual) 333{ 334 efx_mcdi_rpc_start(efx, cmd, inbuf, inlen); 335 return efx_mcdi_rpc_finish(efx, cmd, inlen, 336 outbuf, outlen, outlen_actual); 337} 338 339void efx_mcdi_rpc_start(struct efx_nic *efx, unsigned cmd, const u8 *inbuf, 340 size_t inlen) 341{ 342 struct efx_mcdi_iface *mcdi = efx_mcdi(efx); 343 344 BUG_ON(efx_nic_rev(efx) < EFX_REV_SIENA_A0); 345 346 efx_mcdi_acquire(mcdi); 347 348 /* Serialise with efx_mcdi_ev_cpl() and efx_mcdi_ev_death() */ 349 spin_lock_bh(&mcdi->iface_lock); 350 ++mcdi->seqno; 351 spin_unlock_bh(&mcdi->iface_lock); 352 353 efx_mcdi_copyin(efx, cmd, inbuf, inlen); 354} 355 356int efx_mcdi_rpc_finish(struct efx_nic *efx, unsigned cmd, size_t inlen, 357 u8 *outbuf, size_t outlen, size_t *outlen_actual) 358{ 359 struct efx_mcdi_iface *mcdi = efx_mcdi(efx); 360 int rc; 361 362 BUG_ON(efx_nic_rev(efx) < EFX_REV_SIENA_A0); 363 364 if (mcdi->mode == MCDI_MODE_POLL) 365 rc = efx_mcdi_poll(efx); 366 else 367 rc = efx_mcdi_await_completion(efx); 368 369 if (rc != 0) { 370 /* Close the race with efx_mcdi_ev_cpl() executing just too late 371 * and completing a request we've just cancelled, by ensuring 372 * that the seqno check therein fails. 373 */ 374 spin_lock_bh(&mcdi->iface_lock); 375 ++mcdi->seqno; 376 ++mcdi->credits; 377 spin_unlock_bh(&mcdi->iface_lock); 378 379 netif_err(efx, hw, efx->net_dev, 380 "MC command 0x%x inlen %d mode %d timed out\n", 381 cmd, (int)inlen, mcdi->mode); 382 } else { 383 size_t resplen; 384 385 /* At the very least we need a memory barrier here to ensure 386 * we pick up changes from efx_mcdi_ev_cpl(). Protect against 387 * a spurious efx_mcdi_ev_cpl() running concurrently by 388 * acquiring the iface_lock. */ 389 spin_lock_bh(&mcdi->iface_lock); 390 rc = -mcdi->resprc; 391 resplen = mcdi->resplen; 392 spin_unlock_bh(&mcdi->iface_lock); 393 394 if (rc == 0) { 395 efx_mcdi_copyout(efx, outbuf, 396 min(outlen, mcdi->resplen + 3) & ~0x3); 397 if (outlen_actual != NULL) 398 *outlen_actual = resplen; 399 } else if (cmd == MC_CMD_REBOOT && rc == -EIO) 400 ; /* Don't reset if MC_CMD_REBOOT returns EIO */ 401 else if (rc == -EIO || rc == -EINTR) { 402 netif_err(efx, hw, efx->net_dev, "MC fatal error %d\n", 403 -rc); 404 efx_schedule_reset(efx, RESET_TYPE_MC_FAILURE); 405 } else 406 netif_dbg(efx, hw, efx->net_dev, 407 "MC command 0x%x inlen %d failed rc=%d\n", 408 cmd, (int)inlen, -rc); 409 410 if (rc == -EIO || rc == -EINTR) { 411 msleep(MCDI_STATUS_SLEEP_MS); 412 efx_mcdi_poll_reboot(efx); 413 } 414 } 415 416 efx_mcdi_release(mcdi); 417 return rc; 418} 419 420void efx_mcdi_mode_poll(struct efx_nic *efx) 421{ 422 struct efx_mcdi_iface *mcdi; 423 424 if (efx_nic_rev(efx) < EFX_REV_SIENA_A0) 425 return; 426 427 mcdi = efx_mcdi(efx); 428 if (mcdi->mode == MCDI_MODE_POLL) 429 return; 430 431 /* We can switch from event completion to polled completion, because 432 * mcdi requests are always completed in shared memory. We do this by 433 * switching the mode to POLL'd then completing the request. 434 * efx_mcdi_await_completion() will then call efx_mcdi_poll(). 435 * 436 * We need an smp_wmb() to synchronise with efx_mcdi_await_completion(), 437 * which efx_mcdi_complete() provides for us. 438 */ 439 mcdi->mode = MCDI_MODE_POLL; 440 441 efx_mcdi_complete(mcdi); 442} 443 444void efx_mcdi_mode_event(struct efx_nic *efx) 445{ 446 struct efx_mcdi_iface *mcdi; 447 448 if (efx_nic_rev(efx) < EFX_REV_SIENA_A0) 449 return; 450 451 mcdi = efx_mcdi(efx); 452 453 if (mcdi->mode == MCDI_MODE_EVENTS) 454 return; 455 456 /* We can't switch from polled to event completion in the middle of a 457 * request, because the completion method is specified in the request. 458 * So acquire the interface to serialise the requestors. We don't need 459 * to acquire the iface_lock to change the mode here, but we do need a 460 * write memory barrier ensure that efx_mcdi_rpc() sees it, which 461 * efx_mcdi_acquire() provides. 462 */ 463 efx_mcdi_acquire(mcdi); 464 mcdi->mode = MCDI_MODE_EVENTS; 465 efx_mcdi_release(mcdi); 466} 467 468static void efx_mcdi_ev_death(struct efx_nic *efx, int rc) 469{ 470 struct efx_mcdi_iface *mcdi = efx_mcdi(efx); 471 472 /* If there is an outstanding MCDI request, it has been terminated 473 * either by a BADASSERT or REBOOT event. If the mcdi interface is 474 * in polled mode, then do nothing because the MC reboot handler will 475 * set the header correctly. However, if the mcdi interface is waiting 476 * for a CMDDONE event it won't receive it [and since all MCDI events 477 * are sent to the same queue, we can't be racing with 478 * efx_mcdi_ev_cpl()] 479 * 480 * There's a race here with efx_mcdi_rpc(), because we might receive 481 * a REBOOT event *before* the request has been copied out. In polled 482 * mode (during startup) this is irrelevant, because efx_mcdi_complete() 483 * is ignored. In event mode, this condition is just an edge-case of 484 * receiving a REBOOT event after posting the MCDI request. Did the mc 485 * reboot before or after the copyout? The best we can do always is 486 * just return failure. 487 */ 488 spin_lock(&mcdi->iface_lock); 489 if (efx_mcdi_complete(mcdi)) { 490 if (mcdi->mode == MCDI_MODE_EVENTS) { 491 mcdi->resprc = rc; 492 mcdi->resplen = 0; 493 ++mcdi->credits; 494 } 495 } else { 496 int count; 497 498 /* Nobody was waiting for an MCDI request, so trigger a reset */ 499 efx_schedule_reset(efx, RESET_TYPE_MC_FAILURE); 500 501 /* Consume the status word since efx_mcdi_rpc_finish() won't */ 502 for (count = 0; count < MCDI_STATUS_DELAY_COUNT; ++count) { 503 if (efx_mcdi_poll_reboot(efx)) 504 break; 505 udelay(MCDI_STATUS_DELAY_US); 506 } 507 } 508 509 spin_unlock(&mcdi->iface_lock); 510} 511 512static unsigned int efx_mcdi_event_link_speed[] = { 513 [MCDI_EVENT_LINKCHANGE_SPEED_100M] = 100, 514 [MCDI_EVENT_LINKCHANGE_SPEED_1G] = 1000, 515 [MCDI_EVENT_LINKCHANGE_SPEED_10G] = 10000, 516}; 517 518 519static void efx_mcdi_process_link_change(struct efx_nic *efx, efx_qword_t *ev) 520{ 521 u32 flags, fcntl, speed, lpa; 522 523 speed = EFX_QWORD_FIELD(*ev, MCDI_EVENT_LINKCHANGE_SPEED); 524 EFX_BUG_ON_PARANOID(speed >= ARRAY_SIZE(efx_mcdi_event_link_speed)); 525 speed = efx_mcdi_event_link_speed[speed]; 526 527 flags = EFX_QWORD_FIELD(*ev, MCDI_EVENT_LINKCHANGE_LINK_FLAGS); 528 fcntl = EFX_QWORD_FIELD(*ev, MCDI_EVENT_LINKCHANGE_FCNTL); 529 lpa = EFX_QWORD_FIELD(*ev, MCDI_EVENT_LINKCHANGE_LP_CAP); 530 531 /* efx->link_state is only modified by efx_mcdi_phy_get_link(), 532 * which is only run after flushing the event queues. Therefore, it 533 * is safe to modify the link state outside of the mac_lock here. 534 */ 535 efx_mcdi_phy_decode_link(efx, &efx->link_state, speed, flags, fcntl); 536 537 efx_mcdi_phy_check_fcntl(efx, lpa); 538 539 efx_link_status_changed(efx); 540} 541 542/* Called from falcon_process_eventq for MCDI events */ 543void efx_mcdi_process_event(struct efx_channel *channel, 544 efx_qword_t *event) 545{ 546 struct efx_nic *efx = channel->efx; 547 int code = EFX_QWORD_FIELD(*event, MCDI_EVENT_CODE); 548 u32 data = EFX_QWORD_FIELD(*event, MCDI_EVENT_DATA); 549 550 switch (code) { 551 case MCDI_EVENT_CODE_BADSSERT: 552 netif_err(efx, hw, efx->net_dev, 553 "MC watchdog or assertion failure at 0x%x\n", data); 554 efx_mcdi_ev_death(efx, EINTR); 555 break; 556 557 case MCDI_EVENT_CODE_PMNOTICE: 558 netif_info(efx, wol, efx->net_dev, "MCDI PM event.\n"); 559 break; 560 561 case MCDI_EVENT_CODE_CMDDONE: 562 efx_mcdi_ev_cpl(efx, 563 MCDI_EVENT_FIELD(*event, CMDDONE_SEQ), 564 MCDI_EVENT_FIELD(*event, CMDDONE_DATALEN), 565 MCDI_EVENT_FIELD(*event, CMDDONE_ERRNO)); 566 break; 567 568 case MCDI_EVENT_CODE_LINKCHANGE: 569 efx_mcdi_process_link_change(efx, event); 570 break; 571 case MCDI_EVENT_CODE_SENSOREVT: 572 efx_mcdi_sensor_event(efx, event); 573 break; 574 case MCDI_EVENT_CODE_SCHEDERR: 575 netif_info(efx, hw, efx->net_dev, 576 "MC Scheduler error address=0x%x\n", data); 577 break; 578 case MCDI_EVENT_CODE_REBOOT: 579 netif_info(efx, hw, efx->net_dev, "MC Reboot\n"); 580 efx_mcdi_ev_death(efx, EIO); 581 break; 582 case MCDI_EVENT_CODE_MAC_STATS_DMA: 583 /* MAC stats are gather lazily. We can ignore this. */ 584 break; 585 case MCDI_EVENT_CODE_FLR: 586 efx_sriov_flr(efx, MCDI_EVENT_FIELD(*event, FLR_VF)); 587 break; 588 case MCDI_EVENT_CODE_PTP_RX: 589 case MCDI_EVENT_CODE_PTP_FAULT: 590 case MCDI_EVENT_CODE_PTP_PPS: 591 efx_ptp_event(efx, event); 592 break; 593 594 default: 595 netif_err(efx, hw, efx->net_dev, "Unknown MCDI event 0x%x\n", 596 code); 597 } 598} 599 600/************************************************************************** 601 * 602 * Specific request functions 603 * 604 ************************************************************************** 605 */ 606 607void efx_mcdi_print_fwver(struct efx_nic *efx, char *buf, size_t len) 608{ 609 MCDI_DECLARE_BUF(outbuf, MC_CMD_GET_VERSION_OUT_LEN); 610 size_t outlength; 611 const __le16 *ver_words; 612 int rc; 613 614 BUILD_BUG_ON(MC_CMD_GET_VERSION_IN_LEN != 0); 615 616 rc = efx_mcdi_rpc(efx, MC_CMD_GET_VERSION, NULL, 0, 617 outbuf, sizeof(outbuf), &outlength); 618 if (rc) 619 goto fail; 620 621 if (outlength < MC_CMD_GET_VERSION_OUT_LEN) { 622 rc = -EIO; 623 goto fail; 624 } 625 626 ver_words = (__le16 *)MCDI_PTR(outbuf, GET_VERSION_OUT_VERSION); 627 snprintf(buf, len, "%u.%u.%u.%u", 628 le16_to_cpu(ver_words[0]), le16_to_cpu(ver_words[1]), 629 le16_to_cpu(ver_words[2]), le16_to_cpu(ver_words[3])); 630 return; 631 632fail: 633 netif_err(efx, probe, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 634 buf[0] = 0; 635} 636 637int efx_mcdi_drv_attach(struct efx_nic *efx, bool driver_operating, 638 bool *was_attached) 639{ 640 MCDI_DECLARE_BUF(inbuf, MC_CMD_DRV_ATTACH_IN_LEN); 641 MCDI_DECLARE_BUF(outbuf, MC_CMD_DRV_ATTACH_OUT_LEN); 642 size_t outlen; 643 int rc; 644 645 MCDI_SET_DWORD(inbuf, DRV_ATTACH_IN_NEW_STATE, 646 driver_operating ? 1 : 0); 647 MCDI_SET_DWORD(inbuf, DRV_ATTACH_IN_UPDATE, 1); 648 649 rc = efx_mcdi_rpc(efx, MC_CMD_DRV_ATTACH, inbuf, sizeof(inbuf), 650 outbuf, sizeof(outbuf), &outlen); 651 if (rc) 652 goto fail; 653 if (outlen < MC_CMD_DRV_ATTACH_OUT_LEN) { 654 rc = -EIO; 655 goto fail; 656 } 657 658 if (was_attached != NULL) 659 *was_attached = MCDI_DWORD(outbuf, DRV_ATTACH_OUT_OLD_STATE); 660 return 0; 661 662fail: 663 netif_err(efx, probe, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 664 return rc; 665} 666 667int efx_mcdi_get_board_cfg(struct efx_nic *efx, u8 *mac_address, 668 u16 *fw_subtype_list, u32 *capabilities) 669{ 670 MCDI_DECLARE_BUF(outbuf, MC_CMD_GET_BOARD_CFG_OUT_LENMAX); 671 size_t outlen, offset, i; 672 int port_num = efx_port_num(efx); 673 int rc; 674 675 BUILD_BUG_ON(MC_CMD_GET_BOARD_CFG_IN_LEN != 0); 676 677 rc = efx_mcdi_rpc(efx, MC_CMD_GET_BOARD_CFG, NULL, 0, 678 outbuf, sizeof(outbuf), &outlen); 679 if (rc) 680 goto fail; 681 682 if (outlen < MC_CMD_GET_BOARD_CFG_OUT_LENMIN) { 683 rc = -EIO; 684 goto fail; 685 } 686 687 offset = (port_num) 688 ? MC_CMD_GET_BOARD_CFG_OUT_MAC_ADDR_BASE_PORT1_OFST 689 : MC_CMD_GET_BOARD_CFG_OUT_MAC_ADDR_BASE_PORT0_OFST; 690 if (mac_address) 691 memcpy(mac_address, outbuf + offset, ETH_ALEN); 692 if (fw_subtype_list) { 693 /* Byte-swap and truncate or zero-pad as necessary */ 694 offset = MC_CMD_GET_BOARD_CFG_OUT_FW_SUBTYPE_LIST_OFST; 695 for (i = 0; 696 i < MC_CMD_GET_BOARD_CFG_OUT_FW_SUBTYPE_LIST_MAXNUM; 697 i++) { 698 fw_subtype_list[i] = 699 (offset + 2 <= outlen) ? 700 le16_to_cpup((__le16 *)(outbuf + offset)) : 0; 701 offset += 2; 702 } 703 } 704 if (capabilities) { 705 if (port_num) 706 *capabilities = MCDI_DWORD(outbuf, 707 GET_BOARD_CFG_OUT_CAPABILITIES_PORT1); 708 else 709 *capabilities = MCDI_DWORD(outbuf, 710 GET_BOARD_CFG_OUT_CAPABILITIES_PORT0); 711 } 712 713 return 0; 714 715fail: 716 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d len=%d\n", 717 __func__, rc, (int)outlen); 718 719 return rc; 720} 721 722int efx_mcdi_log_ctrl(struct efx_nic *efx, bool evq, bool uart, u32 dest_evq) 723{ 724 MCDI_DECLARE_BUF(inbuf, MC_CMD_LOG_CTRL_IN_LEN); 725 u32 dest = 0; 726 int rc; 727 728 if (uart) 729 dest |= MC_CMD_LOG_CTRL_IN_LOG_DEST_UART; 730 if (evq) 731 dest |= MC_CMD_LOG_CTRL_IN_LOG_DEST_EVQ; 732 733 MCDI_SET_DWORD(inbuf, LOG_CTRL_IN_LOG_DEST, dest); 734 MCDI_SET_DWORD(inbuf, LOG_CTRL_IN_LOG_DEST_EVQ, dest_evq); 735 736 BUILD_BUG_ON(MC_CMD_LOG_CTRL_OUT_LEN != 0); 737 738 rc = efx_mcdi_rpc(efx, MC_CMD_LOG_CTRL, inbuf, sizeof(inbuf), 739 NULL, 0, NULL); 740 if (rc) 741 goto fail; 742 743 return 0; 744 745fail: 746 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 747 return rc; 748} 749 750int efx_mcdi_nvram_types(struct efx_nic *efx, u32 *nvram_types_out) 751{ 752 MCDI_DECLARE_BUF(outbuf, MC_CMD_NVRAM_TYPES_OUT_LEN); 753 size_t outlen; 754 int rc; 755 756 BUILD_BUG_ON(MC_CMD_NVRAM_TYPES_IN_LEN != 0); 757 758 rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_TYPES, NULL, 0, 759 outbuf, sizeof(outbuf), &outlen); 760 if (rc) 761 goto fail; 762 if (outlen < MC_CMD_NVRAM_TYPES_OUT_LEN) { 763 rc = -EIO; 764 goto fail; 765 } 766 767 *nvram_types_out = MCDI_DWORD(outbuf, NVRAM_TYPES_OUT_TYPES); 768 return 0; 769 770fail: 771 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", 772 __func__, rc); 773 return rc; 774} 775 776int efx_mcdi_nvram_info(struct efx_nic *efx, unsigned int type, 777 size_t *size_out, size_t *erase_size_out, 778 bool *protected_out) 779{ 780 MCDI_DECLARE_BUF(inbuf, MC_CMD_NVRAM_INFO_IN_LEN); 781 MCDI_DECLARE_BUF(outbuf, MC_CMD_NVRAM_INFO_OUT_LEN); 782 size_t outlen; 783 int rc; 784 785 MCDI_SET_DWORD(inbuf, NVRAM_INFO_IN_TYPE, type); 786 787 rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_INFO, inbuf, sizeof(inbuf), 788 outbuf, sizeof(outbuf), &outlen); 789 if (rc) 790 goto fail; 791 if (outlen < MC_CMD_NVRAM_INFO_OUT_LEN) { 792 rc = -EIO; 793 goto fail; 794 } 795 796 *size_out = MCDI_DWORD(outbuf, NVRAM_INFO_OUT_SIZE); 797 *erase_size_out = MCDI_DWORD(outbuf, NVRAM_INFO_OUT_ERASESIZE); 798 *protected_out = !!(MCDI_DWORD(outbuf, NVRAM_INFO_OUT_FLAGS) & 799 (1 << MC_CMD_NVRAM_INFO_OUT_PROTECTED_LBN)); 800 return 0; 801 802fail: 803 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 804 return rc; 805} 806 807int efx_mcdi_nvram_update_start(struct efx_nic *efx, unsigned int type) 808{ 809 MCDI_DECLARE_BUF(inbuf, MC_CMD_NVRAM_UPDATE_START_IN_LEN); 810 int rc; 811 812 MCDI_SET_DWORD(inbuf, NVRAM_UPDATE_START_IN_TYPE, type); 813 814 BUILD_BUG_ON(MC_CMD_NVRAM_UPDATE_START_OUT_LEN != 0); 815 816 rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_UPDATE_START, inbuf, sizeof(inbuf), 817 NULL, 0, NULL); 818 if (rc) 819 goto fail; 820 821 return 0; 822 823fail: 824 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 825 return rc; 826} 827 828int efx_mcdi_nvram_read(struct efx_nic *efx, unsigned int type, 829 loff_t offset, u8 *buffer, size_t length) 830{ 831 MCDI_DECLARE_BUF(inbuf, MC_CMD_NVRAM_READ_IN_LEN); 832 MCDI_DECLARE_BUF(outbuf, 833 MC_CMD_NVRAM_READ_OUT_LEN(EFX_MCDI_NVRAM_LEN_MAX)); 834 size_t outlen; 835 int rc; 836 837 MCDI_SET_DWORD(inbuf, NVRAM_READ_IN_TYPE, type); 838 MCDI_SET_DWORD(inbuf, NVRAM_READ_IN_OFFSET, offset); 839 MCDI_SET_DWORD(inbuf, NVRAM_READ_IN_LENGTH, length); 840 841 rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_READ, inbuf, sizeof(inbuf), 842 outbuf, sizeof(outbuf), &outlen); 843 if (rc) 844 goto fail; 845 846 memcpy(buffer, MCDI_PTR(outbuf, NVRAM_READ_OUT_READ_BUFFER), length); 847 return 0; 848 849fail: 850 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 851 return rc; 852} 853 854int efx_mcdi_nvram_write(struct efx_nic *efx, unsigned int type, 855 loff_t offset, const u8 *buffer, size_t length) 856{ 857 MCDI_DECLARE_BUF(inbuf, 858 MC_CMD_NVRAM_WRITE_IN_LEN(EFX_MCDI_NVRAM_LEN_MAX)); 859 int rc; 860 861 MCDI_SET_DWORD(inbuf, NVRAM_WRITE_IN_TYPE, type); 862 MCDI_SET_DWORD(inbuf, NVRAM_WRITE_IN_OFFSET, offset); 863 MCDI_SET_DWORD(inbuf, NVRAM_WRITE_IN_LENGTH, length); 864 memcpy(MCDI_PTR(inbuf, NVRAM_WRITE_IN_WRITE_BUFFER), buffer, length); 865 866 BUILD_BUG_ON(MC_CMD_NVRAM_WRITE_OUT_LEN != 0); 867 868 rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_WRITE, inbuf, 869 ALIGN(MC_CMD_NVRAM_WRITE_IN_LEN(length), 4), 870 NULL, 0, NULL); 871 if (rc) 872 goto fail; 873 874 return 0; 875 876fail: 877 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 878 return rc; 879} 880 881int efx_mcdi_nvram_erase(struct efx_nic *efx, unsigned int type, 882 loff_t offset, size_t length) 883{ 884 MCDI_DECLARE_BUF(inbuf, MC_CMD_NVRAM_ERASE_IN_LEN); 885 int rc; 886 887 MCDI_SET_DWORD(inbuf, NVRAM_ERASE_IN_TYPE, type); 888 MCDI_SET_DWORD(inbuf, NVRAM_ERASE_IN_OFFSET, offset); 889 MCDI_SET_DWORD(inbuf, NVRAM_ERASE_IN_LENGTH, length); 890 891 BUILD_BUG_ON(MC_CMD_NVRAM_ERASE_OUT_LEN != 0); 892 893 rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_ERASE, inbuf, sizeof(inbuf), 894 NULL, 0, NULL); 895 if (rc) 896 goto fail; 897 898 return 0; 899 900fail: 901 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 902 return rc; 903} 904 905int efx_mcdi_nvram_update_finish(struct efx_nic *efx, unsigned int type) 906{ 907 MCDI_DECLARE_BUF(inbuf, MC_CMD_NVRAM_UPDATE_FINISH_IN_LEN); 908 int rc; 909 910 MCDI_SET_DWORD(inbuf, NVRAM_UPDATE_FINISH_IN_TYPE, type); 911 912 BUILD_BUG_ON(MC_CMD_NVRAM_UPDATE_FINISH_OUT_LEN != 0); 913 914 rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_UPDATE_FINISH, inbuf, sizeof(inbuf), 915 NULL, 0, NULL); 916 if (rc) 917 goto fail; 918 919 return 0; 920 921fail: 922 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 923 return rc; 924} 925 926static int efx_mcdi_nvram_test(struct efx_nic *efx, unsigned int type) 927{ 928 MCDI_DECLARE_BUF(inbuf, MC_CMD_NVRAM_TEST_IN_LEN); 929 MCDI_DECLARE_BUF(outbuf, MC_CMD_NVRAM_TEST_OUT_LEN); 930 int rc; 931 932 MCDI_SET_DWORD(inbuf, NVRAM_TEST_IN_TYPE, type); 933 934 rc = efx_mcdi_rpc(efx, MC_CMD_NVRAM_TEST, inbuf, sizeof(inbuf), 935 outbuf, sizeof(outbuf), NULL); 936 if (rc) 937 return rc; 938 939 switch (MCDI_DWORD(outbuf, NVRAM_TEST_OUT_RESULT)) { 940 case MC_CMD_NVRAM_TEST_PASS: 941 case MC_CMD_NVRAM_TEST_NOTSUPP: 942 return 0; 943 default: 944 return -EIO; 945 } 946} 947 948int efx_mcdi_nvram_test_all(struct efx_nic *efx) 949{ 950 u32 nvram_types; 951 unsigned int type; 952 int rc; 953 954 rc = efx_mcdi_nvram_types(efx, &nvram_types); 955 if (rc) 956 goto fail1; 957 958 type = 0; 959 while (nvram_types != 0) { 960 if (nvram_types & 1) { 961 rc = efx_mcdi_nvram_test(efx, type); 962 if (rc) 963 goto fail2; 964 } 965 type++; 966 nvram_types >>= 1; 967 } 968 969 return 0; 970 971fail2: 972 netif_err(efx, hw, efx->net_dev, "%s: failed type=%u\n", 973 __func__, type); 974fail1: 975 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 976 return rc; 977} 978 979static int efx_mcdi_read_assertion(struct efx_nic *efx) 980{ 981 MCDI_DECLARE_BUF(inbuf, MC_CMD_GET_ASSERTS_IN_LEN); 982 MCDI_DECLARE_BUF(outbuf, MC_CMD_GET_ASSERTS_OUT_LEN); 983 unsigned int flags, index, ofst; 984 const char *reason; 985 size_t outlen; 986 int retry; 987 int rc; 988 989 /* Attempt to read any stored assertion state before we reboot 990 * the mcfw out of the assertion handler. Retry twice, once 991 * because a boot-time assertion might cause this command to fail 992 * with EINTR. And once again because GET_ASSERTS can race with 993 * MC_CMD_REBOOT running on the other port. */ 994 retry = 2; 995 do { 996 MCDI_SET_DWORD(inbuf, GET_ASSERTS_IN_CLEAR, 1); 997 rc = efx_mcdi_rpc(efx, MC_CMD_GET_ASSERTS, 998 inbuf, MC_CMD_GET_ASSERTS_IN_LEN, 999 outbuf, sizeof(outbuf), &outlen); 1000 } while ((rc == -EINTR || rc == -EIO) && retry-- > 0); 1001 1002 if (rc) 1003 return rc; 1004 if (outlen < MC_CMD_GET_ASSERTS_OUT_LEN) 1005 return -EIO; 1006 1007 /* Print out any recorded assertion state */ 1008 flags = MCDI_DWORD(outbuf, GET_ASSERTS_OUT_GLOBAL_FLAGS); 1009 if (flags == MC_CMD_GET_ASSERTS_FLAGS_NO_FAILS) 1010 return 0; 1011 1012 reason = (flags == MC_CMD_GET_ASSERTS_FLAGS_SYS_FAIL) 1013 ? "system-level assertion" 1014 : (flags == MC_CMD_GET_ASSERTS_FLAGS_THR_FAIL) 1015 ? "thread-level assertion" 1016 : (flags == MC_CMD_GET_ASSERTS_FLAGS_WDOG_FIRED) 1017 ? "watchdog reset" 1018 : "unknown assertion"; 1019 netif_err(efx, hw, efx->net_dev, 1020 "MCPU %s at PC = 0x%.8x in thread 0x%.8x\n", reason, 1021 MCDI_DWORD(outbuf, GET_ASSERTS_OUT_SAVED_PC_OFFS), 1022 MCDI_DWORD(outbuf, GET_ASSERTS_OUT_THREAD_OFFS)); 1023 1024 /* Print out the registers */ 1025 ofst = MC_CMD_GET_ASSERTS_OUT_GP_REGS_OFFS_OFST; 1026 for (index = 1; index < 32; index++) { 1027 netif_err(efx, hw, efx->net_dev, "R%.2d (?): 0x%.8x\n", index, 1028 MCDI_DWORD2(outbuf, ofst)); 1029 ofst += sizeof(efx_dword_t); 1030 } 1031 1032 return 0; 1033} 1034 1035static void efx_mcdi_exit_assertion(struct efx_nic *efx) 1036{ 1037 MCDI_DECLARE_BUF(inbuf, MC_CMD_REBOOT_IN_LEN); 1038 1039 /* If the MC is running debug firmware, it might now be 1040 * waiting for a debugger to attach, but we just want it to 1041 * reboot. We set a flag that makes the command a no-op if it 1042 * has already done so. We don't know what return code to 1043 * expect (0 or -EIO), so ignore it. 1044 */ 1045 BUILD_BUG_ON(MC_CMD_REBOOT_OUT_LEN != 0); 1046 MCDI_SET_DWORD(inbuf, REBOOT_IN_FLAGS, 1047 MC_CMD_REBOOT_FLAGS_AFTER_ASSERTION); 1048 (void) efx_mcdi_rpc(efx, MC_CMD_REBOOT, inbuf, MC_CMD_REBOOT_IN_LEN, 1049 NULL, 0, NULL); 1050} 1051 1052int efx_mcdi_handle_assertion(struct efx_nic *efx) 1053{ 1054 int rc; 1055 1056 rc = efx_mcdi_read_assertion(efx); 1057 if (rc) 1058 return rc; 1059 1060 efx_mcdi_exit_assertion(efx); 1061 1062 return 0; 1063} 1064 1065void efx_mcdi_set_id_led(struct efx_nic *efx, enum efx_led_mode mode) 1066{ 1067 MCDI_DECLARE_BUF(inbuf, MC_CMD_SET_ID_LED_IN_LEN); 1068 int rc; 1069 1070 BUILD_BUG_ON(EFX_LED_OFF != MC_CMD_LED_OFF); 1071 BUILD_BUG_ON(EFX_LED_ON != MC_CMD_LED_ON); 1072 BUILD_BUG_ON(EFX_LED_DEFAULT != MC_CMD_LED_DEFAULT); 1073 1074 BUILD_BUG_ON(MC_CMD_SET_ID_LED_OUT_LEN != 0); 1075 1076 MCDI_SET_DWORD(inbuf, SET_ID_LED_IN_STATE, mode); 1077 1078 rc = efx_mcdi_rpc(efx, MC_CMD_SET_ID_LED, inbuf, sizeof(inbuf), 1079 NULL, 0, NULL); 1080 if (rc) 1081 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", 1082 __func__, rc); 1083} 1084 1085int efx_mcdi_reset_port(struct efx_nic *efx) 1086{ 1087 int rc = efx_mcdi_rpc(efx, MC_CMD_ENTITY_RESET, NULL, 0, NULL, 0, NULL); 1088 if (rc) 1089 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", 1090 __func__, rc); 1091 return rc; 1092} 1093 1094int efx_mcdi_reset_mc(struct efx_nic *efx) 1095{ 1096 MCDI_DECLARE_BUF(inbuf, MC_CMD_REBOOT_IN_LEN); 1097 int rc; 1098 1099 BUILD_BUG_ON(MC_CMD_REBOOT_OUT_LEN != 0); 1100 MCDI_SET_DWORD(inbuf, REBOOT_IN_FLAGS, 0); 1101 rc = efx_mcdi_rpc(efx, MC_CMD_REBOOT, inbuf, sizeof(inbuf), 1102 NULL, 0, NULL); 1103 /* White is black, and up is down */ 1104 if (rc == -EIO) 1105 return 0; 1106 if (rc == 0) 1107 rc = -EIO; 1108 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 1109 return rc; 1110} 1111 1112static int efx_mcdi_wol_filter_set(struct efx_nic *efx, u32 type, 1113 const u8 *mac, int *id_out) 1114{ 1115 MCDI_DECLARE_BUF(inbuf, MC_CMD_WOL_FILTER_SET_IN_LEN); 1116 MCDI_DECLARE_BUF(outbuf, MC_CMD_WOL_FILTER_SET_OUT_LEN); 1117 size_t outlen; 1118 int rc; 1119 1120 MCDI_SET_DWORD(inbuf, WOL_FILTER_SET_IN_WOL_TYPE, type); 1121 MCDI_SET_DWORD(inbuf, WOL_FILTER_SET_IN_FILTER_MODE, 1122 MC_CMD_FILTER_MODE_SIMPLE); 1123 memcpy(MCDI_PTR(inbuf, WOL_FILTER_SET_IN_MAGIC_MAC), mac, ETH_ALEN); 1124 1125 rc = efx_mcdi_rpc(efx, MC_CMD_WOL_FILTER_SET, inbuf, sizeof(inbuf), 1126 outbuf, sizeof(outbuf), &outlen); 1127 if (rc) 1128 goto fail; 1129 1130 if (outlen < MC_CMD_WOL_FILTER_SET_OUT_LEN) { 1131 rc = -EIO; 1132 goto fail; 1133 } 1134 1135 *id_out = (int)MCDI_DWORD(outbuf, WOL_FILTER_SET_OUT_FILTER_ID); 1136 1137 return 0; 1138 1139fail: 1140 *id_out = -1; 1141 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 1142 return rc; 1143 1144} 1145 1146 1147int 1148efx_mcdi_wol_filter_set_magic(struct efx_nic *efx, const u8 *mac, int *id_out) 1149{ 1150 return efx_mcdi_wol_filter_set(efx, MC_CMD_WOL_TYPE_MAGIC, mac, id_out); 1151} 1152 1153 1154int efx_mcdi_wol_filter_get_magic(struct efx_nic *efx, int *id_out) 1155{ 1156 MCDI_DECLARE_BUF(outbuf, MC_CMD_WOL_FILTER_GET_OUT_LEN); 1157 size_t outlen; 1158 int rc; 1159 1160 rc = efx_mcdi_rpc(efx, MC_CMD_WOL_FILTER_GET, NULL, 0, 1161 outbuf, sizeof(outbuf), &outlen); 1162 if (rc) 1163 goto fail; 1164 1165 if (outlen < MC_CMD_WOL_FILTER_GET_OUT_LEN) { 1166 rc = -EIO; 1167 goto fail; 1168 } 1169 1170 *id_out = (int)MCDI_DWORD(outbuf, WOL_FILTER_GET_OUT_FILTER_ID); 1171 1172 return 0; 1173 1174fail: 1175 *id_out = -1; 1176 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 1177 return rc; 1178} 1179 1180 1181int efx_mcdi_wol_filter_remove(struct efx_nic *efx, int id) 1182{ 1183 MCDI_DECLARE_BUF(inbuf, MC_CMD_WOL_FILTER_REMOVE_IN_LEN); 1184 int rc; 1185 1186 MCDI_SET_DWORD(inbuf, WOL_FILTER_REMOVE_IN_FILTER_ID, (u32)id); 1187 1188 rc = efx_mcdi_rpc(efx, MC_CMD_WOL_FILTER_REMOVE, inbuf, sizeof(inbuf), 1189 NULL, 0, NULL); 1190 if (rc) 1191 goto fail; 1192 1193 return 0; 1194 1195fail: 1196 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 1197 return rc; 1198} 1199 1200int efx_mcdi_flush_rxqs(struct efx_nic *efx) 1201{ 1202 struct efx_channel *channel; 1203 struct efx_rx_queue *rx_queue; 1204 __le32 *qid; 1205 int rc, count; 1206 1207 BUILD_BUG_ON(EFX_MAX_CHANNELS > 1208 MC_CMD_FLUSH_RX_QUEUES_IN_QID_OFST_MAXNUM); 1209 1210 qid = kmalloc(EFX_MAX_CHANNELS * sizeof(*qid), GFP_KERNEL); 1211 if (qid == NULL) 1212 return -ENOMEM; 1213 1214 count = 0; 1215 efx_for_each_channel(channel, efx) { 1216 efx_for_each_channel_rx_queue(rx_queue, channel) { 1217 if (rx_queue->flush_pending) { 1218 rx_queue->flush_pending = false; 1219 atomic_dec(&efx->rxq_flush_pending); 1220 qid[count++] = cpu_to_le32( 1221 efx_rx_queue_index(rx_queue)); 1222 } 1223 } 1224 } 1225 1226 rc = efx_mcdi_rpc(efx, MC_CMD_FLUSH_RX_QUEUES, (u8 *)qid, 1227 count * sizeof(*qid), NULL, 0, NULL); 1228 WARN_ON(rc < 0); 1229 1230 kfree(qid); 1231 1232 return rc; 1233} 1234 1235int efx_mcdi_wol_filter_reset(struct efx_nic *efx) 1236{ 1237 int rc; 1238 1239 rc = efx_mcdi_rpc(efx, MC_CMD_WOL_FILTER_RESET, NULL, 0, NULL, 0, NULL); 1240 if (rc) 1241 goto fail; 1242 1243 return 0; 1244 1245fail: 1246 netif_err(efx, hw, efx->net_dev, "%s: failed rc=%d\n", __func__, rc); 1247 return rc; 1248} 1249 1250