linux/drivers/net/ethernet/emulex/benet/be_cmds.c
<<
>>
Prefs
   1// SPDX-License-Identifier: GPL-2.0-only
   2/*
   3 * Copyright (C) 2005 - 2016 Broadcom
   4 * All rights reserved.
   5 *
   6 * Contact Information:
   7 * linux-drivers@emulex.com
   8 *
   9 * Emulex
  10 * 3333 Susan Street
  11 * Costa Mesa, CA 92626
  12 */
  13
  14#include <linux/module.h>
  15#include "be.h"
  16#include "be_cmds.h"
  17
  18const char * const be_misconfig_evt_port_state[] = {
  19        "Physical Link is functional",
  20        "Optics faulted/incorrectly installed/not installed - Reseat optics. If issue not resolved, replace.",
  21        "Optics of two types installed – Remove one optic or install matching pair of optics.",
  22        "Incompatible optics – Replace with compatible optics for card to function.",
  23        "Unqualified optics – Replace with Avago optics for Warranty and Technical Support.",
  24        "Uncertified optics – Replace with Avago-certified optics to enable link operation."
  25};
  26
  27static char *be_port_misconfig_evt_severity[] = {
  28        "KERN_WARN",
  29        "KERN_INFO",
  30        "KERN_ERR",
  31        "KERN_WARN"
  32};
  33
  34static char *phy_state_oper_desc[] = {
  35        "Link is non-operational",
  36        "Link is operational",
  37        ""
  38};
  39
  40static struct be_cmd_priv_map cmd_priv_map[] = {
  41        {
  42                OPCODE_ETH_ACPI_WOL_MAGIC_CONFIG,
  43                CMD_SUBSYSTEM_ETH,
  44                BE_PRIV_LNKMGMT | BE_PRIV_VHADM |
  45                BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
  46        },
  47        {
  48                OPCODE_COMMON_GET_FLOW_CONTROL,
  49                CMD_SUBSYSTEM_COMMON,
  50                BE_PRIV_LNKQUERY | BE_PRIV_VHADM |
  51                BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
  52        },
  53        {
  54                OPCODE_COMMON_SET_FLOW_CONTROL,
  55                CMD_SUBSYSTEM_COMMON,
  56                BE_PRIV_LNKMGMT | BE_PRIV_VHADM |
  57                BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
  58        },
  59        {
  60                OPCODE_ETH_GET_PPORT_STATS,
  61                CMD_SUBSYSTEM_ETH,
  62                BE_PRIV_LNKMGMT | BE_PRIV_VHADM |
  63                BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
  64        },
  65        {
  66                OPCODE_COMMON_GET_PHY_DETAILS,
  67                CMD_SUBSYSTEM_COMMON,
  68                BE_PRIV_LNKMGMT | BE_PRIV_VHADM |
  69                BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
  70        },
  71        {
  72                OPCODE_LOWLEVEL_HOST_DDR_DMA,
  73                CMD_SUBSYSTEM_LOWLEVEL,
  74                BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
  75        },
  76        {
  77                OPCODE_LOWLEVEL_LOOPBACK_TEST,
  78                CMD_SUBSYSTEM_LOWLEVEL,
  79                BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
  80        },
  81        {
  82                OPCODE_LOWLEVEL_SET_LOOPBACK_MODE,
  83                CMD_SUBSYSTEM_LOWLEVEL,
  84                BE_PRIV_DEVCFG | BE_PRIV_DEVSEC
  85        },
  86        {
  87                OPCODE_COMMON_SET_HSW_CONFIG,
  88                CMD_SUBSYSTEM_COMMON,
  89                BE_PRIV_DEVCFG | BE_PRIV_VHADM |
  90                BE_PRIV_DEVSEC
  91        },
  92        {
  93                OPCODE_COMMON_GET_EXT_FAT_CAPABILITIES,
  94                CMD_SUBSYSTEM_COMMON,
  95                BE_PRIV_DEVCFG
  96        }
  97};
  98
  99static bool be_cmd_allowed(struct be_adapter *adapter, u8 opcode, u8 subsystem)
 100{
 101        int i;
 102        int num_entries = ARRAY_SIZE(cmd_priv_map);
 103        u32 cmd_privileges = adapter->cmd_privileges;
 104
 105        for (i = 0; i < num_entries; i++)
 106                if (opcode == cmd_priv_map[i].opcode &&
 107                    subsystem == cmd_priv_map[i].subsystem)
 108                        if (!(cmd_privileges & cmd_priv_map[i].priv_mask))
 109                                return false;
 110
 111        return true;
 112}
 113
 114static inline void *embedded_payload(struct be_mcc_wrb *wrb)
 115{
 116        return wrb->payload.embedded_payload;
 117}
 118
 119static int be_mcc_notify(struct be_adapter *adapter)
 120{
 121        struct be_queue_info *mccq = &adapter->mcc_obj.q;
 122        u32 val = 0;
 123
 124        if (be_check_error(adapter, BE_ERROR_ANY))
 125                return -EIO;
 126
 127        val |= mccq->id & DB_MCCQ_RING_ID_MASK;
 128        val |= 1 << DB_MCCQ_NUM_POSTED_SHIFT;
 129
 130        wmb();
 131        iowrite32(val, adapter->db + DB_MCCQ_OFFSET);
 132
 133        return 0;
 134}
 135
 136/* To check if valid bit is set, check the entire word as we don't know
 137 * the endianness of the data (old entry is host endian while a new entry is
 138 * little endian) */
 139static inline bool be_mcc_compl_is_new(struct be_mcc_compl *compl)
 140{
 141        u32 flags;
 142
 143        if (compl->flags != 0) {
 144                flags = le32_to_cpu(compl->flags);
 145                if (flags & CQE_FLAGS_VALID_MASK) {
 146                        compl->flags = flags;
 147                        return true;
 148                }
 149        }
 150        return false;
 151}
 152
 153/* Need to reset the entire word that houses the valid bit */
 154static inline void be_mcc_compl_use(struct be_mcc_compl *compl)
 155{
 156        compl->flags = 0;
 157}
 158
 159static struct be_cmd_resp_hdr *be_decode_resp_hdr(u32 tag0, u32 tag1)
 160{
 161        unsigned long addr;
 162
 163        addr = tag1;
 164        addr = ((addr << 16) << 16) | tag0;
 165        return (void *)addr;
 166}
 167
 168static bool be_skip_err_log(u8 opcode, u16 base_status, u16 addl_status)
 169{
 170        if (base_status == MCC_STATUS_NOT_SUPPORTED ||
 171            base_status == MCC_STATUS_ILLEGAL_REQUEST ||
 172            addl_status == MCC_ADDL_STATUS_TOO_MANY_INTERFACES ||
 173            addl_status == MCC_ADDL_STATUS_INSUFFICIENT_VLANS ||
 174            (opcode == OPCODE_COMMON_WRITE_FLASHROM &&
 175            (base_status == MCC_STATUS_ILLEGAL_FIELD ||
 176             addl_status == MCC_ADDL_STATUS_FLASH_IMAGE_CRC_MISMATCH)))
 177                return true;
 178        else
 179                return false;
 180}
 181
 182/* Place holder for all the async MCC cmds wherein the caller is not in a busy
 183 * loop (has not issued be_mcc_notify_wait())
 184 */
 185static void be_async_cmd_process(struct be_adapter *adapter,
 186                                 struct be_mcc_compl *compl,
 187                                 struct be_cmd_resp_hdr *resp_hdr)
 188{
 189        enum mcc_base_status base_status = base_status(compl->status);
 190        u8 opcode = 0, subsystem = 0;
 191
 192        if (resp_hdr) {
 193                opcode = resp_hdr->opcode;
 194                subsystem = resp_hdr->subsystem;
 195        }
 196
 197        if (opcode == OPCODE_LOWLEVEL_LOOPBACK_TEST &&
 198            subsystem == CMD_SUBSYSTEM_LOWLEVEL) {
 199                complete(&adapter->et_cmd_compl);
 200                return;
 201        }
 202
 203        if (opcode == OPCODE_LOWLEVEL_SET_LOOPBACK_MODE &&
 204            subsystem == CMD_SUBSYSTEM_LOWLEVEL) {
 205                complete(&adapter->et_cmd_compl);
 206                return;
 207        }
 208
 209        if ((opcode == OPCODE_COMMON_WRITE_FLASHROM ||
 210             opcode == OPCODE_COMMON_WRITE_OBJECT) &&
 211            subsystem == CMD_SUBSYSTEM_COMMON) {
 212                adapter->flash_status = compl->status;
 213                complete(&adapter->et_cmd_compl);
 214                return;
 215        }
 216
 217        if ((opcode == OPCODE_ETH_GET_STATISTICS ||
 218             opcode == OPCODE_ETH_GET_PPORT_STATS) &&
 219            subsystem == CMD_SUBSYSTEM_ETH &&
 220            base_status == MCC_STATUS_SUCCESS) {
 221                be_parse_stats(adapter);
 222                adapter->stats_cmd_sent = false;
 223                return;
 224        }
 225
 226        if (opcode == OPCODE_COMMON_GET_CNTL_ADDITIONAL_ATTRIBUTES &&
 227            subsystem == CMD_SUBSYSTEM_COMMON) {
 228                if (base_status == MCC_STATUS_SUCCESS) {
 229                        struct be_cmd_resp_get_cntl_addnl_attribs *resp =
 230                                                        (void *)resp_hdr;
 231                        adapter->hwmon_info.be_on_die_temp =
 232                                                resp->on_die_temperature;
 233                } else {
 234                        adapter->be_get_temp_freq = 0;
 235                        adapter->hwmon_info.be_on_die_temp =
 236                                                BE_INVALID_DIE_TEMP;
 237                }
 238                return;
 239        }
 240}
 241
 242static int be_mcc_compl_process(struct be_adapter *adapter,
 243                                struct be_mcc_compl *compl)
 244{
 245        enum mcc_base_status base_status;
 246        enum mcc_addl_status addl_status;
 247        struct be_cmd_resp_hdr *resp_hdr;
 248        u8 opcode = 0, subsystem = 0;
 249
 250        /* Just swap the status to host endian; mcc tag is opaquely copied
 251         * from mcc_wrb */
 252        be_dws_le_to_cpu(compl, 4);
 253
 254        base_status = base_status(compl->status);
 255        addl_status = addl_status(compl->status);
 256
 257        resp_hdr = be_decode_resp_hdr(compl->tag0, compl->tag1);
 258        if (resp_hdr) {
 259                opcode = resp_hdr->opcode;
 260                subsystem = resp_hdr->subsystem;
 261        }
 262
 263        be_async_cmd_process(adapter, compl, resp_hdr);
 264
 265        if (base_status != MCC_STATUS_SUCCESS &&
 266            !be_skip_err_log(opcode, base_status, addl_status)) {
 267                if (base_status == MCC_STATUS_UNAUTHORIZED_REQUEST ||
 268                    addl_status == MCC_ADDL_STATUS_INSUFFICIENT_PRIVILEGES) {
 269                        dev_warn(&adapter->pdev->dev,
 270                                 "VF is not privileged to issue opcode %d-%d\n",
 271                                 opcode, subsystem);
 272                } else {
 273                        dev_err(&adapter->pdev->dev,
 274                                "opcode %d-%d failed:status %d-%d\n",
 275                                opcode, subsystem, base_status, addl_status);
 276                }
 277        }
 278        return compl->status;
 279}
 280
 281/* Link state evt is a string of bytes; no need for endian swapping */
 282static void be_async_link_state_process(struct be_adapter *adapter,
 283                                        struct be_mcc_compl *compl)
 284{
 285        struct be_async_event_link_state *evt =
 286                        (struct be_async_event_link_state *)compl;
 287
 288        /* When link status changes, link speed must be re-queried from FW */
 289        adapter->phy.link_speed = -1;
 290
 291        /* On BEx the FW does not send a separate link status
 292         * notification for physical and logical link.
 293         * On other chips just process the logical link
 294         * status notification
 295         */
 296        if (!BEx_chip(adapter) &&
 297            !(evt->port_link_status & LOGICAL_LINK_STATUS_MASK))
 298                return;
 299
 300        /* For the initial link status do not rely on the ASYNC event as
 301         * it may not be received in some cases.
 302         */
 303        if (adapter->flags & BE_FLAGS_LINK_STATUS_INIT)
 304                be_link_status_update(adapter,
 305                                      evt->port_link_status & LINK_STATUS_MASK);
 306}
 307
 308static void be_async_port_misconfig_event_process(struct be_adapter *adapter,
 309                                                  struct be_mcc_compl *compl)
 310{
 311        struct be_async_event_misconfig_port *evt =
 312                        (struct be_async_event_misconfig_port *)compl;
 313        u32 sfp_misconfig_evt_word1 = le32_to_cpu(evt->event_data_word1);
 314        u32 sfp_misconfig_evt_word2 = le32_to_cpu(evt->event_data_word2);
 315        u8 phy_oper_state = PHY_STATE_OPER_MSG_NONE;
 316        struct device *dev = &adapter->pdev->dev;
 317        u8 msg_severity = DEFAULT_MSG_SEVERITY;
 318        u8 phy_state_info;
 319        u8 new_phy_state;
 320
 321        new_phy_state =
 322                (sfp_misconfig_evt_word1 >> (adapter->hba_port_num * 8)) & 0xff;
 323
 324        if (new_phy_state == adapter->phy_state)
 325                return;
 326
 327        adapter->phy_state = new_phy_state;
 328
 329        /* for older fw that doesn't populate link effect data */
 330        if (!sfp_misconfig_evt_word2)
 331                goto log_message;
 332
 333        phy_state_info =
 334                (sfp_misconfig_evt_word2 >> (adapter->hba_port_num * 8)) & 0xff;
 335
 336        if (phy_state_info & PHY_STATE_INFO_VALID) {
 337                msg_severity = (phy_state_info & PHY_STATE_MSG_SEVERITY) >> 1;
 338
 339                if (be_phy_unqualified(new_phy_state))
 340                        phy_oper_state = (phy_state_info & PHY_STATE_OPER);
 341        }
 342
 343log_message:
 344        /* Log an error message that would allow a user to determine
 345         * whether the SFPs have an issue
 346         */
 347        if (be_phy_state_unknown(new_phy_state))
 348                dev_printk(be_port_misconfig_evt_severity[msg_severity], dev,
 349                           "Port %c: Unrecognized Optics state: 0x%x. %s",
 350                           adapter->port_name,
 351                           new_phy_state,
 352                           phy_state_oper_desc[phy_oper_state]);
 353        else
 354                dev_printk(be_port_misconfig_evt_severity[msg_severity], dev,
 355                           "Port %c: %s %s",
 356                           adapter->port_name,
 357                           be_misconfig_evt_port_state[new_phy_state],
 358                           phy_state_oper_desc[phy_oper_state]);
 359
 360        /* Log Vendor name and part no. if a misconfigured SFP is detected */
 361        if (be_phy_misconfigured(new_phy_state))
 362                adapter->flags |= BE_FLAGS_PHY_MISCONFIGURED;
 363}
 364
 365/* Grp5 CoS Priority evt */
 366static void be_async_grp5_cos_priority_process(struct be_adapter *adapter,
 367                                               struct be_mcc_compl *compl)
 368{
 369        struct be_async_event_grp5_cos_priority *evt =
 370                        (struct be_async_event_grp5_cos_priority *)compl;
 371
 372        if (evt->valid) {
 373                adapter->vlan_prio_bmap = evt->available_priority_bmap;
 374                adapter->recommended_prio_bits =
 375                        evt->reco_default_priority << VLAN_PRIO_SHIFT;
 376        }
 377}
 378
 379/* Grp5 QOS Speed evt: qos_link_speed is in units of 10 Mbps */
 380static void be_async_grp5_qos_speed_process(struct be_adapter *adapter,
 381                                            struct be_mcc_compl *compl)
 382{
 383        struct be_async_event_grp5_qos_link_speed *evt =
 384                        (struct be_async_event_grp5_qos_link_speed *)compl;
 385
 386        if (adapter->phy.link_speed >= 0 &&
 387            evt->physical_port == adapter->port_num)
 388                adapter->phy.link_speed = le16_to_cpu(evt->qos_link_speed) * 10;
 389}
 390
 391/*Grp5 PVID evt*/
 392static void be_async_grp5_pvid_state_process(struct be_adapter *adapter,
 393                                             struct be_mcc_compl *compl)
 394{
 395        struct be_async_event_grp5_pvid_state *evt =
 396                        (struct be_async_event_grp5_pvid_state *)compl;
 397
 398        if (evt->enabled) {
 399                adapter->pvid = le16_to_cpu(evt->tag) & VLAN_VID_MASK;
 400                dev_info(&adapter->pdev->dev, "LPVID: %d\n", adapter->pvid);
 401        } else {
 402                adapter->pvid = 0;
 403        }
 404}
 405
 406#define MGMT_ENABLE_MASK        0x4
 407static void be_async_grp5_fw_control_process(struct be_adapter *adapter,
 408                                             struct be_mcc_compl *compl)
 409{
 410        struct be_async_fw_control *evt = (struct be_async_fw_control *)compl;
 411        u32 evt_dw1 = le32_to_cpu(evt->event_data_word1);
 412
 413        if (evt_dw1 & MGMT_ENABLE_MASK) {
 414                adapter->flags |= BE_FLAGS_OS2BMC;
 415                adapter->bmc_filt_mask = le32_to_cpu(evt->event_data_word2);
 416        } else {
 417                adapter->flags &= ~BE_FLAGS_OS2BMC;
 418        }
 419}
 420
 421static void be_async_grp5_evt_process(struct be_adapter *adapter,
 422                                      struct be_mcc_compl *compl)
 423{
 424        u8 event_type = (compl->flags >> ASYNC_EVENT_TYPE_SHIFT) &
 425                                ASYNC_EVENT_TYPE_MASK;
 426
 427        switch (event_type) {
 428        case ASYNC_EVENT_COS_PRIORITY:
 429                be_async_grp5_cos_priority_process(adapter, compl);
 430                break;
 431        case ASYNC_EVENT_QOS_SPEED:
 432                be_async_grp5_qos_speed_process(adapter, compl);
 433                break;
 434        case ASYNC_EVENT_PVID_STATE:
 435                be_async_grp5_pvid_state_process(adapter, compl);
 436                break;
 437        /* Async event to disable/enable os2bmc and/or mac-learning */
 438        case ASYNC_EVENT_FW_CONTROL:
 439                be_async_grp5_fw_control_process(adapter, compl);
 440                break;
 441        default:
 442                break;
 443        }
 444}
 445
 446static void be_async_dbg_evt_process(struct be_adapter *adapter,
 447                                     struct be_mcc_compl *cmp)
 448{
 449        u8 event_type = 0;
 450        struct be_async_event_qnq *evt = (struct be_async_event_qnq *)cmp;
 451
 452        event_type = (cmp->flags >> ASYNC_EVENT_TYPE_SHIFT) &
 453                        ASYNC_EVENT_TYPE_MASK;
 454
 455        switch (event_type) {
 456        case ASYNC_DEBUG_EVENT_TYPE_QNQ:
 457                if (evt->valid)
 458                        adapter->qnq_vid = le16_to_cpu(evt->vlan_tag);
 459                adapter->flags |= BE_FLAGS_QNQ_ASYNC_EVT_RCVD;
 460        break;
 461        default:
 462                dev_warn(&adapter->pdev->dev, "Unknown debug event 0x%x!\n",
 463                         event_type);
 464        break;
 465        }
 466}
 467
 468static void be_async_sliport_evt_process(struct be_adapter *adapter,
 469                                         struct be_mcc_compl *cmp)
 470{
 471        u8 event_type = (cmp->flags >> ASYNC_EVENT_TYPE_SHIFT) &
 472                        ASYNC_EVENT_TYPE_MASK;
 473
 474        if (event_type == ASYNC_EVENT_PORT_MISCONFIG)
 475                be_async_port_misconfig_event_process(adapter, cmp);
 476}
 477
 478static inline bool is_link_state_evt(u32 flags)
 479{
 480        return ((flags >> ASYNC_EVENT_CODE_SHIFT) & ASYNC_EVENT_CODE_MASK) ==
 481                        ASYNC_EVENT_CODE_LINK_STATE;
 482}
 483
 484static inline bool is_grp5_evt(u32 flags)
 485{
 486        return ((flags >> ASYNC_EVENT_CODE_SHIFT) & ASYNC_EVENT_CODE_MASK) ==
 487                        ASYNC_EVENT_CODE_GRP_5;
 488}
 489
 490static inline bool is_dbg_evt(u32 flags)
 491{
 492        return ((flags >> ASYNC_EVENT_CODE_SHIFT) & ASYNC_EVENT_CODE_MASK) ==
 493                        ASYNC_EVENT_CODE_QNQ;
 494}
 495
 496static inline bool is_sliport_evt(u32 flags)
 497{
 498        return ((flags >> ASYNC_EVENT_CODE_SHIFT) & ASYNC_EVENT_CODE_MASK) ==
 499                ASYNC_EVENT_CODE_SLIPORT;
 500}
 501
 502static void be_mcc_event_process(struct be_adapter *adapter,
 503                                 struct be_mcc_compl *compl)
 504{
 505        if (is_link_state_evt(compl->flags))
 506                be_async_link_state_process(adapter, compl);
 507        else if (is_grp5_evt(compl->flags))
 508                be_async_grp5_evt_process(adapter, compl);
 509        else if (is_dbg_evt(compl->flags))
 510                be_async_dbg_evt_process(adapter, compl);
 511        else if (is_sliport_evt(compl->flags))
 512                be_async_sliport_evt_process(adapter, compl);
 513}
 514
 515static struct be_mcc_compl *be_mcc_compl_get(struct be_adapter *adapter)
 516{
 517        struct be_queue_info *mcc_cq = &adapter->mcc_obj.cq;
 518        struct be_mcc_compl *compl = queue_tail_node(mcc_cq);
 519
 520        if (be_mcc_compl_is_new(compl)) {
 521                queue_tail_inc(mcc_cq);
 522                return compl;
 523        }
 524        return NULL;
 525}
 526
 527void be_async_mcc_enable(struct be_adapter *adapter)
 528{
 529        spin_lock_bh(&adapter->mcc_cq_lock);
 530
 531        be_cq_notify(adapter, adapter->mcc_obj.cq.id, true, 0);
 532        adapter->mcc_obj.rearm_cq = true;
 533
 534        spin_unlock_bh(&adapter->mcc_cq_lock);
 535}
 536
 537void be_async_mcc_disable(struct be_adapter *adapter)
 538{
 539        spin_lock_bh(&adapter->mcc_cq_lock);
 540
 541        adapter->mcc_obj.rearm_cq = false;
 542        be_cq_notify(adapter, adapter->mcc_obj.cq.id, false, 0);
 543
 544        spin_unlock_bh(&adapter->mcc_cq_lock);
 545}
 546
 547int be_process_mcc(struct be_adapter *adapter)
 548{
 549        struct be_mcc_compl *compl;
 550        int num = 0, status = 0;
 551        struct be_mcc_obj *mcc_obj = &adapter->mcc_obj;
 552
 553        spin_lock(&adapter->mcc_cq_lock);
 554
 555        while ((compl = be_mcc_compl_get(adapter))) {
 556                if (compl->flags & CQE_FLAGS_ASYNC_MASK) {
 557                        be_mcc_event_process(adapter, compl);
 558                } else if (compl->flags & CQE_FLAGS_COMPLETED_MASK) {
 559                        status = be_mcc_compl_process(adapter, compl);
 560                        atomic_dec(&mcc_obj->q.used);
 561                }
 562                be_mcc_compl_use(compl);
 563                num++;
 564        }
 565
 566        if (num)
 567                be_cq_notify(adapter, mcc_obj->cq.id, mcc_obj->rearm_cq, num);
 568
 569        spin_unlock(&adapter->mcc_cq_lock);
 570        return status;
 571}
 572
 573/* Wait till no more pending mcc requests are present */
 574static int be_mcc_wait_compl(struct be_adapter *adapter)
 575{
 576#define mcc_timeout             12000 /* 12s timeout */
 577        int i, status = 0;
 578        struct be_mcc_obj *mcc_obj = &adapter->mcc_obj;
 579
 580        for (i = 0; i < mcc_timeout; i++) {
 581                if (be_check_error(adapter, BE_ERROR_ANY))
 582                        return -EIO;
 583
 584                local_bh_disable();
 585                status = be_process_mcc(adapter);
 586                local_bh_enable();
 587
 588                if (atomic_read(&mcc_obj->q.used) == 0)
 589                        break;
 590                usleep_range(500, 1000);
 591        }
 592        if (i == mcc_timeout) {
 593                dev_err(&adapter->pdev->dev, "FW not responding\n");
 594                be_set_error(adapter, BE_ERROR_FW);
 595                return -EIO;
 596        }
 597        return status;
 598}
 599
 600/* Notify MCC requests and wait for completion */
 601static int be_mcc_notify_wait(struct be_adapter *adapter)
 602{
 603        int status;
 604        struct be_mcc_wrb *wrb;
 605        struct be_mcc_obj *mcc_obj = &adapter->mcc_obj;
 606        u32 index = mcc_obj->q.head;
 607        struct be_cmd_resp_hdr *resp;
 608
 609        index_dec(&index, mcc_obj->q.len);
 610        wrb = queue_index_node(&mcc_obj->q, index);
 611
 612        resp = be_decode_resp_hdr(wrb->tag0, wrb->tag1);
 613
 614        status = be_mcc_notify(adapter);
 615        if (status)
 616                goto out;
 617
 618        status = be_mcc_wait_compl(adapter);
 619        if (status == -EIO)
 620                goto out;
 621
 622        status = (resp->base_status |
 623                  ((resp->addl_status & CQE_ADDL_STATUS_MASK) <<
 624                   CQE_ADDL_STATUS_SHIFT));
 625out:
 626        return status;
 627}
 628
 629static int be_mbox_db_ready_wait(struct be_adapter *adapter, void __iomem *db)
 630{
 631        int msecs = 0;
 632        u32 ready;
 633
 634        do {
 635                if (be_check_error(adapter, BE_ERROR_ANY))
 636                        return -EIO;
 637
 638                ready = ioread32(db);
 639                if (ready == 0xffffffff)
 640                        return -1;
 641
 642                ready &= MPU_MAILBOX_DB_RDY_MASK;
 643                if (ready)
 644                        break;
 645
 646                if (msecs > 4000) {
 647                        dev_err(&adapter->pdev->dev, "FW not responding\n");
 648                        be_set_error(adapter, BE_ERROR_FW);
 649                        be_detect_error(adapter);
 650                        return -1;
 651                }
 652
 653                msleep(1);
 654                msecs++;
 655        } while (true);
 656
 657        return 0;
 658}
 659
 660/*
 661 * Insert the mailbox address into the doorbell in two steps
 662 * Polls on the mbox doorbell till a command completion (or a timeout) occurs
 663 */
 664static int be_mbox_notify_wait(struct be_adapter *adapter)
 665{
 666        int status;
 667        u32 val = 0;
 668        void __iomem *db = adapter->db + MPU_MAILBOX_DB_OFFSET;
 669        struct be_dma_mem *mbox_mem = &adapter->mbox_mem;
 670        struct be_mcc_mailbox *mbox = mbox_mem->va;
 671        struct be_mcc_compl *compl = &mbox->compl;
 672
 673        /* wait for ready to be set */
 674        status = be_mbox_db_ready_wait(adapter, db);
 675        if (status != 0)
 676                return status;
 677
 678        val |= MPU_MAILBOX_DB_HI_MASK;
 679        /* at bits 2 - 31 place mbox dma addr msb bits 34 - 63 */
 680        val |= (upper_32_bits(mbox_mem->dma) >> 2) << 2;
 681        iowrite32(val, db);
 682
 683        /* wait for ready to be set */
 684        status = be_mbox_db_ready_wait(adapter, db);
 685        if (status != 0)
 686                return status;
 687
 688        val = 0;
 689        /* at bits 2 - 31 place mbox dma addr lsb bits 4 - 33 */
 690        val |= (u32)(mbox_mem->dma >> 4) << 2;
 691        iowrite32(val, db);
 692
 693        status = be_mbox_db_ready_wait(adapter, db);
 694        if (status != 0)
 695                return status;
 696
 697        /* A cq entry has been made now */
 698        if (be_mcc_compl_is_new(compl)) {
 699                status = be_mcc_compl_process(adapter, &mbox->compl);
 700                be_mcc_compl_use(compl);
 701                if (status)
 702                        return status;
 703        } else {
 704                dev_err(&adapter->pdev->dev, "invalid mailbox completion\n");
 705                return -1;
 706        }
 707        return 0;
 708}
 709
 710u16 be_POST_stage_get(struct be_adapter *adapter)
 711{
 712        u32 sem;
 713
 714        if (BEx_chip(adapter))
 715                sem  = ioread32(adapter->csr + SLIPORT_SEMAPHORE_OFFSET_BEx);
 716        else
 717                pci_read_config_dword(adapter->pdev,
 718                                      SLIPORT_SEMAPHORE_OFFSET_SH, &sem);
 719
 720        return sem & POST_STAGE_MASK;
 721}
 722
 723static int lancer_wait_ready(struct be_adapter *adapter)
 724{
 725#define SLIPORT_READY_TIMEOUT 30
 726        u32 sliport_status;
 727        int i;
 728
 729        for (i = 0; i < SLIPORT_READY_TIMEOUT; i++) {
 730                sliport_status = ioread32(adapter->db + SLIPORT_STATUS_OFFSET);
 731                if (sliport_status & SLIPORT_STATUS_RDY_MASK)
 732                        return 0;
 733
 734                if (sliport_status & SLIPORT_STATUS_ERR_MASK &&
 735                    !(sliport_status & SLIPORT_STATUS_RN_MASK))
 736                        return -EIO;
 737
 738                msleep(1000);
 739        }
 740
 741        return sliport_status ? : -1;
 742}
 743
 744int be_fw_wait_ready(struct be_adapter *adapter)
 745{
 746        u16 stage;
 747        int status, timeout = 0;
 748        struct device *dev = &adapter->pdev->dev;
 749
 750        if (lancer_chip(adapter)) {
 751                status = lancer_wait_ready(adapter);
 752                if (status) {
 753                        stage = status;
 754                        goto err;
 755                }
 756                return 0;
 757        }
 758
 759        do {
 760                /* There's no means to poll POST state on BE2/3 VFs */
 761                if (BEx_chip(adapter) && be_virtfn(adapter))
 762                        return 0;
 763
 764                stage = be_POST_stage_get(adapter);
 765                if (stage == POST_STAGE_ARMFW_RDY)
 766                        return 0;
 767
 768                dev_info(dev, "Waiting for POST, %ds elapsed\n", timeout);
 769                if (msleep_interruptible(2000)) {
 770                        dev_err(dev, "Waiting for POST aborted\n");
 771                        return -EINTR;
 772                }
 773                timeout += 2;
 774        } while (timeout < 60);
 775
 776err:
 777        dev_err(dev, "POST timeout; stage=%#x\n", stage);
 778        return -ETIMEDOUT;
 779}
 780
 781static inline struct be_sge *nonembedded_sgl(struct be_mcc_wrb *wrb)
 782{
 783        return &wrb->payload.sgl[0];
 784}
 785
 786static inline void fill_wrb_tags(struct be_mcc_wrb *wrb, unsigned long addr)
 787{
 788        wrb->tag0 = addr & 0xFFFFFFFF;
 789        wrb->tag1 = upper_32_bits(addr);
 790}
 791
 792/* Don't touch the hdr after it's prepared */
 793/* mem will be NULL for embedded commands */
 794static void be_wrb_cmd_hdr_prepare(struct be_cmd_req_hdr *req_hdr,
 795                                   u8 subsystem, u8 opcode, int cmd_len,
 796                                   struct be_mcc_wrb *wrb,
 797                                   struct be_dma_mem *mem)
 798{
 799        struct be_sge *sge;
 800
 801        req_hdr->opcode = opcode;
 802        req_hdr->subsystem = subsystem;
 803        req_hdr->request_length = cpu_to_le32(cmd_len - sizeof(*req_hdr));
 804        req_hdr->version = 0;
 805        fill_wrb_tags(wrb, (ulong) req_hdr);
 806        wrb->payload_length = cmd_len;
 807        if (mem) {
 808                wrb->embedded |= (1 & MCC_WRB_SGE_CNT_MASK) <<
 809                        MCC_WRB_SGE_CNT_SHIFT;
 810                sge = nonembedded_sgl(wrb);
 811                sge->pa_hi = cpu_to_le32(upper_32_bits(mem->dma));
 812                sge->pa_lo = cpu_to_le32(mem->dma & 0xFFFFFFFF);
 813                sge->len = cpu_to_le32(mem->size);
 814        } else
 815                wrb->embedded |= MCC_WRB_EMBEDDED_MASK;
 816        be_dws_cpu_to_le(wrb, 8);
 817}
 818
 819static void be_cmd_page_addrs_prepare(struct phys_addr *pages, u32 max_pages,
 820                                      struct be_dma_mem *mem)
 821{
 822        int i, buf_pages = min(PAGES_4K_SPANNED(mem->va, mem->size), max_pages);
 823        u64 dma = (u64)mem->dma;
 824
 825        for (i = 0; i < buf_pages; i++) {
 826                pages[i].lo = cpu_to_le32(dma & 0xFFFFFFFF);
 827                pages[i].hi = cpu_to_le32(upper_32_bits(dma));
 828                dma += PAGE_SIZE_4K;
 829        }
 830}
 831
 832static inline struct be_mcc_wrb *wrb_from_mbox(struct be_adapter *adapter)
 833{
 834        struct be_dma_mem *mbox_mem = &adapter->mbox_mem;
 835        struct be_mcc_wrb *wrb
 836                = &((struct be_mcc_mailbox *)(mbox_mem->va))->wrb;
 837        memset(wrb, 0, sizeof(*wrb));
 838        return wrb;
 839}
 840
 841static struct be_mcc_wrb *wrb_from_mccq(struct be_adapter *adapter)
 842{
 843        struct be_queue_info *mccq = &adapter->mcc_obj.q;
 844        struct be_mcc_wrb *wrb;
 845
 846        if (!mccq->created)
 847                return NULL;
 848
 849        if (atomic_read(&mccq->used) >= mccq->len)
 850                return NULL;
 851
 852        wrb = queue_head_node(mccq);
 853        queue_head_inc(mccq);
 854        atomic_inc(&mccq->used);
 855        memset(wrb, 0, sizeof(*wrb));
 856        return wrb;
 857}
 858
 859static bool use_mcc(struct be_adapter *adapter)
 860{
 861        return adapter->mcc_obj.q.created;
 862}
 863
 864/* Must be used only in process context */
 865static int be_cmd_lock(struct be_adapter *adapter)
 866{
 867        if (use_mcc(adapter)) {
 868                mutex_lock(&adapter->mcc_lock);
 869                return 0;
 870        } else {
 871                return mutex_lock_interruptible(&adapter->mbox_lock);
 872        }
 873}
 874
 875/* Must be used only in process context */
 876static void be_cmd_unlock(struct be_adapter *adapter)
 877{
 878        if (use_mcc(adapter))
 879                return mutex_unlock(&adapter->mcc_lock);
 880        else
 881                return mutex_unlock(&adapter->mbox_lock);
 882}
 883
 884static struct be_mcc_wrb *be_cmd_copy(struct be_adapter *adapter,
 885                                      struct be_mcc_wrb *wrb)
 886{
 887        struct be_mcc_wrb *dest_wrb;
 888
 889        if (use_mcc(adapter)) {
 890                dest_wrb = wrb_from_mccq(adapter);
 891                if (!dest_wrb)
 892                        return NULL;
 893        } else {
 894                dest_wrb = wrb_from_mbox(adapter);
 895        }
 896
 897        memcpy(dest_wrb, wrb, sizeof(*wrb));
 898        if (wrb->embedded & cpu_to_le32(MCC_WRB_EMBEDDED_MASK))
 899                fill_wrb_tags(dest_wrb, (ulong) embedded_payload(wrb));
 900
 901        return dest_wrb;
 902}
 903
 904/* Must be used only in process context */
 905static int be_cmd_notify_wait(struct be_adapter *adapter,
 906                              struct be_mcc_wrb *wrb)
 907{
 908        struct be_mcc_wrb *dest_wrb;
 909        int status;
 910
 911        status = be_cmd_lock(adapter);
 912        if (status)
 913                return status;
 914
 915        dest_wrb = be_cmd_copy(adapter, wrb);
 916        if (!dest_wrb) {
 917                status = -EBUSY;
 918                goto unlock;
 919        }
 920
 921        if (use_mcc(adapter))
 922                status = be_mcc_notify_wait(adapter);
 923        else
 924                status = be_mbox_notify_wait(adapter);
 925
 926        if (!status)
 927                memcpy(wrb, dest_wrb, sizeof(*wrb));
 928
 929unlock:
 930        be_cmd_unlock(adapter);
 931        return status;
 932}
 933
 934/* Tell fw we're about to start firing cmds by writing a
 935 * special pattern across the wrb hdr; uses mbox
 936 */
 937int be_cmd_fw_init(struct be_adapter *adapter)
 938{
 939        u8 *wrb;
 940        int status;
 941
 942        if (lancer_chip(adapter))
 943                return 0;
 944
 945        if (mutex_lock_interruptible(&adapter->mbox_lock))
 946                return -1;
 947
 948        wrb = (u8 *)wrb_from_mbox(adapter);
 949        *wrb++ = 0xFF;
 950        *wrb++ = 0x12;
 951        *wrb++ = 0x34;
 952        *wrb++ = 0xFF;
 953        *wrb++ = 0xFF;
 954        *wrb++ = 0x56;
 955        *wrb++ = 0x78;
 956        *wrb = 0xFF;
 957
 958        status = be_mbox_notify_wait(adapter);
 959
 960        mutex_unlock(&adapter->mbox_lock);
 961        return status;
 962}
 963
 964/* Tell fw we're done with firing cmds by writing a
 965 * special pattern across the wrb hdr; uses mbox
 966 */
 967int be_cmd_fw_clean(struct be_adapter *adapter)
 968{
 969        u8 *wrb;
 970        int status;
 971
 972        if (lancer_chip(adapter))
 973                return 0;
 974
 975        if (mutex_lock_interruptible(&adapter->mbox_lock))
 976                return -1;
 977
 978        wrb = (u8 *)wrb_from_mbox(adapter);
 979        *wrb++ = 0xFF;
 980        *wrb++ = 0xAA;
 981        *wrb++ = 0xBB;
 982        *wrb++ = 0xFF;
 983        *wrb++ = 0xFF;
 984        *wrb++ = 0xCC;
 985        *wrb++ = 0xDD;
 986        *wrb = 0xFF;
 987
 988        status = be_mbox_notify_wait(adapter);
 989
 990        mutex_unlock(&adapter->mbox_lock);
 991        return status;
 992}
 993
 994int be_cmd_eq_create(struct be_adapter *adapter, struct be_eq_obj *eqo)
 995{
 996        struct be_mcc_wrb *wrb;
 997        struct be_cmd_req_eq_create *req;
 998        struct be_dma_mem *q_mem = &eqo->q.dma_mem;
 999        int status, ver = 0;
1000
1001        if (mutex_lock_interruptible(&adapter->mbox_lock))
1002                return -1;
1003
1004        wrb = wrb_from_mbox(adapter);
1005        req = embedded_payload(wrb);
1006
1007        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1008                               OPCODE_COMMON_EQ_CREATE, sizeof(*req), wrb,
1009                               NULL);
1010
1011        /* Support for EQ_CREATEv2 available only SH-R onwards */
1012        if (!(BEx_chip(adapter) || lancer_chip(adapter)))
1013                ver = 2;
1014
1015        req->hdr.version = ver;
1016        req->num_pages =  cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size));
1017
1018        AMAP_SET_BITS(struct amap_eq_context, valid, req->context, 1);
1019        /* 4byte eqe*/
1020        AMAP_SET_BITS(struct amap_eq_context, size, req->context, 0);
1021        AMAP_SET_BITS(struct amap_eq_context, count, req->context,
1022                      __ilog2_u32(eqo->q.len / 256));
1023        be_dws_cpu_to_le(req->context, sizeof(req->context));
1024
1025        be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1026
1027        status = be_mbox_notify_wait(adapter);
1028        if (!status) {
1029                struct be_cmd_resp_eq_create *resp = embedded_payload(wrb);
1030
1031                eqo->q.id = le16_to_cpu(resp->eq_id);
1032                eqo->msix_idx =
1033                        (ver == 2) ? le16_to_cpu(resp->msix_idx) : eqo->idx;
1034                eqo->q.created = true;
1035        }
1036
1037        mutex_unlock(&adapter->mbox_lock);
1038        return status;
1039}
1040
1041/* Use MCC */
1042int be_cmd_mac_addr_query(struct be_adapter *adapter, u8 *mac_addr,
1043                          bool permanent, u32 if_handle, u32 pmac_id)
1044{
1045        struct be_mcc_wrb *wrb;
1046        struct be_cmd_req_mac_query *req;
1047        int status;
1048
1049        mutex_lock(&adapter->mcc_lock);
1050
1051        wrb = wrb_from_mccq(adapter);
1052        if (!wrb) {
1053                status = -EBUSY;
1054                goto err;
1055        }
1056        req = embedded_payload(wrb);
1057
1058        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1059                               OPCODE_COMMON_NTWK_MAC_QUERY, sizeof(*req), wrb,
1060                               NULL);
1061        req->type = MAC_ADDRESS_TYPE_NETWORK;
1062        if (permanent) {
1063                req->permanent = 1;
1064        } else {
1065                req->if_id = cpu_to_le16((u16)if_handle);
1066                req->pmac_id = cpu_to_le32(pmac_id);
1067                req->permanent = 0;
1068        }
1069
1070        status = be_mcc_notify_wait(adapter);
1071        if (!status) {
1072                struct be_cmd_resp_mac_query *resp = embedded_payload(wrb);
1073
1074                memcpy(mac_addr, resp->mac.addr, ETH_ALEN);
1075        }
1076
1077err:
1078        mutex_unlock(&adapter->mcc_lock);
1079        return status;
1080}
1081
1082/* Uses synchronous MCCQ */
1083int be_cmd_pmac_add(struct be_adapter *adapter, u8 *mac_addr,
1084                    u32 if_id, u32 *pmac_id, u32 domain)
1085{
1086        struct be_mcc_wrb *wrb;
1087        struct be_cmd_req_pmac_add *req;
1088        int status;
1089
1090        mutex_lock(&adapter->mcc_lock);
1091
1092        wrb = wrb_from_mccq(adapter);
1093        if (!wrb) {
1094                status = -EBUSY;
1095                goto err;
1096        }
1097        req = embedded_payload(wrb);
1098
1099        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1100                               OPCODE_COMMON_NTWK_PMAC_ADD, sizeof(*req), wrb,
1101                               NULL);
1102
1103        req->hdr.domain = domain;
1104        req->if_id = cpu_to_le32(if_id);
1105        memcpy(req->mac_address, mac_addr, ETH_ALEN);
1106
1107        status = be_mcc_notify_wait(adapter);
1108        if (!status) {
1109                struct be_cmd_resp_pmac_add *resp = embedded_payload(wrb);
1110
1111                *pmac_id = le32_to_cpu(resp->pmac_id);
1112        }
1113
1114err:
1115        mutex_unlock(&adapter->mcc_lock);
1116
1117         if (base_status(status) == MCC_STATUS_UNAUTHORIZED_REQUEST)
1118                status = -EPERM;
1119
1120        return status;
1121}
1122
1123/* Uses synchronous MCCQ */
1124int be_cmd_pmac_del(struct be_adapter *adapter, u32 if_id, int pmac_id, u32 dom)
1125{
1126        struct be_mcc_wrb *wrb;
1127        struct be_cmd_req_pmac_del *req;
1128        int status;
1129
1130        if (pmac_id == -1)
1131                return 0;
1132
1133        mutex_lock(&adapter->mcc_lock);
1134
1135        wrb = wrb_from_mccq(adapter);
1136        if (!wrb) {
1137                status = -EBUSY;
1138                goto err;
1139        }
1140        req = embedded_payload(wrb);
1141
1142        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1143                               OPCODE_COMMON_NTWK_PMAC_DEL, sizeof(*req),
1144                               wrb, NULL);
1145
1146        req->hdr.domain = dom;
1147        req->if_id = cpu_to_le32(if_id);
1148        req->pmac_id = cpu_to_le32(pmac_id);
1149
1150        status = be_mcc_notify_wait(adapter);
1151
1152err:
1153        mutex_unlock(&adapter->mcc_lock);
1154        return status;
1155}
1156
1157/* Uses Mbox */
1158int be_cmd_cq_create(struct be_adapter *adapter, struct be_queue_info *cq,
1159                     struct be_queue_info *eq, bool no_delay, int coalesce_wm)
1160{
1161        struct be_mcc_wrb *wrb;
1162        struct be_cmd_req_cq_create *req;
1163        struct be_dma_mem *q_mem = &cq->dma_mem;
1164        void *ctxt;
1165        int status;
1166
1167        if (mutex_lock_interruptible(&adapter->mbox_lock))
1168                return -1;
1169
1170        wrb = wrb_from_mbox(adapter);
1171        req = embedded_payload(wrb);
1172        ctxt = &req->context;
1173
1174        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1175                               OPCODE_COMMON_CQ_CREATE, sizeof(*req), wrb,
1176                               NULL);
1177
1178        req->num_pages =  cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size));
1179
1180        if (BEx_chip(adapter)) {
1181                AMAP_SET_BITS(struct amap_cq_context_be, coalescwm, ctxt,
1182                              coalesce_wm);
1183                AMAP_SET_BITS(struct amap_cq_context_be, nodelay,
1184                              ctxt, no_delay);
1185                AMAP_SET_BITS(struct amap_cq_context_be, count, ctxt,
1186                              __ilog2_u32(cq->len / 256));
1187                AMAP_SET_BITS(struct amap_cq_context_be, valid, ctxt, 1);
1188                AMAP_SET_BITS(struct amap_cq_context_be, eventable, ctxt, 1);
1189                AMAP_SET_BITS(struct amap_cq_context_be, eqid, ctxt, eq->id);
1190        } else {
1191                req->hdr.version = 2;
1192                req->page_size = 1; /* 1 for 4K */
1193
1194                /* coalesce-wm field in this cmd is not relevant to Lancer.
1195                 * Lancer uses COMMON_MODIFY_CQ to set this field
1196                 */
1197                if (!lancer_chip(adapter))
1198                        AMAP_SET_BITS(struct amap_cq_context_v2, coalescwm,
1199                                      ctxt, coalesce_wm);
1200                AMAP_SET_BITS(struct amap_cq_context_v2, nodelay, ctxt,
1201                              no_delay);
1202                AMAP_SET_BITS(struct amap_cq_context_v2, count, ctxt,
1203                              __ilog2_u32(cq->len / 256));
1204                AMAP_SET_BITS(struct amap_cq_context_v2, valid, ctxt, 1);
1205                AMAP_SET_BITS(struct amap_cq_context_v2, eventable, ctxt, 1);
1206                AMAP_SET_BITS(struct amap_cq_context_v2, eqid, ctxt, eq->id);
1207        }
1208
1209        be_dws_cpu_to_le(ctxt, sizeof(req->context));
1210
1211        be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1212
1213        status = be_mbox_notify_wait(adapter);
1214        if (!status) {
1215                struct be_cmd_resp_cq_create *resp = embedded_payload(wrb);
1216
1217                cq->id = le16_to_cpu(resp->cq_id);
1218                cq->created = true;
1219        }
1220
1221        mutex_unlock(&adapter->mbox_lock);
1222
1223        return status;
1224}
1225
1226static u32 be_encoded_q_len(int q_len)
1227{
1228        u32 len_encoded = fls(q_len); /* log2(len) + 1 */
1229
1230        if (len_encoded == 16)
1231                len_encoded = 0;
1232        return len_encoded;
1233}
1234
1235static int be_cmd_mccq_ext_create(struct be_adapter *adapter,
1236                                  struct be_queue_info *mccq,
1237                                  struct be_queue_info *cq)
1238{
1239        struct be_mcc_wrb *wrb;
1240        struct be_cmd_req_mcc_ext_create *req;
1241        struct be_dma_mem *q_mem = &mccq->dma_mem;
1242        void *ctxt;
1243        int status;
1244
1245        if (mutex_lock_interruptible(&adapter->mbox_lock))
1246                return -1;
1247
1248        wrb = wrb_from_mbox(adapter);
1249        req = embedded_payload(wrb);
1250        ctxt = &req->context;
1251
1252        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1253                               OPCODE_COMMON_MCC_CREATE_EXT, sizeof(*req), wrb,
1254                               NULL);
1255
1256        req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size));
1257        if (BEx_chip(adapter)) {
1258                AMAP_SET_BITS(struct amap_mcc_context_be, valid, ctxt, 1);
1259                AMAP_SET_BITS(struct amap_mcc_context_be, ring_size, ctxt,
1260                              be_encoded_q_len(mccq->len));
1261                AMAP_SET_BITS(struct amap_mcc_context_be, cq_id, ctxt, cq->id);
1262        } else {
1263                req->hdr.version = 1;
1264                req->cq_id = cpu_to_le16(cq->id);
1265
1266                AMAP_SET_BITS(struct amap_mcc_context_v1, ring_size, ctxt,
1267                              be_encoded_q_len(mccq->len));
1268                AMAP_SET_BITS(struct amap_mcc_context_v1, valid, ctxt, 1);
1269                AMAP_SET_BITS(struct amap_mcc_context_v1, async_cq_id,
1270                              ctxt, cq->id);
1271                AMAP_SET_BITS(struct amap_mcc_context_v1, async_cq_valid,
1272                              ctxt, 1);
1273        }
1274
1275        /* Subscribe to Link State, Sliport Event and Group 5 Events
1276         * (bits 1, 5 and 17 set)
1277         */
1278        req->async_event_bitmap[0] =
1279                        cpu_to_le32(BIT(ASYNC_EVENT_CODE_LINK_STATE) |
1280                                    BIT(ASYNC_EVENT_CODE_GRP_5) |
1281                                    BIT(ASYNC_EVENT_CODE_QNQ) |
1282                                    BIT(ASYNC_EVENT_CODE_SLIPORT));
1283
1284        be_dws_cpu_to_le(ctxt, sizeof(req->context));
1285
1286        be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1287
1288        status = be_mbox_notify_wait(adapter);
1289        if (!status) {
1290                struct be_cmd_resp_mcc_create *resp = embedded_payload(wrb);
1291
1292                mccq->id = le16_to_cpu(resp->id);
1293                mccq->created = true;
1294        }
1295        mutex_unlock(&adapter->mbox_lock);
1296
1297        return status;
1298}
1299
1300static int be_cmd_mccq_org_create(struct be_adapter *adapter,
1301                                  struct be_queue_info *mccq,
1302                                  struct be_queue_info *cq)
1303{
1304        struct be_mcc_wrb *wrb;
1305        struct be_cmd_req_mcc_create *req;
1306        struct be_dma_mem *q_mem = &mccq->dma_mem;
1307        void *ctxt;
1308        int status;
1309
1310        if (mutex_lock_interruptible(&adapter->mbox_lock))
1311                return -1;
1312
1313        wrb = wrb_from_mbox(adapter);
1314        req = embedded_payload(wrb);
1315        ctxt = &req->context;
1316
1317        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1318                               OPCODE_COMMON_MCC_CREATE, sizeof(*req), wrb,
1319                               NULL);
1320
1321        req->num_pages = cpu_to_le16(PAGES_4K_SPANNED(q_mem->va, q_mem->size));
1322
1323        AMAP_SET_BITS(struct amap_mcc_context_be, valid, ctxt, 1);
1324        AMAP_SET_BITS(struct amap_mcc_context_be, ring_size, ctxt,
1325                      be_encoded_q_len(mccq->len));
1326        AMAP_SET_BITS(struct amap_mcc_context_be, cq_id, ctxt, cq->id);
1327
1328        be_dws_cpu_to_le(ctxt, sizeof(req->context));
1329
1330        be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1331
1332        status = be_mbox_notify_wait(adapter);
1333        if (!status) {
1334                struct be_cmd_resp_mcc_create *resp = embedded_payload(wrb);
1335
1336                mccq->id = le16_to_cpu(resp->id);
1337                mccq->created = true;
1338        }
1339
1340        mutex_unlock(&adapter->mbox_lock);
1341        return status;
1342}
1343
1344int be_cmd_mccq_create(struct be_adapter *adapter,
1345                       struct be_queue_info *mccq, struct be_queue_info *cq)
1346{
1347        int status;
1348
1349        status = be_cmd_mccq_ext_create(adapter, mccq, cq);
1350        if (status && BEx_chip(adapter)) {
1351                dev_warn(&adapter->pdev->dev, "Upgrade to F/W ver 2.102.235.0 "
1352                        "or newer to avoid conflicting priorities between NIC "
1353                        "and FCoE traffic");
1354                status = be_cmd_mccq_org_create(adapter, mccq, cq);
1355        }
1356        return status;
1357}
1358
1359int be_cmd_txq_create(struct be_adapter *adapter, struct be_tx_obj *txo)
1360{
1361        struct be_mcc_wrb wrb = {0};
1362        struct be_cmd_req_eth_tx_create *req;
1363        struct be_queue_info *txq = &txo->q;
1364        struct be_queue_info *cq = &txo->cq;
1365        struct be_dma_mem *q_mem = &txq->dma_mem;
1366        int status, ver = 0;
1367
1368        req = embedded_payload(&wrb);
1369        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
1370                               OPCODE_ETH_TX_CREATE, sizeof(*req), &wrb, NULL);
1371
1372        if (lancer_chip(adapter)) {
1373                req->hdr.version = 1;
1374        } else if (BEx_chip(adapter)) {
1375                if (adapter->function_caps & BE_FUNCTION_CAPS_SUPER_NIC)
1376                        req->hdr.version = 2;
1377        } else { /* For SH */
1378                req->hdr.version = 2;
1379        }
1380
1381        if (req->hdr.version > 0)
1382                req->if_id = cpu_to_le16(adapter->if_handle);
1383        req->num_pages = PAGES_4K_SPANNED(q_mem->va, q_mem->size);
1384        req->ulp_num = BE_ULP1_NUM;
1385        req->type = BE_ETH_TX_RING_TYPE_STANDARD;
1386        req->cq_id = cpu_to_le16(cq->id);
1387        req->queue_size = be_encoded_q_len(txq->len);
1388        be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1389        ver = req->hdr.version;
1390
1391        status = be_cmd_notify_wait(adapter, &wrb);
1392        if (!status) {
1393                struct be_cmd_resp_eth_tx_create *resp = embedded_payload(&wrb);
1394
1395                txq->id = le16_to_cpu(resp->cid);
1396                if (ver == 2)
1397                        txo->db_offset = le32_to_cpu(resp->db_offset);
1398                else
1399                        txo->db_offset = DB_TXULP1_OFFSET;
1400                txq->created = true;
1401        }
1402
1403        return status;
1404}
1405
1406/* Uses MCC */
1407int be_cmd_rxq_create(struct be_adapter *adapter,
1408                      struct be_queue_info *rxq, u16 cq_id, u16 frag_size,
1409                      u32 if_id, u32 rss, u8 *rss_id)
1410{
1411        struct be_mcc_wrb *wrb;
1412        struct be_cmd_req_eth_rx_create *req;
1413        struct be_dma_mem *q_mem = &rxq->dma_mem;
1414        int status;
1415
1416        mutex_lock(&adapter->mcc_lock);
1417
1418        wrb = wrb_from_mccq(adapter);
1419        if (!wrb) {
1420                status = -EBUSY;
1421                goto err;
1422        }
1423        req = embedded_payload(wrb);
1424
1425        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
1426                               OPCODE_ETH_RX_CREATE, sizeof(*req), wrb, NULL);
1427
1428        req->cq_id = cpu_to_le16(cq_id);
1429        req->frag_size = fls(frag_size) - 1;
1430        req->num_pages = 2;
1431        be_cmd_page_addrs_prepare(req->pages, ARRAY_SIZE(req->pages), q_mem);
1432        req->interface_id = cpu_to_le32(if_id);
1433        req->max_frame_size = cpu_to_le16(BE_MAX_JUMBO_FRAME_SIZE);
1434        req->rss_queue = cpu_to_le32(rss);
1435
1436        status = be_mcc_notify_wait(adapter);
1437        if (!status) {
1438                struct be_cmd_resp_eth_rx_create *resp = embedded_payload(wrb);
1439
1440                rxq->id = le16_to_cpu(resp->id);
1441                rxq->created = true;
1442                *rss_id = resp->rss_id;
1443        }
1444
1445err:
1446        mutex_unlock(&adapter->mcc_lock);
1447        return status;
1448}
1449
1450/* Generic destroyer function for all types of queues
1451 * Uses Mbox
1452 */
1453int be_cmd_q_destroy(struct be_adapter *adapter, struct be_queue_info *q,
1454                     int queue_type)
1455{
1456        struct be_mcc_wrb *wrb;
1457        struct be_cmd_req_q_destroy *req;
1458        u8 subsys = 0, opcode = 0;
1459        int status;
1460
1461        if (mutex_lock_interruptible(&adapter->mbox_lock))
1462                return -1;
1463
1464        wrb = wrb_from_mbox(adapter);
1465        req = embedded_payload(wrb);
1466
1467        switch (queue_type) {
1468        case QTYPE_EQ:
1469                subsys = CMD_SUBSYSTEM_COMMON;
1470                opcode = OPCODE_COMMON_EQ_DESTROY;
1471                break;
1472        case QTYPE_CQ:
1473                subsys = CMD_SUBSYSTEM_COMMON;
1474                opcode = OPCODE_COMMON_CQ_DESTROY;
1475                break;
1476        case QTYPE_TXQ:
1477                subsys = CMD_SUBSYSTEM_ETH;
1478                opcode = OPCODE_ETH_TX_DESTROY;
1479                break;
1480        case QTYPE_RXQ:
1481                subsys = CMD_SUBSYSTEM_ETH;
1482                opcode = OPCODE_ETH_RX_DESTROY;
1483                break;
1484        case QTYPE_MCCQ:
1485                subsys = CMD_SUBSYSTEM_COMMON;
1486                opcode = OPCODE_COMMON_MCC_DESTROY;
1487                break;
1488        default:
1489                BUG();
1490        }
1491
1492        be_wrb_cmd_hdr_prepare(&req->hdr, subsys, opcode, sizeof(*req), wrb,
1493                               NULL);
1494        req->id = cpu_to_le16(q->id);
1495
1496        status = be_mbox_notify_wait(adapter);
1497        q->created = false;
1498
1499        mutex_unlock(&adapter->mbox_lock);
1500        return status;
1501}
1502
1503/* Uses MCC */
1504int be_cmd_rxq_destroy(struct be_adapter *adapter, struct be_queue_info *q)
1505{
1506        struct be_mcc_wrb *wrb;
1507        struct be_cmd_req_q_destroy *req;
1508        int status;
1509
1510        mutex_lock(&adapter->mcc_lock);
1511
1512        wrb = wrb_from_mccq(adapter);
1513        if (!wrb) {
1514                status = -EBUSY;
1515                goto err;
1516        }
1517        req = embedded_payload(wrb);
1518
1519        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
1520                               OPCODE_ETH_RX_DESTROY, sizeof(*req), wrb, NULL);
1521        req->id = cpu_to_le16(q->id);
1522
1523        status = be_mcc_notify_wait(adapter);
1524        q->created = false;
1525
1526err:
1527        mutex_unlock(&adapter->mcc_lock);
1528        return status;
1529}
1530
1531/* Create an rx filtering policy configuration on an i/f
1532 * Will use MBOX only if MCCQ has not been created.
1533 */
1534int be_cmd_if_create(struct be_adapter *adapter, u32 cap_flags, u32 en_flags,
1535                     u32 *if_handle, u32 domain)
1536{
1537        struct be_mcc_wrb wrb = {0};
1538        struct be_cmd_req_if_create *req;
1539        int status;
1540
1541        req = embedded_payload(&wrb);
1542        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1543                               OPCODE_COMMON_NTWK_INTERFACE_CREATE,
1544                               sizeof(*req), &wrb, NULL);
1545        req->hdr.domain = domain;
1546        req->capability_flags = cpu_to_le32(cap_flags);
1547        req->enable_flags = cpu_to_le32(en_flags);
1548        req->pmac_invalid = true;
1549
1550        status = be_cmd_notify_wait(adapter, &wrb);
1551        if (!status) {
1552                struct be_cmd_resp_if_create *resp = embedded_payload(&wrb);
1553
1554                *if_handle = le32_to_cpu(resp->interface_id);
1555
1556                /* Hack to retrieve VF's pmac-id on BE3 */
1557                if (BE3_chip(adapter) && be_virtfn(adapter))
1558                        adapter->pmac_id[0] = le32_to_cpu(resp->pmac_id);
1559        }
1560        return status;
1561}
1562
1563/* Uses MCCQ if available else MBOX */
1564int be_cmd_if_destroy(struct be_adapter *adapter, int interface_id, u32 domain)
1565{
1566        struct be_mcc_wrb wrb = {0};
1567        struct be_cmd_req_if_destroy *req;
1568        int status;
1569
1570        if (interface_id == -1)
1571                return 0;
1572
1573        req = embedded_payload(&wrb);
1574
1575        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1576                               OPCODE_COMMON_NTWK_INTERFACE_DESTROY,
1577                               sizeof(*req), &wrb, NULL);
1578        req->hdr.domain = domain;
1579        req->interface_id = cpu_to_le32(interface_id);
1580
1581        status = be_cmd_notify_wait(adapter, &wrb);
1582        return status;
1583}
1584
1585/* Get stats is a non embedded command: the request is not embedded inside
1586 * WRB but is a separate dma memory block
1587 * Uses asynchronous MCC
1588 */
1589int be_cmd_get_stats(struct be_adapter *adapter, struct be_dma_mem *nonemb_cmd)
1590{
1591        struct be_mcc_wrb *wrb;
1592        struct be_cmd_req_hdr *hdr;
1593        int status = 0;
1594
1595        mutex_lock(&adapter->mcc_lock);
1596
1597        wrb = wrb_from_mccq(adapter);
1598        if (!wrb) {
1599                status = -EBUSY;
1600                goto err;
1601        }
1602        hdr = nonemb_cmd->va;
1603
1604        be_wrb_cmd_hdr_prepare(hdr, CMD_SUBSYSTEM_ETH,
1605                               OPCODE_ETH_GET_STATISTICS, nonemb_cmd->size, wrb,
1606                               nonemb_cmd);
1607
1608        /* version 1 of the cmd is not supported only by BE2 */
1609        if (BE2_chip(adapter))
1610                hdr->version = 0;
1611        if (BE3_chip(adapter) || lancer_chip(adapter))
1612                hdr->version = 1;
1613        else
1614                hdr->version = 2;
1615
1616        status = be_mcc_notify(adapter);
1617        if (status)
1618                goto err;
1619
1620        adapter->stats_cmd_sent = true;
1621
1622err:
1623        mutex_unlock(&adapter->mcc_lock);
1624        return status;
1625}
1626
1627/* Lancer Stats */
1628int lancer_cmd_get_pport_stats(struct be_adapter *adapter,
1629                               struct be_dma_mem *nonemb_cmd)
1630{
1631        struct be_mcc_wrb *wrb;
1632        struct lancer_cmd_req_pport_stats *req;
1633        int status = 0;
1634
1635        if (!be_cmd_allowed(adapter, OPCODE_ETH_GET_PPORT_STATS,
1636                            CMD_SUBSYSTEM_ETH))
1637                return -EPERM;
1638
1639        mutex_lock(&adapter->mcc_lock);
1640
1641        wrb = wrb_from_mccq(adapter);
1642        if (!wrb) {
1643                status = -EBUSY;
1644                goto err;
1645        }
1646        req = nonemb_cmd->va;
1647
1648        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
1649                               OPCODE_ETH_GET_PPORT_STATS, nonemb_cmd->size,
1650                               wrb, nonemb_cmd);
1651
1652        req->cmd_params.params.pport_num = cpu_to_le16(adapter->hba_port_num);
1653        req->cmd_params.params.reset_stats = 0;
1654
1655        status = be_mcc_notify(adapter);
1656        if (status)
1657                goto err;
1658
1659        adapter->stats_cmd_sent = true;
1660
1661err:
1662        mutex_unlock(&adapter->mcc_lock);
1663        return status;
1664}
1665
1666static int be_mac_to_link_speed(int mac_speed)
1667{
1668        switch (mac_speed) {
1669        case PHY_LINK_SPEED_ZERO:
1670                return 0;
1671        case PHY_LINK_SPEED_10MBPS:
1672                return 10;
1673        case PHY_LINK_SPEED_100MBPS:
1674                return 100;
1675        case PHY_LINK_SPEED_1GBPS:
1676                return 1000;
1677        case PHY_LINK_SPEED_10GBPS:
1678                return 10000;
1679        case PHY_LINK_SPEED_20GBPS:
1680                return 20000;
1681        case PHY_LINK_SPEED_25GBPS:
1682                return 25000;
1683        case PHY_LINK_SPEED_40GBPS:
1684                return 40000;
1685        }
1686        return 0;
1687}
1688
1689/* Uses synchronous mcc
1690 * Returns link_speed in Mbps
1691 */
1692int be_cmd_link_status_query(struct be_adapter *adapter, u16 *link_speed,
1693                             u8 *link_status, u32 dom)
1694{
1695        struct be_mcc_wrb *wrb;
1696        struct be_cmd_req_link_status *req;
1697        int status;
1698
1699        mutex_lock(&adapter->mcc_lock);
1700
1701        if (link_status)
1702                *link_status = LINK_DOWN;
1703
1704        wrb = wrb_from_mccq(adapter);
1705        if (!wrb) {
1706                status = -EBUSY;
1707                goto err;
1708        }
1709        req = embedded_payload(wrb);
1710
1711        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1712                               OPCODE_COMMON_NTWK_LINK_STATUS_QUERY,
1713                               sizeof(*req), wrb, NULL);
1714
1715        /* version 1 of the cmd is not supported only by BE2 */
1716        if (!BE2_chip(adapter))
1717                req->hdr.version = 1;
1718
1719        req->hdr.domain = dom;
1720
1721        status = be_mcc_notify_wait(adapter);
1722        if (!status) {
1723                struct be_cmd_resp_link_status *resp = embedded_payload(wrb);
1724
1725                if (link_speed) {
1726                        *link_speed = resp->link_speed ?
1727                                      le16_to_cpu(resp->link_speed) * 10 :
1728                                      be_mac_to_link_speed(resp->mac_speed);
1729
1730                        if (!resp->logical_link_status)
1731                                *link_speed = 0;
1732                }
1733                if (link_status)
1734                        *link_status = resp->logical_link_status;
1735        }
1736
1737err:
1738        mutex_unlock(&adapter->mcc_lock);
1739        return status;
1740}
1741
1742/* Uses synchronous mcc */
1743int be_cmd_get_die_temperature(struct be_adapter *adapter)
1744{
1745        struct be_mcc_wrb *wrb;
1746        struct be_cmd_req_get_cntl_addnl_attribs *req;
1747        int status = 0;
1748
1749        mutex_lock(&adapter->mcc_lock);
1750
1751        wrb = wrb_from_mccq(adapter);
1752        if (!wrb) {
1753                status = -EBUSY;
1754                goto err;
1755        }
1756        req = embedded_payload(wrb);
1757
1758        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1759                               OPCODE_COMMON_GET_CNTL_ADDITIONAL_ATTRIBUTES,
1760                               sizeof(*req), wrb, NULL);
1761
1762        status = be_mcc_notify(adapter);
1763err:
1764        mutex_unlock(&adapter->mcc_lock);
1765        return status;
1766}
1767
1768/* Uses synchronous mcc */
1769int be_cmd_get_fat_dump_len(struct be_adapter *adapter, u32 *dump_size)
1770{
1771        struct be_mcc_wrb wrb = {0};
1772        struct be_cmd_req_get_fat *req;
1773        int status;
1774
1775        req = embedded_payload(&wrb);
1776
1777        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1778                               OPCODE_COMMON_MANAGE_FAT, sizeof(*req),
1779                               &wrb, NULL);
1780        req->fat_operation = cpu_to_le32(QUERY_FAT);
1781        status = be_cmd_notify_wait(adapter, &wrb);
1782        if (!status) {
1783                struct be_cmd_resp_get_fat *resp = embedded_payload(&wrb);
1784
1785                if (dump_size && resp->log_size)
1786                        *dump_size = le32_to_cpu(resp->log_size) -
1787                                        sizeof(u32);
1788        }
1789        return status;
1790}
1791
1792int be_cmd_get_fat_dump(struct be_adapter *adapter, u32 buf_len, void *buf)
1793{
1794        struct be_dma_mem get_fat_cmd;
1795        struct be_mcc_wrb *wrb;
1796        struct be_cmd_req_get_fat *req;
1797        u32 offset = 0, total_size, buf_size,
1798                                log_offset = sizeof(u32), payload_len;
1799        int status;
1800
1801        if (buf_len == 0)
1802                return 0;
1803
1804        total_size = buf_len;
1805
1806        get_fat_cmd.size = sizeof(struct be_cmd_req_get_fat) + 60*1024;
1807        get_fat_cmd.va = dma_alloc_coherent(&adapter->pdev->dev,
1808                                            get_fat_cmd.size,
1809                                            &get_fat_cmd.dma, GFP_ATOMIC);
1810        if (!get_fat_cmd.va)
1811                return -ENOMEM;
1812
1813        mutex_lock(&adapter->mcc_lock);
1814
1815        while (total_size) {
1816                buf_size = min(total_size, (u32)60*1024);
1817                total_size -= buf_size;
1818
1819                wrb = wrb_from_mccq(adapter);
1820                if (!wrb) {
1821                        status = -EBUSY;
1822                        goto err;
1823                }
1824                req = get_fat_cmd.va;
1825
1826                payload_len = sizeof(struct be_cmd_req_get_fat) + buf_size;
1827                be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1828                                       OPCODE_COMMON_MANAGE_FAT, payload_len,
1829                                       wrb, &get_fat_cmd);
1830
1831                req->fat_operation = cpu_to_le32(RETRIEVE_FAT);
1832                req->read_log_offset = cpu_to_le32(log_offset);
1833                req->read_log_length = cpu_to_le32(buf_size);
1834                req->data_buffer_size = cpu_to_le32(buf_size);
1835
1836                status = be_mcc_notify_wait(adapter);
1837                if (!status) {
1838                        struct be_cmd_resp_get_fat *resp = get_fat_cmd.va;
1839
1840                        memcpy(buf + offset,
1841                               resp->data_buffer,
1842                               le32_to_cpu(resp->read_log_length));
1843                } else {
1844                        dev_err(&adapter->pdev->dev, "FAT Table Retrieve error\n");
1845                        goto err;
1846                }
1847                offset += buf_size;
1848                log_offset += buf_size;
1849        }
1850err:
1851        dma_free_coherent(&adapter->pdev->dev, get_fat_cmd.size,
1852                          get_fat_cmd.va, get_fat_cmd.dma);
1853        mutex_unlock(&adapter->mcc_lock);
1854        return status;
1855}
1856
1857/* Uses synchronous mcc */
1858int be_cmd_get_fw_ver(struct be_adapter *adapter)
1859{
1860        struct be_mcc_wrb *wrb;
1861        struct be_cmd_req_get_fw_version *req;
1862        int status;
1863
1864        mutex_lock(&adapter->mcc_lock);
1865
1866        wrb = wrb_from_mccq(adapter);
1867        if (!wrb) {
1868                status = -EBUSY;
1869                goto err;
1870        }
1871
1872        req = embedded_payload(wrb);
1873
1874        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1875                               OPCODE_COMMON_GET_FW_VERSION, sizeof(*req), wrb,
1876                               NULL);
1877        status = be_mcc_notify_wait(adapter);
1878        if (!status) {
1879                struct be_cmd_resp_get_fw_version *resp = embedded_payload(wrb);
1880
1881                strlcpy(adapter->fw_ver, resp->firmware_version_string,
1882                        sizeof(adapter->fw_ver));
1883                strlcpy(adapter->fw_on_flash, resp->fw_on_flash_version_string,
1884                        sizeof(adapter->fw_on_flash));
1885        }
1886err:
1887        mutex_unlock(&adapter->mcc_lock);
1888        return status;
1889}
1890
1891/* set the EQ delay interval of an EQ to specified value
1892 * Uses async mcc
1893 */
1894static int __be_cmd_modify_eqd(struct be_adapter *adapter,
1895                               struct be_set_eqd *set_eqd, int num)
1896{
1897        struct be_mcc_wrb *wrb;
1898        struct be_cmd_req_modify_eq_delay *req;
1899        int status = 0, i;
1900
1901        mutex_lock(&adapter->mcc_lock);
1902
1903        wrb = wrb_from_mccq(adapter);
1904        if (!wrb) {
1905                status = -EBUSY;
1906                goto err;
1907        }
1908        req = embedded_payload(wrb);
1909
1910        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1911                               OPCODE_COMMON_MODIFY_EQ_DELAY, sizeof(*req), wrb,
1912                               NULL);
1913
1914        req->num_eq = cpu_to_le32(num);
1915        for (i = 0; i < num; i++) {
1916                req->set_eqd[i].eq_id = cpu_to_le32(set_eqd[i].eq_id);
1917                req->set_eqd[i].phase = 0;
1918                req->set_eqd[i].delay_multiplier =
1919                                cpu_to_le32(set_eqd[i].delay_multiplier);
1920        }
1921
1922        status = be_mcc_notify(adapter);
1923err:
1924        mutex_unlock(&adapter->mcc_lock);
1925        return status;
1926}
1927
1928int be_cmd_modify_eqd(struct be_adapter *adapter, struct be_set_eqd *set_eqd,
1929                      int num)
1930{
1931        int num_eqs, i = 0;
1932
1933        while (num) {
1934                num_eqs = min(num, 8);
1935                __be_cmd_modify_eqd(adapter, &set_eqd[i], num_eqs);
1936                i += num_eqs;
1937                num -= num_eqs;
1938        }
1939
1940        return 0;
1941}
1942
1943/* Uses sycnhronous mcc */
1944int be_cmd_vlan_config(struct be_adapter *adapter, u32 if_id, u16 *vtag_array,
1945                       u32 num, u32 domain)
1946{
1947        struct be_mcc_wrb *wrb;
1948        struct be_cmd_req_vlan_config *req;
1949        int status;
1950
1951        mutex_lock(&adapter->mcc_lock);
1952
1953        wrb = wrb_from_mccq(adapter);
1954        if (!wrb) {
1955                status = -EBUSY;
1956                goto err;
1957        }
1958        req = embedded_payload(wrb);
1959
1960        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1961                               OPCODE_COMMON_NTWK_VLAN_CONFIG, sizeof(*req),
1962                               wrb, NULL);
1963        req->hdr.domain = domain;
1964
1965        req->interface_id = if_id;
1966        req->untagged = BE_IF_FLAGS_UNTAGGED & be_if_cap_flags(adapter) ? 1 : 0;
1967        req->num_vlan = num;
1968        memcpy(req->normal_vlan, vtag_array,
1969               req->num_vlan * sizeof(vtag_array[0]));
1970
1971        status = be_mcc_notify_wait(adapter);
1972err:
1973        mutex_unlock(&adapter->mcc_lock);
1974        return status;
1975}
1976
1977static int __be_cmd_rx_filter(struct be_adapter *adapter, u32 flags, u32 value)
1978{
1979        struct be_mcc_wrb *wrb;
1980        struct be_dma_mem *mem = &adapter->rx_filter;
1981        struct be_cmd_req_rx_filter *req = mem->va;
1982        int status;
1983
1984        mutex_lock(&adapter->mcc_lock);
1985
1986        wrb = wrb_from_mccq(adapter);
1987        if (!wrb) {
1988                status = -EBUSY;
1989                goto err;
1990        }
1991        memset(req, 0, sizeof(*req));
1992        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
1993                               OPCODE_COMMON_NTWK_RX_FILTER, sizeof(*req),
1994                               wrb, mem);
1995
1996        req->if_id = cpu_to_le32(adapter->if_handle);
1997        req->if_flags_mask = cpu_to_le32(flags);
1998        req->if_flags = (value == ON) ? req->if_flags_mask : 0;
1999
2000        if (flags & BE_IF_FLAGS_MULTICAST) {
2001                int i;
2002
2003                /* Reset mcast promisc mode if already set by setting mask
2004                 * and not setting flags field
2005                 */
2006                req->if_flags_mask |=
2007                        cpu_to_le32(BE_IF_FLAGS_MCAST_PROMISCUOUS &
2008                                    be_if_cap_flags(adapter));
2009                req->mcast_num = cpu_to_le32(adapter->mc_count);
2010                for (i = 0; i < adapter->mc_count; i++)
2011                        ether_addr_copy(req->mcast_mac[i].byte,
2012                                        adapter->mc_list[i].mac);
2013        }
2014
2015        status = be_mcc_notify_wait(adapter);
2016err:
2017        mutex_unlock(&adapter->mcc_lock);
2018        return status;
2019}
2020
2021int be_cmd_rx_filter(struct be_adapter *adapter, u32 flags, u32 value)
2022{
2023        struct device *dev = &adapter->pdev->dev;
2024
2025        if ((flags & be_if_cap_flags(adapter)) != flags) {
2026                dev_warn(dev, "Cannot set rx filter flags 0x%x\n", flags);
2027                dev_warn(dev, "Interface is capable of 0x%x flags only\n",
2028                         be_if_cap_flags(adapter));
2029        }
2030        flags &= be_if_cap_flags(adapter);
2031        if (!flags)
2032                return -ENOTSUPP;
2033
2034        return __be_cmd_rx_filter(adapter, flags, value);
2035}
2036
2037/* Uses synchrounous mcc */
2038int be_cmd_set_flow_control(struct be_adapter *adapter, u32 tx_fc, u32 rx_fc)
2039{
2040        struct be_mcc_wrb *wrb;
2041        struct be_cmd_req_set_flow_control *req;
2042        int status;
2043
2044        if (!be_cmd_allowed(adapter, OPCODE_COMMON_SET_FLOW_CONTROL,
2045                            CMD_SUBSYSTEM_COMMON))
2046                return -EPERM;
2047
2048        mutex_lock(&adapter->mcc_lock);
2049
2050        wrb = wrb_from_mccq(adapter);
2051        if (!wrb) {
2052                status = -EBUSY;
2053                goto err;
2054        }
2055        req = embedded_payload(wrb);
2056
2057        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2058                               OPCODE_COMMON_SET_FLOW_CONTROL, sizeof(*req),
2059                               wrb, NULL);
2060
2061        req->hdr.version = 1;
2062        req->tx_flow_control = cpu_to_le16((u16)tx_fc);
2063        req->rx_flow_control = cpu_to_le16((u16)rx_fc);
2064
2065        status = be_mcc_notify_wait(adapter);
2066
2067err:
2068        mutex_unlock(&adapter->mcc_lock);
2069
2070        if (base_status(status) == MCC_STATUS_FEATURE_NOT_SUPPORTED)
2071                return  -EOPNOTSUPP;
2072
2073        return status;
2074}
2075
2076/* Uses sycn mcc */
2077int be_cmd_get_flow_control(struct be_adapter *adapter, u32 *tx_fc, u32 *rx_fc)
2078{
2079        struct be_mcc_wrb *wrb;
2080        struct be_cmd_req_get_flow_control *req;
2081        int status;
2082
2083        if (!be_cmd_allowed(adapter, OPCODE_COMMON_GET_FLOW_CONTROL,
2084                            CMD_SUBSYSTEM_COMMON))
2085                return -EPERM;
2086
2087        mutex_lock(&adapter->mcc_lock);
2088
2089        wrb = wrb_from_mccq(adapter);
2090        if (!wrb) {
2091                status = -EBUSY;
2092                goto err;
2093        }
2094        req = embedded_payload(wrb);
2095
2096        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2097                               OPCODE_COMMON_GET_FLOW_CONTROL, sizeof(*req),
2098                               wrb, NULL);
2099
2100        status = be_mcc_notify_wait(adapter);
2101        if (!status) {
2102                struct be_cmd_resp_get_flow_control *resp =
2103                                                embedded_payload(wrb);
2104
2105                *tx_fc = le16_to_cpu(resp->tx_flow_control);
2106                *rx_fc = le16_to_cpu(resp->rx_flow_control);
2107        }
2108
2109err:
2110        mutex_unlock(&adapter->mcc_lock);
2111        return status;
2112}
2113
2114/* Uses mbox */
2115int be_cmd_query_fw_cfg(struct be_adapter *adapter)
2116{
2117        struct be_mcc_wrb *wrb;
2118        struct be_cmd_req_query_fw_cfg *req;
2119        int status;
2120
2121        if (mutex_lock_interruptible(&adapter->mbox_lock))
2122                return -1;
2123
2124        wrb = wrb_from_mbox(adapter);
2125        req = embedded_payload(wrb);
2126
2127        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2128                               OPCODE_COMMON_QUERY_FIRMWARE_CONFIG,
2129                               sizeof(*req), wrb, NULL);
2130
2131        status = be_mbox_notify_wait(adapter);
2132        if (!status) {
2133                struct be_cmd_resp_query_fw_cfg *resp = embedded_payload(wrb);
2134
2135                adapter->port_num = le32_to_cpu(resp->phys_port);
2136                adapter->function_mode = le32_to_cpu(resp->function_mode);
2137                adapter->function_caps = le32_to_cpu(resp->function_caps);
2138                adapter->asic_rev = le32_to_cpu(resp->asic_revision) & 0xFF;
2139                dev_info(&adapter->pdev->dev,
2140                         "FW config: function_mode=0x%x, function_caps=0x%x\n",
2141                         adapter->function_mode, adapter->function_caps);
2142        }
2143
2144        mutex_unlock(&adapter->mbox_lock);
2145        return status;
2146}
2147
2148/* Uses mbox */
2149int be_cmd_reset_function(struct be_adapter *adapter)
2150{
2151        struct be_mcc_wrb *wrb;
2152        struct be_cmd_req_hdr *req;
2153        int status;
2154
2155        if (lancer_chip(adapter)) {
2156                iowrite32(SLI_PORT_CONTROL_IP_MASK,
2157                          adapter->db + SLIPORT_CONTROL_OFFSET);
2158                status = lancer_wait_ready(adapter);
2159                if (status)
2160                        dev_err(&adapter->pdev->dev,
2161                                "Adapter in non recoverable error\n");
2162                return status;
2163        }
2164
2165        if (mutex_lock_interruptible(&adapter->mbox_lock))
2166                return -1;
2167
2168        wrb = wrb_from_mbox(adapter);
2169        req = embedded_payload(wrb);
2170
2171        be_wrb_cmd_hdr_prepare(req, CMD_SUBSYSTEM_COMMON,
2172                               OPCODE_COMMON_FUNCTION_RESET, sizeof(*req), wrb,
2173                               NULL);
2174
2175        status = be_mbox_notify_wait(adapter);
2176
2177        mutex_unlock(&adapter->mbox_lock);
2178        return status;
2179}
2180
2181int be_cmd_rss_config(struct be_adapter *adapter, u8 *rsstable,
2182                      u32 rss_hash_opts, u16 table_size, const u8 *rss_hkey)
2183{
2184        struct be_mcc_wrb *wrb;
2185        struct be_cmd_req_rss_config *req;
2186        int status;
2187
2188        if (!(be_if_cap_flags(adapter) & BE_IF_FLAGS_RSS))
2189                return 0;
2190
2191        mutex_lock(&adapter->mcc_lock);
2192
2193        wrb = wrb_from_mccq(adapter);
2194        if (!wrb) {
2195                status = -EBUSY;
2196                goto err;
2197        }
2198        req = embedded_payload(wrb);
2199
2200        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
2201                               OPCODE_ETH_RSS_CONFIG, sizeof(*req), wrb, NULL);
2202
2203        req->if_id = cpu_to_le32(adapter->if_handle);
2204        req->enable_rss = cpu_to_le16(rss_hash_opts);
2205        req->cpu_table_size_log2 = cpu_to_le16(fls(table_size) - 1);
2206
2207        if (!BEx_chip(adapter))
2208                req->hdr.version = 1;
2209
2210        memcpy(req->cpu_table, rsstable, table_size);
2211        memcpy(req->hash, rss_hkey, RSS_HASH_KEY_LEN);
2212        be_dws_cpu_to_le(req->hash, sizeof(req->hash));
2213
2214        status = be_mcc_notify_wait(adapter);
2215err:
2216        mutex_unlock(&adapter->mcc_lock);
2217        return status;
2218}
2219
2220/* Uses sync mcc */
2221int be_cmd_set_beacon_state(struct be_adapter *adapter, u8 port_num,
2222                            u8 bcn, u8 sts, u8 state)
2223{
2224        struct be_mcc_wrb *wrb;
2225        struct be_cmd_req_enable_disable_beacon *req;
2226        int status;
2227
2228        mutex_lock(&adapter->mcc_lock);
2229
2230        wrb = wrb_from_mccq(adapter);
2231        if (!wrb) {
2232                status = -EBUSY;
2233                goto err;
2234        }
2235        req = embedded_payload(wrb);
2236
2237        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2238                               OPCODE_COMMON_ENABLE_DISABLE_BEACON,
2239                               sizeof(*req), wrb, NULL);
2240
2241        req->port_num = port_num;
2242        req->beacon_state = state;
2243        req->beacon_duration = bcn;
2244        req->status_duration = sts;
2245
2246        status = be_mcc_notify_wait(adapter);
2247
2248err:
2249        mutex_unlock(&adapter->mcc_lock);
2250        return status;
2251}
2252
2253/* Uses sync mcc */
2254int be_cmd_get_beacon_state(struct be_adapter *adapter, u8 port_num, u32 *state)
2255{
2256        struct be_mcc_wrb *wrb;
2257        struct be_cmd_req_get_beacon_state *req;
2258        int status;
2259
2260        mutex_lock(&adapter->mcc_lock);
2261
2262        wrb = wrb_from_mccq(adapter);
2263        if (!wrb) {
2264                status = -EBUSY;
2265                goto err;
2266        }
2267        req = embedded_payload(wrb);
2268
2269        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2270                               OPCODE_COMMON_GET_BEACON_STATE, sizeof(*req),
2271                               wrb, NULL);
2272
2273        req->port_num = port_num;
2274
2275        status = be_mcc_notify_wait(adapter);
2276        if (!status) {
2277                struct be_cmd_resp_get_beacon_state *resp =
2278                                                embedded_payload(wrb);
2279
2280                *state = resp->beacon_state;
2281        }
2282
2283err:
2284        mutex_unlock(&adapter->mcc_lock);
2285        return status;
2286}
2287
2288/* Uses sync mcc */
2289int be_cmd_read_port_transceiver_data(struct be_adapter *adapter,
2290                                      u8 page_num, u8 *data)
2291{
2292        struct be_dma_mem cmd;
2293        struct be_mcc_wrb *wrb;
2294        struct be_cmd_req_port_type *req;
2295        int status;
2296
2297        if (page_num > TR_PAGE_A2)
2298                return -EINVAL;
2299
2300        cmd.size = sizeof(struct be_cmd_resp_port_type);
2301        cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
2302                                    GFP_ATOMIC);
2303        if (!cmd.va) {
2304                dev_err(&adapter->pdev->dev, "Memory allocation failed\n");
2305                return -ENOMEM;
2306        }
2307
2308        mutex_lock(&adapter->mcc_lock);
2309
2310        wrb = wrb_from_mccq(adapter);
2311        if (!wrb) {
2312                status = -EBUSY;
2313                goto err;
2314        }
2315        req = cmd.va;
2316
2317        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2318                               OPCODE_COMMON_READ_TRANSRECV_DATA,
2319                               cmd.size, wrb, &cmd);
2320
2321        req->port = cpu_to_le32(adapter->hba_port_num);
2322        req->page_num = cpu_to_le32(page_num);
2323        status = be_mcc_notify_wait(adapter);
2324        if (!status) {
2325                struct be_cmd_resp_port_type *resp = cmd.va;
2326
2327                memcpy(data, resp->page_data, PAGE_DATA_LEN);
2328        }
2329err:
2330        mutex_unlock(&adapter->mcc_lock);
2331        dma_free_coherent(&adapter->pdev->dev, cmd.size, cmd.va, cmd.dma);
2332        return status;
2333}
2334
2335static int lancer_cmd_write_object(struct be_adapter *adapter,
2336                                   struct be_dma_mem *cmd, u32 data_size,
2337                                   u32 data_offset, const char *obj_name,
2338                                   u32 *data_written, u8 *change_status,
2339                                   u8 *addn_status)
2340{
2341        struct be_mcc_wrb *wrb;
2342        struct lancer_cmd_req_write_object *req;
2343        struct lancer_cmd_resp_write_object *resp;
2344        void *ctxt = NULL;
2345        int status;
2346
2347        mutex_lock(&adapter->mcc_lock);
2348        adapter->flash_status = 0;
2349
2350        wrb = wrb_from_mccq(adapter);
2351        if (!wrb) {
2352                status = -EBUSY;
2353                goto err_unlock;
2354        }
2355
2356        req = embedded_payload(wrb);
2357
2358        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2359                               OPCODE_COMMON_WRITE_OBJECT,
2360                               sizeof(struct lancer_cmd_req_write_object), wrb,
2361                               NULL);
2362
2363        ctxt = &req->context;
2364        AMAP_SET_BITS(struct amap_lancer_write_obj_context,
2365                      write_length, ctxt, data_size);
2366
2367        if (data_size == 0)
2368                AMAP_SET_BITS(struct amap_lancer_write_obj_context,
2369                              eof, ctxt, 1);
2370        else
2371                AMAP_SET_BITS(struct amap_lancer_write_obj_context,
2372                              eof, ctxt, 0);
2373
2374        be_dws_cpu_to_le(ctxt, sizeof(req->context));
2375        req->write_offset = cpu_to_le32(data_offset);
2376        strlcpy(req->object_name, obj_name, sizeof(req->object_name));
2377        req->descriptor_count = cpu_to_le32(1);
2378        req->buf_len = cpu_to_le32(data_size);
2379        req->addr_low = cpu_to_le32((cmd->dma +
2380                                     sizeof(struct lancer_cmd_req_write_object))
2381                                    & 0xFFFFFFFF);
2382        req->addr_high = cpu_to_le32(upper_32_bits(cmd->dma +
2383                                sizeof(struct lancer_cmd_req_write_object)));
2384
2385        status = be_mcc_notify(adapter);
2386        if (status)
2387                goto err_unlock;
2388
2389        mutex_unlock(&adapter->mcc_lock);
2390
2391        if (!wait_for_completion_timeout(&adapter->et_cmd_compl,
2392                                         msecs_to_jiffies(60000)))
2393                status = -ETIMEDOUT;
2394        else
2395                status = adapter->flash_status;
2396
2397        resp = embedded_payload(wrb);
2398        if (!status) {
2399                *data_written = le32_to_cpu(resp->actual_write_len);
2400                *change_status = resp->change_status;
2401        } else {
2402                *addn_status = resp->additional_status;
2403        }
2404
2405        return status;
2406
2407err_unlock:
2408        mutex_unlock(&adapter->mcc_lock);
2409        return status;
2410}
2411
2412int be_cmd_query_cable_type(struct be_adapter *adapter)
2413{
2414        u8 page_data[PAGE_DATA_LEN];
2415        int status;
2416
2417        status = be_cmd_read_port_transceiver_data(adapter, TR_PAGE_A0,
2418                                                   page_data);
2419        if (!status) {
2420                switch (adapter->phy.interface_type) {
2421                case PHY_TYPE_QSFP:
2422                        adapter->phy.cable_type =
2423                                page_data[QSFP_PLUS_CABLE_TYPE_OFFSET];
2424                        break;
2425                case PHY_TYPE_SFP_PLUS_10GB:
2426                        adapter->phy.cable_type =
2427                                page_data[SFP_PLUS_CABLE_TYPE_OFFSET];
2428                        break;
2429                default:
2430                        adapter->phy.cable_type = 0;
2431                        break;
2432                }
2433        }
2434        return status;
2435}
2436
2437int be_cmd_query_sfp_info(struct be_adapter *adapter)
2438{
2439        u8 page_data[PAGE_DATA_LEN];
2440        int status;
2441
2442        status = be_cmd_read_port_transceiver_data(adapter, TR_PAGE_A0,
2443                                                   page_data);
2444        if (!status) {
2445                strlcpy(adapter->phy.vendor_name, page_data +
2446                        SFP_VENDOR_NAME_OFFSET, SFP_VENDOR_NAME_LEN - 1);
2447                strlcpy(adapter->phy.vendor_pn,
2448                        page_data + SFP_VENDOR_PN_OFFSET,
2449                        SFP_VENDOR_NAME_LEN - 1);
2450        }
2451
2452        return status;
2453}
2454
2455static int lancer_cmd_delete_object(struct be_adapter *adapter,
2456                                    const char *obj_name)
2457{
2458        struct lancer_cmd_req_delete_object *req;
2459        struct be_mcc_wrb *wrb;
2460        int status;
2461
2462        mutex_lock(&adapter->mcc_lock);
2463
2464        wrb = wrb_from_mccq(adapter);
2465        if (!wrb) {
2466                status = -EBUSY;
2467                goto err;
2468        }
2469
2470        req = embedded_payload(wrb);
2471
2472        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2473                               OPCODE_COMMON_DELETE_OBJECT,
2474                               sizeof(*req), wrb, NULL);
2475
2476        strlcpy(req->object_name, obj_name, sizeof(req->object_name));
2477
2478        status = be_mcc_notify_wait(adapter);
2479err:
2480        mutex_unlock(&adapter->mcc_lock);
2481        return status;
2482}
2483
2484int lancer_cmd_read_object(struct be_adapter *adapter, struct be_dma_mem *cmd,
2485                           u32 data_size, u32 data_offset, const char *obj_name,
2486                           u32 *data_read, u32 *eof, u8 *addn_status)
2487{
2488        struct be_mcc_wrb *wrb;
2489        struct lancer_cmd_req_read_object *req;
2490        struct lancer_cmd_resp_read_object *resp;
2491        int status;
2492
2493        mutex_lock(&adapter->mcc_lock);
2494
2495        wrb = wrb_from_mccq(adapter);
2496        if (!wrb) {
2497                status = -EBUSY;
2498                goto err_unlock;
2499        }
2500
2501        req = embedded_payload(wrb);
2502
2503        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2504                               OPCODE_COMMON_READ_OBJECT,
2505                               sizeof(struct lancer_cmd_req_read_object), wrb,
2506                               NULL);
2507
2508        req->desired_read_len = cpu_to_le32(data_size);
2509        req->read_offset = cpu_to_le32(data_offset);
2510        strcpy(req->object_name, obj_name);
2511        req->descriptor_count = cpu_to_le32(1);
2512        req->buf_len = cpu_to_le32(data_size);
2513        req->addr_low = cpu_to_le32((cmd->dma & 0xFFFFFFFF));
2514        req->addr_high = cpu_to_le32(upper_32_bits(cmd->dma));
2515
2516        status = be_mcc_notify_wait(adapter);
2517
2518        resp = embedded_payload(wrb);
2519        if (!status) {
2520                *data_read = le32_to_cpu(resp->actual_read_len);
2521                *eof = le32_to_cpu(resp->eof);
2522        } else {
2523                *addn_status = resp->additional_status;
2524        }
2525
2526err_unlock:
2527        mutex_unlock(&adapter->mcc_lock);
2528        return status;
2529}
2530
2531static int be_cmd_write_flashrom(struct be_adapter *adapter,
2532                                 struct be_dma_mem *cmd, u32 flash_type,
2533                                 u32 flash_opcode, u32 img_offset, u32 buf_size)
2534{
2535        struct be_mcc_wrb *wrb;
2536        struct be_cmd_write_flashrom *req;
2537        int status;
2538
2539        mutex_lock(&adapter->mcc_lock);
2540        adapter->flash_status = 0;
2541
2542        wrb = wrb_from_mccq(adapter);
2543        if (!wrb) {
2544                status = -EBUSY;
2545                goto err_unlock;
2546        }
2547        req = cmd->va;
2548
2549        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2550                               OPCODE_COMMON_WRITE_FLASHROM, cmd->size, wrb,
2551                               cmd);
2552
2553        req->params.op_type = cpu_to_le32(flash_type);
2554        if (flash_type == OPTYPE_OFFSET_SPECIFIED)
2555                req->params.offset = cpu_to_le32(img_offset);
2556
2557        req->params.op_code = cpu_to_le32(flash_opcode);
2558        req->params.data_buf_size = cpu_to_le32(buf_size);
2559
2560        status = be_mcc_notify(adapter);
2561        if (status)
2562                goto err_unlock;
2563
2564        mutex_unlock(&adapter->mcc_lock);
2565
2566        if (!wait_for_completion_timeout(&adapter->et_cmd_compl,
2567                                         msecs_to_jiffies(40000)))
2568                status = -ETIMEDOUT;
2569        else
2570                status = adapter->flash_status;
2571
2572        return status;
2573
2574err_unlock:
2575        mutex_unlock(&adapter->mcc_lock);
2576        return status;
2577}
2578
2579static int be_cmd_get_flash_crc(struct be_adapter *adapter, u8 *flashed_crc,
2580                                u16 img_optype, u32 img_offset, u32 crc_offset)
2581{
2582        struct be_cmd_read_flash_crc *req;
2583        struct be_mcc_wrb *wrb;
2584        int status;
2585
2586        mutex_lock(&adapter->mcc_lock);
2587
2588        wrb = wrb_from_mccq(adapter);
2589        if (!wrb) {
2590                status = -EBUSY;
2591                goto err;
2592        }
2593        req = embedded_payload(wrb);
2594
2595        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
2596                               OPCODE_COMMON_READ_FLASHROM, sizeof(*req),
2597                               wrb, NULL);
2598
2599        req->params.op_type = cpu_to_le32(img_optype);
2600        if (img_optype == OPTYPE_OFFSET_SPECIFIED)
2601                req->params.offset = cpu_to_le32(img_offset + crc_offset);
2602        else
2603                req->params.offset = cpu_to_le32(crc_offset);
2604
2605        req->params.op_code = cpu_to_le32(FLASHROM_OPER_REPORT);
2606        req->params.data_buf_size = cpu_to_le32(0x4);
2607
2608        status = be_mcc_notify_wait(adapter);
2609        if (!status)
2610                memcpy(flashed_crc, req->crc, 4);
2611
2612err:
2613        mutex_unlock(&adapter->mcc_lock);
2614        return status;
2615}
2616
2617static char flash_cookie[2][16] = {"*** SE FLAS", "H DIRECTORY *** "};
2618
2619static bool phy_flashing_required(struct be_adapter *adapter)
2620{
2621        return (adapter->phy.phy_type == PHY_TYPE_TN_8022 &&
2622                adapter->phy.interface_type == PHY_TYPE_BASET_10GB);
2623}
2624
2625static bool is_comp_in_ufi(struct be_adapter *adapter,
2626                           struct flash_section_info *fsec, int type)
2627{
2628        int i = 0, img_type = 0;
2629        struct flash_section_info_g2 *fsec_g2 = NULL;
2630
2631        if (BE2_chip(adapter))
2632                fsec_g2 = (struct flash_section_info_g2 *)fsec;
2633
2634        for (i = 0; i < MAX_FLASH_COMP; i++) {
2635                if (fsec_g2)
2636                        img_type = le32_to_cpu(fsec_g2->fsec_entry[i].type);
2637                else
2638                        img_type = le32_to_cpu(fsec->fsec_entry[i].type);
2639
2640                if (img_type == type)
2641                        return true;
2642        }
2643        return false;
2644}
2645
2646static struct flash_section_info *get_fsec_info(struct be_adapter *adapter,
2647                                                int header_size,
2648                                                const struct firmware *fw)
2649{
2650        struct flash_section_info *fsec = NULL;
2651        const u8 *p = fw->data;
2652
2653        p += header_size;
2654        while (p < (fw->data + fw->size)) {
2655                fsec = (struct flash_section_info *)p;
2656                if (!memcmp(flash_cookie, fsec->cookie, sizeof(flash_cookie)))
2657                        return fsec;
2658                p += 32;
2659        }
2660        return NULL;
2661}
2662
2663static int be_check_flash_crc(struct be_adapter *adapter, const u8 *p,
2664                              u32 img_offset, u32 img_size, int hdr_size,
2665                              u16 img_optype, bool *crc_match)
2666{
2667        u32 crc_offset;
2668        int status;
2669        u8 crc[4];
2670
2671        status = be_cmd_get_flash_crc(adapter, crc, img_optype, img_offset,
2672                                      img_size - 4);
2673        if (status)
2674                return status;
2675
2676        crc_offset = hdr_size + img_offset + img_size - 4;
2677
2678        /* Skip flashing, if crc of flashed region matches */
2679        if (!memcmp(crc, p + crc_offset, 4))
2680                *crc_match = true;
2681        else
2682                *crc_match = false;
2683
2684        return status;
2685}
2686
2687static int be_flash(struct be_adapter *adapter, const u8 *img,
2688                    struct be_dma_mem *flash_cmd, int optype, int img_size,
2689                    u32 img_offset)
2690{
2691        u32 flash_op, num_bytes, total_bytes = img_size, bytes_sent = 0;
2692        struct be_cmd_write_flashrom *req = flash_cmd->va;
2693        int status;
2694
2695        while (total_bytes) {
2696                num_bytes = min_t(u32, 32 * 1024, total_bytes);
2697
2698                total_bytes -= num_bytes;
2699
2700                if (!total_bytes) {
2701                        if (optype == OPTYPE_PHY_FW)
2702                                flash_op = FLASHROM_OPER_PHY_FLASH;
2703                        else
2704                                flash_op = FLASHROM_OPER_FLASH;
2705                } else {
2706                        if (optype == OPTYPE_PHY_FW)
2707                                flash_op = FLASHROM_OPER_PHY_SAVE;
2708                        else
2709                                flash_op = FLASHROM_OPER_SAVE;
2710                }
2711
2712                memcpy(req->data_buf, img, num_bytes);
2713                img += num_bytes;
2714                status = be_cmd_write_flashrom(adapter, flash_cmd, optype,
2715                                               flash_op, img_offset +
2716                                               bytes_sent, num_bytes);
2717                if (base_status(status) == MCC_STATUS_ILLEGAL_REQUEST &&
2718                    optype == OPTYPE_PHY_FW)
2719                        break;
2720                else if (status)
2721                        return status;
2722
2723                bytes_sent += num_bytes;
2724        }
2725        return 0;
2726}
2727
2728#define NCSI_UPDATE_LOG "NCSI section update is not supported in FW ver %s\n"
2729static bool be_fw_ncsi_supported(char *ver)
2730{
2731        int v1[4] = {3, 102, 148, 0}; /* Min ver that supports NCSI FW */
2732        int v2[4];
2733        int i;
2734
2735        if (sscanf(ver, "%d.%d.%d.%d", &v2[0], &v2[1], &v2[2], &v2[3]) != 4)
2736                return false;
2737
2738        for (i = 0; i < 4; i++) {
2739                if (v1[i] < v2[i])
2740                        return true;
2741                else if (v1[i] > v2[i])
2742                        return false;
2743        }
2744
2745        return true;
2746}
2747
2748/* For BE2, BE3 and BE3-R */
2749static int be_flash_BEx(struct be_adapter *adapter,
2750                        const struct firmware *fw,
2751                        struct be_dma_mem *flash_cmd, int num_of_images)
2752{
2753        int img_hdrs_size = (num_of_images * sizeof(struct image_hdr));
2754        struct device *dev = &adapter->pdev->dev;
2755        struct flash_section_info *fsec = NULL;
2756        int status, i, filehdr_size, num_comp;
2757        const struct flash_comp *pflashcomp;
2758        bool crc_match;
2759        const u8 *p;
2760
2761        static const struct flash_comp gen3_flash_types[] = {
2762                { BE3_ISCSI_PRIMARY_IMAGE_START, OPTYPE_ISCSI_ACTIVE,
2763                        BE3_COMP_MAX_SIZE, IMAGE_FIRMWARE_ISCSI},
2764                { BE3_REDBOOT_START, OPTYPE_REDBOOT,
2765                        BE3_REDBOOT_COMP_MAX_SIZE, IMAGE_BOOT_CODE},
2766                { BE3_ISCSI_BIOS_START, OPTYPE_BIOS,
2767                        BE3_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_ISCSI},
2768                { BE3_PXE_BIOS_START, OPTYPE_PXE_BIOS,
2769                        BE3_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_PXE},
2770                { BE3_FCOE_BIOS_START, OPTYPE_FCOE_BIOS,
2771                        BE3_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_FCOE},
2772                { BE3_ISCSI_BACKUP_IMAGE_START, OPTYPE_ISCSI_BACKUP,
2773                        BE3_COMP_MAX_SIZE, IMAGE_FIRMWARE_BACKUP_ISCSI},
2774                { BE3_FCOE_PRIMARY_IMAGE_START, OPTYPE_FCOE_FW_ACTIVE,
2775                        BE3_COMP_MAX_SIZE, IMAGE_FIRMWARE_FCOE},
2776                { BE3_FCOE_BACKUP_IMAGE_START, OPTYPE_FCOE_FW_BACKUP,
2777                        BE3_COMP_MAX_SIZE, IMAGE_FIRMWARE_BACKUP_FCOE},
2778                { BE3_NCSI_START, OPTYPE_NCSI_FW,
2779                        BE3_NCSI_COMP_MAX_SIZE, IMAGE_NCSI},
2780                { BE3_PHY_FW_START, OPTYPE_PHY_FW,
2781                        BE3_PHY_FW_COMP_MAX_SIZE, IMAGE_FIRMWARE_PHY}
2782        };
2783
2784        static const struct flash_comp gen2_flash_types[] = {
2785                { BE2_ISCSI_PRIMARY_IMAGE_START, OPTYPE_ISCSI_ACTIVE,
2786                        BE2_COMP_MAX_SIZE, IMAGE_FIRMWARE_ISCSI},
2787                { BE2_REDBOOT_START, OPTYPE_REDBOOT,
2788                        BE2_REDBOOT_COMP_MAX_SIZE, IMAGE_BOOT_CODE},
2789                { BE2_ISCSI_BIOS_START, OPTYPE_BIOS,
2790                        BE2_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_ISCSI},
2791                { BE2_PXE_BIOS_START, OPTYPE_PXE_BIOS,
2792                        BE2_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_PXE},
2793                { BE2_FCOE_BIOS_START, OPTYPE_FCOE_BIOS,
2794                        BE2_BIOS_COMP_MAX_SIZE, IMAGE_OPTION_ROM_FCOE},
2795                { BE2_ISCSI_BACKUP_IMAGE_START, OPTYPE_ISCSI_BACKUP,
2796                        BE2_COMP_MAX_SIZE, IMAGE_FIRMWARE_BACKUP_ISCSI},
2797                { BE2_FCOE_PRIMARY_IMAGE_START, OPTYPE_FCOE_FW_ACTIVE,
2798                        BE2_COMP_MAX_SIZE, IMAGE_FIRMWARE_FCOE},
2799                { BE2_FCOE_BACKUP_IMAGE_START, OPTYPE_FCOE_FW_BACKUP,
2800                         BE2_COMP_MAX_SIZE, IMAGE_FIRMWARE_BACKUP_FCOE}
2801        };
2802
2803        if (BE3_chip(adapter)) {
2804                pflashcomp = gen3_flash_types;
2805                filehdr_size = sizeof(struct flash_file_hdr_g3);
2806                num_comp = ARRAY_SIZE(gen3_flash_types);
2807        } else {
2808                pflashcomp = gen2_flash_types;
2809                filehdr_size = sizeof(struct flash_file_hdr_g2);
2810                num_comp = ARRAY_SIZE(gen2_flash_types);
2811                img_hdrs_size = 0;
2812        }
2813
2814        /* Get flash section info*/
2815        fsec = get_fsec_info(adapter, filehdr_size + img_hdrs_size, fw);
2816        if (!fsec) {
2817                dev_err(dev, "Invalid Cookie. FW image may be corrupted\n");
2818                return -1;
2819        }
2820        for (i = 0; i < num_comp; i++) {
2821                if (!is_comp_in_ufi(adapter, fsec, pflashcomp[i].img_type))
2822                        continue;
2823
2824                if ((pflashcomp[i].optype == OPTYPE_NCSI_FW) &&
2825                    !be_fw_ncsi_supported(adapter->fw_ver)) {
2826                        dev_info(dev, NCSI_UPDATE_LOG, adapter->fw_ver);
2827                        continue;
2828                }
2829
2830                if (pflashcomp[i].optype == OPTYPE_PHY_FW  &&
2831                    !phy_flashing_required(adapter))
2832                        continue;
2833
2834                if (pflashcomp[i].optype == OPTYPE_REDBOOT) {
2835                        status = be_check_flash_crc(adapter, fw->data,
2836                                                    pflashcomp[i].offset,
2837                                                    pflashcomp[i].size,
2838                                                    filehdr_size +
2839                                                    img_hdrs_size,
2840                                                    OPTYPE_REDBOOT, &crc_match);
2841                        if (status) {
2842                                dev_err(dev,
2843                                        "Could not get CRC for 0x%x region\n",
2844                                        pflashcomp[i].optype);
2845                                continue;
2846                        }
2847
2848                        if (crc_match)
2849                                continue;
2850                }
2851
2852                p = fw->data + filehdr_size + pflashcomp[i].offset +
2853                        img_hdrs_size;
2854                if (p + pflashcomp[i].size > fw->data + fw->size)
2855                        return -1;
2856
2857                status = be_flash(adapter, p, flash_cmd, pflashcomp[i].optype,
2858                                  pflashcomp[i].size, 0);
2859                if (status) {
2860                        dev_err(dev, "Flashing section type 0x%x failed\n",
2861                                pflashcomp[i].img_type);
2862                        return status;
2863                }
2864        }
2865        return 0;
2866}
2867
2868static u16 be_get_img_optype(struct flash_section_entry fsec_entry)
2869{
2870        u32 img_type = le32_to_cpu(fsec_entry.type);
2871        u16 img_optype = le16_to_cpu(fsec_entry.optype);
2872
2873        if (img_optype != 0xFFFF)
2874                return img_optype;
2875
2876        switch (img_type) {
2877        case IMAGE_FIRMWARE_ISCSI:
2878                img_optype = OPTYPE_ISCSI_ACTIVE;
2879                break;
2880        case IMAGE_BOOT_CODE:
2881                img_optype = OPTYPE_REDBOOT;
2882                break;
2883        case IMAGE_OPTION_ROM_ISCSI:
2884                img_optype = OPTYPE_BIOS;
2885                break;
2886        case IMAGE_OPTION_ROM_PXE:
2887                img_optype = OPTYPE_PXE_BIOS;
2888                break;
2889        case IMAGE_OPTION_ROM_FCOE:
2890                img_optype = OPTYPE_FCOE_BIOS;
2891                break;
2892        case IMAGE_FIRMWARE_BACKUP_ISCSI:
2893                img_optype = OPTYPE_ISCSI_BACKUP;
2894                break;
2895        case IMAGE_NCSI:
2896                img_optype = OPTYPE_NCSI_FW;
2897                break;
2898        case IMAGE_FLASHISM_JUMPVECTOR:
2899                img_optype = OPTYPE_FLASHISM_JUMPVECTOR;
2900                break;
2901        case IMAGE_FIRMWARE_PHY:
2902                img_optype = OPTYPE_SH_PHY_FW;
2903                break;
2904        case IMAGE_REDBOOT_DIR:
2905                img_optype = OPTYPE_REDBOOT_DIR;
2906                break;
2907        case IMAGE_REDBOOT_CONFIG:
2908                img_optype = OPTYPE_REDBOOT_CONFIG;
2909                break;
2910        case IMAGE_UFI_DIR:
2911                img_optype = OPTYPE_UFI_DIR;
2912                break;
2913        default:
2914                break;
2915        }
2916
2917        return img_optype;
2918}
2919
2920static int be_flash_skyhawk(struct be_adapter *adapter,
2921                            const struct firmware *fw,
2922                            struct be_dma_mem *flash_cmd, int num_of_images)
2923{
2924        int img_hdrs_size = num_of_images * sizeof(struct image_hdr);
2925        bool crc_match, old_fw_img, flash_offset_support = true;
2926        struct device *dev = &adapter->pdev->dev;
2927        struct flash_section_info *fsec = NULL;
2928        u32 img_offset, img_size, img_type;
2929        u16 img_optype, flash_optype;
2930        int status, i, filehdr_size;
2931        const u8 *p;
2932
2933        filehdr_size = sizeof(struct flash_file_hdr_g3);
2934        fsec = get_fsec_info(adapter, filehdr_size + img_hdrs_size, fw);
2935        if (!fsec) {
2936                dev_err(dev, "Invalid Cookie. FW image may be corrupted\n");
2937                return -EINVAL;
2938        }
2939
2940retry_flash:
2941        for (i = 0; i < le32_to_cpu(fsec->fsec_hdr.num_images); i++) {
2942                img_offset = le32_to_cpu(fsec->fsec_entry[i].offset);
2943                img_size   = le32_to_cpu(fsec->fsec_entry[i].pad_size);
2944                img_type   = le32_to_cpu(fsec->fsec_entry[i].type);
2945                img_optype = be_get_img_optype(fsec->fsec_entry[i]);
2946                old_fw_img = fsec->fsec_entry[i].optype == 0xFFFF;
2947
2948                if (img_optype == 0xFFFF)
2949                        continue;
2950
2951                if (flash_offset_support)
2952                        flash_optype = OPTYPE_OFFSET_SPECIFIED;
2953                else
2954                        flash_optype = img_optype;
2955
2956                /* Don't bother verifying CRC if an old FW image is being
2957                 * flashed
2958                 */
2959                if (old_fw_img)
2960                        goto flash;
2961
2962                status = be_check_flash_crc(adapter, fw->data, img_offset,
2963                                            img_size, filehdr_size +
2964                                            img_hdrs_size, flash_optype,
2965                                            &crc_match);
2966                if (base_status(status) == MCC_STATUS_ILLEGAL_REQUEST ||
2967                    base_status(status) == MCC_STATUS_ILLEGAL_FIELD) {
2968                        /* The current FW image on the card does not support
2969                         * OFFSET based flashing. Retry using older mechanism
2970                         * of OPTYPE based flashing
2971                         */
2972                        if (flash_optype == OPTYPE_OFFSET_SPECIFIED) {
2973                                flash_offset_support = false;
2974                                goto retry_flash;
2975                        }
2976
2977                        /* The current FW image on the card does not recognize
2978                         * the new FLASH op_type. The FW download is partially
2979                         * complete. Reboot the server now to enable FW image
2980                         * to recognize the new FLASH op_type. To complete the
2981                         * remaining process, download the same FW again after
2982                         * the reboot.
2983                         */
2984                        dev_err(dev, "Flash incomplete. Reset the server\n");
2985                        dev_err(dev, "Download FW image again after reset\n");
2986                        return -EAGAIN;
2987                } else if (status) {
2988                        dev_err(dev, "Could not get CRC for 0x%x region\n",
2989                                img_optype);
2990                        return -EFAULT;
2991                }
2992
2993                if (crc_match)
2994                        continue;
2995
2996flash:
2997                p = fw->data + filehdr_size + img_offset + img_hdrs_size;
2998                if (p + img_size > fw->data + fw->size)
2999                        return -1;
3000
3001                status = be_flash(adapter, p, flash_cmd, flash_optype, img_size,
3002                                  img_offset);
3003
3004                /* The current FW image on the card does not support OFFSET
3005                 * based flashing. Retry using older mechanism of OPTYPE based
3006                 * flashing
3007                 */
3008                if (base_status(status) == MCC_STATUS_ILLEGAL_FIELD &&
3009                    flash_optype == OPTYPE_OFFSET_SPECIFIED) {
3010                        flash_offset_support = false;
3011                        goto retry_flash;
3012                }
3013
3014                /* For old FW images ignore ILLEGAL_FIELD error or errors on
3015                 * UFI_DIR region
3016                 */
3017                if (old_fw_img &&
3018                    (base_status(status) == MCC_STATUS_ILLEGAL_FIELD ||
3019                     (img_optype == OPTYPE_UFI_DIR &&
3020                      base_status(status) == MCC_STATUS_FAILED))) {
3021                        continue;
3022                } else if (status) {
3023                        dev_err(dev, "Flashing section type 0x%x failed\n",
3024                                img_type);
3025
3026                        switch (addl_status(status)) {
3027                        case MCC_ADDL_STATUS_MISSING_SIGNATURE:
3028                                dev_err(dev,
3029                                        "Digital signature missing in FW\n");
3030                                return -EINVAL;
3031                        case MCC_ADDL_STATUS_INVALID_SIGNATURE:
3032                                dev_err(dev,
3033                                        "Invalid digital signature in FW\n");
3034                                return -EINVAL;
3035                        default:
3036                                return -EFAULT;
3037                        }
3038                }
3039        }
3040        return 0;
3041}
3042
3043int lancer_fw_download(struct be_adapter *adapter,
3044                       const struct firmware *fw)
3045{
3046        struct device *dev = &adapter->pdev->dev;
3047        struct be_dma_mem flash_cmd;
3048        const u8 *data_ptr = NULL;
3049        u8 *dest_image_ptr = NULL;
3050        size_t image_size = 0;
3051        u32 chunk_size = 0;
3052        u32 data_written = 0;
3053        u32 offset = 0;
3054        int status = 0;
3055        u8 add_status = 0;
3056        u8 change_status;
3057
3058        if (!IS_ALIGNED(fw->size, sizeof(u32))) {
3059                dev_err(dev, "FW image size should be multiple of 4\n");
3060                return -EINVAL;
3061        }
3062
3063        flash_cmd.size = sizeof(struct lancer_cmd_req_write_object)
3064                                + LANCER_FW_DOWNLOAD_CHUNK;
3065        flash_cmd.va = dma_alloc_coherent(dev, flash_cmd.size, &flash_cmd.dma,
3066                                          GFP_KERNEL);
3067        if (!flash_cmd.va)
3068                return -ENOMEM;
3069
3070        dest_image_ptr = flash_cmd.va +
3071                                sizeof(struct lancer_cmd_req_write_object);
3072        image_size = fw->size;
3073        data_ptr = fw->data;
3074
3075        while (image_size) {
3076                chunk_size = min_t(u32, image_size, LANCER_FW_DOWNLOAD_CHUNK);
3077
3078                /* Copy the image chunk content. */
3079                memcpy(dest_image_ptr, data_ptr, chunk_size);
3080
3081                status = lancer_cmd_write_object(adapter, &flash_cmd,
3082                                                 chunk_size, offset,
3083                                                 LANCER_FW_DOWNLOAD_LOCATION,
3084                                                 &data_written, &change_status,
3085                                                 &add_status);
3086                if (status)
3087                        break;
3088
3089                offset += data_written;
3090                data_ptr += data_written;
3091                image_size -= data_written;
3092        }
3093
3094        if (!status) {
3095                /* Commit the FW written */
3096                status = lancer_cmd_write_object(adapter, &flash_cmd,
3097                                                 0, offset,
3098                                                 LANCER_FW_DOWNLOAD_LOCATION,
3099                                                 &data_written, &change_status,
3100                                                 &add_status);
3101        }
3102
3103        dma_free_coherent(dev, flash_cmd.size, flash_cmd.va, flash_cmd.dma);
3104        if (status) {
3105                dev_err(dev, "Firmware load error\n");
3106                return be_cmd_status(status);
3107        }
3108
3109        dev_info(dev, "Firmware flashed successfully\n");
3110
3111        if (change_status == LANCER_FW_RESET_NEEDED) {
3112                dev_info(dev, "Resetting adapter to activate new FW\n");
3113                status = lancer_physdev_ctrl(adapter,
3114                                             PHYSDEV_CONTROL_FW_RESET_MASK);
3115                if (status) {
3116                        dev_err(dev, "Adapter busy, could not reset FW\n");
3117                        dev_err(dev, "Reboot server to activate new FW\n");
3118                }
3119        } else if (change_status != LANCER_NO_RESET_NEEDED) {
3120                dev_info(dev, "Reboot server to activate new FW\n");
3121        }
3122
3123        return 0;
3124}
3125
3126/* Check if the flash image file is compatible with the adapter that
3127 * is being flashed.
3128 */
3129static bool be_check_ufi_compatibility(struct be_adapter *adapter,
3130                                       struct flash_file_hdr_g3 *fhdr)
3131{
3132        if (!fhdr) {
3133                dev_err(&adapter->pdev->dev, "Invalid FW UFI file");
3134                return false;
3135        }
3136
3137        /* First letter of the build version is used to identify
3138         * which chip this image file is meant for.
3139         */
3140        switch (fhdr->build[0]) {
3141        case BLD_STR_UFI_TYPE_SH:
3142                if (!skyhawk_chip(adapter))
3143                        return false;
3144                break;
3145        case BLD_STR_UFI_TYPE_BE3:
3146                if (!BE3_chip(adapter))
3147                        return false;
3148                break;
3149        case BLD_STR_UFI_TYPE_BE2:
3150                if (!BE2_chip(adapter))
3151                        return false;
3152                break;
3153        default:
3154                return false;
3155        }
3156
3157        /* In BE3 FW images the "asic_type_rev" field doesn't track the
3158         * asic_rev of the chips it is compatible with.
3159         * When asic_type_rev is 0 the image is compatible only with
3160         * pre-BE3-R chips (asic_rev < 0x10)
3161         */
3162        if (BEx_chip(adapter) && fhdr->asic_type_rev == 0)
3163                return adapter->asic_rev < 0x10;
3164        else
3165                return (fhdr->asic_type_rev >= adapter->asic_rev);
3166}
3167
3168int be_fw_download(struct be_adapter *adapter, const struct firmware *fw)
3169{
3170        struct device *dev = &adapter->pdev->dev;
3171        struct flash_file_hdr_g3 *fhdr3;
3172        struct image_hdr *img_hdr_ptr;
3173        int status = 0, i, num_imgs;
3174        struct be_dma_mem flash_cmd;
3175
3176        fhdr3 = (struct flash_file_hdr_g3 *)fw->data;
3177        if (!be_check_ufi_compatibility(adapter, fhdr3)) {
3178                dev_err(dev, "Flash image is not compatible with adapter\n");
3179                return -EINVAL;
3180        }
3181
3182        flash_cmd.size = sizeof(struct be_cmd_write_flashrom);
3183        flash_cmd.va = dma_alloc_coherent(dev, flash_cmd.size, &flash_cmd.dma,
3184                                          GFP_KERNEL);
3185        if (!flash_cmd.va)
3186                return -ENOMEM;
3187
3188        num_imgs = le32_to_cpu(fhdr3->num_imgs);
3189        for (i = 0; i < num_imgs; i++) {
3190                img_hdr_ptr = (struct image_hdr *)(fw->data +
3191                                (sizeof(struct flash_file_hdr_g3) +
3192                                 i * sizeof(struct image_hdr)));
3193                if (!BE2_chip(adapter) &&
3194                    le32_to_cpu(img_hdr_ptr->imageid) != 1)
3195                        continue;
3196
3197                if (skyhawk_chip(adapter))
3198                        status = be_flash_skyhawk(adapter, fw, &flash_cmd,
3199                                                  num_imgs);
3200                else
3201                        status = be_flash_BEx(adapter, fw, &flash_cmd,
3202                                              num_imgs);
3203        }
3204
3205        dma_free_coherent(dev, flash_cmd.size, flash_cmd.va, flash_cmd.dma);
3206        if (!status)
3207                dev_info(dev, "Firmware flashed successfully\n");
3208
3209        return status;
3210}
3211
3212int be_cmd_enable_magic_wol(struct be_adapter *adapter, u8 *mac,
3213                            struct be_dma_mem *nonemb_cmd)
3214{
3215        struct be_mcc_wrb *wrb;
3216        struct be_cmd_req_acpi_wol_magic_config *req;
3217        int status;
3218
3219        mutex_lock(&adapter->mcc_lock);
3220
3221        wrb = wrb_from_mccq(adapter);
3222        if (!wrb) {
3223                status = -EBUSY;
3224                goto err;
3225        }
3226        req = nonemb_cmd->va;
3227
3228        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_ETH,
3229                               OPCODE_ETH_ACPI_WOL_MAGIC_CONFIG, sizeof(*req),
3230                               wrb, nonemb_cmd);
3231        memcpy(req->magic_mac, mac, ETH_ALEN);
3232
3233        status = be_mcc_notify_wait(adapter);
3234
3235err:
3236        mutex_unlock(&adapter->mcc_lock);
3237        return status;
3238}
3239
3240int be_cmd_set_loopback(struct be_adapter *adapter, u8 port_num,
3241                        u8 loopback_type, u8 enable)
3242{
3243        struct be_mcc_wrb *wrb;
3244        struct be_cmd_req_set_lmode *req;
3245        int status;
3246
3247        if (!be_cmd_allowed(adapter, OPCODE_LOWLEVEL_SET_LOOPBACK_MODE,
3248                            CMD_SUBSYSTEM_LOWLEVEL))
3249                return -EPERM;
3250
3251        mutex_lock(&adapter->mcc_lock);
3252
3253        wrb = wrb_from_mccq(adapter);
3254        if (!wrb) {
3255                status = -EBUSY;
3256                goto err_unlock;
3257        }
3258
3259        req = embedded_payload(wrb);
3260
3261        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_LOWLEVEL,
3262                               OPCODE_LOWLEVEL_SET_LOOPBACK_MODE, sizeof(*req),
3263                               wrb, NULL);
3264
3265        req->src_port = port_num;
3266        req->dest_port = port_num;
3267        req->loopback_type = loopback_type;
3268        req->loopback_state = enable;
3269
3270        status = be_mcc_notify(adapter);
3271        if (status)
3272                goto err_unlock;
3273
3274        mutex_unlock(&adapter->mcc_lock);
3275
3276        if (!wait_for_completion_timeout(&adapter->et_cmd_compl,
3277                                         msecs_to_jiffies(SET_LB_MODE_TIMEOUT)))
3278                status = -ETIMEDOUT;
3279
3280        return status;
3281
3282err_unlock:
3283        mutex_unlock(&adapter->mcc_lock);
3284        return status;
3285}
3286
3287int be_cmd_loopback_test(struct be_adapter *adapter, u32 port_num,
3288                         u32 loopback_type, u32 pkt_size, u32 num_pkts,
3289                         u64 pattern)
3290{
3291        struct be_mcc_wrb *wrb;
3292        struct be_cmd_req_loopback_test *req;
3293        struct be_cmd_resp_loopback_test *resp;
3294        int status;
3295
3296        if (!be_cmd_allowed(adapter, OPCODE_LOWLEVEL_LOOPBACK_TEST,
3297                            CMD_SUBSYSTEM_LOWLEVEL))
3298                return -EPERM;
3299
3300        mutex_lock(&adapter->mcc_lock);
3301
3302        wrb = wrb_from_mccq(adapter);
3303        if (!wrb) {
3304                status = -EBUSY;
3305                goto err;
3306        }
3307
3308        req = embedded_payload(wrb);
3309
3310        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_LOWLEVEL,
3311                               OPCODE_LOWLEVEL_LOOPBACK_TEST, sizeof(*req), wrb,
3312                               NULL);
3313
3314        req->hdr.timeout = cpu_to_le32(15);
3315        req->pattern = cpu_to_le64(pattern);
3316        req->src_port = cpu_to_le32(port_num);
3317        req->dest_port = cpu_to_le32(port_num);
3318        req->pkt_size = cpu_to_le32(pkt_size);
3319        req->num_pkts = cpu_to_le32(num_pkts);
3320        req->loopback_type = cpu_to_le32(loopback_type);
3321
3322        status = be_mcc_notify(adapter);
3323        if (status)
3324                goto err;
3325
3326        mutex_unlock(&adapter->mcc_lock);
3327
3328        wait_for_completion(&adapter->et_cmd_compl);
3329        resp = embedded_payload(wrb);
3330        status = le32_to_cpu(resp->status);
3331
3332        return status;
3333err:
3334        mutex_unlock(&adapter->mcc_lock);
3335        return status;
3336}
3337
3338int be_cmd_ddr_dma_test(struct be_adapter *adapter, u64 pattern,
3339                        u32 byte_cnt, struct be_dma_mem *cmd)
3340{
3341        struct be_mcc_wrb *wrb;
3342        struct be_cmd_req_ddrdma_test *req;
3343        int status;
3344        int i, j = 0;
3345
3346        if (!be_cmd_allowed(adapter, OPCODE_LOWLEVEL_HOST_DDR_DMA,
3347                            CMD_SUBSYSTEM_LOWLEVEL))
3348                return -EPERM;
3349
3350        mutex_lock(&adapter->mcc_lock);
3351
3352        wrb = wrb_from_mccq(adapter);
3353        if (!wrb) {
3354                status = -EBUSY;
3355                goto err;
3356        }
3357        req = cmd->va;
3358        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_LOWLEVEL,
3359                               OPCODE_LOWLEVEL_HOST_DDR_DMA, cmd->size, wrb,
3360                               cmd);
3361
3362        req->pattern = cpu_to_le64(pattern);
3363        req->byte_count = cpu_to_le32(byte_cnt);
3364        for (i = 0; i < byte_cnt; i++) {
3365                req->snd_buff[i] = (u8)(pattern >> (j*8));
3366                j++;
3367                if (j > 7)
3368                        j = 0;
3369        }
3370
3371        status = be_mcc_notify_wait(adapter);
3372
3373        if (!status) {
3374                struct be_cmd_resp_ddrdma_test *resp;
3375
3376                resp = cmd->va;
3377                if ((memcmp(resp->rcv_buff, req->snd_buff, byte_cnt) != 0) ||
3378                    resp->snd_err) {
3379                        status = -1;
3380                }
3381        }
3382
3383err:
3384        mutex_unlock(&adapter->mcc_lock);
3385        return status;
3386}
3387
3388int be_cmd_get_seeprom_data(struct be_adapter *adapter,
3389                            struct be_dma_mem *nonemb_cmd)
3390{
3391        struct be_mcc_wrb *wrb;
3392        struct be_cmd_req_seeprom_read *req;
3393        int status;
3394
3395        mutex_lock(&adapter->mcc_lock);
3396
3397        wrb = wrb_from_mccq(adapter);
3398        if (!wrb) {
3399                status = -EBUSY;
3400                goto err;
3401        }
3402        req = nonemb_cmd->va;
3403
3404        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3405                               OPCODE_COMMON_SEEPROM_READ, sizeof(*req), wrb,
3406                               nonemb_cmd);
3407
3408        status = be_mcc_notify_wait(adapter);
3409
3410err:
3411        mutex_unlock(&adapter->mcc_lock);
3412        return status;
3413}
3414
3415int be_cmd_get_phy_info(struct be_adapter *adapter)
3416{
3417        struct be_mcc_wrb *wrb;
3418        struct be_cmd_req_get_phy_info *req;
3419        struct be_dma_mem cmd;
3420        int status;
3421
3422        if (!be_cmd_allowed(adapter, OPCODE_COMMON_GET_PHY_DETAILS,
3423                            CMD_SUBSYSTEM_COMMON))
3424                return -EPERM;
3425
3426        mutex_lock(&adapter->mcc_lock);
3427
3428        wrb = wrb_from_mccq(adapter);
3429        if (!wrb) {
3430                status = -EBUSY;
3431                goto err;
3432        }
3433        cmd.size = sizeof(struct be_cmd_req_get_phy_info);
3434        cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
3435                                    GFP_ATOMIC);
3436        if (!cmd.va) {
3437                dev_err(&adapter->pdev->dev, "Memory alloc failure\n");
3438                status = -ENOMEM;
3439                goto err;
3440        }
3441
3442        req = cmd.va;
3443
3444        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3445                               OPCODE_COMMON_GET_PHY_DETAILS, sizeof(*req),
3446                               wrb, &cmd);
3447
3448        status = be_mcc_notify_wait(adapter);
3449        if (!status) {
3450                struct be_phy_info *resp_phy_info =
3451                                cmd.va + sizeof(struct be_cmd_req_hdr);
3452
3453                adapter->phy.phy_type = le16_to_cpu(resp_phy_info->phy_type);
3454                adapter->phy.interface_type =
3455                        le16_to_cpu(resp_phy_info->interface_type);
3456                adapter->phy.auto_speeds_supported =
3457                        le16_to_cpu(resp_phy_info->auto_speeds_supported);
3458                adapter->phy.fixed_speeds_supported =
3459                        le16_to_cpu(resp_phy_info->fixed_speeds_supported);
3460                adapter->phy.misc_params =
3461                        le32_to_cpu(resp_phy_info->misc_params);
3462
3463                if (BE2_chip(adapter)) {
3464                        adapter->phy.fixed_speeds_supported =
3465                                BE_SUPPORTED_SPEED_10GBPS |
3466                                BE_SUPPORTED_SPEED_1GBPS;
3467                }
3468        }
3469        dma_free_coherent(&adapter->pdev->dev, cmd.size, cmd.va, cmd.dma);
3470err:
3471        mutex_unlock(&adapter->mcc_lock);
3472        return status;
3473}
3474
3475static int be_cmd_set_qos(struct be_adapter *adapter, u32 bps, u32 domain)
3476{
3477        struct be_mcc_wrb *wrb;
3478        struct be_cmd_req_set_qos *req;
3479        int status;
3480
3481        mutex_lock(&adapter->mcc_lock);
3482
3483        wrb = wrb_from_mccq(adapter);
3484        if (!wrb) {
3485                status = -EBUSY;
3486                goto err;
3487        }
3488
3489        req = embedded_payload(wrb);
3490
3491        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3492                               OPCODE_COMMON_SET_QOS, sizeof(*req), wrb, NULL);
3493
3494        req->hdr.domain = domain;
3495        req->valid_bits = cpu_to_le32(BE_QOS_BITS_NIC);
3496        req->max_bps_nic = cpu_to_le32(bps);
3497
3498        status = be_mcc_notify_wait(adapter);
3499
3500err:
3501        mutex_unlock(&adapter->mcc_lock);
3502        return status;
3503}
3504
3505int be_cmd_get_cntl_attributes(struct be_adapter *adapter)
3506{
3507        struct be_mcc_wrb *wrb;
3508        struct be_cmd_req_cntl_attribs *req;
3509        struct be_cmd_resp_cntl_attribs *resp;
3510        int status, i;
3511        int payload_len = max(sizeof(*req), sizeof(*resp));
3512        struct mgmt_controller_attrib *attribs;
3513        struct be_dma_mem attribs_cmd;
3514        u32 *serial_num;
3515
3516        if (mutex_lock_interruptible(&adapter->mbox_lock))
3517                return -1;
3518
3519        memset(&attribs_cmd, 0, sizeof(struct be_dma_mem));
3520        attribs_cmd.size = sizeof(struct be_cmd_resp_cntl_attribs);
3521        attribs_cmd.va = dma_alloc_coherent(&adapter->pdev->dev,
3522                                            attribs_cmd.size,
3523                                            &attribs_cmd.dma, GFP_ATOMIC);
3524        if (!attribs_cmd.va) {
3525                dev_err(&adapter->pdev->dev, "Memory allocation failure\n");
3526                status = -ENOMEM;
3527                goto err;
3528        }
3529
3530        wrb = wrb_from_mbox(adapter);
3531        if (!wrb) {
3532                status = -EBUSY;
3533                goto err;
3534        }
3535        req = attribs_cmd.va;
3536
3537        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3538                               OPCODE_COMMON_GET_CNTL_ATTRIBUTES, payload_len,
3539                               wrb, &attribs_cmd);
3540
3541        status = be_mbox_notify_wait(adapter);
3542        if (!status) {
3543                attribs = attribs_cmd.va + sizeof(struct be_cmd_resp_hdr);
3544                adapter->hba_port_num = attribs->hba_attribs.phy_port;
3545                serial_num = attribs->hba_attribs.controller_serial_number;
3546                for (i = 0; i < CNTL_SERIAL_NUM_WORDS; i++)
3547                        adapter->serial_num[i] = le32_to_cpu(serial_num[i]) &
3548                                (BIT_MASK(16) - 1);
3549                /* For BEx, since GET_FUNC_CONFIG command is not
3550                 * supported, we read funcnum here as a workaround.
3551                 */
3552                if (BEx_chip(adapter))
3553                        adapter->pf_num = attribs->hba_attribs.pci_funcnum;
3554        }
3555
3556err:
3557        mutex_unlock(&adapter->mbox_lock);
3558        if (attribs_cmd.va)
3559                dma_free_coherent(&adapter->pdev->dev, attribs_cmd.size,
3560                                  attribs_cmd.va, attribs_cmd.dma);
3561        return status;
3562}
3563
3564/* Uses mbox */
3565int be_cmd_req_native_mode(struct be_adapter *adapter)
3566{
3567        struct be_mcc_wrb *wrb;
3568        struct be_cmd_req_set_func_cap *req;
3569        int status;
3570
3571        if (mutex_lock_interruptible(&adapter->mbox_lock))
3572                return -1;
3573
3574        wrb = wrb_from_mbox(adapter);
3575        if (!wrb) {
3576                status = -EBUSY;
3577                goto err;
3578        }
3579
3580        req = embedded_payload(wrb);
3581
3582        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3583                               OPCODE_COMMON_SET_DRIVER_FUNCTION_CAP,
3584                               sizeof(*req), wrb, NULL);
3585
3586        req->valid_cap_flags = cpu_to_le32(CAPABILITY_SW_TIMESTAMPS |
3587                                CAPABILITY_BE3_NATIVE_ERX_API);
3588        req->cap_flags = cpu_to_le32(CAPABILITY_BE3_NATIVE_ERX_API);
3589
3590        status = be_mbox_notify_wait(adapter);
3591        if (!status) {
3592                struct be_cmd_resp_set_func_cap *resp = embedded_payload(wrb);
3593
3594                adapter->be3_native = le32_to_cpu(resp->cap_flags) &
3595                                        CAPABILITY_BE3_NATIVE_ERX_API;
3596                if (!adapter->be3_native)
3597                        dev_warn(&adapter->pdev->dev,
3598                                 "adapter not in advanced mode\n");
3599        }
3600err:
3601        mutex_unlock(&adapter->mbox_lock);
3602        return status;
3603}
3604
3605/* Get privilege(s) for a function */
3606int be_cmd_get_fn_privileges(struct be_adapter *adapter, u32 *privilege,
3607                             u32 domain)
3608{
3609        struct be_mcc_wrb *wrb;
3610        struct be_cmd_req_get_fn_privileges *req;
3611        int status;
3612
3613        mutex_lock(&adapter->mcc_lock);
3614
3615        wrb = wrb_from_mccq(adapter);
3616        if (!wrb) {
3617                status = -EBUSY;
3618                goto err;
3619        }
3620
3621        req = embedded_payload(wrb);
3622
3623        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3624                               OPCODE_COMMON_GET_FN_PRIVILEGES, sizeof(*req),
3625                               wrb, NULL);
3626
3627        req->hdr.domain = domain;
3628
3629        status = be_mcc_notify_wait(adapter);
3630        if (!status) {
3631                struct be_cmd_resp_get_fn_privileges *resp =
3632                                                embedded_payload(wrb);
3633
3634                *privilege = le32_to_cpu(resp->privilege_mask);
3635
3636                /* In UMC mode FW does not return right privileges.
3637                 * Override with correct privilege equivalent to PF.
3638                 */
3639                if (BEx_chip(adapter) && be_is_mc(adapter) &&
3640                    be_physfn(adapter))
3641                        *privilege = MAX_PRIVILEGES;
3642        }
3643
3644err:
3645        mutex_unlock(&adapter->mcc_lock);
3646        return status;
3647}
3648
3649/* Set privilege(s) for a function */
3650int be_cmd_set_fn_privileges(struct be_adapter *adapter, u32 privileges,
3651                             u32 domain)
3652{
3653        struct be_mcc_wrb *wrb;
3654        struct be_cmd_req_set_fn_privileges *req;
3655        int status;
3656
3657        mutex_lock(&adapter->mcc_lock);
3658
3659        wrb = wrb_from_mccq(adapter);
3660        if (!wrb) {
3661                status = -EBUSY;
3662                goto err;
3663        }
3664
3665        req = embedded_payload(wrb);
3666        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3667                               OPCODE_COMMON_SET_FN_PRIVILEGES, sizeof(*req),
3668                               wrb, NULL);
3669        req->hdr.domain = domain;
3670        if (lancer_chip(adapter))
3671                req->privileges_lancer = cpu_to_le32(privileges);
3672        else
3673                req->privileges = cpu_to_le32(privileges);
3674
3675        status = be_mcc_notify_wait(adapter);
3676err:
3677        mutex_unlock(&adapter->mcc_lock);
3678        return status;
3679}
3680
3681/* pmac_id_valid: true => pmac_id is supplied and MAC address is requested.
3682 * pmac_id_valid: false => pmac_id or MAC address is requested.
3683 *                If pmac_id is returned, pmac_id_valid is returned as true
3684 */
3685int be_cmd_get_mac_from_list(struct be_adapter *adapter, u8 *mac,
3686                             bool *pmac_id_valid, u32 *pmac_id, u32 if_handle,
3687                             u8 domain)
3688{
3689        struct be_mcc_wrb *wrb;
3690        struct be_cmd_req_get_mac_list *req;
3691        int status;
3692        int mac_count;
3693        struct be_dma_mem get_mac_list_cmd;
3694        int i;
3695
3696        memset(&get_mac_list_cmd, 0, sizeof(struct be_dma_mem));
3697        get_mac_list_cmd.size = sizeof(struct be_cmd_resp_get_mac_list);
3698        get_mac_list_cmd.va = dma_alloc_coherent(&adapter->pdev->dev,
3699                                                 get_mac_list_cmd.size,
3700                                                 &get_mac_list_cmd.dma,
3701                                                 GFP_ATOMIC);
3702
3703        if (!get_mac_list_cmd.va) {
3704                dev_err(&adapter->pdev->dev,
3705                        "Memory allocation failure during GET_MAC_LIST\n");
3706                return -ENOMEM;
3707        }
3708
3709        mutex_lock(&adapter->mcc_lock);
3710
3711        wrb = wrb_from_mccq(adapter);
3712        if (!wrb) {
3713                status = -EBUSY;
3714                goto out;
3715        }
3716
3717        req = get_mac_list_cmd.va;
3718
3719        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,
3720                               OPCODE_COMMON_GET_MAC_LIST,
3721                               get_mac_list_cmd.size, wrb, &get_mac_list_cmd);
3722        req->hdr.domain = domain;
3723        req->mac_type = MAC_ADDRESS_TYPE_NETWORK;
3724        if (*pmac_id_valid) {
3725                req->mac_id = cpu_to_le32(*pmac_id);
3726                req->iface_id = cpu_to_le16(if_handle);
3727                req->perm_override = 0;
3728        } else {
3729                req->perm_override = 1;
3730        }
3731
3732        status = be_mcc_notify_wait(adapter);
3733        if (!status) {
3734                struct be_cmd_resp_get_mac_list *resp =
3735                                                get_mac_list_cmd.va;
3736
3737                if (*pmac_id_valid) {
3738                        memcpy(mac, resp->macid_macaddr.mac_addr_id.macaddr,
3739                               ETH_ALEN);
3740                        goto out;
3741                }
3742
3743                mac_count = resp->true_mac_count + resp->pseudo_mac_count;
3744                /* Mac list returned could contain one or more active mac_ids
3745                 * or one or more true or pseudo permanent mac addresses.
3746                 * If an active mac_id is present, return first active mac_id
3747                 * found.
3748                 */
3749                for (i = 0; i < mac_count; i++) {
3750                        struct get_list_macaddr *mac_entry;
3751                        u16 mac_addr_size;
3752                        u32 mac_id;
3753
3754                        mac_entry = &resp->macaddr_list[i];
3755                        mac_addr_size = le16_to_cpu(mac_entry->mac_addr_size);
3756                        /* mac_id is a 32 bit value and mac_addr size
3757                         * is 6 bytes
3758                         */
3759                        if (mac_addr_size == sizeof(u32)) {
3760                                *pmac_id_valid = true;
3761                                mac_id = mac_entry->mac_addr_id.s_mac_id.mac_id;
3762                                *pmac_id = le32_to_cpu(mac_id);
3763                                goto out;
3764                        }
3765                }
3766                /* If no active mac_id found, return first mac addr */
3767                *pmac_id_valid = false;
3768                memcpy(mac, resp->macaddr_list[0].mac_addr_id.macaddr,
3769                       ETH_ALEN);
3770        }
3771
3772out:
3773        mutex_unlock(&adapter->mcc_lock);
3774        dma_free_coherent(&adapter->pdev->dev, get_mac_list_cmd.size,
3775                          get_mac_list_cmd.va, get_mac_list_cmd.dma);
3776        return status;
3777}
3778
3779int be_cmd_get_active_mac(struct be_adapter *adapter, u32 curr_pmac_id,
3780                          u8 *mac, u32 if_handle, bool active, u32 domain)
3781{
3782        if (!active)
3783                be_cmd_get_mac_from_list(adapter, mac, &active, &curr_pmac_id,
3784                                         if_handle, domain);
3785        if (BEx_chip(adapter))
3786                return be_cmd_mac_addr_query(adapter, mac, false,
3787                                             if_handle, curr_pmac_id);
3788        else
3789                /* Fetch the MAC address using pmac_id */
3790                return be_cmd_get_mac_from_list(adapter, mac, &active,
3791                                                &curr_pmac_id,
3792                                                if_handle, domain);
3793}
3794
3795int be_cmd_get_perm_mac(struct be_adapter *adapter, u8 *mac)
3796{
3797        int status;
3798        bool pmac_valid = false;
3799
3800        eth_zero_addr(mac);
3801
3802        if (BEx_chip(adapter)) {
3803                if (be_physfn(adapter))
3804                        status = be_cmd_mac_addr_query(adapter, mac, true, 0,
3805                                                       0);
3806                else
3807                        status = be_cmd_mac_addr_query(adapter, mac, false,
3808                                                       adapter->if_handle, 0);
3809        } else {
3810                status = be_cmd_get_mac_from_list(adapter, mac, &pmac_valid,
3811                                                  NULL, adapter->if_handle, 0);
3812        }
3813
3814        return status;
3815}
3816
3817/* Uses synchronous MCCQ */
3818int be_cmd_set_mac_list(struct be_adapter *adapter, u8 *mac_array,
3819                        u8 mac_count, u32 domain)
3820{
3821        struct be_mcc_wrb *wrb;
3822        struct be_cmd_req_set_mac_list *req;
3823        int status;
3824        struct be_dma_mem cmd;
3825
3826        memset(&cmd, 0, sizeof(struct be_dma_mem));
3827        cmd.size = sizeof(struct be_cmd_req_set_mac_list);
3828        cmd.va = dma_alloc_coherent(&adapter->pdev->dev, cmd.size, &cmd.dma,
3829                                    GFP_KERNEL);
3830        if (!cmd.va)
3831                return -ENOMEM;
3832
3833        mutex_lock(&adapter->mcc_lock);
3834
3835        wrb = wrb_from_mccq(adapter);
3836        if (!wrb) {
3837                status = -EBUSY;
3838                goto err;
3839        }
3840
3841        req = cmd.va;
3842        be_wrb_cmd_hdr_prepare(&req->hdr, CMD_SUBSYSTEM_COMMON,