/* SPDX-License-Identifier: BSD-3-Clause * Copyright(c) 2013 - 2015 Intel Corporation */ #include "iavf_type.h" #include "iavf_adminq.h" #include "iavf_prototype.h" #include "virtchnl.h" /** * iavf_set_mac_type - Sets MAC type * @hw: pointer to the HW structure * * This function sets the mac type of the adapter based on the * vendor ID and device ID stored in the hw structure. **/ enum iavf_status_code iavf_set_mac_type(struct iavf_hw *hw) { enum iavf_status_code status = IAVF_SUCCESS; DEBUGFUNC("iavf_set_mac_type\n"); if (hw->vendor_id == IAVF_INTEL_VENDOR_ID) { switch (hw->device_id) { /* TODO: remove undefined device ID now, need to think how to * remove them in share code */ case IAVF_DEV_ID_ADAPTIVE_VF: hw->mac.type = IAVF_MAC_VF; break; default: hw->mac.type = IAVF_MAC_GENERIC; break; } } else { status = IAVF_ERR_DEVICE_NOT_SUPPORTED; } DEBUGOUT2("iavf_set_mac_type found mac: %d, returns: %d\n", hw->mac.type, status); return status; } /** * iavf_aq_str - convert AQ err code to a string * @hw: pointer to the HW structure * @aq_err: the AQ error code to convert **/ const char *iavf_aq_str(struct iavf_hw *hw, enum iavf_admin_queue_err aq_err) { switch (aq_err) { case IAVF_AQ_RC_OK: return "OK"; case IAVF_AQ_RC_EPERM: return "IAVF_AQ_RC_EPERM"; case IAVF_AQ_RC_ENOENT: return "IAVF_AQ_RC_ENOENT"; case IAVF_AQ_RC_ESRCH: return "IAVF_AQ_RC_ESRCH"; case IAVF_AQ_RC_EINTR: return "IAVF_AQ_RC_EINTR"; case IAVF_AQ_RC_EIO: return "IAVF_AQ_RC_EIO"; case IAVF_AQ_RC_ENXIO: return "IAVF_AQ_RC_ENXIO"; case IAVF_AQ_RC_E2BIG: return "IAVF_AQ_RC_E2BIG"; case IAVF_AQ_RC_EAGAIN: return "IAVF_AQ_RC_EAGAIN"; case IAVF_AQ_RC_ENOMEM: return "IAVF_AQ_RC_ENOMEM"; case IAVF_AQ_RC_EACCES: return "IAVF_AQ_RC_EACCES"; case IAVF_AQ_RC_EFAULT: return "IAVF_AQ_RC_EFAULT"; case IAVF_AQ_RC_EBUSY: return "IAVF_AQ_RC_EBUSY"; case IAVF_AQ_RC_EEXIST: return "IAVF_AQ_RC_EEXIST"; case IAVF_AQ_RC_EINVAL: return "IAVF_AQ_RC_EINVAL"; case IAVF_AQ_RC_ENOTTY: return "IAVF_AQ_RC_ENOTTY"; case IAVF_AQ_RC_ENOSPC: return "IAVF_AQ_RC_ENOSPC"; case IAVF_AQ_RC_ENOSYS: return "IAVF_AQ_RC_ENOSYS"; case IAVF_AQ_RC_ERANGE: return "IAVF_AQ_RC_ERANGE"; case IAVF_AQ_RC_EFLUSHED: return "IAVF_AQ_RC_EFLUSHED"; case IAVF_AQ_RC_BAD_ADDR: return "IAVF_AQ_RC_BAD_ADDR"; case IAVF_AQ_RC_EMODE: return "IAVF_AQ_RC_EMODE"; case IAVF_AQ_RC_EFBIG: return "IAVF_AQ_RC_EFBIG"; } snprintf(hw->err_str, sizeof(hw->err_str), "%d", aq_err); return hw->err_str; } /** * iavf_stat_str - convert status err code to a string * @hw: pointer to the HW structure * @stat_err: the status error code to convert **/ const char *iavf_stat_str(struct iavf_hw *hw, enum iavf_status_code stat_err) { switch (stat_err) { case IAVF_SUCCESS: return "OK"; case IAVF_ERR_NVM: return "IAVF_ERR_NVM"; case IAVF_ERR_NVM_CHECKSUM: return "IAVF_ERR_NVM_CHECKSUM"; case IAVF_ERR_PHY: return "IAVF_ERR_PHY"; case IAVF_ERR_CONFIG: return "IAVF_ERR_CONFIG"; case IAVF_ERR_PARAM: return "IAVF_ERR_PARAM"; case IAVF_ERR_MAC_TYPE: return "IAVF_ERR_MAC_TYPE"; case IAVF_ERR_UNKNOWN_PHY: return "IAVF_ERR_UNKNOWN_PHY"; case IAVF_ERR_LINK_SETUP: return "IAVF_ERR_LINK_SETUP"; case IAVF_ERR_ADAPTER_STOPPED: return "IAVF_ERR_ADAPTER_STOPPED"; case IAVF_ERR_INVALID_MAC_ADDR: return "IAVF_ERR_INVALID_MAC_ADDR"; case IAVF_ERR_DEVICE_NOT_SUPPORTED: return "IAVF_ERR_DEVICE_NOT_SUPPORTED"; case IAVF_ERR_MASTER_REQUESTS_PENDING: return "IAVF_ERR_MASTER_REQUESTS_PENDING"; case IAVF_ERR_INVALID_LINK_SETTINGS: return "IAVF_ERR_INVALID_LINK_SETTINGS"; case IAVF_ERR_AUTONEG_NOT_COMPLETE: return "IAVF_ERR_AUTONEG_NOT_COMPLETE"; case IAVF_ERR_RESET_FAILED: return "IAVF_ERR_RESET_FAILED"; case IAVF_ERR_SWFW_SYNC: return "IAVF_ERR_SWFW_SYNC"; case IAVF_ERR_NO_AVAILABLE_VSI: return "IAVF_ERR_NO_AVAILABLE_VSI"; case IAVF_ERR_NO_MEMORY: return "IAVF_ERR_NO_MEMORY"; case IAVF_ERR_BAD_PTR: return "IAVF_ERR_BAD_PTR"; case IAVF_ERR_RING_FULL: return "IAVF_ERR_RING_FULL"; case IAVF_ERR_INVALID_PD_ID: return "IAVF_ERR_INVALID_PD_ID"; case IAVF_ERR_INVALID_QP_ID: return "IAVF_ERR_INVALID_QP_ID"; case IAVF_ERR_INVALID_CQ_ID: return "IAVF_ERR_INVALID_CQ_ID"; case IAVF_ERR_INVALID_CEQ_ID: return "IAVF_ERR_INVALID_CEQ_ID"; case IAVF_ERR_INVALID_AEQ_ID: return "IAVF_ERR_INVALID_AEQ_ID"; case IAVF_ERR_INVALID_SIZE: return "IAVF_ERR_INVALID_SIZE"; case IAVF_ERR_INVALID_ARP_INDEX: return "IAVF_ERR_INVALID_ARP_INDEX"; case IAVF_ERR_INVALID_FPM_FUNC_ID: return "IAVF_ERR_INVALID_FPM_FUNC_ID"; case IAVF_ERR_QP_INVALID_MSG_SIZE: return "IAVF_ERR_QP_INVALID_MSG_SIZE"; case IAVF_ERR_QP_TOOMANY_WRS_POSTED: return "IAVF_ERR_QP_TOOMANY_WRS_POSTED"; case IAVF_ERR_INVALID_FRAG_COUNT: return "IAVF_ERR_INVALID_FRAG_COUNT"; case IAVF_ERR_QUEUE_EMPTY: return "IAVF_ERR_QUEUE_EMPTY"; case IAVF_ERR_INVALID_ALIGNMENT: return "IAVF_ERR_INVALID_ALIGNMENT"; case IAVF_ERR_FLUSHED_QUEUE: return "IAVF_ERR_FLUSHED_QUEUE"; case IAVF_ERR_INVALID_PUSH_PAGE_INDEX: return "IAVF_ERR_INVALID_PUSH_PAGE_INDEX"; case IAVF_ERR_INVALID_IMM_DATA_SIZE: return "IAVF_ERR_INVALID_IMM_DATA_SIZE"; case IAVF_ERR_TIMEOUT: return "IAVF_ERR_TIMEOUT"; case IAVF_ERR_OPCODE_MISMATCH: return "IAVF_ERR_OPCODE_MISMATCH"; case IAVF_ERR_CQP_COMPL_ERROR: return "IAVF_ERR_CQP_COMPL_ERROR"; case IAVF_ERR_INVALID_VF_ID: return "IAVF_ERR_INVALID_VF_ID"; case IAVF_ERR_INVALID_HMCFN_ID: return "IAVF_ERR_INVALID_HMCFN_ID"; case IAVF_ERR_BACKING_PAGE_ERROR: return "IAVF_ERR_BACKING_PAGE_ERROR"; case IAVF_ERR_NO_PBLCHUNKS_AVAILABLE: return "IAVF_ERR_NO_PBLCHUNKS_AVAILABLE"; case IAVF_ERR_INVALID_PBLE_INDEX: return "IAVF_ERR_INVALID_PBLE_INDEX"; case IAVF_ERR_INVALID_SD_INDEX: return "IAVF_ERR_INVALID_SD_INDEX"; case IAVF_ERR_INVALID_PAGE_DESC_INDEX: return "IAVF_ERR_INVALID_PAGE_DESC_INDEX"; case IAVF_ERR_INVALID_SD_TYPE: return "IAVF_ERR_INVALID_SD_TYPE"; case IAVF_ERR_MEMCPY_FAILED: return "IAVF_ERR_MEMCPY_FAILED"; case IAVF_ERR_INVALID_HMC_OBJ_INDEX: return "IAVF_ERR_INVALID_HMC_OBJ_INDEX"; case IAVF_ERR_INVALID_HMC_OBJ_COUNT: return "IAVF_ERR_INVALID_HMC_OBJ_COUNT"; case IAVF_ERR_INVALID_SRQ_ARM_LIMIT: return "IAVF_ERR_INVALID_SRQ_ARM_LIMIT"; case IAVF_ERR_SRQ_ENABLED: return "IAVF_ERR_SRQ_ENABLED"; case IAVF_ERR_ADMIN_QUEUE_ERROR: return "IAVF_ERR_ADMIN_QUEUE_ERROR"; case IAVF_ERR_ADMIN_QUEUE_TIMEOUT: return "IAVF_ERR_ADMIN_QUEUE_TIMEOUT"; case IAVF_ERR_BUF_TOO_SHORT: return "IAVF_ERR_BUF_TOO_SHORT"; case IAVF_ERR_ADMIN_QUEUE_FULL: return "IAVF_ERR_ADMIN_QUEUE_FULL"; case IAVF_ERR_ADMIN_QUEUE_NO_WORK: return "IAVF_ERR_ADMIN_QUEUE_NO_WORK"; case IAVF_ERR_BAD_IWARP_CQE: return "IAVF_ERR_BAD_IWARP_CQE"; case IAVF_ERR_NVM_BLANK_MODE: return "IAVF_ERR_NVM_BLANK_MODE"; case IAVF_ERR_NOT_IMPLEMENTED: return "IAVF_ERR_NOT_IMPLEMENTED"; case IAVF_ERR_PE_DOORBELL_NOT_ENABLED: return "IAVF_ERR_PE_DOORBELL_NOT_ENABLED"; case IAVF_ERR_DIAG_TEST_FAILED: return "IAVF_ERR_DIAG_TEST_FAILED"; case IAVF_ERR_NOT_READY: return "IAVF_ERR_NOT_READY"; case IAVF_NOT_SUPPORTED: return "IAVF_NOT_SUPPORTED"; case IAVF_ERR_FIRMWARE_API_VERSION: return "IAVF_ERR_FIRMWARE_API_VERSION"; case IAVF_ERR_ADMIN_QUEUE_CRITICAL_ERROR: return "IAVF_ERR_ADMIN_QUEUE_CRITICAL_ERROR"; } snprintf(hw->err_str, sizeof(hw->err_str), "%d", stat_err); return hw->err_str; } /** * iavf_debug_aq * @hw: debug mask related to admin queue * @mask: debug mask * @desc: pointer to admin queue descriptor * @buffer: pointer to command buffer * @buf_len: max length of buffer * * Dumps debug log about adminq command with descriptor contents. **/ void iavf_debug_aq(struct iavf_hw *hw, enum iavf_debug_mask mask, void *desc, void *buffer, u16 buf_len) { struct iavf_aq_desc *aq_desc = (struct iavf_aq_desc *)desc; u8 *buf = (u8 *)buffer; u16 len; u16 i = 0; if ((!(mask & hw->debug_mask)) || (desc == NULL)) return; len = LE16_TO_CPU(aq_desc->datalen); iavf_debug(hw, mask, "AQ CMD: opcode 0x%04X, flags 0x%04X, datalen 0x%04X, retval 0x%04X\n", LE16_TO_CPU(aq_desc->opcode), LE16_TO_CPU(aq_desc->flags), LE16_TO_CPU(aq_desc->datalen), LE16_TO_CPU(aq_desc->retval)); iavf_debug(hw, mask, "\tcookie (h,l) 0x%08X 0x%08X\n", LE32_TO_CPU(aq_desc->cookie_high), LE32_TO_CPU(aq_desc->cookie_low)); iavf_debug(hw, mask, "\tparam (0,1) 0x%08X 0x%08X\n", LE32_TO_CPU(aq_desc->params.internal.param0), LE32_TO_CPU(aq_desc->params.internal.param1)); iavf_debug(hw, mask, "\taddr (h,l) 0x%08X 0x%08X\n", LE32_TO_CPU(aq_desc->params.external.addr_high), LE32_TO_CPU(aq_desc->params.external.addr_low)); if ((buffer != NULL) && (aq_desc->datalen != 0)) { iavf_debug(hw, mask, "AQ CMD Buffer:\n"); if (buf_len < len) len = buf_len; /* write the full 16-byte chunks */ for (i = 0; i < (len - 16); i += 16) iavf_debug(hw, mask, "\t0x%04X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X\n", i, buf[i], buf[i+1], buf[i+2], buf[i+3], buf[i+4], buf[i+5], buf[i+6], buf[i+7], buf[i+8], buf[i+9], buf[i+10], buf[i+11], buf[i+12], buf[i+13], buf[i+14], buf[i+15]); /* the most we could have left is 16 bytes, pad with zeros */ if (i < len) { char d_buf[16]; int j, i_sav; i_sav = i; memset(d_buf, 0, sizeof(d_buf)); for (j = 0; i < len; j++, i++) d_buf[j] = buf[i]; iavf_debug(hw, mask, "\t0x%04X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X %02X\n", i_sav, d_buf[0], d_buf[1], d_buf[2], d_buf[3], d_buf[4], d_buf[5], d_buf[6], d_buf[7], d_buf[8], d_buf[9], d_buf[10], d_buf[11], d_buf[12], d_buf[13], d_buf[14], d_buf[15]); } } } /** * iavf_check_asq_alive * @hw: pointer to the hw struct * * Returns true if Queue is enabled else false. **/ bool iavf_check_asq_alive(struct iavf_hw *hw) { if (hw->aq.asq.len) #ifdef INTEGRATED_VF if (iavf_is_vf(hw)) return !!(rd32(hw, hw->aq.asq.len) & IAVF_ATQLEN1_ATQENABLE_MASK); #else return !!(rd32(hw, hw->aq.asq.len) & IAVF_ATQLEN1_ATQENABLE_MASK); #endif /* INTEGRATED_VF */ return false; } /** * iavf_aq_queue_shutdown * @hw: pointer to the hw struct * @unloading: is the driver unloading itself * * Tell the Firmware that we're shutting down the AdminQ and whether * or not the driver is unloading as well. **/ enum iavf_status_code iavf_aq_queue_shutdown(struct iavf_hw *hw, bool unloading) { struct iavf_aq_desc desc; struct iavf_aqc_queue_shutdown *cmd = (struct iavf_aqc_queue_shutdown *)&desc.params.raw; enum iavf_status_code status; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_queue_shutdown); if (unloading) cmd->driver_unloading = CPU_TO_LE32(IAVF_AQ_DRIVER_UNLOADING); status = iavf_asq_send_command(hw, &desc, NULL, 0, NULL); return status; } /** * iavf_aq_get_set_rss_lut * @hw: pointer to the hardware structure * @vsi_id: vsi fw index * @pf_lut: for PF table set true, for VSI table set false * @lut: pointer to the lut buffer provided by the caller * @lut_size: size of the lut buffer * @set: set true to set the table, false to get the table * * Internal function to get or set RSS look up table **/ STATIC enum iavf_status_code iavf_aq_get_set_rss_lut(struct iavf_hw *hw, u16 vsi_id, bool pf_lut, u8 *lut, u16 lut_size, bool set) { enum iavf_status_code status; struct iavf_aq_desc desc; struct iavf_aqc_get_set_rss_lut *cmd_resp = (struct iavf_aqc_get_set_rss_lut *)&desc.params.raw; if (set) iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_set_rss_lut); else iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_get_rss_lut); /* Indirect command */ desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_BUF); desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_RD); cmd_resp->vsi_id = CPU_TO_LE16((u16)((vsi_id << IAVF_AQC_SET_RSS_LUT_VSI_ID_SHIFT) & IAVF_AQC_SET_RSS_LUT_VSI_ID_MASK)); cmd_resp->vsi_id |= CPU_TO_LE16((u16)IAVF_AQC_SET_RSS_LUT_VSI_VALID); if (pf_lut) cmd_resp->flags |= CPU_TO_LE16((u16) ((IAVF_AQC_SET_RSS_LUT_TABLE_TYPE_PF << IAVF_AQC_SET_RSS_LUT_TABLE_TYPE_SHIFT) & IAVF_AQC_SET_RSS_LUT_TABLE_TYPE_MASK)); else cmd_resp->flags |= CPU_TO_LE16((u16) ((IAVF_AQC_SET_RSS_LUT_TABLE_TYPE_VSI << IAVF_AQC_SET_RSS_LUT_TABLE_TYPE_SHIFT) & IAVF_AQC_SET_RSS_LUT_TABLE_TYPE_MASK)); status = iavf_asq_send_command(hw, &desc, lut, lut_size, NULL); return status; } /** * iavf_aq_get_rss_lut * @hw: pointer to the hardware structure * @vsi_id: vsi fw index * @pf_lut: for PF table set true, for VSI table set false * @lut: pointer to the lut buffer provided by the caller * @lut_size: size of the lut buffer * * get the RSS lookup table, PF or VSI type **/ enum iavf_status_code iavf_aq_get_rss_lut(struct iavf_hw *hw, u16 vsi_id, bool pf_lut, u8 *lut, u16 lut_size) { return iavf_aq_get_set_rss_lut(hw, vsi_id, pf_lut, lut, lut_size, false); } /** * iavf_aq_set_rss_lut * @hw: pointer to the hardware structure * @vsi_id: vsi fw index * @pf_lut: for PF table set true, for VSI table set false * @lut: pointer to the lut buffer provided by the caller * @lut_size: size of the lut buffer * * set the RSS lookup table, PF or VSI type **/ enum iavf_status_code iavf_aq_set_rss_lut(struct iavf_hw *hw, u16 vsi_id, bool pf_lut, u8 *lut, u16 lut_size) { return iavf_aq_get_set_rss_lut(hw, vsi_id, pf_lut, lut, lut_size, true); } /** * iavf_aq_get_set_rss_key * @hw: pointer to the hw struct * @vsi_id: vsi fw index * @key: pointer to key info struct * @set: set true to set the key, false to get the key * * get the RSS key per VSI **/ STATIC enum iavf_status_code iavf_aq_get_set_rss_key(struct iavf_hw *hw, u16 vsi_id, struct iavf_aqc_get_set_rss_key_data *key, bool set) { enum iavf_status_code status; struct iavf_aq_desc desc; struct iavf_aqc_get_set_rss_key *cmd_resp = (struct iavf_aqc_get_set_rss_key *)&desc.params.raw; u16 key_size = sizeof(struct iavf_aqc_get_set_rss_key_data); if (set) iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_set_rss_key); else iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_get_rss_key); /* Indirect command */ desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_BUF); desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_RD); cmd_resp->vsi_id = CPU_TO_LE16((u16)((vsi_id << IAVF_AQC_SET_RSS_KEY_VSI_ID_SHIFT) & IAVF_AQC_SET_RSS_KEY_VSI_ID_MASK)); cmd_resp->vsi_id |= CPU_TO_LE16((u16)IAVF_AQC_SET_RSS_KEY_VSI_VALID); status = iavf_asq_send_command(hw, &desc, key, key_size, NULL); return status; } /** * iavf_aq_get_rss_key * @hw: pointer to the hw struct * @vsi_id: vsi fw index * @key: pointer to key info struct * **/ enum iavf_status_code iavf_aq_get_rss_key(struct iavf_hw *hw, u16 vsi_id, struct iavf_aqc_get_set_rss_key_data *key) { return iavf_aq_get_set_rss_key(hw, vsi_id, key, false); } /** * iavf_aq_set_rss_key * @hw: pointer to the hw struct * @vsi_id: vsi fw index * @key: pointer to key info struct * * set the RSS key per VSI **/ enum iavf_status_code iavf_aq_set_rss_key(struct iavf_hw *hw, u16 vsi_id, struct iavf_aqc_get_set_rss_key_data *key) { return iavf_aq_get_set_rss_key(hw, vsi_id, key, true); } /* The iavf_ptype_lookup table is used to convert from the 8-bit ptype in the * hardware to a bit-field that can be used by SW to more easily determine the * packet type. * * Macros are used to shorten the table lines and make this table human * readable. * * We store the PTYPE in the top byte of the bit field - this is just so that * we can check that the table doesn't have a row missing, as the index into * the table should be the PTYPE. * * Typical work flow: * * IF NOT iavf_ptype_lookup[ptype].known * THEN * Packet is unknown * ELSE IF iavf_ptype_lookup[ptype].outer_ip == IAVF_RX_PTYPE_OUTER_IP * Use the rest of the fields to look at the tunnels, inner protocols, etc * ELSE * Use the enum iavf_rx_l2_ptype to decode the packet type * ENDIF */ /* macro to make the table lines short */ #define IAVF_PTT(PTYPE, OUTER_IP, OUTER_IP_VER, OUTER_FRAG, T, TE, TEF, I, PL)\ { PTYPE, \ 1, \ IAVF_RX_PTYPE_OUTER_##OUTER_IP, \ IAVF_RX_PTYPE_OUTER_##OUTER_IP_VER, \ IAVF_RX_PTYPE_##OUTER_FRAG, \ IAVF_RX_PTYPE_TUNNEL_##T, \ IAVF_RX_PTYPE_TUNNEL_END_##TE, \ IAVF_RX_PTYPE_##TEF, \ IAVF_RX_PTYPE_INNER_PROT_##I, \ IAVF_RX_PTYPE_PAYLOAD_LAYER_##PL } #define IAVF_PTT_UNUSED_ENTRY(PTYPE) \ { PTYPE, 0, 0, 0, 0, 0, 0, 0, 0, 0 } /* shorter macros makes the table fit but are terse */ #define IAVF_RX_PTYPE_NOF IAVF_RX_PTYPE_NOT_FRAG #define IAVF_RX_PTYPE_FRG IAVF_RX_PTYPE_FRAG #define IAVF_RX_PTYPE_INNER_PROT_TS IAVF_RX_PTYPE_INNER_PROT_TIMESYNC /* Lookup table mapping the HW PTYPE to the bit field for decoding */ struct iavf_rx_ptype_decoded iavf_ptype_lookup[] = { /* L2 Packet types */ IAVF_PTT_UNUSED_ENTRY(0), IAVF_PTT(1, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY2), IAVF_PTT(2, L2, NONE, NOF, NONE, NONE, NOF, TS, PAY2), IAVF_PTT(3, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY2), IAVF_PTT_UNUSED_ENTRY(4), IAVF_PTT_UNUSED_ENTRY(5), IAVF_PTT(6, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY2), IAVF_PTT(7, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY2), IAVF_PTT_UNUSED_ENTRY(8), IAVF_PTT_UNUSED_ENTRY(9), IAVF_PTT(10, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY2), IAVF_PTT(11, L2, NONE, NOF, NONE, NONE, NOF, NONE, NONE), IAVF_PTT(12, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(13, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(14, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(15, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(16, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(17, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(18, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(19, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(20, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(21, L2, NONE, NOF, NONE, NONE, NOF, NONE, PAY3), /* Non Tunneled IPv4 */ IAVF_PTT(22, IP, IPV4, FRG, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(23, IP, IPV4, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(24, IP, IPV4, NOF, NONE, NONE, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(25), IAVF_PTT(26, IP, IPV4, NOF, NONE, NONE, NOF, TCP, PAY4), IAVF_PTT(27, IP, IPV4, NOF, NONE, NONE, NOF, SCTP, PAY4), IAVF_PTT(28, IP, IPV4, NOF, NONE, NONE, NOF, ICMP, PAY4), /* IPv4 --> IPv4 */ IAVF_PTT(29, IP, IPV4, NOF, IP_IP, IPV4, FRG, NONE, PAY3), IAVF_PTT(30, IP, IPV4, NOF, IP_IP, IPV4, NOF, NONE, PAY3), IAVF_PTT(31, IP, IPV4, NOF, IP_IP, IPV4, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(32), IAVF_PTT(33, IP, IPV4, NOF, IP_IP, IPV4, NOF, TCP, PAY4), IAVF_PTT(34, IP, IPV4, NOF, IP_IP, IPV4, NOF, SCTP, PAY4), IAVF_PTT(35, IP, IPV4, NOF, IP_IP, IPV4, NOF, ICMP, PAY4), /* IPv4 --> IPv6 */ IAVF_PTT(36, IP, IPV4, NOF, IP_IP, IPV6, FRG, NONE, PAY3), IAVF_PTT(37, IP, IPV4, NOF, IP_IP, IPV6, NOF, NONE, PAY3), IAVF_PTT(38, IP, IPV4, NOF, IP_IP, IPV6, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(39), IAVF_PTT(40, IP, IPV4, NOF, IP_IP, IPV6, NOF, TCP, PAY4), IAVF_PTT(41, IP, IPV4, NOF, IP_IP, IPV6, NOF, SCTP, PAY4), IAVF_PTT(42, IP, IPV4, NOF, IP_IP, IPV6, NOF, ICMP, PAY4), /* IPv4 --> GRE/NAT */ IAVF_PTT(43, IP, IPV4, NOF, IP_GRENAT, NONE, NOF, NONE, PAY3), /* IPv4 --> GRE/NAT --> IPv4 */ IAVF_PTT(44, IP, IPV4, NOF, IP_GRENAT, IPV4, FRG, NONE, PAY3), IAVF_PTT(45, IP, IPV4, NOF, IP_GRENAT, IPV4, NOF, NONE, PAY3), IAVF_PTT(46, IP, IPV4, NOF, IP_GRENAT, IPV4, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(47), IAVF_PTT(48, IP, IPV4, NOF, IP_GRENAT, IPV4, NOF, TCP, PAY4), IAVF_PTT(49, IP, IPV4, NOF, IP_GRENAT, IPV4, NOF, SCTP, PAY4), IAVF_PTT(50, IP, IPV4, NOF, IP_GRENAT, IPV4, NOF, ICMP, PAY4), /* IPv4 --> GRE/NAT --> IPv6 */ IAVF_PTT(51, IP, IPV4, NOF, IP_GRENAT, IPV6, FRG, NONE, PAY3), IAVF_PTT(52, IP, IPV4, NOF, IP_GRENAT, IPV6, NOF, NONE, PAY3), IAVF_PTT(53, IP, IPV4, NOF, IP_GRENAT, IPV6, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(54), IAVF_PTT(55, IP, IPV4, NOF, IP_GRENAT, IPV6, NOF, TCP, PAY4), IAVF_PTT(56, IP, IPV4, NOF, IP_GRENAT, IPV6, NOF, SCTP, PAY4), IAVF_PTT(57, IP, IPV4, NOF, IP_GRENAT, IPV6, NOF, ICMP, PAY4), /* IPv4 --> GRE/NAT --> MAC */ IAVF_PTT(58, IP, IPV4, NOF, IP_GRENAT_MAC, NONE, NOF, NONE, PAY3), /* IPv4 --> GRE/NAT --> MAC --> IPv4 */ IAVF_PTT(59, IP, IPV4, NOF, IP_GRENAT_MAC, IPV4, FRG, NONE, PAY3), IAVF_PTT(60, IP, IPV4, NOF, IP_GRENAT_MAC, IPV4, NOF, NONE, PAY3), IAVF_PTT(61, IP, IPV4, NOF, IP_GRENAT_MAC, IPV4, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(62), IAVF_PTT(63, IP, IPV4, NOF, IP_GRENAT_MAC, IPV4, NOF, TCP, PAY4), IAVF_PTT(64, IP, IPV4, NOF, IP_GRENAT_MAC, IPV4, NOF, SCTP, PAY4), IAVF_PTT(65, IP, IPV4, NOF, IP_GRENAT_MAC, IPV4, NOF, ICMP, PAY4), /* IPv4 --> GRE/NAT -> MAC --> IPv6 */ IAVF_PTT(66, IP, IPV4, NOF, IP_GRENAT_MAC, IPV6, FRG, NONE, PAY3), IAVF_PTT(67, IP, IPV4, NOF, IP_GRENAT_MAC, IPV6, NOF, NONE, PAY3), IAVF_PTT(68, IP, IPV4, NOF, IP_GRENAT_MAC, IPV6, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(69), IAVF_PTT(70, IP, IPV4, NOF, IP_GRENAT_MAC, IPV6, NOF, TCP, PAY4), IAVF_PTT(71, IP, IPV4, NOF, IP_GRENAT_MAC, IPV6, NOF, SCTP, PAY4), IAVF_PTT(72, IP, IPV4, NOF, IP_GRENAT_MAC, IPV6, NOF, ICMP, PAY4), /* IPv4 --> GRE/NAT --> MAC/VLAN */ IAVF_PTT(73, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, NONE, NOF, NONE, PAY3), /* IPv4 ---> GRE/NAT -> MAC/VLAN --> IPv4 */ IAVF_PTT(74, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV4, FRG, NONE, PAY3), IAVF_PTT(75, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV4, NOF, NONE, PAY3), IAVF_PTT(76, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV4, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(77), IAVF_PTT(78, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV4, NOF, TCP, PAY4), IAVF_PTT(79, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV4, NOF, SCTP, PAY4), IAVF_PTT(80, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV4, NOF, ICMP, PAY4), /* IPv4 -> GRE/NAT -> MAC/VLAN --> IPv6 */ IAVF_PTT(81, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV6, FRG, NONE, PAY3), IAVF_PTT(82, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV6, NOF, NONE, PAY3), IAVF_PTT(83, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV6, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(84), IAVF_PTT(85, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV6, NOF, TCP, PAY4), IAVF_PTT(86, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV6, NOF, SCTP, PAY4), IAVF_PTT(87, IP, IPV4, NOF, IP_GRENAT_MAC_VLAN, IPV6, NOF, ICMP, PAY4), /* Non Tunneled IPv6 */ IAVF_PTT(88, IP, IPV6, FRG, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(89, IP, IPV6, NOF, NONE, NONE, NOF, NONE, PAY3), IAVF_PTT(90, IP, IPV6, NOF, NONE, NONE, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(91), IAVF_PTT(92, IP, IPV6, NOF, NONE, NONE, NOF, TCP, PAY4), IAVF_PTT(93, IP, IPV6, NOF, NONE, NONE, NOF, SCTP, PAY4), IAVF_PTT(94, IP, IPV6, NOF, NONE, NONE, NOF, ICMP, PAY4), /* IPv6 --> IPv4 */ IAVF_PTT(95, IP, IPV6, NOF, IP_IP, IPV4, FRG, NONE, PAY3), IAVF_PTT(96, IP, IPV6, NOF, IP_IP, IPV4, NOF, NONE, PAY3), IAVF_PTT(97, IP, IPV6, NOF, IP_IP, IPV4, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(98), IAVF_PTT(99, IP, IPV6, NOF, IP_IP, IPV4, NOF, TCP, PAY4), IAVF_PTT(100, IP, IPV6, NOF, IP_IP, IPV4, NOF, SCTP, PAY4), IAVF_PTT(101, IP, IPV6, NOF, IP_IP, IPV4, NOF, ICMP, PAY4), /* IPv6 --> IPv6 */ IAVF_PTT(102, IP, IPV6, NOF, IP_IP, IPV6, FRG, NONE, PAY3), IAVF_PTT(103, IP, IPV6, NOF, IP_IP, IPV6, NOF, NONE, PAY3), IAVF_PTT(104, IP, IPV6, NOF, IP_IP, IPV6, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(105), IAVF_PTT(106, IP, IPV6, NOF, IP_IP, IPV6, NOF, TCP, PAY4), IAVF_PTT(107, IP, IPV6, NOF, IP_IP, IPV6, NOF, SCTP, PAY4), IAVF_PTT(108, IP, IPV6, NOF, IP_IP, IPV6, NOF, ICMP, PAY4), /* IPv6 --> GRE/NAT */ IAVF_PTT(109, IP, IPV6, NOF, IP_GRENAT, NONE, NOF, NONE, PAY3), /* IPv6 --> GRE/NAT -> IPv4 */ IAVF_PTT(110, IP, IPV6, NOF, IP_GRENAT, IPV4, FRG, NONE, PAY3), IAVF_PTT(111, IP, IPV6, NOF, IP_GRENAT, IPV4, NOF, NONE, PAY3), IAVF_PTT(112, IP, IPV6, NOF, IP_GRENAT, IPV4, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(113), IAVF_PTT(114, IP, IPV6, NOF, IP_GRENAT, IPV4, NOF, TCP, PAY4), IAVF_PTT(115, IP, IPV6, NOF, IP_GRENAT, IPV4, NOF, SCTP, PAY4), IAVF_PTT(116, IP, IPV6, NOF, IP_GRENAT, IPV4, NOF, ICMP, PAY4), /* IPv6 --> GRE/NAT -> IPv6 */ IAVF_PTT(117, IP, IPV6, NOF, IP_GRENAT, IPV6, FRG, NONE, PAY3), IAVF_PTT(118, IP, IPV6, NOF, IP_GRENAT, IPV6, NOF, NONE, PAY3), IAVF_PTT(119, IP, IPV6, NOF, IP_GRENAT, IPV6, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(120), IAVF_PTT(121, IP, IPV6, NOF, IP_GRENAT, IPV6, NOF, TCP, PAY4), IAVF_PTT(122, IP, IPV6, NOF, IP_GRENAT, IPV6, NOF, SCTP, PAY4), IAVF_PTT(123, IP, IPV6, NOF, IP_GRENAT, IPV6, NOF, ICMP, PAY4), /* IPv6 --> GRE/NAT -> MAC */ IAVF_PTT(124, IP, IPV6, NOF, IP_GRENAT_MAC, NONE, NOF, NONE, PAY3), /* IPv6 --> GRE/NAT -> MAC -> IPv4 */ IAVF_PTT(125, IP, IPV6, NOF, IP_GRENAT_MAC, IPV4, FRG, NONE, PAY3), IAVF_PTT(126, IP, IPV6, NOF, IP_GRENAT_MAC, IPV4, NOF, NONE, PAY3), IAVF_PTT(127, IP, IPV6, NOF, IP_GRENAT_MAC, IPV4, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(128), IAVF_PTT(129, IP, IPV6, NOF, IP_GRENAT_MAC, IPV4, NOF, TCP, PAY4), IAVF_PTT(130, IP, IPV6, NOF, IP_GRENAT_MAC, IPV4, NOF, SCTP, PAY4), IAVF_PTT(131, IP, IPV6, NOF, IP_GRENAT_MAC, IPV4, NOF, ICMP, PAY4), /* IPv6 --> GRE/NAT -> MAC -> IPv6 */ IAVF_PTT(132, IP, IPV6, NOF, IP_GRENAT_MAC, IPV6, FRG, NONE, PAY3), IAVF_PTT(133, IP, IPV6, NOF, IP_GRENAT_MAC, IPV6, NOF, NONE, PAY3), IAVF_PTT(134, IP, IPV6, NOF, IP_GRENAT_MAC, IPV6, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(135), IAVF_PTT(136, IP, IPV6, NOF, IP_GRENAT_MAC, IPV6, NOF, TCP, PAY4), IAVF_PTT(137, IP, IPV6, NOF, IP_GRENAT_MAC, IPV6, NOF, SCTP, PAY4), IAVF_PTT(138, IP, IPV6, NOF, IP_GRENAT_MAC, IPV6, NOF, ICMP, PAY4), /* IPv6 --> GRE/NAT -> MAC/VLAN */ IAVF_PTT(139, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, NONE, NOF, NONE, PAY3), /* IPv6 --> GRE/NAT -> MAC/VLAN --> IPv4 */ IAVF_PTT(140, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV4, FRG, NONE, PAY3), IAVF_PTT(141, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV4, NOF, NONE, PAY3), IAVF_PTT(142, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV4, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(143), IAVF_PTT(144, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV4, NOF, TCP, PAY4), IAVF_PTT(145, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV4, NOF, SCTP, PAY4), IAVF_PTT(146, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV4, NOF, ICMP, PAY4), /* IPv6 --> GRE/NAT -> MAC/VLAN --> IPv6 */ IAVF_PTT(147, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV6, FRG, NONE, PAY3), IAVF_PTT(148, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV6, NOF, NONE, PAY3), IAVF_PTT(149, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV6, NOF, UDP, PAY4), IAVF_PTT_UNUSED_ENTRY(150), IAVF_PTT(151, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV6, NOF, TCP, PAY4), IAVF_PTT(152, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV6, NOF, SCTP, PAY4), IAVF_PTT(153, IP, IPV6, NOF, IP_GRENAT_MAC_VLAN, IPV6, NOF, ICMP, PAY4), /* unused entries */ IAVF_PTT_UNUSED_ENTRY(154), IAVF_PTT_UNUSED_ENTRY(155), IAVF_PTT_UNUSED_ENTRY(156), IAVF_PTT_UNUSED_ENTRY(157), IAVF_PTT_UNUSED_ENTRY(158), IAVF_PTT_UNUSED_ENTRY(159), IAVF_PTT_UNUSED_ENTRY(160), IAVF_PTT_UNUSED_ENTRY(161), IAVF_PTT_UNUSED_ENTRY(162), IAVF_PTT_UNUSED_ENTRY(163), IAVF_PTT_UNUSED_ENTRY(164), IAVF_PTT_UNUSED_ENTRY(165), IAVF_PTT_UNUSED_ENTRY(166), IAVF_PTT_UNUSED_ENTRY(167), IAVF_PTT_UNUSED_ENTRY(168), IAVF_PTT_UNUSED_ENTRY(169), IAVF_PTT_UNUSED_ENTRY(170), IAVF_PTT_UNUSED_ENTRY(171), IAVF_PTT_UNUSED_ENTRY(172), IAVF_PTT_UNUSED_ENTRY(173), IAVF_PTT_UNUSED_ENTRY(174), IAVF_PTT_UNUSED_ENTRY(175), IAVF_PTT_UNUSED_ENTRY(176), IAVF_PTT_UNUSED_ENTRY(177), IAVF_PTT_UNUSED_ENTRY(178), IAVF_PTT_UNUSED_ENTRY(179), IAVF_PTT_UNUSED_ENTRY(180), IAVF_PTT_UNUSED_ENTRY(181), IAVF_PTT_UNUSED_ENTRY(182), IAVF_PTT_UNUSED_ENTRY(183), IAVF_PTT_UNUSED_ENTRY(184), IAVF_PTT_UNUSED_ENTRY(185), IAVF_PTT_UNUSED_ENTRY(186), IAVF_PTT_UNUSED_ENTRY(187), IAVF_PTT_UNUSED_ENTRY(188), IAVF_PTT_UNUSED_ENTRY(189), IAVF_PTT_UNUSED_ENTRY(190), IAVF_PTT_UNUSED_ENTRY(191), IAVF_PTT_UNUSED_ENTRY(192), IAVF_PTT_UNUSED_ENTRY(193), IAVF_PTT_UNUSED_ENTRY(194), IAVF_PTT_UNUSED_ENTRY(195), IAVF_PTT_UNUSED_ENTRY(196), IAVF_PTT_UNUSED_ENTRY(197), IAVF_PTT_UNUSED_ENTRY(198), IAVF_PTT_UNUSED_ENTRY(199), IAVF_PTT_UNUSED_ENTRY(200), IAVF_PTT_UNUSED_ENTRY(201), IAVF_PTT_UNUSED_ENTRY(202), IAVF_PTT_UNUSED_ENTRY(203), IAVF_PTT_UNUSED_ENTRY(204), IAVF_PTT_UNUSED_ENTRY(205), IAVF_PTT_UNUSED_ENTRY(206), IAVF_PTT_UNUSED_ENTRY(207), IAVF_PTT_UNUSED_ENTRY(208), IAVF_PTT_UNUSED_ENTRY(209), IAVF_PTT_UNUSED_ENTRY(210), IAVF_PTT_UNUSED_ENTRY(211), IAVF_PTT_UNUSED_ENTRY(212), IAVF_PTT_UNUSED_ENTRY(213), IAVF_PTT_UNUSED_ENTRY(214), IAVF_PTT_UNUSED_ENTRY(215), IAVF_PTT_UNUSED_ENTRY(216), IAVF_PTT_UNUSED_ENTRY(217), IAVF_PTT_UNUSED_ENTRY(218), IAVF_PTT_UNUSED_ENTRY(219), IAVF_PTT_UNUSED_ENTRY(220), IAVF_PTT_UNUSED_ENTRY(221), IAVF_PTT_UNUSED_ENTRY(222), IAVF_PTT_UNUSED_ENTRY(223), IAVF_PTT_UNUSED_ENTRY(224), IAVF_PTT_UNUSED_ENTRY(225), IAVF_PTT_UNUSED_ENTRY(226), IAVF_PTT_UNUSED_ENTRY(227), IAVF_PTT_UNUSED_ENTRY(228), IAVF_PTT_UNUSED_ENTRY(229), IAVF_PTT_UNUSED_ENTRY(230), IAVF_PTT_UNUSED_ENTRY(231), IAVF_PTT_UNUSED_ENTRY(232), IAVF_PTT_UNUSED_ENTRY(233), IAVF_PTT_UNUSED_ENTRY(234), IAVF_PTT_UNUSED_ENTRY(235), IAVF_PTT_UNUSED_ENTRY(236), IAVF_PTT_UNUSED_ENTRY(237), IAVF_PTT_UNUSED_ENTRY(238), IAVF_PTT_UNUSED_ENTRY(239), IAVF_PTT_UNUSED_ENTRY(240), IAVF_PTT_UNUSED_ENTRY(241), IAVF_PTT_UNUSED_ENTRY(242), IAVF_PTT_UNUSED_ENTRY(243), IAVF_PTT_UNUSED_ENTRY(244), IAVF_PTT_UNUSED_ENTRY(245), IAVF_PTT_UNUSED_ENTRY(246), IAVF_PTT_UNUSED_ENTRY(247), IAVF_PTT_UNUSED_ENTRY(248), IAVF_PTT_UNUSED_ENTRY(249), IAVF_PTT_UNUSED_ENTRY(250), IAVF_PTT_UNUSED_ENTRY(251), IAVF_PTT_UNUSED_ENTRY(252), IAVF_PTT_UNUSED_ENTRY(253), IAVF_PTT_UNUSED_ENTRY(254), IAVF_PTT_UNUSED_ENTRY(255) }; /** * iavf_validate_mac_addr - Validate unicast MAC address * @mac_addr: pointer to MAC address * * Tests a MAC address to ensure it is a valid Individual Address **/ enum iavf_status_code iavf_validate_mac_addr(u8 *mac_addr) { enum iavf_status_code status = IAVF_SUCCESS; DEBUGFUNC("iavf_validate_mac_addr"); /* Broadcast addresses ARE multicast addresses * Make sure it is not a multicast address * Reject the zero address */ if (IAVF_IS_MULTICAST(mac_addr) || (mac_addr[0] == 0 && mac_addr[1] == 0 && mac_addr[2] == 0 && mac_addr[3] == 0 && mac_addr[4] == 0 && mac_addr[5] == 0)) status = IAVF_ERR_INVALID_MAC_ADDR; return status; } /** * iavf_aq_rx_ctl_read_register - use FW to read from an Rx control register * @hw: pointer to the hw struct * @reg_addr: register address * @reg_val: ptr to register value * @cmd_details: pointer to command details structure or NULL * * Use the firmware to read the Rx control register, * especially useful if the Rx unit is under heavy pressure **/ enum iavf_status_code iavf_aq_rx_ctl_read_register(struct iavf_hw *hw, u32 reg_addr, u32 *reg_val, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; struct iavf_aqc_rx_ctl_reg_read_write *cmd_resp = (struct iavf_aqc_rx_ctl_reg_read_write *)&desc.params.raw; enum iavf_status_code status; if (reg_val == NULL) return IAVF_ERR_PARAM; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_rx_ctl_reg_read); cmd_resp->address = CPU_TO_LE32(reg_addr); status = iavf_asq_send_command(hw, &desc, NULL, 0, cmd_details); if (status == IAVF_SUCCESS) *reg_val = LE32_TO_CPU(cmd_resp->value); return status; } /** * iavf_read_rx_ctl - read from an Rx control register * @hw: pointer to the hw struct * @reg_addr: register address **/ u32 iavf_read_rx_ctl(struct iavf_hw *hw, u32 reg_addr) { enum iavf_status_code status = IAVF_SUCCESS; bool use_register; int retry = 5; u32 val = 0; use_register = (((hw->aq.api_maj_ver == 1) && (hw->aq.api_min_ver < 5)) || (hw->mac.type == IAVF_MAC_X722)); if (!use_register) { do_retry: status = iavf_aq_rx_ctl_read_register(hw, reg_addr, &val, NULL); if (hw->aq.asq_last_status == IAVF_AQ_RC_EAGAIN && retry) { iavf_msec_delay(1); retry--; goto do_retry; } } /* if the AQ access failed, try the old-fashioned way */ if (status || use_register) val = rd32(hw, reg_addr); return val; } /** * iavf_aq_rx_ctl_write_register * @hw: pointer to the hw struct * @reg_addr: register address * @reg_val: register value * @cmd_details: pointer to command details structure or NULL * * Use the firmware to write to an Rx control register, * especially useful if the Rx unit is under heavy pressure **/ enum iavf_status_code iavf_aq_rx_ctl_write_register(struct iavf_hw *hw, u32 reg_addr, u32 reg_val, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; struct iavf_aqc_rx_ctl_reg_read_write *cmd = (struct iavf_aqc_rx_ctl_reg_read_write *)&desc.params.raw; enum iavf_status_code status; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_rx_ctl_reg_write); cmd->address = CPU_TO_LE32(reg_addr); cmd->value = CPU_TO_LE32(reg_val); status = iavf_asq_send_command(hw, &desc, NULL, 0, cmd_details); return status; } /** * iavf_write_rx_ctl - write to an Rx control register * @hw: pointer to the hw struct * @reg_addr: register address * @reg_val: register value **/ void iavf_write_rx_ctl(struct iavf_hw *hw, u32 reg_addr, u32 reg_val) { enum iavf_status_code status = IAVF_SUCCESS; bool use_register; int retry = 5; use_register = (((hw->aq.api_maj_ver == 1) && (hw->aq.api_min_ver < 5)) || (hw->mac.type == IAVF_MAC_X722)); if (!use_register) { do_retry: status = iavf_aq_rx_ctl_write_register(hw, reg_addr, reg_val, NULL); if (hw->aq.asq_last_status == IAVF_AQ_RC_EAGAIN && retry) { iavf_msec_delay(1); retry--; goto do_retry; } } /* if the AQ access failed, try the old-fashioned way */ if (status || use_register) wr32(hw, reg_addr, reg_val); } /** * iavf_aq_set_phy_register * @hw: pointer to the hw struct * @phy_select: select which phy should be accessed * @dev_addr: PHY device address * @reg_addr: PHY register address * @reg_val: new register value * @cmd_details: pointer to command details structure or NULL * * Write the external PHY register. **/ enum iavf_status_code iavf_aq_set_phy_register(struct iavf_hw *hw, u8 phy_select, u8 dev_addr, u32 reg_addr, u32 reg_val, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; struct iavf_aqc_phy_register_access *cmd = (struct iavf_aqc_phy_register_access *)&desc.params.raw; enum iavf_status_code status; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_set_phy_register); cmd->phy_interface = phy_select; cmd->dev_addres = dev_addr; cmd->reg_address = CPU_TO_LE32(reg_addr); cmd->reg_value = CPU_TO_LE32(reg_val); status = iavf_asq_send_command(hw, &desc, NULL, 0, cmd_details); return status; } /** * iavf_aq_get_phy_register * @hw: pointer to the hw struct * @phy_select: select which phy should be accessed * @dev_addr: PHY device address * @reg_addr: PHY register address * @reg_val: read register value * @cmd_details: pointer to command details structure or NULL * * Read the external PHY register. **/ enum iavf_status_code iavf_aq_get_phy_register(struct iavf_hw *hw, u8 phy_select, u8 dev_addr, u32 reg_addr, u32 *reg_val, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; struct iavf_aqc_phy_register_access *cmd = (struct iavf_aqc_phy_register_access *)&desc.params.raw; enum iavf_status_code status; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_get_phy_register); cmd->phy_interface = phy_select; cmd->dev_addres = dev_addr; cmd->reg_address = CPU_TO_LE32(reg_addr); status = iavf_asq_send_command(hw, &desc, NULL, 0, cmd_details); if (!status) *reg_val = LE32_TO_CPU(cmd->reg_value); return status; } /** * iavf_aq_send_msg_to_pf * @hw: pointer to the hardware structure * @v_opcode: opcodes for VF-PF communication * @v_retval: return error code * @msg: pointer to the msg buffer * @msglen: msg length * @cmd_details: pointer to command details * * Send message to PF driver using admin queue. By default, this message * is sent asynchronously, i.e. iavf_asq_send_command() does not wait for * completion before returning. **/ enum iavf_status_code iavf_aq_send_msg_to_pf(struct iavf_hw *hw, enum virtchnl_ops v_opcode, enum iavf_status_code v_retval, u8 *msg, u16 msglen, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; struct iavf_asq_cmd_details details; enum iavf_status_code status; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_send_msg_to_pf); desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_SI); desc.cookie_high = CPU_TO_LE32(v_opcode); desc.cookie_low = CPU_TO_LE32(v_retval); if (msglen) { desc.flags |= CPU_TO_LE16((u16)(IAVF_AQ_FLAG_BUF | IAVF_AQ_FLAG_RD)); if (msglen > IAVF_AQ_LARGE_BUF) desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_LB); desc.datalen = CPU_TO_LE16(msglen); } if (!cmd_details) { iavf_memset(&details, 0, sizeof(details), IAVF_NONDMA_MEM); details.async = true; cmd_details = &details; } status = iavf_asq_send_command(hw, (struct iavf_aq_desc *)&desc, msg, msglen, cmd_details); return status; } /** * iavf_parse_hw_config * @hw: pointer to the hardware structure * @msg: pointer to the virtual channel VF resource structure * * Given a VF resource message from the PF, populate the hw struct * with appropriate information. **/ void iavf_parse_hw_config(struct iavf_hw *hw, struct virtchnl_vf_resource *msg) { struct virtchnl_vsi_resource *vsi_res; int i; vsi_res = &msg->vsi_res[0]; hw->dev_caps.num_vsis = msg->num_vsis; hw->dev_caps.num_rx_qp = msg->num_queue_pairs; hw->dev_caps.num_tx_qp = msg->num_queue_pairs; hw->dev_caps.num_msix_vectors_vf = msg->max_vectors; hw->dev_caps.dcb = msg->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_L2; hw->dev_caps.iwarp = (msg->vf_cap_flags & VIRTCHNL_VF_OFFLOAD_IWARP) ? 1 : 0; for (i = 0; i < msg->num_vsis; i++) { if (vsi_res->vsi_type == VIRTCHNL_VSI_SRIOV) { iavf_memcpy(hw->mac.perm_addr, vsi_res->default_mac_addr, ETH_ALEN, IAVF_NONDMA_TO_NONDMA); iavf_memcpy(hw->mac.addr, vsi_res->default_mac_addr, ETH_ALEN, IAVF_NONDMA_TO_NONDMA); } vsi_res++; } } /** * iavf_reset * @hw: pointer to the hardware structure * * Send a VF_RESET message to the PF. Does not wait for response from PF * as none will be forthcoming. Immediately after calling this function, * the admin queue should be shut down and (optionally) reinitialized. **/ enum iavf_status_code iavf_reset(struct iavf_hw *hw) { return iavf_aq_send_msg_to_pf(hw, VIRTCHNL_OP_RESET_VF, IAVF_SUCCESS, NULL, 0, NULL); } /** * iavf_aq_set_arp_proxy_config * @hw: pointer to the HW structure * @proxy_config: pointer to proxy config command table struct * @cmd_details: pointer to command details * * Set ARP offload parameters from pre-populated * iavf_aqc_arp_proxy_data struct **/ enum iavf_status_code iavf_aq_set_arp_proxy_config(struct iavf_hw *hw, struct iavf_aqc_arp_proxy_data *proxy_config, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; enum iavf_status_code status; if (!proxy_config) return IAVF_ERR_PARAM; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_set_proxy_config); desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_BUF); desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_RD); desc.params.external.addr_high = CPU_TO_LE32(IAVF_HI_DWORD((u64)proxy_config)); desc.params.external.addr_low = CPU_TO_LE32(IAVF_LO_DWORD((u64)proxy_config)); desc.datalen = CPU_TO_LE16(sizeof(struct iavf_aqc_arp_proxy_data)); status = iavf_asq_send_command(hw, &desc, proxy_config, sizeof(struct iavf_aqc_arp_proxy_data), cmd_details); return status; } /** * iavf_aq_opc_set_ns_proxy_table_entry * @hw: pointer to the HW structure * @ns_proxy_table_entry: pointer to NS table entry command struct * @cmd_details: pointer to command details * * Set IPv6 Neighbor Solicitation (NS) protocol offload parameters * from pre-populated iavf_aqc_ns_proxy_data struct **/ enum iavf_status_code iavf_aq_set_ns_proxy_table_entry(struct iavf_hw *hw, struct iavf_aqc_ns_proxy_data *ns_proxy_table_entry, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; enum iavf_status_code status; if (!ns_proxy_table_entry) return IAVF_ERR_PARAM; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_set_ns_proxy_table_entry); desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_BUF); desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_RD); desc.params.external.addr_high = CPU_TO_LE32(IAVF_HI_DWORD((u64)ns_proxy_table_entry)); desc.params.external.addr_low = CPU_TO_LE32(IAVF_LO_DWORD((u64)ns_proxy_table_entry)); desc.datalen = CPU_TO_LE16(sizeof(struct iavf_aqc_ns_proxy_data)); status = iavf_asq_send_command(hw, &desc, ns_proxy_table_entry, sizeof(struct iavf_aqc_ns_proxy_data), cmd_details); return status; } /** * iavf_aq_set_clear_wol_filter * @hw: pointer to the hw struct * @filter_index: index of filter to modify (0-7) * @filter: buffer containing filter to be set * @set_filter: true to set filter, false to clear filter * @no_wol_tco: if true, pass through packets cannot cause wake-up * if false, pass through packets may cause wake-up * @filter_valid: true if filter action is valid * @no_wol_tco_valid: true if no WoL in TCO traffic action valid * @cmd_details: pointer to command details structure or NULL * * Set or clear WoL filter for port attached to the PF **/ enum iavf_status_code iavf_aq_set_clear_wol_filter(struct iavf_hw *hw, u8 filter_index, struct iavf_aqc_set_wol_filter_data *filter, bool set_filter, bool no_wol_tco, bool filter_valid, bool no_wol_tco_valid, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; struct iavf_aqc_set_wol_filter *cmd = (struct iavf_aqc_set_wol_filter *)&desc.params.raw; enum iavf_status_code status; u16 cmd_flags = 0; u16 valid_flags = 0; u16 buff_len = 0; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_set_wol_filter); if (filter_index >= IAVF_AQC_MAX_NUM_WOL_FILTERS) return IAVF_ERR_PARAM; cmd->filter_index = CPU_TO_LE16(filter_index); if (set_filter) { if (!filter) return IAVF_ERR_PARAM; cmd_flags |= IAVF_AQC_SET_WOL_FILTER; cmd_flags |= IAVF_AQC_SET_WOL_FILTER_WOL_PRESERVE_ON_PFR; } if (no_wol_tco) cmd_flags |= IAVF_AQC_SET_WOL_FILTER_NO_TCO_WOL; cmd->cmd_flags = CPU_TO_LE16(cmd_flags); if (filter_valid) valid_flags |= IAVF_AQC_SET_WOL_FILTER_ACTION_VALID; if (no_wol_tco_valid) valid_flags |= IAVF_AQC_SET_WOL_FILTER_NO_TCO_ACTION_VALID; cmd->valid_flags = CPU_TO_LE16(valid_flags); buff_len = sizeof(*filter); desc.datalen = CPU_TO_LE16(buff_len); desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_BUF); desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_RD); cmd->address_high = CPU_TO_LE32(IAVF_HI_DWORD((u64)filter)); cmd->address_low = CPU_TO_LE32(IAVF_LO_DWORD((u64)filter)); status = iavf_asq_send_command(hw, &desc, filter, buff_len, cmd_details); return status; } /** * iavf_aq_get_wake_event_reason * @hw: pointer to the hw struct * @wake_reason: return value, index of matching filter * @cmd_details: pointer to command details structure or NULL * * Get information for the reason of a Wake Up event **/ enum iavf_status_code iavf_aq_get_wake_event_reason(struct iavf_hw *hw, u16 *wake_reason, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; struct iavf_aqc_get_wake_reason_completion *resp = (struct iavf_aqc_get_wake_reason_completion *)&desc.params.raw; enum iavf_status_code status; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_get_wake_reason); status = iavf_asq_send_command(hw, &desc, NULL, 0, cmd_details); if (status == IAVF_SUCCESS) *wake_reason = LE16_TO_CPU(resp->wake_reason); return status; } /** * iavf_aq_clear_all_wol_filters * @hw: pointer to the hw struct * @cmd_details: pointer to command details structure or NULL * * Get information for the reason of a Wake Up event **/ enum iavf_status_code iavf_aq_clear_all_wol_filters(struct iavf_hw *hw, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; enum iavf_status_code status; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_clear_all_wol_filters); status = iavf_asq_send_command(hw, &desc, NULL, 0, cmd_details); return status; } /** * iavf_aq_write_ddp - Write dynamic device personalization (ddp) * @hw: pointer to the hw struct * @buff: command buffer (size in bytes = buff_size) * @buff_size: buffer size in bytes * @track_id: package tracking id * @error_offset: returns error offset * @error_info: returns error information * @cmd_details: pointer to command details structure or NULL **/ enum iavf_status_code iavf_aq_write_ddp(struct iavf_hw *hw, void *buff, u16 buff_size, u32 track_id, u32 *error_offset, u32 *error_info, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; struct iavf_aqc_write_personalization_profile *cmd = (struct iavf_aqc_write_personalization_profile *) &desc.params.raw; struct iavf_aqc_write_ddp_resp *resp; enum iavf_status_code status; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_write_personalization_profile); desc.flags |= CPU_TO_LE16(IAVF_AQ_FLAG_BUF | IAVF_AQ_FLAG_RD); if (buff_size > IAVF_AQ_LARGE_BUF) desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_LB); desc.datalen = CPU_TO_LE16(buff_size); cmd->profile_track_id = CPU_TO_LE32(track_id); status = iavf_asq_send_command(hw, &desc, buff, buff_size, cmd_details); if (!status) { resp = (struct iavf_aqc_write_ddp_resp *)&desc.params.raw; if (error_offset) *error_offset = LE32_TO_CPU(resp->error_offset); if (error_info) *error_info = LE32_TO_CPU(resp->error_info); } return status; } /** * iavf_aq_get_ddp_list - Read dynamic device personalization (ddp) * @hw: pointer to the hw struct * @buff: command buffer (size in bytes = buff_size) * @buff_size: buffer size in bytes * @flags: AdminQ command flags * @cmd_details: pointer to command details structure or NULL **/ enum iavf_status_code iavf_aq_get_ddp_list(struct iavf_hw *hw, void *buff, u16 buff_size, u8 flags, struct iavf_asq_cmd_details *cmd_details) { struct iavf_aq_desc desc; struct iavf_aqc_get_applied_profiles *cmd = (struct iavf_aqc_get_applied_profiles *)&desc.params.raw; enum iavf_status_code status; iavf_fill_default_direct_cmd_desc(&desc, iavf_aqc_opc_get_personalization_profile_list); desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_BUF); if (buff_size > IAVF_AQ_LARGE_BUF) desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_LB); desc.datalen = CPU_TO_LE16(buff_size); cmd->flags = flags; status = iavf_asq_send_command(hw, &desc, buff, buff_size, cmd_details); return status; } /** * iavf_find_segment_in_package * @segment_type: the segment type to search for (i.e., SEGMENT_TYPE_IAVF) * @pkg_hdr: pointer to the package header to be searched * * This function searches a package file for a particular segment type. On * success it returns a pointer to the segment header, otherwise it will * return NULL. **/ struct iavf_generic_seg_header * iavf_find_segment_in_package(u32 segment_type, struct iavf_package_header *pkg_hdr) { struct iavf_generic_seg_header *segment; u32 i; /* Search all package segments for the requested segment type */ for (i = 0; i < pkg_hdr->segment_count; i++) { segment = (struct iavf_generic_seg_header *)((u8 *)pkg_hdr + pkg_hdr->segment_offset[i]); if (segment->type == segment_type) return segment; } return NULL; } /* Get section table in profile */ #define IAVF_SECTION_TABLE(profile, sec_tbl) \ do { \ struct iavf_profile_segment *p = (profile); \ u32 count; \ u32 *nvm; \ count = p->device_table_count; \ nvm = (u32 *)&p->device_table[count]; \ sec_tbl = (struct iavf_section_table *)&nvm[nvm[0] + 1]; \ } while (0) /* Get section header in profile */ #define IAVF_SECTION_HEADER(profile, offset) \ (struct iavf_profile_section_header *)((u8 *)(profile) + (offset)) /** * iavf_find_section_in_profile * @section_type: the section type to search for (i.e., SECTION_TYPE_NOTE) * @profile: pointer to the iavf segment header to be searched * * This function searches iavf segment for a particular section type. On * success it returns a pointer to the section header, otherwise it will * return NULL. **/ struct iavf_profile_section_header * iavf_find_section_in_profile(u32 section_type, struct iavf_profile_segment *profile) { struct iavf_profile_section_header *sec; struct iavf_section_table *sec_tbl; u32 sec_off; u32 i; if (profile->header.type != SEGMENT_TYPE_IAVF) return NULL; IAVF_SECTION_TABLE(profile, sec_tbl); for (i = 0; i < sec_tbl->section_count; i++) { sec_off = sec_tbl->section_offset[i]; sec = IAVF_SECTION_HEADER(profile, sec_off); if (sec->section.type == section_type) return sec; } return NULL; } /** * iavf_ddp_exec_aq_section - Execute generic AQ for DDP * @hw: pointer to the hw struct * @aq: command buffer containing all data to execute AQ **/ STATIC enum iavf_status_code iavf_ddp_exec_aq_section(struct iavf_hw *hw, struct iavf_profile_aq_section *aq) { enum iavf_status_code status; struct iavf_aq_desc desc; u8 *msg = NULL; u16 msglen; iavf_fill_default_direct_cmd_desc(&desc, aq->opcode); desc.flags |= CPU_TO_LE16(aq->flags); iavf_memcpy(desc.params.raw, aq->param, sizeof(desc.params.raw), IAVF_NONDMA_TO_NONDMA); msglen = aq->datalen; if (msglen) { desc.flags |= CPU_TO_LE16((u16)(IAVF_AQ_FLAG_BUF | IAVF_AQ_FLAG_RD)); if (msglen > IAVF_AQ_LARGE_BUF) desc.flags |= CPU_TO_LE16((u16)IAVF_AQ_FLAG_LB); desc.datalen = CPU_TO_LE16(msglen); msg = &aq->data[0]; } status = iavf_asq_send_command(hw, &desc, msg, msglen, NULL); if (status != IAVF_SUCCESS) { iavf_debug(hw, IAVF_DEBUG_PACKAGE, "unable to exec DDP AQ opcode %u, error %d\n", aq->opcode, status); return status; } /* copy returned desc to aq_buf */ iavf_memcpy(aq->param, desc.params.raw, sizeof(desc.params.raw), IAVF_NONDMA_TO_NONDMA); return IAVF_SUCCESS; } /** * iavf_validate_profile * @hw: pointer to the hardware structure * @profile: pointer to the profile segment of the package to be validated * @track_id: package tracking id * @rollback: flag if the profile is for rollback. * * Validates supported devices and profile's sections. */ STATIC enum iavf_status_code iavf_validate_profile(struct iavf_hw *hw, struct iavf_profile_segment *profile, u32 track_id, bool rollback) { struct iavf_profile_section_header *sec = NULL; enum iavf_status_code status = IAVF_SUCCESS; struct iavf_section_table *sec_tbl; u32 vendor_dev_id; u32 dev_cnt; u32 sec_off; u32 i; if (track_id == IAVF_DDP_TRACKID_INVALID) { iavf_debug(hw, IAVF_DEBUG_PACKAGE, "Invalid track_id\n"); return IAVF_NOT_SUPPORTED; } dev_cnt = profile->device_table_count; for (i = 0; i < dev_cnt; i++) { vendor_dev_id = profile->device_table[i].vendor_dev_id; if ((vendor_dev_id >> 16) == IAVF_INTEL_VENDOR_ID && hw->device_id == (vendor_dev_id & 0xFFFF)) break; } if (dev_cnt && (i == dev_cnt)) { iavf_debug(hw, IAVF_DEBUG_PACKAGE, "Device doesn't support DDP\n"); return IAVF_ERR_DEVICE_NOT_SUPPORTED; } IAVF_SECTION_TABLE(profile, sec_tbl); /* Validate sections types */ for (i = 0; i < sec_tbl->section_count; i++) { sec_off = sec_tbl->section_offset[i]; sec = IAVF_SECTION_HEADER(profile, sec_off); if (rollback) { if (sec->section.type == SECTION_TYPE_MMIO || sec->section.type == SECTION_TYPE_AQ || sec->section.type == SECTION_TYPE_RB_AQ) { iavf_debug(hw, IAVF_DEBUG_PACKAGE, "Not a roll-back package\n"); return IAVF_NOT_SUPPORTED; } } else { if (sec->section.type == SECTION_TYPE_RB_AQ || sec->section.type == SECTION_TYPE_RB_MMIO) { iavf_debug(hw, IAVF_DEBUG_PACKAGE, "Not an original package\n"); return IAVF_NOT_SUPPORTED; } } } return status; } /** * iavf_write_profile * @hw: pointer to the hardware structure * @profile: pointer to the profile segment of the package to be downloaded * @track_id: package tracking id * * Handles the download of a complete package. */ enum iavf_status_code iavf_write_profile(struct iavf_hw *hw, struct iavf_profile_segment *profile, u32 track_id) { enum iavf_status_code status = IAVF_SUCCESS; struct iavf_section_table *sec_tbl; struct iavf_profile_section_header *sec = NULL; struct iavf_profile_aq_section *ddp_aq; u32 section_size = 0; u32 offset = 0, info = 0; u32 sec_off; u32 i; status = iavf_validate_profile(hw, profile, track_id, false); if (status) return status; IAVF_SECTION_TABLE(profile, sec_tbl); for (i = 0; i < sec_tbl->section_count; i++) { sec_off = sec_tbl->section_offset[i]; sec = IAVF_SECTION_HEADER(profile, sec_off); /* Process generic admin command */ if (sec->section.type == SECTION_TYPE_AQ) { ddp_aq = (struct iavf_profile_aq_section *)&sec[1]; status = iavf_ddp_exec_aq_section(hw, ddp_aq); if (status) { iavf_debug(hw, IAVF_DEBUG_PACKAGE, "Failed to execute aq: section %d, opcode %u\n", i, ddp_aq->opcode); break; } sec->section.type = SECTION_TYPE_RB_AQ; } /* Skip any non-mmio sections */ if (sec->section.type != SECTION_TYPE_MMIO) continue; section_size = sec->section.size + sizeof(struct iavf_profile_section_header); /* Write MMIO section */ status = iavf_aq_write_ddp(hw, (void *)sec, (u16)section_size, track_id, &offset, &info, NULL); if (status) { iavf_debug(hw, IAVF_DEBUG_PACKAGE, "Failed to write profile: section %d, offset %d, info %d\n", i, offset, info); break; } } return status; } /** * iavf_rollback_profile * @hw: pointer to the hardware structure * @profile: pointer to the profile segment of the package to be removed * @track_id: package tracking id * * Rolls back previously loaded package. */ enum iavf_status_code iavf_rollback_profile(struct iavf_hw *hw, struct iavf_profile_segment *profile, u32 track_id) { struct iavf_profile_section_header *sec = NULL; enum iavf_status_code status = IAVF_SUCCESS; struct iavf_section_table *sec_tbl; u32 offset = 0, info = 0; u32 section_size = 0; u32 sec_off; int i; status = iavf_validate_profile(hw, profile, track_id, true); if (status) return status; IAVF_SECTION_TABLE(profile, sec_tbl); /* For rollback write sections in reverse */ for (i = sec_tbl->section_count - 1; i >= 0; i--) { sec_off = sec_tbl->section_offset[i]; sec = IAVF_SECTION_HEADER(profile, sec_off); /* Skip any non-rollback sections */ if (sec->section.type != SECTION_TYPE_RB_MMIO) continue; section_size = sec->section.size + sizeof(struct iavf_profile_section_header); /* Write roll-back MMIO section */ status = iavf_aq_write_ddp(hw, (void *)sec, (u16)section_size, track_id, &offset, &info, NULL); if (status) { iavf_debug(hw, IAVF_DEBUG_PACKAGE, "Failed to write profile: section %d, offset %d, info %d\n", i, offset, info); break; } } return status; } /** * iavf_add_pinfo_to_list * @hw: pointer to the hardware structure * @profile: pointer to the profile segment of the package * @profile_info_sec: buffer for information section * @track_id: package tracking id * * Register a profile to the list of loaded profiles. */ enum iavf_status_code iavf_add_pinfo_to_list(struct iavf_hw *hw, struct iavf_profile_segment *profile, u8 *profile_info_sec, u32 track_id) { enum iavf_status_code status = IAVF_SUCCESS; struct iavf_profile_section_header *sec = NULL; struct iavf_profile_info *pinfo; u32 offset = 0, info = 0; sec = (struct iavf_profile_section_header *)profile_info_sec; sec->tbl_size = 1; sec->data_end = sizeof(struct iavf_profile_section_header) + sizeof(struct iavf_profile_info); sec->section.type = SECTION_TYPE_INFO; sec->section.offset = sizeof(struct iavf_profile_section_header); sec->section.size = sizeof(struct iavf_profile_info); pinfo = (struct iavf_profile_info *)(profile_info_sec + sec->section.offset); pinfo->track_id = track_id; pinfo->version = profile->version; pinfo->op = IAVF_DDP_ADD_TRACKID; iavf_memcpy(pinfo->name, profile->name, IAVF_DDP_NAME_SIZE, IAVF_NONDMA_TO_NONDMA); status = iavf_aq_write_ddp(hw, (void *)sec, sec->data_end, track_id, &offset, &info, NULL); return status; }