| /* |
| * Copyright (c) 2005-2009 Brocade Communications Systems, Inc. |
| * All rights reserved |
| * www.brocade.com |
| * |
| * Linux driver for Brocade Fibre Channel Host Bus Adapter. |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms of the GNU General Public License (GPL) Version 2 as |
| * published by the Free Software Foundation |
| * |
| * This program is distributed in the hope that it will be useful, but |
| * WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * General Public License for more details. |
| */ |
| |
| #include <bfa.h> |
| #include <bfa_svc.h> |
| #include <bfi/bfi_pport.h> |
| #include <cs/bfa_debug.h> |
| #include <aen/bfa_aen.h> |
| #include <cs/bfa_plog.h> |
| #include <aen/bfa_aen_port.h> |
| |
| BFA_TRC_FILE(HAL, PPORT); |
| BFA_MODULE(pport); |
| |
| /* |
| * The port is considered disabled if corresponding physical port or IOC are |
| * disabled explicitly |
| */ |
| #define BFA_PORT_IS_DISABLED(bfa) \ |
| ((bfa_pport_is_disabled(bfa) == BFA_TRUE) || \ |
| (bfa_ioc_is_disabled(&bfa->ioc) == BFA_TRUE)) |
| |
| /* |
| * forward declarations |
| */ |
| static bfa_boolean_t bfa_pport_send_enable(struct bfa_pport_s *port); |
| static bfa_boolean_t bfa_pport_send_disable(struct bfa_pport_s *port); |
| static void bfa_pport_update_linkinfo(struct bfa_pport_s *pport); |
| static void bfa_pport_reset_linkinfo(struct bfa_pport_s *pport); |
| static void bfa_pport_set_wwns(struct bfa_pport_s *port); |
| static void __bfa_cb_port_event(void *cbarg, bfa_boolean_t complete); |
| static void __bfa_cb_port_stats(void *cbarg, bfa_boolean_t complete); |
| static void __bfa_cb_port_stats_clr(void *cbarg, bfa_boolean_t complete); |
| static void bfa_port_stats_timeout(void *cbarg); |
| static void bfa_port_stats_clr_timeout(void *cbarg); |
| static void bfa_pport_callback(struct bfa_pport_s *pport, |
| enum bfa_pport_linkstate event); |
| static void bfa_pport_queue_cb(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_linkstate event); |
| /** |
| * bfa_pport_private |
| */ |
| |
| /** |
| * BFA port state machine events |
| */ |
| enum bfa_pport_sm_event { |
| BFA_PPORT_SM_START = 1, /* start port state machine */ |
| BFA_PPORT_SM_STOP = 2, /* stop port state machine */ |
| BFA_PPORT_SM_ENABLE = 3, /* enable port */ |
| BFA_PPORT_SM_DISABLE = 4, /* disable port state machine */ |
| BFA_PPORT_SM_FWRSP = 5, /* firmware enable/disable rsp */ |
| BFA_PPORT_SM_LINKUP = 6, /* firmware linkup event */ |
| BFA_PPORT_SM_LINKDOWN = 7, /* firmware linkup down */ |
| BFA_PPORT_SM_QRESUME = 8, /* CQ space available */ |
| BFA_PPORT_SM_HWFAIL = 9, /* IOC h/w failure */ |
| }; |
| |
| /** |
| * BFA port link notification state machine events |
| */ |
| |
| enum bfa_pport_ln_sm_event { |
| BFA_PPORT_LN_SM_LINKUP = 1, /* linkup event */ |
| BFA_PPORT_LN_SM_LINKDOWN = 2, /* linkdown event */ |
| BFA_PPORT_LN_SM_NOTIFICATION = 3 /* done notification */ |
| }; |
| |
| static void bfa_pport_sm_uninit(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| static void bfa_pport_sm_enabling_qwait(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| static void bfa_pport_sm_enabling(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| static void bfa_pport_sm_linkdown(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| static void bfa_pport_sm_linkup(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| static void bfa_pport_sm_disabling(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| static void bfa_pport_sm_disabling_qwait(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| static void bfa_pport_sm_disabled(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| static void bfa_pport_sm_stopped(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| static void bfa_pport_sm_iocdown(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| static void bfa_pport_sm_iocfail(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event); |
| |
| static void bfa_pport_ln_sm_dn(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event); |
| static void bfa_pport_ln_sm_dn_nf(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event); |
| static void bfa_pport_ln_sm_dn_up_nf(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event); |
| static void bfa_pport_ln_sm_up(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event); |
| static void bfa_pport_ln_sm_up_nf(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event); |
| static void bfa_pport_ln_sm_up_dn_nf(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event); |
| static void bfa_pport_ln_sm_up_dn_up_nf(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event); |
| |
| static struct bfa_sm_table_s hal_pport_sm_table[] = { |
| {BFA_SM(bfa_pport_sm_uninit), BFA_PPORT_ST_UNINIT}, |
| {BFA_SM(bfa_pport_sm_enabling_qwait), BFA_PPORT_ST_ENABLING_QWAIT}, |
| {BFA_SM(bfa_pport_sm_enabling), BFA_PPORT_ST_ENABLING}, |
| {BFA_SM(bfa_pport_sm_linkdown), BFA_PPORT_ST_LINKDOWN}, |
| {BFA_SM(bfa_pport_sm_linkup), BFA_PPORT_ST_LINKUP}, |
| {BFA_SM(bfa_pport_sm_disabling_qwait), |
| BFA_PPORT_ST_DISABLING_QWAIT}, |
| {BFA_SM(bfa_pport_sm_disabling), BFA_PPORT_ST_DISABLING}, |
| {BFA_SM(bfa_pport_sm_disabled), BFA_PPORT_ST_DISABLED}, |
| {BFA_SM(bfa_pport_sm_stopped), BFA_PPORT_ST_STOPPED}, |
| {BFA_SM(bfa_pport_sm_iocdown), BFA_PPORT_ST_IOCDOWN}, |
| {BFA_SM(bfa_pport_sm_iocfail), BFA_PPORT_ST_IOCDOWN}, |
| }; |
| |
| static void |
| bfa_pport_aen_post(struct bfa_pport_s *pport, enum bfa_port_aen_event event) |
| { |
| union bfa_aen_data_u aen_data; |
| struct bfa_log_mod_s *logmod = pport->bfa->logm; |
| wwn_t pwwn = pport->pwwn; |
| char pwwn_ptr[BFA_STRING_32]; |
| struct bfa_ioc_attr_s ioc_attr; |
| |
| memset(&aen_data, 0, sizeof(aen_data)); |
| wwn2str(pwwn_ptr, pwwn); |
| switch (event) { |
| case BFA_PORT_AEN_ONLINE: |
| bfa_log(logmod, BFA_AEN_PORT_ONLINE, pwwn_ptr); |
| break; |
| case BFA_PORT_AEN_OFFLINE: |
| bfa_log(logmod, BFA_AEN_PORT_OFFLINE, pwwn_ptr); |
| break; |
| case BFA_PORT_AEN_ENABLE: |
| bfa_log(logmod, BFA_AEN_PORT_ENABLE, pwwn_ptr); |
| break; |
| case BFA_PORT_AEN_DISABLE: |
| bfa_log(logmod, BFA_AEN_PORT_DISABLE, pwwn_ptr); |
| break; |
| case BFA_PORT_AEN_DISCONNECT: |
| bfa_log(logmod, BFA_AEN_PORT_DISCONNECT, pwwn_ptr); |
| break; |
| case BFA_PORT_AEN_QOS_NEG: |
| bfa_log(logmod, BFA_AEN_PORT_QOS_NEG, pwwn_ptr); |
| break; |
| default: |
| break; |
| } |
| |
| bfa_ioc_get_attr(&pport->bfa->ioc, &ioc_attr); |
| aen_data.port.ioc_type = ioc_attr.ioc_type; |
| aen_data.port.pwwn = pwwn; |
| } |
| |
| static void |
| bfa_pport_sm_uninit(struct bfa_pport_s *pport, enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_START: |
| /** |
| * Start event after IOC is configured and BFA is started. |
| */ |
| if (bfa_pport_send_enable(pport)) |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling); |
| else |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling_qwait); |
| break; |
| |
| case BFA_PPORT_SM_ENABLE: |
| /** |
| * Port is persistently configured to be in enabled state. Do |
| * not change state. Port enabling is done when START event is |
| * received. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_DISABLE: |
| /** |
| * If a port is persistently configured to be disabled, the |
| * first event will a port disable request. |
| */ |
| bfa_sm_set_state(pport, bfa_pport_sm_disabled); |
| break; |
| |
| case BFA_PPORT_SM_HWFAIL: |
| bfa_sm_set_state(pport, bfa_pport_sm_iocdown); |
| break; |
| |
| default: |
| bfa_sm_fault(pport->bfa, event); |
| } |
| } |
| |
| static void |
| bfa_pport_sm_enabling_qwait(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_QRESUME: |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling); |
| bfa_pport_send_enable(pport); |
| break; |
| |
| case BFA_PPORT_SM_STOP: |
| bfa_reqq_wcancel(&pport->reqq_wait); |
| bfa_sm_set_state(pport, bfa_pport_sm_stopped); |
| break; |
| |
| case BFA_PPORT_SM_ENABLE: |
| /** |
| * Already enable is in progress. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_DISABLE: |
| /** |
| * Just send disable request to firmware when room becomes |
| * available in request queue. |
| */ |
| bfa_sm_set_state(pport, bfa_pport_sm_disabled); |
| bfa_reqq_wcancel(&pport->reqq_wait); |
| bfa_plog_str(pport->bfa->plog, BFA_PL_MID_HAL, |
| BFA_PL_EID_PORT_DISABLE, 0, "Port Disable"); |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_DISABLE); |
| break; |
| |
| case BFA_PPORT_SM_LINKUP: |
| case BFA_PPORT_SM_LINKDOWN: |
| /** |
| * Possible to get link events when doing back-to-back |
| * enable/disables. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_HWFAIL: |
| bfa_reqq_wcancel(&pport->reqq_wait); |
| bfa_sm_set_state(pport, bfa_pport_sm_iocdown); |
| break; |
| |
| default: |
| bfa_sm_fault(pport->bfa, event); |
| } |
| } |
| |
| static void |
| bfa_pport_sm_enabling(struct bfa_pport_s *pport, enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_FWRSP: |
| case BFA_PPORT_SM_LINKDOWN: |
| bfa_sm_set_state(pport, bfa_pport_sm_linkdown); |
| break; |
| |
| case BFA_PPORT_SM_LINKUP: |
| bfa_pport_update_linkinfo(pport); |
| bfa_sm_set_state(pport, bfa_pport_sm_linkup); |
| |
| bfa_assert(pport->event_cbfn); |
| bfa_pport_callback(pport, BFA_PPORT_LINKUP); |
| break; |
| |
| case BFA_PPORT_SM_ENABLE: |
| /** |
| * Already being enabled. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_DISABLE: |
| if (bfa_pport_send_disable(pport)) |
| bfa_sm_set_state(pport, bfa_pport_sm_disabling); |
| else |
| bfa_sm_set_state(pport, bfa_pport_sm_disabling_qwait); |
| |
| bfa_plog_str(pport->bfa->plog, BFA_PL_MID_HAL, |
| BFA_PL_EID_PORT_DISABLE, 0, "Port Disable"); |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_DISABLE); |
| break; |
| |
| case BFA_PPORT_SM_STOP: |
| bfa_sm_set_state(pport, bfa_pport_sm_stopped); |
| break; |
| |
| case BFA_PPORT_SM_HWFAIL: |
| bfa_sm_set_state(pport, bfa_pport_sm_iocdown); |
| break; |
| |
| default: |
| bfa_sm_fault(pport->bfa, event); |
| } |
| } |
| |
| static void |
| bfa_pport_sm_linkdown(struct bfa_pport_s *pport, enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_LINKUP: |
| bfa_pport_update_linkinfo(pport); |
| bfa_sm_set_state(pport, bfa_pport_sm_linkup); |
| bfa_assert(pport->event_cbfn); |
| bfa_plog_str(pport->bfa->plog, BFA_PL_MID_HAL, |
| BFA_PL_EID_PORT_ST_CHANGE, 0, "Port Linkup"); |
| bfa_pport_callback(pport, BFA_PPORT_LINKUP); |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_ONLINE); |
| /** |
| * If QoS is enabled and it is not online, |
| * Send a separate event. |
| */ |
| if ((pport->cfg.qos_enabled) |
| && (bfa_os_ntohl(pport->qos_attr.state) != BFA_QOS_ONLINE)) |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_QOS_NEG); |
| |
| break; |
| |
| case BFA_PPORT_SM_LINKDOWN: |
| /** |
| * Possible to get link down event. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_ENABLE: |
| /** |
| * Already enabled. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_DISABLE: |
| if (bfa_pport_send_disable(pport)) |
| bfa_sm_set_state(pport, bfa_pport_sm_disabling); |
| else |
| bfa_sm_set_state(pport, bfa_pport_sm_disabling_qwait); |
| |
| bfa_plog_str(pport->bfa->plog, BFA_PL_MID_HAL, |
| BFA_PL_EID_PORT_DISABLE, 0, "Port Disable"); |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_DISABLE); |
| break; |
| |
| case BFA_PPORT_SM_STOP: |
| bfa_sm_set_state(pport, bfa_pport_sm_stopped); |
| break; |
| |
| case BFA_PPORT_SM_HWFAIL: |
| bfa_sm_set_state(pport, bfa_pport_sm_iocdown); |
| break; |
| |
| default: |
| bfa_sm_fault(pport->bfa, event); |
| } |
| } |
| |
| static void |
| bfa_pport_sm_linkup(struct bfa_pport_s *pport, enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_ENABLE: |
| /** |
| * Already enabled. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_DISABLE: |
| if (bfa_pport_send_disable(pport)) |
| bfa_sm_set_state(pport, bfa_pport_sm_disabling); |
| else |
| bfa_sm_set_state(pport, bfa_pport_sm_disabling_qwait); |
| |
| bfa_pport_reset_linkinfo(pport); |
| bfa_pport_callback(pport, BFA_PPORT_LINKDOWN); |
| bfa_plog_str(pport->bfa->plog, BFA_PL_MID_HAL, |
| BFA_PL_EID_PORT_DISABLE, 0, "Port Disable"); |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_OFFLINE); |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_DISABLE); |
| break; |
| |
| case BFA_PPORT_SM_LINKDOWN: |
| bfa_sm_set_state(pport, bfa_pport_sm_linkdown); |
| bfa_pport_reset_linkinfo(pport); |
| bfa_pport_callback(pport, BFA_PPORT_LINKDOWN); |
| bfa_plog_str(pport->bfa->plog, BFA_PL_MID_HAL, |
| BFA_PL_EID_PORT_ST_CHANGE, 0, "Port Linkdown"); |
| if (BFA_PORT_IS_DISABLED(pport->bfa)) |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_OFFLINE); |
| else |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_DISCONNECT); |
| break; |
| |
| case BFA_PPORT_SM_STOP: |
| bfa_sm_set_state(pport, bfa_pport_sm_stopped); |
| bfa_pport_reset_linkinfo(pport); |
| if (BFA_PORT_IS_DISABLED(pport->bfa)) |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_OFFLINE); |
| else |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_DISCONNECT); |
| break; |
| |
| case BFA_PPORT_SM_HWFAIL: |
| bfa_sm_set_state(pport, bfa_pport_sm_iocdown); |
| bfa_pport_reset_linkinfo(pport); |
| bfa_pport_callback(pport, BFA_PPORT_LINKDOWN); |
| if (BFA_PORT_IS_DISABLED(pport->bfa)) |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_OFFLINE); |
| else |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_DISCONNECT); |
| break; |
| |
| default: |
| bfa_sm_fault(pport->bfa, event); |
| } |
| } |
| |
| static void |
| bfa_pport_sm_disabling_qwait(struct bfa_pport_s *pport, |
| enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_QRESUME: |
| bfa_sm_set_state(pport, bfa_pport_sm_disabling); |
| bfa_pport_send_disable(pport); |
| break; |
| |
| case BFA_PPORT_SM_STOP: |
| bfa_sm_set_state(pport, bfa_pport_sm_stopped); |
| bfa_reqq_wcancel(&pport->reqq_wait); |
| break; |
| |
| case BFA_PPORT_SM_DISABLE: |
| /** |
| * Already being disabled. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_LINKUP: |
| case BFA_PPORT_SM_LINKDOWN: |
| /** |
| * Possible to get link events when doing back-to-back |
| * enable/disables. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_HWFAIL: |
| bfa_sm_set_state(pport, bfa_pport_sm_iocfail); |
| bfa_reqq_wcancel(&pport->reqq_wait); |
| break; |
| |
| default: |
| bfa_sm_fault(pport->bfa, event); |
| } |
| } |
| |
| static void |
| bfa_pport_sm_disabling(struct bfa_pport_s *pport, enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_FWRSP: |
| bfa_sm_set_state(pport, bfa_pport_sm_disabled); |
| break; |
| |
| case BFA_PPORT_SM_DISABLE: |
| /** |
| * Already being disabled. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_ENABLE: |
| if (bfa_pport_send_enable(pport)) |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling); |
| else |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling_qwait); |
| |
| bfa_plog_str(pport->bfa->plog, BFA_PL_MID_HAL, |
| BFA_PL_EID_PORT_ENABLE, 0, "Port Enable"); |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_ENABLE); |
| break; |
| |
| case BFA_PPORT_SM_STOP: |
| bfa_sm_set_state(pport, bfa_pport_sm_stopped); |
| break; |
| |
| case BFA_PPORT_SM_LINKUP: |
| case BFA_PPORT_SM_LINKDOWN: |
| /** |
| * Possible to get link events when doing back-to-back |
| * enable/disables. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_HWFAIL: |
| bfa_sm_set_state(pport, bfa_pport_sm_iocfail); |
| break; |
| |
| default: |
| bfa_sm_fault(pport->bfa, event); |
| } |
| } |
| |
| static void |
| bfa_pport_sm_disabled(struct bfa_pport_s *pport, enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_START: |
| /** |
| * Ignore start event for a port that is disabled. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_STOP: |
| bfa_sm_set_state(pport, bfa_pport_sm_stopped); |
| break; |
| |
| case BFA_PPORT_SM_ENABLE: |
| if (bfa_pport_send_enable(pport)) |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling); |
| else |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling_qwait); |
| |
| bfa_plog_str(pport->bfa->plog, BFA_PL_MID_HAL, |
| BFA_PL_EID_PORT_ENABLE, 0, "Port Enable"); |
| bfa_pport_aen_post(pport, BFA_PORT_AEN_ENABLE); |
| break; |
| |
| case BFA_PPORT_SM_DISABLE: |
| /** |
| * Already disabled. |
| */ |
| break; |
| |
| case BFA_PPORT_SM_HWFAIL: |
| bfa_sm_set_state(pport, bfa_pport_sm_iocfail); |
| break; |
| |
| default: |
| bfa_sm_fault(pport->bfa, event); |
| } |
| } |
| |
| static void |
| bfa_pport_sm_stopped(struct bfa_pport_s *pport, enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_START: |
| if (bfa_pport_send_enable(pport)) |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling); |
| else |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling_qwait); |
| break; |
| |
| default: |
| /** |
| * Ignore all other events. |
| */ |
| ; |
| } |
| } |
| |
| /** |
| * Port is enabled. IOC is down/failed. |
| */ |
| static void |
| bfa_pport_sm_iocdown(struct bfa_pport_s *pport, enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_START: |
| if (bfa_pport_send_enable(pport)) |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling); |
| else |
| bfa_sm_set_state(pport, bfa_pport_sm_enabling_qwait); |
| break; |
| |
| default: |
| /** |
| * Ignore all events. |
| */ |
| ; |
| } |
| } |
| |
| /** |
| * Port is disabled. IOC is down/failed. |
| */ |
| static void |
| bfa_pport_sm_iocfail(struct bfa_pport_s *pport, enum bfa_pport_sm_event event) |
| { |
| bfa_trc(pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_SM_START: |
| bfa_sm_set_state(pport, bfa_pport_sm_disabled); |
| break; |
| |
| case BFA_PPORT_SM_ENABLE: |
| bfa_sm_set_state(pport, bfa_pport_sm_iocdown); |
| break; |
| |
| default: |
| /** |
| * Ignore all events. |
| */ |
| ; |
| } |
| } |
| |
| /** |
| * Link state is down |
| */ |
| static void |
| bfa_pport_ln_sm_dn(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event) |
| { |
| bfa_trc(ln->pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_LN_SM_LINKUP: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_up_nf); |
| bfa_pport_queue_cb(ln, BFA_PPORT_LINKUP); |
| break; |
| |
| default: |
| bfa_sm_fault(ln->pport->bfa, event); |
| } |
| } |
| |
| /** |
| * Link state is waiting for down notification |
| */ |
| static void |
| bfa_pport_ln_sm_dn_nf(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event) |
| { |
| bfa_trc(ln->pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_LN_SM_LINKUP: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_dn_up_nf); |
| break; |
| |
| case BFA_PPORT_LN_SM_NOTIFICATION: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_dn); |
| break; |
| |
| default: |
| bfa_sm_fault(ln->pport->bfa, event); |
| } |
| } |
| |
| /** |
| * Link state is waiting for down notification and there is a pending up |
| */ |
| static void |
| bfa_pport_ln_sm_dn_up_nf(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event) |
| { |
| bfa_trc(ln->pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_LN_SM_LINKDOWN: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_dn_nf); |
| break; |
| |
| case BFA_PPORT_LN_SM_NOTIFICATION: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_up_nf); |
| bfa_pport_queue_cb(ln, BFA_PPORT_LINKUP); |
| break; |
| |
| default: |
| bfa_sm_fault(ln->pport->bfa, event); |
| } |
| } |
| |
| /** |
| * Link state is up |
| */ |
| static void |
| bfa_pport_ln_sm_up(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event) |
| { |
| bfa_trc(ln->pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_LN_SM_LINKDOWN: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_dn_nf); |
| bfa_pport_queue_cb(ln, BFA_PPORT_LINKDOWN); |
| break; |
| |
| default: |
| bfa_sm_fault(ln->pport->bfa, event); |
| } |
| } |
| |
| /** |
| * Link state is waiting for up notification |
| */ |
| static void |
| bfa_pport_ln_sm_up_nf(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event) |
| { |
| bfa_trc(ln->pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_LN_SM_LINKDOWN: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_up_dn_nf); |
| break; |
| |
| case BFA_PPORT_LN_SM_NOTIFICATION: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_up); |
| break; |
| |
| default: |
| bfa_sm_fault(ln->pport->bfa, event); |
| } |
| } |
| |
| /** |
| * Link state is waiting for up notification and there is a pending down |
| */ |
| static void |
| bfa_pport_ln_sm_up_dn_nf(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event) |
| { |
| bfa_trc(ln->pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_LN_SM_LINKUP: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_up_dn_up_nf); |
| break; |
| |
| case BFA_PPORT_LN_SM_NOTIFICATION: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_dn_nf); |
| bfa_pport_queue_cb(ln, BFA_PPORT_LINKDOWN); |
| break; |
| |
| default: |
| bfa_sm_fault(ln->pport->bfa, event); |
| } |
| } |
| |
| /** |
| * Link state is waiting for up notification and there are pending down and up |
| */ |
| static void |
| bfa_pport_ln_sm_up_dn_up_nf(struct bfa_pport_ln_s *ln, |
| enum bfa_pport_ln_sm_event event) |
| { |
| bfa_trc(ln->pport->bfa, event); |
| |
| switch (event) { |
| case BFA_PPORT_LN_SM_LINKDOWN: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_up_dn_nf); |
| break; |
| |
| case BFA_PPORT_LN_SM_NOTIFICATION: |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_dn_up_nf); |
| bfa_pport_queue_cb(ln, BFA_PPORT_LINKDOWN); |
| break; |
| |
| default: |
| bfa_sm_fault(ln->pport->bfa, event); |
| } |
| } |
| |
| /** |
| * bfa_pport_private |
| */ |
| |
| static void |
| __bfa_cb_port_event(void *cbarg, bfa_boolean_t complete) |
| { |
| struct bfa_pport_ln_s *ln = cbarg; |
| |
| if (complete) |
| ln->pport->event_cbfn(ln->pport->event_cbarg, ln->ln_event); |
| else |
| bfa_sm_send_event(ln, BFA_PPORT_LN_SM_NOTIFICATION); |
| } |
| |
| #define PPORT_STATS_DMA_SZ (BFA_ROUNDUP(sizeof(union bfa_pport_stats_u), \ |
| BFA_CACHELINE_SZ)) |
| |
| static void |
| bfa_pport_meminfo(struct bfa_iocfc_cfg_s *cfg, u32 *ndm_len, |
| u32 *dm_len) |
| { |
| *dm_len += PPORT_STATS_DMA_SZ; |
| } |
| |
| static void |
| bfa_pport_qresume(void *cbarg) |
| { |
| struct bfa_pport_s *port = cbarg; |
| |
| bfa_sm_send_event(port, BFA_PPORT_SM_QRESUME); |
| } |
| |
| static void |
| bfa_pport_mem_claim(struct bfa_pport_s *pport, struct bfa_meminfo_s *meminfo) |
| { |
| u8 *dm_kva; |
| u64 dm_pa; |
| |
| dm_kva = bfa_meminfo_dma_virt(meminfo); |
| dm_pa = bfa_meminfo_dma_phys(meminfo); |
| |
| pport->stats_kva = dm_kva; |
| pport->stats_pa = dm_pa; |
| pport->stats = (union bfa_pport_stats_u *)dm_kva; |
| |
| dm_kva += PPORT_STATS_DMA_SZ; |
| dm_pa += PPORT_STATS_DMA_SZ; |
| |
| bfa_meminfo_dma_virt(meminfo) = dm_kva; |
| bfa_meminfo_dma_phys(meminfo) = dm_pa; |
| } |
| |
| /** |
| * Memory initialization. |
| */ |
| static void |
| bfa_pport_attach(struct bfa_s *bfa, void *bfad, struct bfa_iocfc_cfg_s *cfg, |
| struct bfa_meminfo_s *meminfo, struct bfa_pcidev_s *pcidev) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| struct bfa_pport_cfg_s *port_cfg = &pport->cfg; |
| struct bfa_pport_ln_s *ln = &pport->ln; |
| |
| bfa_os_memset(pport, 0, sizeof(struct bfa_pport_s)); |
| pport->bfa = bfa; |
| ln->pport = pport; |
| |
| bfa_pport_mem_claim(pport, meminfo); |
| |
| bfa_sm_set_state(pport, bfa_pport_sm_uninit); |
| bfa_sm_set_state(ln, bfa_pport_ln_sm_dn); |
| |
| /** |
| * initialize and set default configuration |
| */ |
| port_cfg->topology = BFA_PPORT_TOPOLOGY_P2P; |
| port_cfg->speed = BFA_PPORT_SPEED_AUTO; |
| port_cfg->trunked = BFA_FALSE; |
| port_cfg->maxfrsize = 0; |
| |
| port_cfg->trl_def_speed = BFA_PPORT_SPEED_1GBPS; |
| |
| bfa_reqq_winit(&pport->reqq_wait, bfa_pport_qresume, pport); |
| } |
| |
| static void |
| bfa_pport_initdone(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| /** |
| * Initialize port attributes from IOC hardware data. |
| */ |
| bfa_pport_set_wwns(pport); |
| if (pport->cfg.maxfrsize == 0) |
| pport->cfg.maxfrsize = bfa_ioc_maxfrsize(&bfa->ioc); |
| pport->cfg.rx_bbcredit = bfa_ioc_rx_bbcredit(&bfa->ioc); |
| pport->speed_sup = bfa_ioc_speed_sup(&bfa->ioc); |
| |
| bfa_assert(pport->cfg.maxfrsize); |
| bfa_assert(pport->cfg.rx_bbcredit); |
| bfa_assert(pport->speed_sup); |
| } |
| |
| static void |
| bfa_pport_detach(struct bfa_s *bfa) |
| { |
| } |
| |
| /** |
| * Called when IOC is ready. |
| */ |
| static void |
| bfa_pport_start(struct bfa_s *bfa) |
| { |
| bfa_sm_send_event(BFA_PORT_MOD(bfa), BFA_PPORT_SM_START); |
| } |
| |
| /** |
| * Called before IOC is stopped. |
| */ |
| static void |
| bfa_pport_stop(struct bfa_s *bfa) |
| { |
| bfa_sm_send_event(BFA_PORT_MOD(bfa), BFA_PPORT_SM_STOP); |
| } |
| |
| /** |
| * Called when IOC failure is detected. |
| */ |
| static void |
| bfa_pport_iocdisable(struct bfa_s *bfa) |
| { |
| bfa_sm_send_event(BFA_PORT_MOD(bfa), BFA_PPORT_SM_HWFAIL); |
| } |
| |
| static void |
| bfa_pport_update_linkinfo(struct bfa_pport_s *pport) |
| { |
| struct bfi_pport_event_s *pevent = pport->event_arg.i2hmsg.event; |
| |
| pport->speed = pevent->link_state.speed; |
| pport->topology = pevent->link_state.topology; |
| |
| if (pport->topology == BFA_PPORT_TOPOLOGY_LOOP) |
| pport->myalpa = pevent->link_state.tl.loop_info.myalpa; |
| |
| /* |
| * QoS Details |
| */ |
| bfa_os_assign(pport->qos_attr, pevent->link_state.qos_attr); |
| bfa_os_assign(pport->qos_vc_attr, pevent->link_state.qos_vc_attr); |
| |
| bfa_trc(pport->bfa, pport->speed); |
| bfa_trc(pport->bfa, pport->topology); |
| } |
| |
| static void |
| bfa_pport_reset_linkinfo(struct bfa_pport_s *pport) |
| { |
| pport->speed = BFA_PPORT_SPEED_UNKNOWN; |
| pport->topology = BFA_PPORT_TOPOLOGY_NONE; |
| } |
| |
| /** |
| * Send port enable message to firmware. |
| */ |
| static bfa_boolean_t |
| bfa_pport_send_enable(struct bfa_pport_s *port) |
| { |
| struct bfi_pport_enable_req_s *m; |
| |
| /** |
| * Increment message tag before queue check, so that responses to old |
| * requests are discarded. |
| */ |
| port->msgtag++; |
| |
| /** |
| * check for room in queue to send request now |
| */ |
| m = bfa_reqq_next(port->bfa, BFA_REQQ_PORT); |
| if (!m) { |
| bfa_reqq_wait(port->bfa, BFA_REQQ_PORT, &port->reqq_wait); |
| return BFA_FALSE; |
| } |
| |
| bfi_h2i_set(m->mh, BFI_MC_FC_PORT, BFI_PPORT_H2I_ENABLE_REQ, |
| bfa_lpuid(port->bfa)); |
| m->nwwn = port->nwwn; |
| m->pwwn = port->pwwn; |
| m->port_cfg = port->cfg; |
| m->msgtag = port->msgtag; |
| m->port_cfg.maxfrsize = bfa_os_htons(port->cfg.maxfrsize); |
| bfa_dma_be_addr_set(m->stats_dma_addr, port->stats_pa); |
| bfa_trc(port->bfa, m->stats_dma_addr.a32.addr_lo); |
| bfa_trc(port->bfa, m->stats_dma_addr.a32.addr_hi); |
| |
| /** |
| * queue I/O message to firmware |
| */ |
| bfa_reqq_produce(port->bfa, BFA_REQQ_PORT); |
| return BFA_TRUE; |
| } |
| |
| /** |
| * Send port disable message to firmware. |
| */ |
| static bfa_boolean_t |
| bfa_pport_send_disable(struct bfa_pport_s *port) |
| { |
| bfi_pport_disable_req_t *m; |
| |
| /** |
| * Increment message tag before queue check, so that responses to old |
| * requests are discarded. |
| */ |
| port->msgtag++; |
| |
| /** |
| * check for room in queue to send request now |
| */ |
| m = bfa_reqq_next(port->bfa, BFA_REQQ_PORT); |
| if (!m) { |
| bfa_reqq_wait(port->bfa, BFA_REQQ_PORT, &port->reqq_wait); |
| return BFA_FALSE; |
| } |
| |
| bfi_h2i_set(m->mh, BFI_MC_FC_PORT, BFI_PPORT_H2I_DISABLE_REQ, |
| bfa_lpuid(port->bfa)); |
| m->msgtag = port->msgtag; |
| |
| /** |
| * queue I/O message to firmware |
| */ |
| bfa_reqq_produce(port->bfa, BFA_REQQ_PORT); |
| |
| return BFA_TRUE; |
| } |
| |
| static void |
| bfa_pport_set_wwns(struct bfa_pport_s *port) |
| { |
| port->pwwn = bfa_ioc_get_pwwn(&port->bfa->ioc); |
| port->nwwn = bfa_ioc_get_nwwn(&port->bfa->ioc); |
| |
| bfa_trc(port->bfa, port->pwwn); |
| bfa_trc(port->bfa, port->nwwn); |
| } |
| |
| static void |
| bfa_port_send_txcredit(void *port_cbarg) |
| { |
| |
| struct bfa_pport_s *port = port_cbarg; |
| struct bfi_pport_set_svc_params_req_s *m; |
| |
| /** |
| * check for room in queue to send request now |
| */ |
| m = bfa_reqq_next(port->bfa, BFA_REQQ_PORT); |
| if (!m) { |
| bfa_trc(port->bfa, port->cfg.tx_bbcredit); |
| return; |
| } |
| |
| bfi_h2i_set(m->mh, BFI_MC_FC_PORT, BFI_PPORT_H2I_SET_SVC_PARAMS_REQ, |
| bfa_lpuid(port->bfa)); |
| m->tx_bbcredit = bfa_os_htons((u16) port->cfg.tx_bbcredit); |
| |
| /** |
| * queue I/O message to firmware |
| */ |
| bfa_reqq_produce(port->bfa, BFA_REQQ_PORT); |
| } |
| |
| |
| |
| /** |
| * bfa_pport_public |
| */ |
| |
| /** |
| * Firmware message handler. |
| */ |
| void |
| bfa_pport_isr(struct bfa_s *bfa, struct bfi_msg_s *msg) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| union bfi_pport_i2h_msg_u i2hmsg; |
| |
| i2hmsg.msg = msg; |
| pport->event_arg.i2hmsg = i2hmsg; |
| |
| switch (msg->mhdr.msg_id) { |
| case BFI_PPORT_I2H_ENABLE_RSP: |
| if (pport->msgtag == i2hmsg.enable_rsp->msgtag) |
| bfa_sm_send_event(pport, BFA_PPORT_SM_FWRSP); |
| break; |
| |
| case BFI_PPORT_I2H_DISABLE_RSP: |
| if (pport->msgtag == i2hmsg.enable_rsp->msgtag) |
| bfa_sm_send_event(pport, BFA_PPORT_SM_FWRSP); |
| break; |
| |
| case BFI_PPORT_I2H_EVENT: |
| switch (i2hmsg.event->link_state.linkstate) { |
| case BFA_PPORT_LINKUP: |
| bfa_sm_send_event(pport, BFA_PPORT_SM_LINKUP); |
| break; |
| case BFA_PPORT_LINKDOWN: |
| bfa_sm_send_event(pport, BFA_PPORT_SM_LINKDOWN); |
| break; |
| case BFA_PPORT_TRUNK_LINKDOWN: |
| /** todo: event notification */ |
| break; |
| } |
| break; |
| |
| case BFI_PPORT_I2H_GET_STATS_RSP: |
| case BFI_PPORT_I2H_GET_QOS_STATS_RSP: |
| /* |
| * check for timer pop before processing the rsp |
| */ |
| if (pport->stats_busy == BFA_FALSE |
| || pport->stats_status == BFA_STATUS_ETIMER) |
| break; |
| |
| bfa_timer_stop(&pport->timer); |
| pport->stats_status = i2hmsg.getstats_rsp->status; |
| bfa_cb_queue(pport->bfa, &pport->hcb_qe, __bfa_cb_port_stats, |
| pport); |
| break; |
| case BFI_PPORT_I2H_CLEAR_STATS_RSP: |
| case BFI_PPORT_I2H_CLEAR_QOS_STATS_RSP: |
| /* |
| * check for timer pop before processing the rsp |
| */ |
| if (pport->stats_busy == BFA_FALSE |
| || pport->stats_status == BFA_STATUS_ETIMER) |
| break; |
| |
| bfa_timer_stop(&pport->timer); |
| pport->stats_status = BFA_STATUS_OK; |
| bfa_cb_queue(pport->bfa, &pport->hcb_qe, |
| __bfa_cb_port_stats_clr, pport); |
| break; |
| |
| default: |
| bfa_assert(0); |
| } |
| } |
| |
| |
| |
| /** |
| * bfa_pport_api |
| */ |
| |
| /** |
| * Registered callback for port events. |
| */ |
| void |
| bfa_pport_event_register(struct bfa_s *bfa, |
| void (*cbfn) (void *cbarg, bfa_pport_event_t event), |
| void *cbarg) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| pport->event_cbfn = cbfn; |
| pport->event_cbarg = cbarg; |
| } |
| |
| bfa_status_t |
| bfa_pport_enable(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| if (pport->diag_busy) |
| return BFA_STATUS_DIAG_BUSY; |
| else if (bfa_sm_cmp_state |
| (BFA_PORT_MOD(bfa), bfa_pport_sm_disabling_qwait)) |
| return BFA_STATUS_DEVBUSY; |
| |
| bfa_sm_send_event(BFA_PORT_MOD(bfa), BFA_PPORT_SM_ENABLE); |
| return BFA_STATUS_OK; |
| } |
| |
| bfa_status_t |
| bfa_pport_disable(struct bfa_s *bfa) |
| { |
| bfa_sm_send_event(BFA_PORT_MOD(bfa), BFA_PPORT_SM_DISABLE); |
| return BFA_STATUS_OK; |
| } |
| |
| /** |
| * Configure port speed. |
| */ |
| bfa_status_t |
| bfa_pport_cfg_speed(struct bfa_s *bfa, enum bfa_pport_speed speed) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, speed); |
| |
| if ((speed != BFA_PPORT_SPEED_AUTO) && (speed > pport->speed_sup)) { |
| bfa_trc(bfa, pport->speed_sup); |
| return BFA_STATUS_UNSUPP_SPEED; |
| } |
| |
| pport->cfg.speed = speed; |
| |
| return BFA_STATUS_OK; |
| } |
| |
| /** |
| * Get current speed. |
| */ |
| enum bfa_pport_speed |
| bfa_pport_get_speed(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| return port->speed; |
| } |
| |
| /** |
| * Configure port topology. |
| */ |
| bfa_status_t |
| bfa_pport_cfg_topology(struct bfa_s *bfa, enum bfa_pport_topology topology) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, topology); |
| bfa_trc(bfa, pport->cfg.topology); |
| |
| switch (topology) { |
| case BFA_PPORT_TOPOLOGY_P2P: |
| case BFA_PPORT_TOPOLOGY_LOOP: |
| case BFA_PPORT_TOPOLOGY_AUTO: |
| break; |
| |
| default: |
| return BFA_STATUS_EINVAL; |
| } |
| |
| pport->cfg.topology = topology; |
| return BFA_STATUS_OK; |
| } |
| |
| /** |
| * Get current topology. |
| */ |
| enum bfa_pport_topology |
| bfa_pport_get_topology(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| return port->topology; |
| } |
| |
| bfa_status_t |
| bfa_pport_cfg_hardalpa(struct bfa_s *bfa, u8 alpa) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, alpa); |
| bfa_trc(bfa, pport->cfg.cfg_hardalpa); |
| bfa_trc(bfa, pport->cfg.hardalpa); |
| |
| pport->cfg.cfg_hardalpa = BFA_TRUE; |
| pport->cfg.hardalpa = alpa; |
| |
| return BFA_STATUS_OK; |
| } |
| |
| bfa_status_t |
| bfa_pport_clr_hardalpa(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, pport->cfg.cfg_hardalpa); |
| bfa_trc(bfa, pport->cfg.hardalpa); |
| |
| pport->cfg.cfg_hardalpa = BFA_FALSE; |
| return BFA_STATUS_OK; |
| } |
| |
| bfa_boolean_t |
| bfa_pport_get_hardalpa(struct bfa_s *bfa, u8 *alpa) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| *alpa = port->cfg.hardalpa; |
| return port->cfg.cfg_hardalpa; |
| } |
| |
| u8 |
| bfa_pport_get_myalpa(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| return port->myalpa; |
| } |
| |
| bfa_status_t |
| bfa_pport_cfg_maxfrsize(struct bfa_s *bfa, u16 maxfrsize) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, maxfrsize); |
| bfa_trc(bfa, pport->cfg.maxfrsize); |
| |
| /* |
| * with in range |
| */ |
| if ((maxfrsize > FC_MAX_PDUSZ) || (maxfrsize < FC_MIN_PDUSZ)) |
| return BFA_STATUS_INVLD_DFSZ; |
| |
| /* |
| * power of 2, if not the max frame size of 2112 |
| */ |
| if ((maxfrsize != FC_MAX_PDUSZ) && (maxfrsize & (maxfrsize - 1))) |
| return BFA_STATUS_INVLD_DFSZ; |
| |
| pport->cfg.maxfrsize = maxfrsize; |
| return BFA_STATUS_OK; |
| } |
| |
| u16 |
| bfa_pport_get_maxfrsize(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| return port->cfg.maxfrsize; |
| } |
| |
| u32 |
| bfa_pport_mypid(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| return port->mypid; |
| } |
| |
| u8 |
| bfa_pport_get_rx_bbcredit(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| return port->cfg.rx_bbcredit; |
| } |
| |
| void |
| bfa_pport_set_tx_bbcredit(struct bfa_s *bfa, u16 tx_bbcredit) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| port->cfg.tx_bbcredit = (u8) tx_bbcredit; |
| bfa_port_send_txcredit(port); |
| } |
| |
| /** |
| * Get port attributes. |
| */ |
| |
| wwn_t |
| bfa_pport_get_wwn(struct bfa_s *bfa, bfa_boolean_t node) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| if (node) |
| return pport->nwwn; |
| else |
| return pport->pwwn; |
| } |
| |
| void |
| bfa_pport_get_attr(struct bfa_s *bfa, struct bfa_pport_attr_s *attr) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_os_memset(attr, 0, sizeof(struct bfa_pport_attr_s)); |
| |
| attr->nwwn = pport->nwwn; |
| attr->pwwn = pport->pwwn; |
| |
| bfa_os_memcpy(&attr->pport_cfg, &pport->cfg, |
| sizeof(struct bfa_pport_cfg_s)); |
| /* |
| * speed attributes |
| */ |
| attr->pport_cfg.speed = pport->cfg.speed; |
| attr->speed_supported = pport->speed_sup; |
| attr->speed = pport->speed; |
| attr->cos_supported = FC_CLASS_3; |
| |
| /* |
| * topology attributes |
| */ |
| attr->pport_cfg.topology = pport->cfg.topology; |
| attr->topology = pport->topology; |
| |
| /* |
| * beacon attributes |
| */ |
| attr->beacon = pport->beacon; |
| attr->link_e2e_beacon = pport->link_e2e_beacon; |
| attr->plog_enabled = bfa_plog_get_setting(pport->bfa->plog); |
| |
| attr->pport_cfg.path_tov = bfa_fcpim_path_tov_get(bfa); |
| attr->pport_cfg.q_depth = bfa_fcpim_qdepth_get(bfa); |
| attr->port_state = bfa_sm_to_state(hal_pport_sm_table, pport->sm); |
| if (bfa_ioc_is_disabled(&pport->bfa->ioc)) |
| attr->port_state = BFA_PPORT_ST_IOCDIS; |
| else if (bfa_ioc_fw_mismatch(&pport->bfa->ioc)) |
| attr->port_state = BFA_PPORT_ST_FWMISMATCH; |
| } |
| |
| static void |
| bfa_port_stats_query(void *cbarg) |
| { |
| struct bfa_pport_s *port = (struct bfa_pport_s *)cbarg; |
| bfi_pport_get_stats_req_t *msg; |
| |
| msg = bfa_reqq_next(port->bfa, BFA_REQQ_PORT); |
| |
| if (!msg) { |
| port->stats_qfull = BFA_TRUE; |
| bfa_reqq_winit(&port->stats_reqq_wait, bfa_port_stats_query, |
| port); |
| bfa_reqq_wait(port->bfa, BFA_REQQ_PORT, &port->stats_reqq_wait); |
| return; |
| } |
| port->stats_qfull = BFA_FALSE; |
| |
| bfa_os_memset(msg, 0, sizeof(bfi_pport_get_stats_req_t)); |
| bfi_h2i_set(msg->mh, BFI_MC_FC_PORT, BFI_PPORT_H2I_GET_STATS_REQ, |
| bfa_lpuid(port->bfa)); |
| bfa_reqq_produce(port->bfa, BFA_REQQ_PORT); |
| |
| return; |
| } |
| |
| static void |
| bfa_port_stats_clear(void *cbarg) |
| { |
| struct bfa_pport_s *port = (struct bfa_pport_s *)cbarg; |
| bfi_pport_clear_stats_req_t *msg; |
| |
| msg = bfa_reqq_next(port->bfa, BFA_REQQ_PORT); |
| |
| if (!msg) { |
| port->stats_qfull = BFA_TRUE; |
| bfa_reqq_winit(&port->stats_reqq_wait, bfa_port_stats_clear, |
| port); |
| bfa_reqq_wait(port->bfa, BFA_REQQ_PORT, &port->stats_reqq_wait); |
| return; |
| } |
| port->stats_qfull = BFA_FALSE; |
| |
| bfa_os_memset(msg, 0, sizeof(bfi_pport_clear_stats_req_t)); |
| bfi_h2i_set(msg->mh, BFI_MC_FC_PORT, BFI_PPORT_H2I_CLEAR_STATS_REQ, |
| bfa_lpuid(port->bfa)); |
| bfa_reqq_produce(port->bfa, BFA_REQQ_PORT); |
| return; |
| } |
| |
| static void |
| bfa_port_qos_stats_clear(void *cbarg) |
| { |
| struct bfa_pport_s *port = (struct bfa_pport_s *)cbarg; |
| bfi_pport_clear_qos_stats_req_t *msg; |
| |
| msg = bfa_reqq_next(port->bfa, BFA_REQQ_PORT); |
| |
| if (!msg) { |
| port->stats_qfull = BFA_TRUE; |
| bfa_reqq_winit(&port->stats_reqq_wait, bfa_port_qos_stats_clear, |
| port); |
| bfa_reqq_wait(port->bfa, BFA_REQQ_PORT, &port->stats_reqq_wait); |
| return; |
| } |
| port->stats_qfull = BFA_FALSE; |
| |
| bfa_os_memset(msg, 0, sizeof(bfi_pport_clear_qos_stats_req_t)); |
| bfi_h2i_set(msg->mh, BFI_MC_FC_PORT, BFI_PPORT_H2I_CLEAR_QOS_STATS_REQ, |
| bfa_lpuid(port->bfa)); |
| bfa_reqq_produce(port->bfa, BFA_REQQ_PORT); |
| return; |
| } |
| |
| static void |
| bfa_pport_stats_swap(union bfa_pport_stats_u *d, union bfa_pport_stats_u *s) |
| { |
| u32 *dip = (u32 *) d; |
| u32 *sip = (u32 *) s; |
| int i; |
| |
| /* |
| * Do 64 bit fields swap first |
| */ |
| for (i = 0; |
| i < |
| ((sizeof(union bfa_pport_stats_u) - |
| sizeof(struct bfa_qos_stats_s)) / sizeof(u32)); i = i + 2) { |
| #ifdef __BIGENDIAN |
| dip[i] = bfa_os_ntohl(sip[i]); |
| dip[i + 1] = bfa_os_ntohl(sip[i + 1]); |
| #else |
| dip[i] = bfa_os_ntohl(sip[i + 1]); |
| dip[i + 1] = bfa_os_ntohl(sip[i]); |
| #endif |
| } |
| |
| /* |
| * Now swap the 32 bit fields |
| */ |
| for (; i < (sizeof(union bfa_pport_stats_u) / sizeof(u32)); ++i) |
| dip[i] = bfa_os_ntohl(sip[i]); |
| } |
| |
| static void |
| __bfa_cb_port_stats_clr(void *cbarg, bfa_boolean_t complete) |
| { |
| struct bfa_pport_s *port = cbarg; |
| |
| if (complete) { |
| port->stats_cbfn(port->stats_cbarg, port->stats_status); |
| } else { |
| port->stats_busy = BFA_FALSE; |
| port->stats_status = BFA_STATUS_OK; |
| } |
| } |
| |
| static void |
| bfa_port_stats_clr_timeout(void *cbarg) |
| { |
| struct bfa_pport_s *port = (struct bfa_pport_s *)cbarg; |
| |
| bfa_trc(port->bfa, port->stats_qfull); |
| |
| if (port->stats_qfull) { |
| bfa_reqq_wcancel(&port->stats_reqq_wait); |
| port->stats_qfull = BFA_FALSE; |
| } |
| |
| port->stats_status = BFA_STATUS_ETIMER; |
| bfa_cb_queue(port->bfa, &port->hcb_qe, __bfa_cb_port_stats_clr, port); |
| } |
| |
| static void |
| bfa_pport_callback(struct bfa_pport_s *pport, enum bfa_pport_linkstate event) |
| { |
| if (pport->bfa->fcs) { |
| pport->event_cbfn(pport->event_cbarg, event); |
| return; |
| } |
| |
| switch (event) { |
| case BFA_PPORT_LINKUP: |
| bfa_sm_send_event(&pport->ln, BFA_PPORT_LN_SM_LINKUP); |
| break; |
| case BFA_PPORT_LINKDOWN: |
| bfa_sm_send_event(&pport->ln, BFA_PPORT_LN_SM_LINKDOWN); |
| break; |
| default: |
| bfa_assert(0); |
| } |
| } |
| |
| static void |
| bfa_pport_queue_cb(struct bfa_pport_ln_s *ln, enum bfa_pport_linkstate event) |
| { |
| ln->ln_event = event; |
| bfa_cb_queue(ln->pport->bfa, &ln->ln_qe, __bfa_cb_port_event, ln); |
| } |
| |
| static void |
| __bfa_cb_port_stats(void *cbarg, bfa_boolean_t complete) |
| { |
| struct bfa_pport_s *port = cbarg; |
| |
| if (complete) { |
| if (port->stats_status == BFA_STATUS_OK) |
| bfa_pport_stats_swap(port->stats_ret, port->stats); |
| port->stats_cbfn(port->stats_cbarg, port->stats_status); |
| } else { |
| port->stats_busy = BFA_FALSE; |
| port->stats_status = BFA_STATUS_OK; |
| } |
| } |
| |
| static void |
| bfa_port_stats_timeout(void *cbarg) |
| { |
| struct bfa_pport_s *port = (struct bfa_pport_s *)cbarg; |
| |
| bfa_trc(port->bfa, port->stats_qfull); |
| |
| if (port->stats_qfull) { |
| bfa_reqq_wcancel(&port->stats_reqq_wait); |
| port->stats_qfull = BFA_FALSE; |
| } |
| |
| port->stats_status = BFA_STATUS_ETIMER; |
| bfa_cb_queue(port->bfa, &port->hcb_qe, __bfa_cb_port_stats, port); |
| } |
| |
| #define BFA_PORT_STATS_TOV 1000 |
| |
| /** |
| * Fetch port attributes. |
| */ |
| bfa_status_t |
| bfa_pport_get_stats(struct bfa_s *bfa, union bfa_pport_stats_u *stats, |
| bfa_cb_pport_t cbfn, void *cbarg) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| if (port->stats_busy) { |
| bfa_trc(bfa, port->stats_busy); |
| return BFA_STATUS_DEVBUSY; |
| } |
| |
| port->stats_busy = BFA_TRUE; |
| port->stats_ret = stats; |
| port->stats_cbfn = cbfn; |
| port->stats_cbarg = cbarg; |
| |
| bfa_port_stats_query(port); |
| |
| bfa_timer_start(bfa, &port->timer, bfa_port_stats_timeout, port, |
| BFA_PORT_STATS_TOV); |
| return BFA_STATUS_OK; |
| } |
| |
| bfa_status_t |
| bfa_pport_clear_stats(struct bfa_s *bfa, bfa_cb_pport_t cbfn, void *cbarg) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| if (port->stats_busy) { |
| bfa_trc(bfa, port->stats_busy); |
| return BFA_STATUS_DEVBUSY; |
| } |
| |
| port->stats_busy = BFA_TRUE; |
| port->stats_cbfn = cbfn; |
| port->stats_cbarg = cbarg; |
| |
| bfa_port_stats_clear(port); |
| |
| bfa_timer_start(bfa, &port->timer, bfa_port_stats_clr_timeout, port, |
| BFA_PORT_STATS_TOV); |
| return BFA_STATUS_OK; |
| } |
| |
| bfa_status_t |
| bfa_pport_trunk_enable(struct bfa_s *bfa, u8 bitmap) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, bitmap); |
| bfa_trc(bfa, pport->cfg.trunked); |
| bfa_trc(bfa, pport->cfg.trunk_ports); |
| |
| if (!bitmap || (bitmap & (bitmap - 1))) |
| return BFA_STATUS_EINVAL; |
| |
| pport->cfg.trunked = BFA_TRUE; |
| pport->cfg.trunk_ports = bitmap; |
| |
| return BFA_STATUS_OK; |
| } |
| |
| void |
| bfa_pport_qos_get_attr(struct bfa_s *bfa, struct bfa_qos_attr_s *qos_attr) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| qos_attr->state = bfa_os_ntohl(pport->qos_attr.state); |
| qos_attr->total_bb_cr = bfa_os_ntohl(pport->qos_attr.total_bb_cr); |
| } |
| |
| void |
| bfa_pport_qos_get_vc_attr(struct bfa_s *bfa, |
| struct bfa_qos_vc_attr_s *qos_vc_attr) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| struct bfa_qos_vc_attr_s *bfa_vc_attr = &pport->qos_vc_attr; |
| u32 i = 0; |
| |
| qos_vc_attr->total_vc_count = bfa_os_ntohs(bfa_vc_attr->total_vc_count); |
| qos_vc_attr->shared_credit = bfa_os_ntohs(bfa_vc_attr->shared_credit); |
| qos_vc_attr->elp_opmode_flags = |
| bfa_os_ntohl(bfa_vc_attr->elp_opmode_flags); |
| |
| /* |
| * Individual VC info |
| */ |
| while (i < qos_vc_attr->total_vc_count) { |
| qos_vc_attr->vc_info[i].vc_credit = |
| bfa_vc_attr->vc_info[i].vc_credit; |
| qos_vc_attr->vc_info[i].borrow_credit = |
| bfa_vc_attr->vc_info[i].borrow_credit; |
| qos_vc_attr->vc_info[i].priority = |
| bfa_vc_attr->vc_info[i].priority; |
| ++i; |
| } |
| } |
| |
| /** |
| * Fetch QoS Stats. |
| */ |
| bfa_status_t |
| bfa_pport_get_qos_stats(struct bfa_s *bfa, union bfa_pport_stats_u *stats, |
| bfa_cb_pport_t cbfn, void *cbarg) |
| { |
| /* |
| * QoS stats is embedded in port stats |
| */ |
| return bfa_pport_get_stats(bfa, stats, cbfn, cbarg); |
| } |
| |
| bfa_status_t |
| bfa_pport_clear_qos_stats(struct bfa_s *bfa, bfa_cb_pport_t cbfn, void *cbarg) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| if (port->stats_busy) { |
| bfa_trc(bfa, port->stats_busy); |
| return BFA_STATUS_DEVBUSY; |
| } |
| |
| port->stats_busy = BFA_TRUE; |
| port->stats_cbfn = cbfn; |
| port->stats_cbarg = cbarg; |
| |
| bfa_port_qos_stats_clear(port); |
| |
| bfa_timer_start(bfa, &port->timer, bfa_port_stats_clr_timeout, port, |
| BFA_PORT_STATS_TOV); |
| return BFA_STATUS_OK; |
| } |
| |
| /** |
| * Fetch port attributes. |
| */ |
| bfa_status_t |
| bfa_pport_trunk_disable(struct bfa_s *bfa) |
| { |
| return BFA_STATUS_OK; |
| } |
| |
| bfa_boolean_t |
| bfa_pport_trunk_query(struct bfa_s *bfa, u32 *bitmap) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| *bitmap = port->cfg.trunk_ports; |
| return port->cfg.trunked; |
| } |
| |
| bfa_boolean_t |
| bfa_pport_is_disabled(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *port = BFA_PORT_MOD(bfa); |
| |
| return bfa_sm_to_state(hal_pport_sm_table, port->sm) == |
| BFA_PPORT_ST_DISABLED; |
| |
| } |
| |
| bfa_boolean_t |
| bfa_pport_is_ratelim(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| return pport->cfg.ratelimit ? BFA_TRUE : BFA_FALSE; |
| |
| } |
| |
| void |
| bfa_pport_cfg_qos(struct bfa_s *bfa, bfa_boolean_t on_off) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, on_off); |
| bfa_trc(bfa, pport->cfg.qos_enabled); |
| |
| pport->cfg.qos_enabled = on_off; |
| } |
| |
| void |
| bfa_pport_cfg_ratelim(struct bfa_s *bfa, bfa_boolean_t on_off) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, on_off); |
| bfa_trc(bfa, pport->cfg.ratelimit); |
| |
| pport->cfg.ratelimit = on_off; |
| if (pport->cfg.trl_def_speed == BFA_PPORT_SPEED_UNKNOWN) |
| pport->cfg.trl_def_speed = BFA_PPORT_SPEED_1GBPS; |
| } |
| |
| /** |
| * Configure default minimum ratelim speed |
| */ |
| bfa_status_t |
| bfa_pport_cfg_ratelim_speed(struct bfa_s *bfa, enum bfa_pport_speed speed) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, speed); |
| |
| /* |
| * Auto and speeds greater than the supported speed, are invalid |
| */ |
| if ((speed == BFA_PPORT_SPEED_AUTO) || (speed > pport->speed_sup)) { |
| bfa_trc(bfa, pport->speed_sup); |
| return BFA_STATUS_UNSUPP_SPEED; |
| } |
| |
| pport->cfg.trl_def_speed = speed; |
| |
| return BFA_STATUS_OK; |
| } |
| |
| /** |
| * Get default minimum ratelim speed |
| */ |
| enum bfa_pport_speed |
| bfa_pport_get_ratelim_speed(struct bfa_s *bfa) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, pport->cfg.trl_def_speed); |
| return pport->cfg.trl_def_speed; |
| |
| } |
| |
| void |
| bfa_pport_busy(struct bfa_s *bfa, bfa_boolean_t status) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, status); |
| bfa_trc(bfa, pport->diag_busy); |
| |
| pport->diag_busy = status; |
| } |
| |
| void |
| bfa_pport_beacon(struct bfa_s *bfa, bfa_boolean_t beacon, |
| bfa_boolean_t link_e2e_beacon) |
| { |
| struct bfa_pport_s *pport = BFA_PORT_MOD(bfa); |
| |
| bfa_trc(bfa, beacon); |
| bfa_trc(bfa, link_e2e_beacon); |
| bfa_trc(bfa, pport->beacon); |
| bfa_trc(bfa, pport->link_e2e_beacon); |
| |
| pport->beacon = beacon; |
| pport->link_e2e_beacon = link_e2e_beacon; |
| } |
| |
| bfa_boolean_t |
| bfa_pport_is_linkup(struct bfa_s *bfa) |
| { |
| return bfa_sm_cmp_state(BFA_PORT_MOD(bfa), bfa_pport_sm_linkup); |
| } |
| |
| |