| /* |
| * Copyright (c) 2005-2009 Brocade Communications Systems, Inc. |
| * All rights reserved |
| * www.brocade.com |
| * |
| * Linux driver for Brocade Fibre Channel Host Bus Adapter. |
| * |
| * This program is free software; you can redistribute it and/or modify it |
| * under the terms of the GNU General Public License (GPL) Version 2 as |
| * published by the Free Software Foundation |
| * |
| * This program is distributed in the hope that it will be useful, but |
| * WITHOUT ANY WARRANTY; without even the implied warranty of |
| * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the GNU |
| * General Public License for more details. |
| */ |
| |
| #include <bfa.h> |
| #include <cs/bfa_debug.h> |
| #include <bfa_cb_ioim_macros.h> |
| |
| BFA_TRC_FILE(HAL, IOIM); |
| |
| /* |
| * forward declarations. |
| */ |
| static bfa_boolean_t bfa_ioim_send_ioreq(struct bfa_ioim_s *ioim); |
| static bfa_boolean_t bfa_ioim_sge_setup(struct bfa_ioim_s *ioim); |
| static void bfa_ioim_sgpg_setup(struct bfa_ioim_s *ioim); |
| static bfa_boolean_t bfa_ioim_send_abort(struct bfa_ioim_s *ioim); |
| static void bfa_ioim_notify_cleanup(struct bfa_ioim_s *ioim); |
| static void __bfa_cb_ioim_good_comp(void *cbarg, bfa_boolean_t complete); |
| static void __bfa_cb_ioim_comp(void *cbarg, bfa_boolean_t complete); |
| static void __bfa_cb_ioim_abort(void *cbarg, bfa_boolean_t complete); |
| static void __bfa_cb_ioim_failed(void *cbarg, bfa_boolean_t complete); |
| static void __bfa_cb_ioim_pathtov(void *cbarg, bfa_boolean_t complete); |
| |
| /** |
| * bfa_ioim_sm |
| */ |
| |
| /** |
| * IO state machine events |
| */ |
| enum bfa_ioim_event { |
| BFA_IOIM_SM_START = 1, /* io start request from host */ |
| BFA_IOIM_SM_COMP_GOOD = 2, /* io good comp, resource free */ |
| BFA_IOIM_SM_COMP = 3, /* io comp, resource is free */ |
| BFA_IOIM_SM_COMP_UTAG = 4, /* io comp, resource is free */ |
| BFA_IOIM_SM_DONE = 5, /* io comp, resource not free */ |
| BFA_IOIM_SM_FREE = 6, /* io resource is freed */ |
| BFA_IOIM_SM_ABORT = 7, /* abort request from scsi stack */ |
| BFA_IOIM_SM_ABORT_COMP = 8, /* abort from f/w */ |
| BFA_IOIM_SM_ABORT_DONE = 9, /* abort completion from f/w */ |
| BFA_IOIM_SM_QRESUME = 10, /* CQ space available to queue IO */ |
| BFA_IOIM_SM_SGALLOCED = 11, /* SG page allocation successful */ |
| BFA_IOIM_SM_SQRETRY = 12, /* sequence recovery retry */ |
| BFA_IOIM_SM_HCB = 13, /* bfa callback complete */ |
| BFA_IOIM_SM_CLEANUP = 14, /* IO cleanup from itnim */ |
| BFA_IOIM_SM_TMSTART = 15, /* IO cleanup from tskim */ |
| BFA_IOIM_SM_TMDONE = 16, /* IO cleanup from tskim */ |
| BFA_IOIM_SM_HWFAIL = 17, /* IOC h/w failure event */ |
| BFA_IOIM_SM_IOTOV = 18, /* ITN offline TOV */ |
| }; |
| |
| /* |
| * forward declaration of IO state machine |
| */ |
| static void bfa_ioim_sm_uninit(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| static void bfa_ioim_sm_sgalloc(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| static void bfa_ioim_sm_active(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| static void bfa_ioim_sm_abort(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| static void bfa_ioim_sm_cleanup(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| static void bfa_ioim_sm_qfull(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| static void bfa_ioim_sm_abort_qfull(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| static void bfa_ioim_sm_cleanup_qfull(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| static void bfa_ioim_sm_hcb(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| static void bfa_ioim_sm_hcb_free(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| static void bfa_ioim_sm_resfree(struct bfa_ioim_s *ioim, |
| enum bfa_ioim_event event); |
| |
| /** |
| * IO is not started (unallocated). |
| */ |
| static void |
| bfa_ioim_sm_uninit(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc_fp(ioim->bfa, ioim->iotag); |
| bfa_trc_fp(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_START: |
| if (!bfa_itnim_is_online(ioim->itnim)) { |
| if (!bfa_itnim_hold_io(ioim->itnim)) { |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, |
| &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, |
| __bfa_cb_ioim_pathtov, ioim); |
| } else { |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, |
| &ioim->itnim->pending_q); |
| } |
| break; |
| } |
| |
| if (ioim->nsges > BFI_SGE_INLINE) { |
| if (!bfa_ioim_sge_setup(ioim)) { |
| bfa_sm_set_state(ioim, bfa_ioim_sm_sgalloc); |
| return; |
| } |
| } |
| |
| if (!bfa_ioim_send_ioreq(ioim)) { |
| bfa_sm_set_state(ioim, bfa_ioim_sm_qfull); |
| break; |
| } |
| |
| bfa_sm_set_state(ioim, bfa_ioim_sm_active); |
| break; |
| |
| case BFA_IOIM_SM_IOTOV: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, |
| __bfa_cb_ioim_pathtov, ioim); |
| break; |
| |
| case BFA_IOIM_SM_ABORT: |
| /** |
| * IO in pending queue can get abort requests. Complete abort |
| * requests immediately. |
| */ |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_assert(bfa_q_is_on_q(&ioim->itnim->pending_q, ioim)); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_abort, |
| ioim); |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| /** |
| * IO is waiting for SG pages. |
| */ |
| static void |
| bfa_ioim_sm_sgalloc(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_trc(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_SGALLOCED: |
| if (!bfa_ioim_send_ioreq(ioim)) { |
| bfa_sm_set_state(ioim, bfa_ioim_sm_qfull); |
| break; |
| } |
| bfa_sm_set_state(ioim, bfa_ioim_sm_active); |
| break; |
| |
| case BFA_IOIM_SM_CLEANUP: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_sgpg_wcancel(ioim->bfa, &ioim->iosp->sgpg_wqe); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_failed, |
| ioim); |
| bfa_ioim_notify_cleanup(ioim); |
| break; |
| |
| case BFA_IOIM_SM_ABORT: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_sgpg_wcancel(ioim->bfa, &ioim->iosp->sgpg_wqe); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_abort, |
| ioim); |
| break; |
| |
| case BFA_IOIM_SM_HWFAIL: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_sgpg_wcancel(ioim->bfa, &ioim->iosp->sgpg_wqe); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_failed, |
| ioim); |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| /** |
| * IO is active. |
| */ |
| static void |
| bfa_ioim_sm_active(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc_fp(ioim->bfa, ioim->iotag); |
| bfa_trc_fp(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_COMP_GOOD: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, |
| __bfa_cb_ioim_good_comp, ioim); |
| break; |
| |
| case BFA_IOIM_SM_COMP: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_comp, |
| ioim); |
| break; |
| |
| case BFA_IOIM_SM_DONE: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb_free); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_comp, |
| ioim); |
| break; |
| |
| case BFA_IOIM_SM_ABORT: |
| ioim->iosp->abort_explicit = BFA_TRUE; |
| ioim->io_cbfn = __bfa_cb_ioim_abort; |
| |
| if (bfa_ioim_send_abort(ioim)) |
| bfa_sm_set_state(ioim, bfa_ioim_sm_abort); |
| else { |
| bfa_sm_set_state(ioim, bfa_ioim_sm_abort_qfull); |
| bfa_reqq_wait(ioim->bfa, ioim->reqq, |
| &ioim->iosp->reqq_wait); |
| } |
| break; |
| |
| case BFA_IOIM_SM_CLEANUP: |
| ioim->iosp->abort_explicit = BFA_FALSE; |
| ioim->io_cbfn = __bfa_cb_ioim_failed; |
| |
| if (bfa_ioim_send_abort(ioim)) |
| bfa_sm_set_state(ioim, bfa_ioim_sm_cleanup); |
| else { |
| bfa_sm_set_state(ioim, bfa_ioim_sm_cleanup_qfull); |
| bfa_reqq_wait(ioim->bfa, ioim->reqq, |
| &ioim->iosp->reqq_wait); |
| } |
| break; |
| |
| case BFA_IOIM_SM_HWFAIL: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_failed, |
| ioim); |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| /** |
| * IO is being aborted, waiting for completion from firmware. |
| */ |
| static void |
| bfa_ioim_sm_abort(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_trc(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_COMP_GOOD: |
| case BFA_IOIM_SM_COMP: |
| case BFA_IOIM_SM_DONE: |
| case BFA_IOIM_SM_FREE: |
| break; |
| |
| case BFA_IOIM_SM_ABORT_DONE: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb_free); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_abort, |
| ioim); |
| break; |
| |
| case BFA_IOIM_SM_ABORT_COMP: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_abort, |
| ioim); |
| break; |
| |
| case BFA_IOIM_SM_COMP_UTAG: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_abort, |
| ioim); |
| break; |
| |
| case BFA_IOIM_SM_CLEANUP: |
| bfa_assert(ioim->iosp->abort_explicit == BFA_TRUE); |
| ioim->iosp->abort_explicit = BFA_FALSE; |
| |
| if (bfa_ioim_send_abort(ioim)) |
| bfa_sm_set_state(ioim, bfa_ioim_sm_cleanup); |
| else { |
| bfa_sm_set_state(ioim, bfa_ioim_sm_cleanup_qfull); |
| bfa_reqq_wait(ioim->bfa, ioim->reqq, |
| &ioim->iosp->reqq_wait); |
| } |
| break; |
| |
| case BFA_IOIM_SM_HWFAIL: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_failed, |
| ioim); |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| /** |
| * IO is being cleaned up (implicit abort), waiting for completion from |
| * firmware. |
| */ |
| static void |
| bfa_ioim_sm_cleanup(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_trc(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_COMP_GOOD: |
| case BFA_IOIM_SM_COMP: |
| case BFA_IOIM_SM_DONE: |
| case BFA_IOIM_SM_FREE: |
| break; |
| |
| case BFA_IOIM_SM_ABORT: |
| /** |
| * IO is already being aborted implicitly |
| */ |
| ioim->io_cbfn = __bfa_cb_ioim_abort; |
| break; |
| |
| case BFA_IOIM_SM_ABORT_DONE: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb_free); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, ioim->io_cbfn, ioim); |
| bfa_ioim_notify_cleanup(ioim); |
| break; |
| |
| case BFA_IOIM_SM_ABORT_COMP: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, ioim->io_cbfn, ioim); |
| bfa_ioim_notify_cleanup(ioim); |
| break; |
| |
| case BFA_IOIM_SM_COMP_UTAG: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, ioim->io_cbfn, ioim); |
| bfa_ioim_notify_cleanup(ioim); |
| break; |
| |
| case BFA_IOIM_SM_HWFAIL: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_failed, |
| ioim); |
| break; |
| |
| case BFA_IOIM_SM_CLEANUP: |
| /** |
| * IO can be in cleanup state already due to TM command. 2nd cleanup |
| * request comes from ITN offline event. |
| */ |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| /** |
| * IO is waiting for room in request CQ |
| */ |
| static void |
| bfa_ioim_sm_qfull(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_trc(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_QRESUME: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_active); |
| bfa_ioim_send_ioreq(ioim); |
| break; |
| |
| case BFA_IOIM_SM_ABORT: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_reqq_wcancel(&ioim->iosp->reqq_wait); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_abort, |
| ioim); |
| break; |
| |
| case BFA_IOIM_SM_CLEANUP: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_reqq_wcancel(&ioim->iosp->reqq_wait); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_failed, |
| ioim); |
| bfa_ioim_notify_cleanup(ioim); |
| break; |
| |
| case BFA_IOIM_SM_HWFAIL: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_reqq_wcancel(&ioim->iosp->reqq_wait); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_failed, |
| ioim); |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| /** |
| * Active IO is being aborted, waiting for room in request CQ. |
| */ |
| static void |
| bfa_ioim_sm_abort_qfull(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_trc(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_QRESUME: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_abort); |
| bfa_ioim_send_abort(ioim); |
| break; |
| |
| case BFA_IOIM_SM_CLEANUP: |
| bfa_assert(ioim->iosp->abort_explicit == BFA_TRUE); |
| ioim->iosp->abort_explicit = BFA_FALSE; |
| bfa_sm_set_state(ioim, bfa_ioim_sm_cleanup_qfull); |
| break; |
| |
| case BFA_IOIM_SM_COMP_GOOD: |
| case BFA_IOIM_SM_COMP: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_reqq_wcancel(&ioim->iosp->reqq_wait); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_abort, |
| ioim); |
| break; |
| |
| case BFA_IOIM_SM_DONE: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb_free); |
| bfa_reqq_wcancel(&ioim->iosp->reqq_wait); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_abort, |
| ioim); |
| break; |
| |
| case BFA_IOIM_SM_HWFAIL: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_reqq_wcancel(&ioim->iosp->reqq_wait); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_failed, |
| ioim); |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| /** |
| * Active IO is being cleaned up, waiting for room in request CQ. |
| */ |
| static void |
| bfa_ioim_sm_cleanup_qfull(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_trc(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_QRESUME: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_cleanup); |
| bfa_ioim_send_abort(ioim); |
| break; |
| |
| case BFA_IOIM_SM_ABORT: |
| /** |
| * IO is already being cleaned up implicitly |
| */ |
| ioim->io_cbfn = __bfa_cb_ioim_abort; |
| break; |
| |
| case BFA_IOIM_SM_COMP_GOOD: |
| case BFA_IOIM_SM_COMP: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_reqq_wcancel(&ioim->iosp->reqq_wait); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, ioim->io_cbfn, ioim); |
| bfa_ioim_notify_cleanup(ioim); |
| break; |
| |
| case BFA_IOIM_SM_DONE: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb_free); |
| bfa_reqq_wcancel(&ioim->iosp->reqq_wait); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, ioim->io_cbfn, ioim); |
| bfa_ioim_notify_cleanup(ioim); |
| break; |
| |
| case BFA_IOIM_SM_HWFAIL: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| bfa_reqq_wcancel(&ioim->iosp->reqq_wait); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, __bfa_cb_ioim_failed, |
| ioim); |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| /** |
| * IO bfa callback is pending. |
| */ |
| static void |
| bfa_ioim_sm_hcb(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc_fp(ioim->bfa, ioim->iotag); |
| bfa_trc_fp(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_HCB: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_uninit); |
| bfa_ioim_free(ioim); |
| bfa_cb_ioim_resfree(ioim->bfa->bfad); |
| break; |
| |
| case BFA_IOIM_SM_CLEANUP: |
| bfa_ioim_notify_cleanup(ioim); |
| break; |
| |
| case BFA_IOIM_SM_HWFAIL: |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| /** |
| * IO bfa callback is pending. IO resource cannot be freed. |
| */ |
| static void |
| bfa_ioim_sm_hcb_free(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_trc(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_HCB: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_resfree); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_resfree_q); |
| break; |
| |
| case BFA_IOIM_SM_FREE: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| break; |
| |
| case BFA_IOIM_SM_CLEANUP: |
| bfa_ioim_notify_cleanup(ioim); |
| break; |
| |
| case BFA_IOIM_SM_HWFAIL: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_hcb); |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| /** |
| * IO is completed, waiting resource free from firmware. |
| */ |
| static void |
| bfa_ioim_sm_resfree(struct bfa_ioim_s *ioim, enum bfa_ioim_event event) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_trc(ioim->bfa, event); |
| |
| switch (event) { |
| case BFA_IOIM_SM_FREE: |
| bfa_sm_set_state(ioim, bfa_ioim_sm_uninit); |
| bfa_ioim_free(ioim); |
| bfa_cb_ioim_resfree(ioim->bfa->bfad); |
| break; |
| |
| case BFA_IOIM_SM_CLEANUP: |
| bfa_ioim_notify_cleanup(ioim); |
| break; |
| |
| case BFA_IOIM_SM_HWFAIL: |
| break; |
| |
| default: |
| bfa_sm_fault(ioim->bfa, event); |
| } |
| } |
| |
| |
| |
| /** |
| * bfa_ioim_private |
| */ |
| |
| static void |
| __bfa_cb_ioim_good_comp(void *cbarg, bfa_boolean_t complete) |
| { |
| struct bfa_ioim_s *ioim = cbarg; |
| |
| if (!complete) { |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_HCB); |
| return; |
| } |
| |
| bfa_cb_ioim_good_comp(ioim->bfa->bfad, ioim->dio); |
| } |
| |
| static void |
| __bfa_cb_ioim_comp(void *cbarg, bfa_boolean_t complete) |
| { |
| struct bfa_ioim_s *ioim = cbarg; |
| struct bfi_ioim_rsp_s *m; |
| u8 *snsinfo = NULL; |
| u8 sns_len = 0; |
| s32 residue = 0; |
| |
| if (!complete) { |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_HCB); |
| return; |
| } |
| |
| m = (struct bfi_ioim_rsp_s *) &ioim->iosp->comp_rspmsg; |
| if (m->io_status == BFI_IOIM_STS_OK) { |
| /** |
| * setup sense information, if present |
| */ |
| if (m->scsi_status == SCSI_STATUS_CHECK_CONDITION |
| && m->sns_len) { |
| sns_len = m->sns_len; |
| snsinfo = ioim->iosp->snsinfo; |
| } |
| |
| /** |
| * setup residue value correctly for normal completions |
| */ |
| if (m->resid_flags == FCP_RESID_UNDER) |
| residue = bfa_os_ntohl(m->residue); |
| if (m->resid_flags == FCP_RESID_OVER) { |
| residue = bfa_os_ntohl(m->residue); |
| residue = -residue; |
| } |
| } |
| |
| bfa_cb_ioim_done(ioim->bfa->bfad, ioim->dio, m->io_status, |
| m->scsi_status, sns_len, snsinfo, residue); |
| } |
| |
| static void |
| __bfa_cb_ioim_failed(void *cbarg, bfa_boolean_t complete) |
| { |
| struct bfa_ioim_s *ioim = cbarg; |
| |
| if (!complete) { |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_HCB); |
| return; |
| } |
| |
| bfa_cb_ioim_done(ioim->bfa->bfad, ioim->dio, BFI_IOIM_STS_ABORTED, |
| 0, 0, NULL, 0); |
| } |
| |
| static void |
| __bfa_cb_ioim_pathtov(void *cbarg, bfa_boolean_t complete) |
| { |
| struct bfa_ioim_s *ioim = cbarg; |
| |
| if (!complete) { |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_HCB); |
| return; |
| } |
| |
| bfa_cb_ioim_done(ioim->bfa->bfad, ioim->dio, BFI_IOIM_STS_PATHTOV, |
| 0, 0, NULL, 0); |
| } |
| |
| static void |
| __bfa_cb_ioim_abort(void *cbarg, bfa_boolean_t complete) |
| { |
| struct bfa_ioim_s *ioim = cbarg; |
| |
| if (!complete) { |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_HCB); |
| return; |
| } |
| |
| bfa_cb_ioim_abort(ioim->bfa->bfad, ioim->dio); |
| } |
| |
| static void |
| bfa_ioim_sgpg_alloced(void *cbarg) |
| { |
| struct bfa_ioim_s *ioim = cbarg; |
| |
| ioim->nsgpgs = BFA_SGPG_NPAGE(ioim->nsges); |
| list_splice_tail_init(&ioim->iosp->sgpg_wqe.sgpg_q, &ioim->sgpg_q); |
| bfa_ioim_sgpg_setup(ioim); |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_SGALLOCED); |
| } |
| |
| /** |
| * Send I/O request to firmware. |
| */ |
| static bfa_boolean_t |
| bfa_ioim_send_ioreq(struct bfa_ioim_s *ioim) |
| { |
| struct bfa_itnim_s *itnim = ioim->itnim; |
| struct bfi_ioim_req_s *m; |
| static struct fcp_cmnd_s cmnd_z0 = { 0 }; |
| struct bfi_sge_s *sge; |
| u32 pgdlen = 0; |
| u64 addr; |
| struct scatterlist *sg; |
| struct scsi_cmnd *cmnd = (struct scsi_cmnd *) ioim->dio; |
| |
| /** |
| * check for room in queue to send request now |
| */ |
| m = bfa_reqq_next(ioim->bfa, ioim->reqq); |
| if (!m) { |
| bfa_reqq_wait(ioim->bfa, ioim->reqq, |
| &ioim->iosp->reqq_wait); |
| return BFA_FALSE; |
| } |
| |
| /** |
| * build i/o request message next |
| */ |
| m->io_tag = bfa_os_htons(ioim->iotag); |
| m->rport_hdl = ioim->itnim->rport->fw_handle; |
| m->io_timeout = bfa_cb_ioim_get_timeout(ioim->dio); |
| |
| /** |
| * build inline IO SG element here |
| */ |
| sge = &m->sges[0]; |
| if (ioim->nsges) { |
| sg = (struct scatterlist *)scsi_sglist(cmnd); |
| addr = bfa_os_sgaddr(sg_dma_address(sg)); |
| sge->sga = *(union bfi_addr_u *) &addr; |
| pgdlen = sg_dma_len(sg); |
| sge->sg_len = pgdlen; |
| sge->flags = (ioim->nsges > BFI_SGE_INLINE) ? |
| BFI_SGE_DATA_CPL : BFI_SGE_DATA_LAST; |
| bfa_sge_to_be(sge); |
| sge++; |
| } |
| |
| if (ioim->nsges > BFI_SGE_INLINE) { |
| sge->sga = ioim->sgpg->sgpg_pa; |
| } else { |
| sge->sga.a32.addr_lo = 0; |
| sge->sga.a32.addr_hi = 0; |
| } |
| sge->sg_len = pgdlen; |
| sge->flags = BFI_SGE_PGDLEN; |
| bfa_sge_to_be(sge); |
| |
| /** |
| * set up I/O command parameters |
| */ |
| bfa_os_assign(m->cmnd, cmnd_z0); |
| m->cmnd.lun = bfa_cb_ioim_get_lun(ioim->dio); |
| m->cmnd.iodir = bfa_cb_ioim_get_iodir(ioim->dio); |
| bfa_os_assign(m->cmnd.cdb, |
| *(struct scsi_cdb_s *)bfa_cb_ioim_get_cdb(ioim->dio)); |
| m->cmnd.fcp_dl = bfa_os_htonl(bfa_cb_ioim_get_size(ioim->dio)); |
| |
| /** |
| * set up I/O message header |
| */ |
| switch (m->cmnd.iodir) { |
| case FCP_IODIR_READ: |
| bfi_h2i_set(m->mh, BFI_MC_IOIM_READ, 0, bfa_lpuid(ioim->bfa)); |
| bfa_stats(itnim, input_reqs); |
| break; |
| case FCP_IODIR_WRITE: |
| bfi_h2i_set(m->mh, BFI_MC_IOIM_WRITE, 0, bfa_lpuid(ioim->bfa)); |
| bfa_stats(itnim, output_reqs); |
| break; |
| case FCP_IODIR_RW: |
| bfa_stats(itnim, input_reqs); |
| bfa_stats(itnim, output_reqs); |
| default: |
| bfi_h2i_set(m->mh, BFI_MC_IOIM_IO, 0, bfa_lpuid(ioim->bfa)); |
| } |
| if (itnim->seq_rec || |
| (bfa_cb_ioim_get_size(ioim->dio) & (sizeof(u32) - 1))) |
| bfi_h2i_set(m->mh, BFI_MC_IOIM_IO, 0, bfa_lpuid(ioim->bfa)); |
| |
| #ifdef IOIM_ADVANCED |
| m->cmnd.crn = bfa_cb_ioim_get_crn(ioim->dio); |
| m->cmnd.priority = bfa_cb_ioim_get_priority(ioim->dio); |
| m->cmnd.taskattr = bfa_cb_ioim_get_taskattr(ioim->dio); |
| |
| /** |
| * Handle large CDB (>16 bytes). |
| */ |
| m->cmnd.addl_cdb_len = (bfa_cb_ioim_get_cdblen(ioim->dio) - |
| FCP_CMND_CDB_LEN) / sizeof(u32); |
| if (m->cmnd.addl_cdb_len) { |
| bfa_os_memcpy(&m->cmnd.cdb + 1, (struct scsi_cdb_s *) |
| bfa_cb_ioim_get_cdb(ioim->dio) + 1, |
| m->cmnd.addl_cdb_len * sizeof(u32)); |
| fcp_cmnd_fcpdl(&m->cmnd) = |
| bfa_os_htonl(bfa_cb_ioim_get_size(ioim->dio)); |
| } |
| #endif |
| |
| /** |
| * queue I/O message to firmware |
| */ |
| bfa_reqq_produce(ioim->bfa, ioim->reqq); |
| return BFA_TRUE; |
| } |
| |
| /** |
| * Setup any additional SG pages needed.Inline SG element is setup |
| * at queuing time. |
| */ |
| static bfa_boolean_t |
| bfa_ioim_sge_setup(struct bfa_ioim_s *ioim) |
| { |
| u16 nsgpgs; |
| |
| bfa_assert(ioim->nsges > BFI_SGE_INLINE); |
| |
| /** |
| * allocate SG pages needed |
| */ |
| nsgpgs = BFA_SGPG_NPAGE(ioim->nsges); |
| if (!nsgpgs) |
| return BFA_TRUE; |
| |
| if (bfa_sgpg_malloc(ioim->bfa, &ioim->sgpg_q, nsgpgs) |
| != BFA_STATUS_OK) { |
| bfa_sgpg_wait(ioim->bfa, &ioim->iosp->sgpg_wqe, nsgpgs); |
| return BFA_FALSE; |
| } |
| |
| ioim->nsgpgs = nsgpgs; |
| bfa_ioim_sgpg_setup(ioim); |
| |
| return BFA_TRUE; |
| } |
| |
| static void |
| bfa_ioim_sgpg_setup(struct bfa_ioim_s *ioim) |
| { |
| int sgeid, nsges, i; |
| struct bfi_sge_s *sge; |
| struct bfa_sgpg_s *sgpg; |
| u32 pgcumsz; |
| u64 addr; |
| struct scatterlist *sg; |
| struct scsi_cmnd *cmnd = (struct scsi_cmnd *) ioim->dio; |
| |
| sgeid = BFI_SGE_INLINE; |
| ioim->sgpg = sgpg = bfa_q_first(&ioim->sgpg_q); |
| |
| sg = scsi_sglist(cmnd); |
| sg = sg_next(sg); |
| |
| do { |
| sge = sgpg->sgpg->sges; |
| nsges = ioim->nsges - sgeid; |
| if (nsges > BFI_SGPG_DATA_SGES) |
| nsges = BFI_SGPG_DATA_SGES; |
| |
| pgcumsz = 0; |
| for (i = 0; i < nsges; i++, sge++, sgeid++, sg = sg_next(sg)) { |
| addr = bfa_os_sgaddr(sg_dma_address(sg)); |
| sge->sga = *(union bfi_addr_u *) &addr; |
| sge->sg_len = sg_dma_len(sg); |
| pgcumsz += sge->sg_len; |
| |
| /** |
| * set flags |
| */ |
| if (i < (nsges - 1)) |
| sge->flags = BFI_SGE_DATA; |
| else if (sgeid < (ioim->nsges - 1)) |
| sge->flags = BFI_SGE_DATA_CPL; |
| else |
| sge->flags = BFI_SGE_DATA_LAST; |
| } |
| |
| sgpg = (struct bfa_sgpg_s *) bfa_q_next(sgpg); |
| |
| /** |
| * set the link element of each page |
| */ |
| if (sgeid == ioim->nsges) { |
| sge->flags = BFI_SGE_PGDLEN; |
| sge->sga.a32.addr_lo = 0; |
| sge->sga.a32.addr_hi = 0; |
| } else { |
| sge->flags = BFI_SGE_LINK; |
| sge->sga = sgpg->sgpg_pa; |
| } |
| sge->sg_len = pgcumsz; |
| } while (sgeid < ioim->nsges); |
| } |
| |
| /** |
| * Send I/O abort request to firmware. |
| */ |
| static bfa_boolean_t |
| bfa_ioim_send_abort(struct bfa_ioim_s *ioim) |
| { |
| struct bfi_ioim_abort_req_s *m; |
| enum bfi_ioim_h2i msgop; |
| |
| /** |
| * check for room in queue to send request now |
| */ |
| m = bfa_reqq_next(ioim->bfa, ioim->reqq); |
| if (!m) |
| return BFA_FALSE; |
| |
| /** |
| * build i/o request message next |
| */ |
| if (ioim->iosp->abort_explicit) |
| msgop = BFI_IOIM_H2I_IOABORT_REQ; |
| else |
| msgop = BFI_IOIM_H2I_IOCLEANUP_REQ; |
| |
| bfi_h2i_set(m->mh, BFI_MC_IOIM, msgop, bfa_lpuid(ioim->bfa)); |
| m->io_tag = bfa_os_htons(ioim->iotag); |
| m->abort_tag = ++ioim->abort_tag; |
| |
| /** |
| * queue I/O message to firmware |
| */ |
| bfa_reqq_produce(ioim->bfa, ioim->reqq); |
| return BFA_TRUE; |
| } |
| |
| /** |
| * Call to resume any I/O requests waiting for room in request queue. |
| */ |
| static void |
| bfa_ioim_qresume(void *cbarg) |
| { |
| struct bfa_ioim_s *ioim = cbarg; |
| |
| bfa_fcpim_stats(ioim->fcpim, qresumes); |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_QRESUME); |
| } |
| |
| |
| static void |
| bfa_ioim_notify_cleanup(struct bfa_ioim_s *ioim) |
| { |
| /** |
| * Move IO from itnim queue to fcpim global queue since itnim will be |
| * freed. |
| */ |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| |
| if (!ioim->iosp->tskim) { |
| if (ioim->fcpim->delay_comp && ioim->itnim->iotov_active) { |
| bfa_cb_dequeue(&ioim->hcb_qe); |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->itnim->delay_comp_q); |
| } |
| bfa_itnim_iodone(ioim->itnim); |
| } else |
| bfa_tskim_iodone(ioim->iosp->tskim); |
| } |
| |
| /** |
| * or after the link comes back. |
| */ |
| void |
| bfa_ioim_delayed_comp(struct bfa_ioim_s *ioim, bfa_boolean_t iotov) |
| { |
| /** |
| * If path tov timer expired, failback with PATHTOV status - these |
| * IO requests are not normally retried by IO stack. |
| * |
| * Otherwise device cameback online and fail it with normal failed |
| * status so that IO stack retries these failed IO requests. |
| */ |
| if (iotov) |
| ioim->io_cbfn = __bfa_cb_ioim_pathtov; |
| else |
| ioim->io_cbfn = __bfa_cb_ioim_failed; |
| |
| bfa_cb_queue(ioim->bfa, &ioim->hcb_qe, ioim->io_cbfn, ioim); |
| |
| /** |
| * Move IO to fcpim global queue since itnim will be |
| * freed. |
| */ |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &ioim->fcpim->ioim_comp_q); |
| } |
| |
| |
| |
| /** |
| * bfa_ioim_friend |
| */ |
| |
| /** |
| * Memory allocation and initialization. |
| */ |
| void |
| bfa_ioim_attach(struct bfa_fcpim_mod_s *fcpim, struct bfa_meminfo_s *minfo) |
| { |
| struct bfa_ioim_s *ioim; |
| struct bfa_ioim_sp_s *iosp; |
| u16 i; |
| u8 *snsinfo; |
| u32 snsbufsz; |
| |
| /** |
| * claim memory first |
| */ |
| ioim = (struct bfa_ioim_s *) bfa_meminfo_kva(minfo); |
| fcpim->ioim_arr = ioim; |
| bfa_meminfo_kva(minfo) = (u8 *) (ioim + fcpim->num_ioim_reqs); |
| |
| iosp = (struct bfa_ioim_sp_s *) bfa_meminfo_kva(minfo); |
| fcpim->ioim_sp_arr = iosp; |
| bfa_meminfo_kva(minfo) = (u8 *) (iosp + fcpim->num_ioim_reqs); |
| |
| /** |
| * Claim DMA memory for per IO sense data. |
| */ |
| snsbufsz = fcpim->num_ioim_reqs * BFI_IOIM_SNSLEN; |
| fcpim->snsbase.pa = bfa_meminfo_dma_phys(minfo); |
| bfa_meminfo_dma_phys(minfo) += snsbufsz; |
| |
| fcpim->snsbase.kva = bfa_meminfo_dma_virt(minfo); |
| bfa_meminfo_dma_virt(minfo) += snsbufsz; |
| snsinfo = fcpim->snsbase.kva; |
| bfa_iocfc_set_snsbase(fcpim->bfa, fcpim->snsbase.pa); |
| |
| /** |
| * Initialize ioim free queues |
| */ |
| INIT_LIST_HEAD(&fcpim->ioim_free_q); |
| INIT_LIST_HEAD(&fcpim->ioim_resfree_q); |
| INIT_LIST_HEAD(&fcpim->ioim_comp_q); |
| |
| for (i = 0; i < fcpim->num_ioim_reqs; |
| i++, ioim++, iosp++, snsinfo += BFI_IOIM_SNSLEN) { |
| /* |
| * initialize IOIM |
| */ |
| bfa_os_memset(ioim, 0, sizeof(struct bfa_ioim_s)); |
| ioim->iotag = i; |
| ioim->bfa = fcpim->bfa; |
| ioim->fcpim = fcpim; |
| ioim->iosp = iosp; |
| iosp->snsinfo = snsinfo; |
| INIT_LIST_HEAD(&ioim->sgpg_q); |
| bfa_reqq_winit(&ioim->iosp->reqq_wait, |
| bfa_ioim_qresume, ioim); |
| bfa_sgpg_winit(&ioim->iosp->sgpg_wqe, |
| bfa_ioim_sgpg_alloced, ioim); |
| bfa_sm_set_state(ioim, bfa_ioim_sm_uninit); |
| |
| list_add_tail(&ioim->qe, &fcpim->ioim_free_q); |
| } |
| } |
| |
| /** |
| * Driver detach time call. |
| */ |
| void |
| bfa_ioim_detach(struct bfa_fcpim_mod_s *fcpim) |
| { |
| } |
| |
| void |
| bfa_ioim_isr(struct bfa_s *bfa, struct bfi_msg_s *m) |
| { |
| struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); |
| struct bfi_ioim_rsp_s *rsp = (struct bfi_ioim_rsp_s *) m; |
| struct bfa_ioim_s *ioim; |
| u16 iotag; |
| enum bfa_ioim_event evt = BFA_IOIM_SM_COMP; |
| |
| iotag = bfa_os_ntohs(rsp->io_tag); |
| |
| ioim = BFA_IOIM_FROM_TAG(fcpim, iotag); |
| bfa_assert(ioim->iotag == iotag); |
| |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_trc(ioim->bfa, rsp->io_status); |
| bfa_trc(ioim->bfa, rsp->reuse_io_tag); |
| |
| if (bfa_sm_cmp_state(ioim, bfa_ioim_sm_active)) |
| bfa_os_assign(ioim->iosp->comp_rspmsg, *m); |
| |
| switch (rsp->io_status) { |
| case BFI_IOIM_STS_OK: |
| bfa_fcpim_stats(fcpim, iocomp_ok); |
| if (rsp->reuse_io_tag == 0) |
| evt = BFA_IOIM_SM_DONE; |
| else |
| evt = BFA_IOIM_SM_COMP; |
| break; |
| |
| case BFI_IOIM_STS_TIMEDOUT: |
| case BFI_IOIM_STS_ABORTED: |
| rsp->io_status = BFI_IOIM_STS_ABORTED; |
| bfa_fcpim_stats(fcpim, iocomp_aborted); |
| if (rsp->reuse_io_tag == 0) |
| evt = BFA_IOIM_SM_DONE; |
| else |
| evt = BFA_IOIM_SM_COMP; |
| break; |
| |
| case BFI_IOIM_STS_PROTO_ERR: |
| bfa_fcpim_stats(fcpim, iocom_proto_err); |
| bfa_assert(rsp->reuse_io_tag); |
| evt = BFA_IOIM_SM_COMP; |
| break; |
| |
| case BFI_IOIM_STS_SQER_NEEDED: |
| bfa_fcpim_stats(fcpim, iocom_sqer_needed); |
| bfa_assert(rsp->reuse_io_tag == 0); |
| evt = BFA_IOIM_SM_SQRETRY; |
| break; |
| |
| case BFI_IOIM_STS_RES_FREE: |
| bfa_fcpim_stats(fcpim, iocom_res_free); |
| evt = BFA_IOIM_SM_FREE; |
| break; |
| |
| case BFI_IOIM_STS_HOST_ABORTED: |
| bfa_fcpim_stats(fcpim, iocom_hostabrts); |
| if (rsp->abort_tag != ioim->abort_tag) { |
| bfa_trc(ioim->bfa, rsp->abort_tag); |
| bfa_trc(ioim->bfa, ioim->abort_tag); |
| return; |
| } |
| |
| if (rsp->reuse_io_tag) |
| evt = BFA_IOIM_SM_ABORT_COMP; |
| else |
| evt = BFA_IOIM_SM_ABORT_DONE; |
| break; |
| |
| case BFI_IOIM_STS_UTAG: |
| bfa_fcpim_stats(fcpim, iocom_utags); |
| evt = BFA_IOIM_SM_COMP_UTAG; |
| break; |
| |
| default: |
| bfa_assert(0); |
| } |
| |
| bfa_sm_send_event(ioim, evt); |
| } |
| |
| void |
| bfa_ioim_good_comp_isr(struct bfa_s *bfa, struct bfi_msg_s *m) |
| { |
| struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); |
| struct bfi_ioim_rsp_s *rsp = (struct bfi_ioim_rsp_s *) m; |
| struct bfa_ioim_s *ioim; |
| u16 iotag; |
| |
| iotag = bfa_os_ntohs(rsp->io_tag); |
| |
| ioim = BFA_IOIM_FROM_TAG(fcpim, iotag); |
| bfa_assert(ioim->iotag == iotag); |
| |
| bfa_trc_fp(ioim->bfa, ioim->iotag); |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_COMP_GOOD); |
| } |
| |
| /** |
| * Called by itnim to clean up IO while going offline. |
| */ |
| void |
| bfa_ioim_cleanup(struct bfa_ioim_s *ioim) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_fcpim_stats(ioim->fcpim, io_cleanups); |
| |
| ioim->iosp->tskim = NULL; |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_CLEANUP); |
| } |
| |
| void |
| bfa_ioim_cleanup_tm(struct bfa_ioim_s *ioim, struct bfa_tskim_s *tskim) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_fcpim_stats(ioim->fcpim, io_tmaborts); |
| |
| ioim->iosp->tskim = tskim; |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_CLEANUP); |
| } |
| |
| /** |
| * IOC failure handling. |
| */ |
| void |
| bfa_ioim_iocdisable(struct bfa_ioim_s *ioim) |
| { |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_HWFAIL); |
| } |
| |
| /** |
| * IO offline TOV popped. Fail the pending IO. |
| */ |
| void |
| bfa_ioim_tov(struct bfa_ioim_s *ioim) |
| { |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_IOTOV); |
| } |
| |
| |
| |
| /** |
| * bfa_ioim_api |
| */ |
| |
| /** |
| * Allocate IOIM resource for initiator mode I/O request. |
| */ |
| struct bfa_ioim_s * |
| bfa_ioim_alloc(struct bfa_s *bfa, struct bfad_ioim_s *dio, |
| struct bfa_itnim_s *itnim, u16 nsges) |
| { |
| struct bfa_fcpim_mod_s *fcpim = BFA_FCPIM_MOD(bfa); |
| struct bfa_ioim_s *ioim; |
| |
| /** |
| * alocate IOIM resource |
| */ |
| bfa_q_deq(&fcpim->ioim_free_q, &ioim); |
| if (!ioim) { |
| bfa_fcpim_stats(fcpim, no_iotags); |
| return NULL; |
| } |
| |
| ioim->dio = dio; |
| ioim->itnim = itnim; |
| ioim->nsges = nsges; |
| ioim->nsgpgs = 0; |
| |
| bfa_stats(fcpim, total_ios); |
| bfa_stats(itnim, ios); |
| fcpim->ios_active++; |
| |
| list_add_tail(&ioim->qe, &itnim->io_q); |
| bfa_trc_fp(ioim->bfa, ioim->iotag); |
| |
| return ioim; |
| } |
| |
| void |
| bfa_ioim_free(struct bfa_ioim_s *ioim) |
| { |
| struct bfa_fcpim_mod_s *fcpim = ioim->fcpim; |
| |
| bfa_trc_fp(ioim->bfa, ioim->iotag); |
| bfa_assert_fp(bfa_sm_cmp_state(ioim, bfa_ioim_sm_uninit)); |
| |
| bfa_assert_fp(list_empty(&ioim->sgpg_q) |
| || (ioim->nsges > BFI_SGE_INLINE)); |
| |
| if (ioim->nsgpgs > 0) |
| bfa_sgpg_mfree(ioim->bfa, &ioim->sgpg_q, ioim->nsgpgs); |
| |
| bfa_stats(ioim->itnim, io_comps); |
| fcpim->ios_active--; |
| |
| list_del(&ioim->qe); |
| list_add_tail(&ioim->qe, &fcpim->ioim_free_q); |
| } |
| |
| void |
| bfa_ioim_start(struct bfa_ioim_s *ioim) |
| { |
| bfa_trc_fp(ioim->bfa, ioim->iotag); |
| |
| /** |
| * Obtain the queue over which this request has to be issued |
| */ |
| ioim->reqq = bfa_fcpim_ioredirect_enabled(ioim->bfa) ? |
| bfa_cb_ioim_get_reqq(ioim->dio) : |
| bfa_itnim_get_reqq(ioim); |
| |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_START); |
| } |
| |
| /** |
| * Driver I/O abort request. |
| */ |
| void |
| bfa_ioim_abort(struct bfa_ioim_s *ioim) |
| { |
| bfa_trc(ioim->bfa, ioim->iotag); |
| bfa_fcpim_stats(ioim->fcpim, io_aborts); |
| bfa_sm_send_event(ioim, BFA_IOIM_SM_ABORT); |
| } |
| |
| |