diff --git a/sys/dev/mpt/mpt.c b/sys/dev/mpt/mpt.c index 12fb146694cc..f780e8825c08 100644 --- a/sys/dev/mpt/mpt.c +++ b/sys/dev/mpt/mpt.c @@ -100,6 +100,8 @@ __FBSDID("$FreeBSD$"); #include #include +#include +#include #include @@ -123,7 +125,7 @@ static int mpt_send_event_request(struct mpt_softc *mpt, int onoff); static int mpt_soft_reset(struct mpt_softc *mpt); static void mpt_hard_reset(struct mpt_softc *mpt); static int mpt_configure_ioc(struct mpt_softc *mpt); -static int mpt_enable_ioc(struct mpt_softc *mpt); +static int mpt_enable_ioc(struct mpt_softc *mpt, int); /************************* Personality Module Support *************************/ /* @@ -150,7 +152,7 @@ mpt_pers_find(struct mpt_softc *mpt, u_int start_at) } /* - * Used infrequenstly, so no need to optimize like a forward + * Used infrequently, so no need to optimize like a forward * traversal where we use the MAX+1 is guaranteed to be NULL * trick. */ @@ -179,6 +181,7 @@ mpt_pers_find_reverse(struct mpt_softc *mpt, u_int start_at) static mpt_load_handler_t mpt_stdload; static mpt_probe_handler_t mpt_stdprobe; static mpt_attach_handler_t mpt_stdattach; +static mpt_enable_handler_t mpt_stdenable; static mpt_event_handler_t mpt_stdevent; static mpt_reset_handler_t mpt_stdreset; static mpt_shutdown_handler_t mpt_stdshutdown; @@ -189,6 +192,7 @@ static struct mpt_personality mpt_default_personality = .load = mpt_stdload, .probe = mpt_stdprobe, .attach = mpt_stdattach, + .enable = mpt_stdenable, .event = mpt_stdevent, .reset = mpt_stdreset, .shutdown = mpt_stdshutdown, @@ -198,6 +202,7 @@ static struct mpt_personality mpt_default_personality = static mpt_load_handler_t mpt_core_load; static mpt_attach_handler_t mpt_core_attach; +static mpt_enable_handler_t mpt_core_enable; static mpt_reset_handler_t mpt_core_ioc_reset; static mpt_event_handler_t mpt_core_event; static mpt_shutdown_handler_t mpt_core_shutdown; @@ -208,6 +213,7 @@ static struct mpt_personality mpt_core_personality = .name = "mpt_core", .load = mpt_core_load, .attach = mpt_core_attach, + .enable = mpt_core_enable, .event = mpt_core_event, .reset = mpt_core_ioc_reset, .shutdown = mpt_core_shutdown, @@ -226,8 +232,7 @@ static moduledata_t mpt_core_mod = { DECLARE_MODULE(mpt_core, mpt_core_mod, SI_SUB_DRIVERS, SI_ORDER_FIRST); MODULE_VERSION(mpt_core, 1); -#define MPT_PERS_ATACHED(pers, mpt) \ - ((mpt)->pers_mask & (0x1 << pers->id)) +#define MPT_PERS_ATTACHED(pers, mpt) ((mpt)->mpt_pers_mask & (0x1 << pers->id)) int @@ -274,8 +279,10 @@ mpt_modevent(module_t mod, int type, void *data) } case MOD_SHUTDOWN: break; +#if __FreeBSD_version >= 500000 case MOD_QUIESCE: break; +#endif case MOD_UNLOAD: error = pers->unload(pers); mpt_personalities[pers->id] = NULL; @@ -308,6 +315,13 @@ mpt_stdattach(struct mpt_softc *mpt) return (0); } +int +mpt_stdenable(struct mpt_softc *mpt) +{ + /* Enable is always successfull. */ + return (0); +} + int mpt_stdevent(struct mpt_softc *mpt, request_t *req, MSG_EVENT_NOTIFY_REPLY *msg) { @@ -385,13 +399,14 @@ mpt_register_handler(struct mpt_softc *mpt, mpt_handler_type type, * that the full table is checked to see if * this handler was previously registered. */ - if (free_cbi == MPT_HANDLER_ID_NONE - && (mpt_reply_handlers[cbi] + if (free_cbi == MPT_HANDLER_ID_NONE && + (mpt_reply_handlers[cbi] == mpt_default_reply_handler)) free_cbi = cbi; } - if (free_cbi == MPT_HANDLER_ID_NONE) + if (free_cbi == MPT_HANDLER_ID_NONE) { return (ENOMEM); + } mpt_reply_handlers[free_cbi] = handler.reply_handler; *phandler_id = MPT_CBI_TO_HID(free_cbi); break; @@ -429,22 +444,23 @@ mpt_deregister_handler(struct mpt_softc *mpt, mpt_handler_type type, static int mpt_default_reply_handler(struct mpt_softc *mpt, request_t *req, - MSG_DEFAULT_REPLY *reply_frame) + uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame) { - mpt_prt(mpt, "XXXX Default Handler Called. Req %p, Frame %p\n", - req, reply_frame); + mpt_prt(mpt, + "Default Handler Called: req=%p:%u reply_descriptor=%x frame=%p\n", + req, req->serno, reply_desc, reply_frame); if (reply_frame != NULL) mpt_dump_reply_frame(mpt, reply_frame); - mpt_prt(mpt, "XXXX Reply Frame Ignored\n"); + mpt_prt(mpt, "Reply Frame Ignored\n"); return (/*free_reply*/TRUE); } static int mpt_config_reply_handler(struct mpt_softc *mpt, request_t *req, - MSG_DEFAULT_REPLY *reply_frame) + uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame) { if (req != NULL) { @@ -471,7 +487,7 @@ mpt_config_reply_handler(struct mpt_softc *mpt, request_t *req, static int mpt_handshake_reply_handler(struct mpt_softc *mpt, request_t *req, - MSG_DEFAULT_REPLY *reply_frame) + uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame) { /* Nothing to be done. */ return (/*free_reply*/TRUE); @@ -479,12 +495,13 @@ mpt_handshake_reply_handler(struct mpt_softc *mpt, request_t *req, static int mpt_event_reply_handler(struct mpt_softc *mpt, request_t *req, - MSG_DEFAULT_REPLY *reply_frame) + uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame) { int free_reply; if (reply_frame == NULL) { - mpt_prt(mpt, "Event Handler: req %p - Unexpected NULL reply\n"); + mpt_prt(mpt, "Event Handler: req %p:%u - Unexpected NULL reply\n", + req, req->serno); return (/*free_reply*/TRUE); } @@ -619,13 +636,15 @@ void mpt_intr(void *arg) { struct mpt_softc *mpt; - uint32_t reply_desc; + uint32_t reply_desc; + int ntrips = 0; mpt = (struct mpt_softc *)arg; while ((reply_desc = mpt_pop_reply_queue(mpt)) != MPT_REPLY_EMPTY) { request_t *req; MSG_DEFAULT_REPLY *reply_frame; uint32_t reply_baddr; + uint32_t ctxt_idx; u_int cb_index; u_int req_index; int free_rf; @@ -635,27 +654,102 @@ mpt_intr(void *arg) reply_baddr = 0; if ((reply_desc & MPI_ADDRESS_REPLY_A_BIT) != 0) { u_int offset; - /* * Insure that the reply frame is coherent. */ reply_baddr = (reply_desc << 1); offset = reply_baddr - (mpt->reply_phys & 0xFFFFFFFF); - bus_dmamap_sync_range(mpt->reply_dmat, mpt->reply_dmap, - offset, MPT_REPLY_SIZE, - BUS_DMASYNC_POSTREAD); + bus_dmamap_sync_range(mpt->reply_dmat, + mpt->reply_dmap, offset, MPT_REPLY_SIZE, + BUS_DMASYNC_POSTREAD); reply_frame = MPT_REPLY_OTOV(mpt, offset); - reply_desc = le32toh(reply_frame->MsgContext); - } - cb_index = MPT_CONTEXT_TO_CBI(reply_desc); - req_index = MPT_CONTEXT_TO_REQI(reply_desc); - if (req_index < MPT_MAX_REQUESTS(mpt)) - req = &mpt->request_pool[req_index]; + ctxt_idx = le32toh(reply_frame->MsgContext); + } else { + uint32_t type; - free_rf = mpt_reply_handlers[cb_index](mpt, req, reply_frame); + type = MPI_GET_CONTEXT_REPLY_TYPE(reply_desc); + ctxt_idx = reply_desc; + mpt_lprt(mpt, MPT_PRT_DEBUG1, "Context Reply: 0x%08x\n", + reply_desc); + + switch (type) { + case MPI_CONTEXT_REPLY_TYPE_SCSI_INIT: + ctxt_idx &= MPI_CONTEXT_REPLY_CONTEXT_MASK; + break; + case MPI_CONTEXT_REPLY_TYPE_SCSI_TARGET: + ctxt_idx = GET_IO_INDEX(reply_desc); + if (mpt->tgt_cmd_ptrs == NULL) { + mpt_prt(mpt, + "mpt_intr: no target cmd ptrs\n"); + reply_desc = MPT_REPLY_EMPTY; + break; + } + if (ctxt_idx >= mpt->tgt_cmds_allocated) { + mpt_prt(mpt, + "mpt_intr: bad tgt cmd ctxt %u\n", + ctxt_idx); + reply_desc = MPT_REPLY_EMPTY; + ntrips = 1000; + break; + } + req = mpt->tgt_cmd_ptrs[ctxt_idx]; + if (req == NULL) { + mpt_prt(mpt, "no request backpointer " + "at index %u", ctxt_idx); + reply_desc = MPT_REPLY_EMPTY; + ntrips = 1000; + break; + } + /* + * Reformulate ctxt_idx to be just as if + * it were another type of context reply + * so the code below will find the request + * via indexing into the pool. + */ + ctxt_idx = + req->index | mpt->scsi_tgt_handler_id; + req = NULL; + break; + case MPI_CONTEXT_REPLY_TYPE_LAN: + mpt_prt(mpt, "LAN CONTEXT REPLY: 0x%08x\n", + reply_desc); + reply_desc = MPT_REPLY_EMPTY; + break; + default: + mpt_prt(mpt, "Context Reply 0x%08x?\n", type); + reply_desc = MPT_REPLY_EMPTY; + break; + } + if (reply_desc == MPT_REPLY_EMPTY) { + if (ntrips++ > 1000) { + break; + } + continue; + } + } + + cb_index = MPT_CONTEXT_TO_CBI(ctxt_idx); + req_index = MPT_CONTEXT_TO_REQI(ctxt_idx); + if (req_index < MPT_MAX_REQUESTS(mpt)) { + req = &mpt->request_pool[req_index]; + } + + free_rf = mpt_reply_handlers[cb_index](mpt, req, + reply_desc, reply_frame); if (reply_frame != NULL && free_rf) mpt_free_reply(mpt, reply_baddr); + + /* + * If we got ourselves disabled, don't get stuck in a loop + */ + if (mpt->disabled) { + mpt_disable_ints(mpt); + break; + } + if (ntrips++ > 1000) { + break; + } } } @@ -678,7 +772,8 @@ mpt_complete_request_chain(struct mpt_softc *mpt, struct req_queue *chain, ioc_status_frame.Function = msg_hdr->Function; ioc_status_frame.MsgContext = msg_hdr->MsgContext; cb_index = MPT_CONTEXT_TO_CBI(le32toh(msg_hdr->MsgContext)); - mpt_reply_handlers[cb_index](mpt, req, &ioc_status_frame); + mpt_reply_handlers[cb_index](mpt, req, msg_hdr->MsgContext, + &ioc_status_frame); } } @@ -718,12 +813,11 @@ mpt_wait_db_ack(struct mpt_softc *mpt) for (i=0; i < MPT_MAX_WAIT; i++) { if (!MPT_DB_IS_BUSY(mpt_rd_intr(mpt))) { maxwait_ack = i > maxwait_ack ? i : maxwait_ack; - return MPT_OK; + return (MPT_OK); } - - DELAY(1000); + DELAY(200); } - return MPT_FAIL; + return (MPT_FAIL); } /* Busy wait for a door bell interrupt */ @@ -738,7 +832,7 @@ mpt_wait_db_int(struct mpt_softc *mpt) } DELAY(100); } - return MPT_FAIL; + return (MPT_FAIL); } /* Wait for IOC to transition to a give state */ @@ -782,7 +876,7 @@ mpt_soft_reset(struct mpt_softc *mpt) /* Have to use hard reset if we are not in Running state */ if (MPT_STATE(mpt_rd_db(mpt)) != MPT_DB_STATE_RUNNING) { mpt_prt(mpt, "soft reset failed: device not running\n"); - return MPT_FAIL; + return (MPT_FAIL); } /* If door bell is in use we don't have a chance of getting @@ -791,7 +885,7 @@ mpt_soft_reset(struct mpt_softc *mpt) */ if (MPT_DB_IS_IN_USE(mpt_rd_db(mpt))) { mpt_prt(mpt, "soft reset failed: doorbell wedged\n"); - return MPT_FAIL; + return (MPT_FAIL); } /* Send the reset request to the IOC */ @@ -799,13 +893,13 @@ mpt_soft_reset(struct mpt_softc *mpt) MPI_FUNCTION_IOC_MESSAGE_UNIT_RESET << MPI_DOORBELL_FUNCTION_SHIFT); if (mpt_wait_db_ack(mpt) != MPT_OK) { mpt_prt(mpt, "soft reset failed: ack timeout\n"); - return MPT_FAIL; + return (MPT_FAIL); } /* Wait for the IOC to reload and come out of reset state */ if (mpt_wait_state(mpt, MPT_DB_STATE_READY) != MPT_OK) { mpt_prt(mpt, "soft reset failed: device did not restart\n"); - return MPT_FAIL; + return (MPT_FAIL); } return MPT_OK; @@ -979,7 +1073,7 @@ mpt_reset(struct mpt_softc *mpt, int reinit) } if (reinit != 0) { - ret = mpt_enable_ioc(mpt); + ret = mpt_enable_ioc(mpt, 1); if (ret == MPT_OK) { mpt_enable_ints(mpt); } @@ -1010,7 +1104,10 @@ mpt_free_request(struct mpt_softc *mpt, request_t *req) req->ccb = NULL; req->state = REQ_STATE_FREE; if (LIST_EMPTY(&mpt->ack_frames)) { - TAILQ_INSERT_HEAD(&mpt->request_free_list, req, links); + /* + * Insert free ones at the tail + */ + TAILQ_INSERT_TAIL(&mpt->request_free_list, req, links); if (mpt->getreqwaiter != 0) { mpt->getreqwaiter = 0; wakeup(&mpt->request_free_list); @@ -1043,9 +1140,7 @@ mpt_get_request(struct mpt_softc *mpt, int sleep_ok) TAILQ_REMOVE(&mpt->request_free_list, req, links); req->state = REQ_STATE_ALLOCATED; req->chain = NULL; - if ((req->serno = ++(mpt->cmd_serno)) == 0) { - req->serno = ++(mpt->cmd_serno); - } + req->serno = mpt->sequence++; } else if (sleep_ok != 0) { mpt->getreqwaiter = 1; mpt_sleep(mpt, &mpt->request_free_list, PUSER, "mptgreq", 0); @@ -1063,8 +1158,13 @@ mpt_send_cmd(struct mpt_softc *mpt, request_t *req) pReq = req->req_vbuf; if (mpt->verbose > MPT_PRT_TRACE) { int offset; - mpt_prt(mpt, "Send Request %d (0x%x):", - req->index, req->req_pbuf); +#if __FreeBSD_version >= 500000 + mpt_prt(mpt, "Send Request %d (%jx):", + req->index, (uintmax_t) req->req_pbuf); +#else + mpt_prt(mpt, "Send Request %d (%llx):", + req->index, (unsigned long long) req->req_pbuf); +#endif for (offset = 0; offset < mpt->request_frame_size; offset++) { if ((offset & 0x7) == 0) { mpt_prtc(mpt, "\n"); @@ -1108,16 +1208,15 @@ mpt_wait_req(struct mpt_softc *mpt, request_t *req, * Convert to ticks or 500us units depending on * our sleep mode. */ - if (sleep_ok != 0) + if (sleep_ok != 0) { timeout = (time_ms * hz) / 1000; - else + } else { timeout = time_ms * 2; + } req->state |= REQ_STATE_NEED_WAKEUP; mask &= ~REQ_STATE_NEED_WAKEUP; saved_cnt = mpt->reset_cnt; - while ((req->state & mask) != state - && mpt->reset_cnt == saved_cnt) { - + while ((req->state & mask) != state && mpt->reset_cnt == saved_cnt) { if (sleep_ok != 0) { error = mpt_sleep(mpt, req, PUSER, "mptreq", timeout); if (error == EWOULDBLOCK) { @@ -1126,7 +1225,6 @@ mpt_wait_req(struct mpt_softc *mpt, request_t *req, } } else { if (time_ms != 0 && --timeout == 0) { - mpt_prt(mpt, "mpt_wait_req timed out\n"); break; } DELAY(500); @@ -1134,10 +1232,14 @@ mpt_wait_req(struct mpt_softc *mpt, request_t *req, } } req->state &= ~REQ_STATE_NEED_WAKEUP; - if (mpt->reset_cnt != saved_cnt) + if (mpt->reset_cnt != saved_cnt) { return (EIO); - if (time_ms && timeout <= 0) + } + if (time_ms && timeout <= 0) { + MSG_REQUEST_HEADER *msg_hdr = req->req_vbuf; + mpt_prt(mpt, "mpt_wait_req(%x) timed out\n", msg_hdr->Function); return (ETIMEDOUT); + } return (0); } @@ -1436,8 +1538,6 @@ mpt_read_cfg_header(struct mpt_softc *mpt, int PageType, int PageNumber, return (error); } -#define CFG_DATA_OFF 128 - int mpt_read_cfg_page(struct mpt_softc *mpt, int Action, uint32_t PageAddress, CONFIG_PAGE_HEADER *hdr, size_t len, int sleep_ok, @@ -1455,7 +1555,7 @@ mpt_read_cfg_page(struct mpt_softc *mpt, int Action, uint32_t PageAddress, error = mpt_issue_cfg_req(mpt, req, Action, hdr->PageVersion, hdr->PageLength, hdr->PageNumber, hdr->PageType & MPI_CONFIG_PAGETYPE_MASK, - PageAddress, req->req_pbuf + CFG_DATA_OFF, + PageAddress, req->req_pbuf + MPT_RQSL(mpt), len, sleep_ok, timeout_ms); if (error != 0) { mpt_prt(mpt, "read_cfg_page(%d) timed out\n", Action); @@ -1470,7 +1570,7 @@ mpt_read_cfg_page(struct mpt_softc *mpt, int Action, uint32_t PageAddress, } bus_dmamap_sync(mpt->request_dmat, mpt->request_dmap, BUS_DMASYNC_POSTREAD); - memcpy(hdr, ((uint8_t *)req->req_vbuf)+CFG_DATA_OFF, len); + memcpy(hdr, ((uint8_t *)req->req_vbuf)+MPT_RQSL(mpt), len); mpt_free_request(mpt, req); return (0); } @@ -1497,14 +1597,14 @@ mpt_write_cfg_page(struct mpt_softc *mpt, int Action, uint32_t PageAddress, if (req == NULL) return (-1); - memcpy(((caddr_t)req->req_vbuf)+CFG_DATA_OFF, hdr, len); + memcpy(((caddr_t)req->req_vbuf)+MPT_RQSL(mpt), hdr, len); /* Restore stripped out attributes */ hdr->PageType |= hdr_attr; error = mpt_issue_cfg_req(mpt, req, Action, hdr->PageVersion, hdr->PageLength, hdr->PageNumber, hdr->PageType & MPI_CONFIG_PAGETYPE_MASK, - PageAddress, req->req_pbuf + CFG_DATA_OFF, + PageAddress, req->req_pbuf + MPT_RQSL(mpt), len, sleep_ok, timeout_ms); if (error != 0) { mpt_prt(mpt, "mpt_write_cfg_page timed out\n"); @@ -1673,236 +1773,6 @@ mpt_read_config_info_ioc(struct mpt_softc *mpt) return (0); } -/* - * Read SCSI configuration information - */ -static int -mpt_read_config_info_spi(struct mpt_softc *mpt) -{ - int rv, i; - - rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 0, - 0, &mpt->mpt_port_page0.Header, - /*sleep_ok*/FALSE, /*timeout_ms*/5000); - if (rv) - return (-1); - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Port Page 0 Header: %x %x %x %x\n", - mpt->mpt_port_page0.Header.PageVersion, - mpt->mpt_port_page0.Header.PageLength, - mpt->mpt_port_page0.Header.PageNumber, - mpt->mpt_port_page0.Header.PageType); - - rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 1, - 0, &mpt->mpt_port_page1.Header, - /*sleep_ok*/FALSE, /*timeout_ms*/5000); - if (rv) - return (-1); - - mpt_lprt(mpt, MPT_PRT_DEBUG, "SPI Port Page 1 Header: %x %x %x %x\n", - mpt->mpt_port_page1.Header.PageVersion, - mpt->mpt_port_page1.Header.PageLength, - mpt->mpt_port_page1.Header.PageNumber, - mpt->mpt_port_page1.Header.PageType); - - rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 2, - /*PageAddress*/0, &mpt->mpt_port_page2.Header, - /*sleep_ok*/FALSE, /*timeout_ms*/5000); - if (rv) - return (-1); - - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Port Page 2 Header: %x %x %x %x\n", - mpt->mpt_port_page1.Header.PageVersion, - mpt->mpt_port_page1.Header.PageLength, - mpt->mpt_port_page1.Header.PageNumber, - mpt->mpt_port_page1.Header.PageType); - - for (i = 0; i < 16; i++) { - rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_DEVICE, - 0, i, &mpt->mpt_dev_page0[i].Header, - /*sleep_ok*/FALSE, /*timeout_ms*/5000); - if (rv) - return (-1); - - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Target %d Device Page 0 Header: %x %x %x %x\n", - i, mpt->mpt_dev_page0[i].Header.PageVersion, - mpt->mpt_dev_page0[i].Header.PageLength, - mpt->mpt_dev_page0[i].Header.PageNumber, - mpt->mpt_dev_page0[i].Header.PageType); - - rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_DEVICE, - 1, i, &mpt->mpt_dev_page1[i].Header, - /*sleep_ok*/FALSE, /*timeout_ms*/5000); - if (rv) - return (-1); - - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Target %d Device Page 1 Header: %x %x %x %x\n", - i, mpt->mpt_dev_page1[i].Header.PageVersion, - mpt->mpt_dev_page1[i].Header.PageLength, - mpt->mpt_dev_page1[i].Header.PageNumber, - mpt->mpt_dev_page1[i].Header.PageType); - } - - /* - * At this point, we don't *have* to fail. As long as we have - * valid config header information, we can (barely) lurch - * along. - */ - - rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0, - &mpt->mpt_port_page0.Header, - sizeof(mpt->mpt_port_page0), - /*sleep_ok*/FALSE, /*timeout_ms*/5000); - if (rv) { - mpt_prt(mpt, "failed to read SPI Port Page 0\n"); - } else { - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Port Page 0: Capabilities %x PhysicalInterface %x\n", - mpt->mpt_port_page0.Capabilities, - mpt->mpt_port_page0.PhysicalInterface); - } - - rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0, - &mpt->mpt_port_page1.Header, - sizeof(mpt->mpt_port_page1), - /*sleep_ok*/FALSE, /*timeout_ms*/5000); - if (rv) { - mpt_prt(mpt, "failed to read SPI Port Page 1\n"); - } else { - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Port Page 1: Configuration %x OnBusTimerValue %x\n", - mpt->mpt_port_page1.Configuration, - mpt->mpt_port_page1.OnBusTimerValue); - } - - rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0, - &mpt->mpt_port_page2.Header, - sizeof(mpt->mpt_port_page2), - /*sleep_ok*/FALSE, /*timeout_ms*/5000); - if (rv) { - mpt_prt(mpt, "failed to read SPI Port Page 2\n"); - } else { - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Port Page 2: Flags %x Settings %x\n", - mpt->mpt_port_page2.PortFlags, - mpt->mpt_port_page2.PortSettings); - for (i = 0; i < 16; i++) { - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Port Page 2 Tgt %d: timo %x SF %x Flags %x\n", - i, mpt->mpt_port_page2.DeviceSettings[i].Timeout, - mpt->mpt_port_page2.DeviceSettings[i].SyncFactor, - mpt->mpt_port_page2.DeviceSettings[i].DeviceFlags); - } - } - - for (i = 0; i < 16; i++) { - rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i, - &mpt->mpt_dev_page0[i].Header, - sizeof(*mpt->mpt_dev_page0), - /*sleep_ok*/FALSE, - /*timeout_ms*/5000); - if (rv) { - mpt_prt(mpt, - "cannot read SPI Tgt %d Device Page 0\n", i); - continue; - } - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Tgt %d Page 0: NParms %x Information %x", - i, mpt->mpt_dev_page0[i].NegotiatedParameters, - mpt->mpt_dev_page0[i].Information); - - rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i, - &mpt->mpt_dev_page1[i].Header, - sizeof(*mpt->mpt_dev_page1), - /*sleep_ok*/FALSE, - /*timeout_ms*/5000); - if (rv) { - mpt_prt(mpt, - "cannot read SPI Tgt %d Device Page 1\n", i); - continue; - } - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Tgt %d Page 1: RParms %x Configuration %x\n", - i, mpt->mpt_dev_page1[i].RequestedParameters, - mpt->mpt_dev_page1[i].Configuration); - } - return (0); -} - -/* - * Validate SPI configuration information. - * - * In particular, validate SPI Port Page 1. - */ -static int -mpt_set_initial_config_spi(struct mpt_softc *mpt) -{ - int i, pp1val = ((1 << mpt->mpt_ini_id) << 16) | mpt->mpt_ini_id; - int error; - - mpt->mpt_disc_enable = 0xff; - mpt->mpt_tag_enable = 0; - - if (mpt->mpt_port_page1.Configuration != pp1val) { - CONFIG_PAGE_SCSI_PORT_1 tmp; - - mpt_prt(mpt, - "SPI Port Page 1 Config value bad (%x)- should be %x\n", - mpt->mpt_port_page1.Configuration, pp1val); - tmp = mpt->mpt_port_page1; - tmp.Configuration = pp1val; - error = mpt_write_cur_cfg_page(mpt, /*PageAddress*/0, - &tmp.Header, sizeof(tmp), - /*sleep_ok*/FALSE, - /*timeout_ms*/5000); - if (error) - return (-1); - error = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0, - &tmp.Header, sizeof(tmp), - /*sleep_ok*/FALSE, - /*timeout_ms*/5000); - if (error) - return (-1); - if (tmp.Configuration != pp1val) { - mpt_prt(mpt, - "failed to reset SPI Port Page 1 Config value\n"); - return (-1); - } - mpt->mpt_port_page1 = tmp; - } - - for (i = 0; i < 16; i++) { - CONFIG_PAGE_SCSI_DEVICE_1 tmp; - tmp = mpt->mpt_dev_page1[i]; - tmp.RequestedParameters = 0; - tmp.Configuration = 0; - mpt_lprt(mpt, MPT_PRT_DEBUG, - "Set Tgt %d SPI DevicePage 1 values to %x 0 %x\n", - i, tmp.RequestedParameters, tmp.Configuration); - error = mpt_write_cur_cfg_page(mpt, /*PageAddress*/i, - &tmp.Header, sizeof(tmp), - /*sleep_ok*/FALSE, - /*timeout_ms*/5000); - if (error) - return (-1); - error = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i, - &tmp.Header, sizeof(tmp), - /*sleep_ok*/FALSE, - /*timeout_ms*/5000); - if (error) - return (-1); - mpt->mpt_dev_page1[i] = tmp; - mpt_lprt(mpt, MPT_PRT_DEBUG, - "SPI Tgt %d Page 1: RParm %x Configuration %x\n", i, - mpt->mpt_dev_page1[i].RequestedParameters, - mpt->mpt_dev_page1[i].Configuration); - } - return (0); -} - /* * Enable IOC port */ @@ -1918,7 +1788,7 @@ mpt_send_port_enable(struct mpt_softc *mpt, int port) return (-1); enable_req = req->req_vbuf; - bzero(enable_req, sizeof *enable_req); + bzero(enable_req, MPT_RQSL(mpt)); enable_req->Function = MPI_FUNCTION_PORT_ENABLE; enable_req->MsgContext = htole32(req->index | MPT_REPLY_HANDLER_CONFIG); @@ -1932,10 +1802,11 @@ mpt_send_port_enable(struct mpt_softc *mpt, int port) /*sleep_ok*/FALSE, /*time_ms*/(mpt->is_sas || mpt->is_fc)? 30000 : 3000); if (error != 0) { - mpt_prt(mpt, "port enable timed out\n"); + mpt_prt(mpt, "port %d enable timed out\n", port); return (-1); } mpt_free_request(mpt, req); + mpt_lprt(mpt, MPT_PRT_DEBUG, "enabled port %d\n", port); return (0); } @@ -1992,27 +1863,28 @@ mpt_disable_ints(struct mpt_softc *mpt) static void mpt_sysctl_attach(struct mpt_softc *mpt) { +#if __FreeBSD_version >= 500000 struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(mpt->dev); struct sysctl_oid *tree = device_get_sysctl_tree(mpt->dev); SYSCTL_ADD_INT(ctx, SYSCTL_CHILDREN(tree), OID_AUTO, "debug", CTLFLAG_RW, &mpt->verbose, 0, "Debugging/Verbose level"); +#endif } int mpt_attach(struct mpt_softc *mpt) { + struct mpt_personality *pers; int i; + int error; for (i = 0; i < MPT_MAX_PERSONALITIES; i++) { - struct mpt_personality *pers; - int error; - pers = mpt_personalities[i]; - if (pers == NULL) + if (pers == NULL) { continue; - + } if (pers->probe(mpt) == 0) { error = pers->attach(mpt); if (error != 0) { @@ -2024,6 +1896,24 @@ mpt_attach(struct mpt_softc *mpt) } } + /* + * Now that we've attached everything, do the enable function + * for all of the personalities. This allows the personalities + * to do setups that are appropriate for them prior to enabling + * any ports. + */ + for (i = 0; i < MPT_MAX_PERSONALITIES; i++) { + pers = mpt_personalities[i]; + if (pers != NULL && MPT_PERS_ATTACHED(pers, mpt) != 0) { + error = pers->enable(mpt); + if (error != 0) { + mpt_prt(mpt, "personality %s attached but would" + " not enable (%d)\n", pers->name, error); + mpt_detach(mpt); + return (error); + } + } + } return (0); } @@ -2032,10 +1922,9 @@ mpt_shutdown(struct mpt_softc *mpt) { struct mpt_personality *pers; - MPT_PERS_FOREACH_REVERSE(mpt, pers) + MPT_PERS_FOREACH_REVERSE(mpt, pers) { pers->shutdown(mpt); - - mpt_reset(mpt, /*reinit*/FALSE); + } return (0); } @@ -2062,8 +1951,9 @@ mpt_core_load(struct mpt_personality *pers) * Setup core handlers and insert the default handler * into all "empty slots". */ - for (i = 0; i < MPT_NUM_REPLY_HANDLERS; i++) + for (i = 0; i < MPT_NUM_REPLY_HANDLERS; i++) { mpt_reply_handlers[i] = mpt_default_reply_handler; + } mpt_reply_handlers[MPT_CBI(MPT_REPLY_HANDLER_EVENTS)] = mpt_event_reply_handler; @@ -2071,7 +1961,6 @@ mpt_core_load(struct mpt_personality *pers) mpt_config_reply_handler; mpt_reply_handlers[MPT_CBI(MPT_REPLY_HANDLER_HANDSHAKE)] = mpt_handshake_reply_handler; - return (0); } @@ -2085,32 +1974,92 @@ mpt_core_attach(struct mpt_softc *mpt) int val; int error; + LIST_INIT(&mpt->ack_frames); /* Put all request buffers on the free list */ TAILQ_INIT(&mpt->request_pending_list); TAILQ_INIT(&mpt->request_free_list); - for (val = 0; val < MPT_MAX_REQUESTS(mpt); val++) + for (val = 0; val < MPT_MAX_REQUESTS(mpt); val++) { mpt_free_request(mpt, &mpt->request_pool[val]); + } + + for (val = 0; val < MPT_MAX_LUNS; val++) { + STAILQ_INIT(&mpt->trt[val].atios); + STAILQ_INIT(&mpt->trt[val].inots); + } + STAILQ_INIT(&mpt->trt_wildcard.atios); + STAILQ_INIT(&mpt->trt_wildcard.inots); + + mpt->scsi_tgt_handler_id = MPT_HANDLER_ID_NONE; mpt_sysctl_attach(mpt); mpt_lprt(mpt, MPT_PRT_DEBUG, "doorbell req = %s\n", - mpt_ioc_diag(mpt_read(mpt, MPT_OFFSET_DOORBELL))); + mpt_ioc_diag(mpt_read(mpt, MPT_OFFSET_DOORBELL))); error = mpt_configure_ioc(mpt); return (error); } +int +mpt_core_enable(struct mpt_softc *mpt) +{ + /* + * We enter with the IOC enabled, but async events + * not enabled, ports not enabled and interrupts + * not enabled. + */ + + /* + * Enable asynchronous event reporting- all personalities + * have attached so that they should be able to now field + * async events. + */ + mpt_send_event_request(mpt, 1); + + /* + * Catch any pending interrupts + * + * This seems to be crucial- otherwise + * the portenable below times out. + */ + mpt_intr(mpt); + + /* + * Enable Interrupts + */ + mpt_enable_ints(mpt); + + /* + * Catch any pending interrupts + * + * This seems to be crucial- otherwise + * the portenable below times out. + */ + mpt_intr(mpt); + + /* + * Enable the port- but only if we are not MPT_ROLE_NONE. + */ + if (mpt_send_port_enable(mpt, 0) != MPT_OK) { + mpt_prt(mpt, "failed to enable port 0\n"); + return (ENXIO); + } + return (0); +} + void mpt_core_shutdown(struct mpt_softc *mpt) { + mpt_disable_ints(mpt); } void mpt_core_detach(struct mpt_softc *mpt) { + mpt_disable_ints(mpt); } int @@ -2262,10 +2211,11 @@ mpt_configure_ioc(struct mpt_softc *mpt) * first channel is ok, the second will not require a hard * reset. */ - if (needreset || (mpt_rd_db(mpt) & MPT_DB_STATE_MASK) != + if (needreset || MPT_STATE(mpt_rd_db(mpt)) != MPT_DB_STATE_READY) { - if (mpt_reset(mpt, /*reinit*/FALSE) != MPT_OK) + if (mpt_reset(mpt, FALSE) != MPT_OK) { continue; + } } needreset = 0; @@ -2408,10 +2358,8 @@ mpt_configure_ioc(struct mpt_softc *mpt) pfp.PortType); return (ENXIO); } - if (!(pfp.ProtocolFlags & MPI_PORTFACTS_PROTOCOL_INITIATOR)) { - mpt_prt(mpt, "initiator role unsupported\n"); - return (ENXIO); - } + mpt->mpt_max_tgtcmds = le16toh(pfp.MaxPostedCmdBuffers); + if (pfp.PortType == MPI_PORTFACTS_PORTTYPE_FC) { mpt->is_fc = 1; mpt->is_sas = 0; @@ -2425,30 +2373,49 @@ mpt_configure_ioc(struct mpt_softc *mpt) mpt->mpt_ini_id = pfp.PortSCSIID; mpt->mpt_max_devices = pfp.MaxDevices; - if (mpt_enable_ioc(mpt) != 0) { - mpt_prt(mpt, "Unable to initialize IOC\n"); + /* + * Match our expected role with what this port supports. + * + * We only do this to meet expectations. That is, if the + * user has specified they want initiator role, and we + * don't support it, that's an error we return back upstream. + */ + + mpt->cap = MPT_ROLE_NONE; + if (pfp.ProtocolFlags & MPI_PORTFACTS_PROTOCOL_INITIATOR) { + mpt->cap |= MPT_ROLE_INITIATOR; + } + if (pfp.ProtocolFlags & MPI_PORTFACTS_PROTOCOL_TARGET) { + mpt->cap |= MPT_ROLE_TARGET; + } + if (mpt->cap == MPT_ROLE_NONE) { + mpt_prt(mpt, "port does not support either target or " + "initiator role\n"); + return (ENXIO); + } + + if ((mpt->role & MPT_ROLE_INITIATOR) && + (mpt->cap & MPT_ROLE_INITIATOR) == 0) { + mpt_prt(mpt, "port does not support initiator role\n"); + return (ENXIO); + } + + if ((mpt->role & MPT_ROLE_TARGET) && + (mpt->cap & MPT_ROLE_TARGET) == 0) { + mpt_prt(mpt, "port does not support target role\n"); + return (ENXIO); + } + + if (mpt_enable_ioc(mpt, 0) != MPT_OK) { + mpt_prt(mpt, "unable to initialize IOC\n"); return (ENXIO); } /* - * Read and set up initial configuration information - * (IOC and SPI only for now) - * - * XXX Should figure out what "personalities" are - * available and defer all initialization junk to - * them. + * Read IOC configuration information. */ mpt_read_config_info_ioc(mpt); - if (mpt->is_fc == 0 && mpt->is_sas == 0) { - if (mpt_read_config_info_spi(mpt)) { - return (EIO); - } - if (mpt_set_initial_config_spi(mpt)) { - return (EIO); - } - } - /* Everything worked */ break; } @@ -2458,14 +2425,11 @@ mpt_configure_ioc(struct mpt_softc *mpt) return (EIO); } - mpt_lprt(mpt, MPT_PRT_DEBUG, "enabling interrupts\n"); - - mpt_enable_ints(mpt); return (0); } static int -mpt_enable_ioc(struct mpt_softc *mpt) +mpt_enable_ioc(struct mpt_softc *mpt, int portenable) { uint32_t pptr; int val; @@ -2496,20 +2460,20 @@ mpt_enable_ioc(struct mpt_softc *mpt) break; } - /* - * Enable asynchronous event reporting - */ - mpt_send_event_request(mpt, 1); /* - * Enable the port + * Enable the port if asked */ - if (mpt_send_port_enable(mpt, 0) != MPT_OK) { - mpt_prt(mpt, "failed to enable port 0\n"); - return (ENXIO); + if (portenable) { + /* + * Enable asynchronous event reporting + */ + mpt_send_event_request(mpt, 1); + + if (mpt_send_port_enable(mpt, 0) != MPT_OK) { + mpt_prt(mpt, "failed to enable port 0\n"); + return (ENXIO); + } } - mpt_lprt(mpt, MPT_PRT_DEBUG, "enabled port 0\n"); - - return (MPT_OK); } diff --git a/sys/dev/mpt/mpt.h b/sys/dev/mpt/mpt.h index b88d94b11729..a38731c957e4 100644 --- a/sys/dev/mpt/mpt.h +++ b/sys/dev/mpt/mpt.h @@ -121,6 +121,10 @@ #include #include +#if __FreeBSD_version < 500000 +#include +#endif + #include #if __FreeBSD_version < 500000 @@ -153,6 +157,12 @@ #define NUM_ELEMENTS(array) (sizeof(array) / sizeof(*array)) +#define MPT_ROLE_NONE 0 +#define MPT_ROLE_INITIATOR 1 +#define MPT_ROLE_TARGET 2 +#define MPT_ROLE_BOTH 3 +#define MPT_ROLE_DEFAULT MPT_ROLE_INITIATOR + /**************************** Forward Declarations ****************************/ struct mpt_softc; struct mpt_personality; @@ -162,6 +172,7 @@ typedef struct req_entry request_t; typedef int mpt_load_handler_t(struct mpt_personality *); typedef int mpt_probe_handler_t(struct mpt_softc *); typedef int mpt_attach_handler_t(struct mpt_softc *); +typedef int mpt_enable_handler_t(struct mpt_softc *); typedef int mpt_event_handler_t(struct mpt_softc *, request_t *, MSG_EVENT_NOTIFY_REPLY *); typedef void mpt_reset_handler_t(struct mpt_softc *, int /*type*/); @@ -179,6 +190,7 @@ struct mpt_personality #define MPT_PERS_FIRST_HANDLER(pers) (&(pers)->load) mpt_probe_handler_t *probe; /* configure personailty */ mpt_attach_handler_t *attach; /* initialize device instance */ + mpt_enable_handler_t *enable; /* enable device */ mpt_event_handler_t *event; /* Handle MPI event. */ mpt_reset_handler_t *reset; /* Re-init after reset. */ mpt_shutdown_handler_t *shutdown; /* Shutdown instance. */ @@ -280,7 +292,7 @@ struct req_entry { mpt_req_state_t state; /* Request State Information */ uint16_t index; /* Index of this entry */ uint16_t IOCStatus; /* Completion status */ - uint32_t serno; /* serial number */ + uint16_t serno; /* serial number */ union ccb *ccb; /* CAM request */ void *req_vbuf; /* Virtual Address of Entry */ void *sense_vbuf; /* Virtual Address of sense data */ @@ -290,6 +302,59 @@ struct req_entry { struct req_entry *chain; /* for SGE overallocations */ }; +/**************************** MPI Target State Info ***************************/ + +typedef struct { + uint32_t reply_desc; /* current reply descriptor */ + uint32_t resid; /* current data residual */ + uint32_t bytes_xfered; /* current relative offset */ + union ccb *ccb; /* pointer to currently active ccb */ + request_t *req; /* pointer to currently active assist request */ + int flags; +#define BOGUS_JO 0x01 + int nxfers; + enum { + TGT_STATE_NIL, + TGT_STATE_LOADED, + TGT_STATE_IN_CAM, + TGT_STATE_SETTING_UP_FOR_DATA, + TGT_STATE_MOVING_DATA, + TGT_STATE_MOVING_DATA_AND_STATUS, + TGT_STATE_SENDING_STATUS + } state; +} mpt_tgt_state_t; + +/* + * When we get an incoming command it has its own tag which is called the + * IoIndex. This is the value we gave that particular command buffer when + * we originally assigned it. It's just a number, really. The FC card uses + * it as an RX_ID. We can use it to index into mpt->tgt_cmd_ptrs, which + * contains pointers the request_t structures related to that IoIndex. + * + * What *we* do is construct a tag out of the index for the target command + * which owns the incoming ATIO plus a rolling sequence number. + */ +#define MPT_MAKE_TAGID(mpt, req, ioindex) \ + ((ioindex << 16) | (mpt->sequence++)) + +#ifdef INVARIANTS +#define MPT_TAG_2_REQ(a, b) mpt_tag_2_req(a, (uint32_t) b) +#else +#define MPT_TAG_2_REQ(mpt, tag) mpt->tgt_cmd_ptrs[tag >> 16] +#endif + +#define MPT_TGT_STATE(mpt, req) ((mpt_tgt_state_t *) \ + (&((uint8_t *)req->req_vbuf)[MPT_RQSL(mpt) - sizeof (mpt_tgt_state_t)])) + +STAILQ_HEAD(mpt_hdr_stailq, ccb_hdr); +#define MPT_MAX_LUNS 256 +typedef struct { + struct mpt_hdr_stailq atios; + struct mpt_hdr_stailq inots; + int enabled; +} tgt_resource_t; +#define MPT_MAX_ELS 8 + /**************************** Handler Registration ****************************/ /* * Global table of registered reply handlers. The @@ -311,12 +376,12 @@ struct req_entry { * all commonly executed handlers fit in a single cache * line. */ -#define MPT_NUM_REPLY_HANDLERS (16) +#define MPT_NUM_REPLY_HANDLERS (32) #define MPT_REPLY_HANDLER_EVENTS MPT_CBI_TO_HID(0) #define MPT_REPLY_HANDLER_CONFIG MPT_CBI_TO_HID(MPT_NUM_REPLY_HANDLERS-1) #define MPT_REPLY_HANDLER_HANDSHAKE MPT_CBI_TO_HID(MPT_NUM_REPLY_HANDLERS-2) typedef int mpt_reply_handler_t(struct mpt_softc *mpt, request_t *request, - MSG_DEFAULT_REPLY *reply_frame); + uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame); typedef union { mpt_reply_handler_t *reply_handler; } mpt_handler_t; @@ -417,21 +482,24 @@ struct mpt_softc { struct mtx mpt_lock; #endif uint32_t mpt_pers_mask; - uint32_t : 14, - is_sas : 1, + uint32_t : 8, + unit : 8, + : 1, + twildcard : 1, + tenabled : 1, + cap : 2, /* none, ini, target, both */ + role : 2, /* none, ini, target, both */ raid_mwce_set : 1, getreqwaiter : 1, shutdwn_raid : 1, shutdwn_recovery: 1, - unit : 8, outofbeer : 1, mpt_locksetup : 1, disabled : 1, - is_fc : 1, - bus : 1; /* FC929/1030 have two busses */ + is_sas : 1, + is_fc : 1; u_int verbose; - uint32_t cmd_serno; /* * IOC Facts @@ -449,6 +517,7 @@ struct mpt_softc { uint16_t mpt_ini_id; uint16_t mpt_port_type; uint16_t mpt_proto_flags; + uint16_t mpt_max_tgtcmds; /* * Device Configuration Information @@ -475,7 +544,8 @@ struct mpt_softc { #define mpt_update_params0 cfg.spi._update_params0 #define mpt_update_params1 cfg.spi._update_params1 struct mpi_fc_cfg { - uint8_t nada; + CONFIG_PAGE_FC_PORT_0 _port_page0; +#define mpt_fcport_page0 cfg.fc._port_page0 } fc; } cfg; @@ -528,7 +598,7 @@ struct mpt_softc { bus_dma_tag_t request_dmat; /* DMA tag for request memroy */ bus_dmamap_t request_dmap; /* DMA map for request memroy */ uint8_t *request; /* KVA of Request memory */ - bus_addr_t request_phys; /* BusADdr of request memory */ + bus_addr_t request_phys; /* BusAddr of request memory */ uint32_t max_seg_cnt; /* calculated after IOC facts */ @@ -560,9 +630,28 @@ struct mpt_softc { struct proc *recovery_thread; request_t *tmf_req; - uint32_t sequence; /* Sequence Number */ - uint32_t timeouts; /* timeout count */ - uint32_t success; /* successes afer timeout */ + /* + * Target Mode Support + */ + uint32_t scsi_tgt_handler_id; + request_t ** tgt_cmd_ptrs; + + /* + * *snork*- this is chosen to be here *just in case* somebody + * forgets to point to it exactly and we index off of trt with + * CAM_LUN_WILDCARD. + */ + tgt_resource_t trt_wildcard; /* wildcard luns */ + tgt_resource_t trt[MPT_MAX_LUNS]; + uint16_t tgt_cmds_allocated; + + /* + * Stuff.. + */ + uint16_t sequence; /* Sequence Number */ + uint16_t timeouts; /* timeout count */ + uint16_t success; /* successes afer timeout */ + /* Opposing port in a 929 or 1030, or NULL */ struct mpt_softc * mpt2; @@ -580,7 +669,7 @@ struct mpt_softc { TAILQ_ENTRY(mpt_softc) links; }; -/***************************** Locking Primatives *****************************/ +/***************************** Locking Primitives *****************************/ #if __FreeBSD_version < 500000 #define MPT_IFLAGS INTR_TYPE_CAM #define MPT_LOCK(mpt) mpt_lockspl(mpt) @@ -631,7 +720,7 @@ mpt_sleep(struct mpt_softc *mpt, void *ident, int priority, saved_spl = mpt->mpt_splsaved; mpt->mpt_islocked = 0; error = tsleep(ident, priority, wmesg, timo); - KASSERT(mpt->mpt_islocked = 0, ("Invalid lock count on wakeup")); + KASSERT(mpt->mpt_islocked == 0, ("Invalid lock count on wakeup")); mpt->mpt_islocked = saved_cnt; mpt->mpt_splsaved = saved_spl; return (error); @@ -709,19 +798,27 @@ mpt_pio_read(struct mpt_softc *mpt, int offset) /* Max MPT Reply we are willing to accept (must be power of 2) */ #define MPT_REPLY_SIZE 256 +/* + * Must be less than 16384 in order for target mode to work + */ #define MPT_MAX_REQUESTS(mpt) 512 #define MPT_REQUEST_AREA 512 #define MPT_SENSE_SIZE 32 /* included in MPT_REQUEST_AREA */ #define MPT_REQ_MEM_SIZE(mpt) (MPT_MAX_REQUESTS(mpt) * MPT_REQUEST_AREA) -#define MPT_CONTEXT_CB_SHIFT (16) -#define MPT_CBI(handle) (handle >> MPT_CONTEXT_CB_SHIFT) +/* + * Currently we try to pack both callbacks and request indices into 14 bits + * so that we don't have to get fancy when we get a target mode context + * reply (which only has 14 bits of IoIndex value) or a normal scsi + * initiator context reply (where we get bits 28..0 of context). + */ +#define MPT_CONTEXT_CB_SHIFT (14) +#define MPT_CBI(handle) (handle >> MPT_CONTEXT_CB_SHIFT) #define MPT_CBI_TO_HID(cbi) ((cbi) << MPT_CONTEXT_CB_SHIFT) #define MPT_CONTEXT_TO_CBI(x) \ (((x) >> MPT_CONTEXT_CB_SHIFT) & (MPT_NUM_REPLY_HANDLERS - 1)) -#define MPT_CONTEXT_REQI_MASK 0xFFFF -#define MPT_CONTEXT_TO_REQI(x) \ - ((x) & MPT_CONTEXT_REQI_MASK) +#define MPT_CONTEXT_REQI_MASK 0x3FFF +#define MPT_CONTEXT_TO_REQI(x) ((x) & MPT_CONTEXT_REQI_MASK) /* * Convert a 32bit physical address returned from IOC to an @@ -786,7 +883,7 @@ void mpt_complete_request_chain(struct mpt_softc *mpt, /***************************** IOC Initialization *****************************/ int mpt_reset(struct mpt_softc *, int /*reinit*/); -/****************************** Debugging/Logging *****************************/ +/****************************** Debugging ************************************/ typedef struct mpt_decode_entry { char *name; u_int value; @@ -804,10 +901,14 @@ enum { MPT_PRT_WARN, MPT_PRT_INFO, MPT_PRT_DEBUG, + MPT_PRT_DEBUG1, + MPT_PRT_DEBUG2, + MPT_PRT_DEBUG3, MPT_PRT_TRACE, MPT_PRT_NONE=100 }; +#if __FreeBSD_version > 500000 #define mpt_lprt(mpt, level, ...) \ do { \ if (level <= (mpt)->verbose) \ @@ -819,9 +920,56 @@ do { \ if (level <= (mpt)->debug_level) \ mpt_prtc(mpt, __VA_ARGS__); \ } while (0) +#else +void mpt_lprt(struct mpt_softc *, int, const char *, ...) + __printflike(3, 4); +void mpt_lprtc(struct mpt_softc *, int, const char *, ...) + __printflike(3, 4); +#endif +void mpt_prt(struct mpt_softc *, const char *, ...) + __printflike(2, 3); +void mpt_prtc(struct mpt_softc *, const char *, ...) + __printflike(2, 3); -void mpt_prt(struct mpt_softc *, const char *, ...); -void mpt_prtc(struct mpt_softc *, const char *, ...); +/**************************** Target Mode Related ***************************/ +static __inline int mpt_cdblen(uint8_t, int); +static __inline int +mpt_cdblen(uint8_t cdb0, int maxlen) +{ + int group = cdb0 >> 5; + switch (group) { + case 0: + return (6); + case 1: + return (10); + case 4: + case 5: + return (12); + default: + return (16); + } +} +#ifdef INVARIANTS +static __inline request_t * mpt_tag_2_req(struct mpt_softc *, uint32_t); +static __inline request_t * +mpt_tag_2_req(struct mpt_softc *mpt, uint32_t tag) +{ + uint16_t rtg = (tag >> 16); + KASSERT(rtg < mpt->tgt_cmds_allocated, ("bad tag %d\n", tag)); + KASSERT(mpt->tgt_cmd_ptrs, ("no cmd backpointer array")); + KASSERT(mpt->tgt_cmd_ptrs[rtg], ("no cmd backpointer")); + return (mpt->tgt_cmd_ptrs[rtg]); +} +#endif + +typedef enum { + MPT_ABORT_TASK_SET=1234, + MPT_CLEAR_TASK_SET, + MPT_TARGET_RESET, + MPT_CLEAR_ACA, + MPT_TERMINATE_TASK, + MPT_NIL_TMT_VALUE=5678 +} mpt_task_mgmt_t; /**************************** Unclassified Routines ***************************/ void mpt_send_cmd(struct mpt_softc *mpt, request_t *req); diff --git a/sys/dev/mpt/mpt_cam.c b/sys/dev/mpt/mpt_cam.c index 7785f7d5fe27..95d8438431c0 100644 --- a/sys/dev/mpt/mpt_cam.c +++ b/sys/dev/mpt/mpt_cam.c @@ -100,6 +100,7 @@ __FBSDID("$FreeBSD$"); #include "dev/mpt/mpilib/mpi_ioc.h" /* XXX Fix Event Handling!!! */ #include "dev/mpt/mpilib/mpi_init.h" #include "dev/mpt/mpilib/mpi_targ.h" +#include "dev/mpt/mpilib/mpi_fc.h" #include #include @@ -112,9 +113,12 @@ static int mpt_setsync(struct mpt_softc *, int, int, int); static void mpt_calc_geometry(struct ccb_calc_geometry *ccg, int extended); static mpt_reply_handler_t mpt_scsi_reply_handler; static mpt_reply_handler_t mpt_scsi_tmf_reply_handler; +static mpt_reply_handler_t mpt_fc_els_reply_handler; +static mpt_reply_handler_t mpt_scsi_tgt_reply_handler; static int mpt_scsi_reply_frame_handler(struct mpt_softc *mpt, request_t *req, MSG_DEFAULT_REPLY *reply_frame); static int mpt_bus_reset(struct mpt_softc *, int /*sleep_ok*/); +static int mpt_fc_reset_link(struct mpt_softc *, int); static int mpt_spawn_recovery_thread(struct mpt_softc *mpt); static void mpt_terminate_recovery_thread(struct mpt_softc *mpt); @@ -123,13 +127,25 @@ static int mpt_scsi_send_tmf(struct mpt_softc *, u_int /*type*/, u_int /*flags*/, u_int /*channel*/, u_int /*target*/, u_int /*lun*/, u_int /*abort_ctx*/, int /*sleep_ok*/); + +static void mpt_fc_add_els(struct mpt_softc *mpt, request_t *); +static void mpt_post_target_command(struct mpt_softc *, request_t *, int); +static void mpt_add_target_commands(struct mpt_softc *mpt); +static int mpt_enable_lun(struct mpt_softc *, target_id_t, lun_id_t); +static int mpt_disable_lun(struct mpt_softc *, target_id_t, lun_id_t); +static void mpt_target_start_io(struct mpt_softc *, union ccb *); +static cam_status mpt_abort_target_ccb(struct mpt_softc *, union ccb *); +static cam_status mpt_abort_target_cmd(struct mpt_softc *, request_t *); + static void mpt_recover_commands(struct mpt_softc *mpt); static uint32_t scsi_io_handler_id = MPT_HANDLER_ID_NONE; static uint32_t scsi_tmf_handler_id = MPT_HANDLER_ID_NONE; +static uint32_t fc_els_handler_id = MPT_HANDLER_ID_NONE; static mpt_probe_handler_t mpt_cam_probe; static mpt_attach_handler_t mpt_cam_attach; +static mpt_enable_handler_t mpt_cam_enable; static mpt_event_handler_t mpt_cam_event; static mpt_reset_handler_t mpt_cam_ioc_reset; static mpt_detach_handler_t mpt_cam_detach; @@ -139,6 +155,7 @@ static struct mpt_personality mpt_cam_personality = .name = "mpt_cam", .probe = mpt_cam_probe, .attach = mpt_cam_attach, + .enable = mpt_cam_enable, .event = mpt_cam_event, .reset = mpt_cam_ioc_reset, .detach = mpt_cam_detach, @@ -150,13 +167,14 @@ int mpt_cam_probe(struct mpt_softc *mpt) { /* - * Only attach to nodes that support the initiator - * role or have RAID physical devices that need - * CAM pass-thru support. + * Only attach to nodes that support the initiator or target + * role or have RAID physical devices that need CAM pass-thru support. */ if ((mpt->mpt_proto_flags & MPI_PORTFACTS_PROTOCOL_INITIATOR) != 0 - || (mpt->ioc_page2 != NULL && mpt->ioc_page2->MaxPhysDisks != 0)) + || (mpt->mpt_proto_flags & MPI_PORTFACTS_PROTOCOL_TARGET) != 0 + || (mpt->ioc_page2 != NULL && mpt->ioc_page2->MaxPhysDisks != 0)) { return (0); + } return (ENODEV); } @@ -168,22 +186,73 @@ mpt_cam_attach(struct mpt_softc *mpt) int maxq; int error; - MPTLOCK_2_CAMLOCK(mpt); TAILQ_INIT(&mpt->request_timeout_list); - mpt->bus = 0; maxq = (mpt->mpt_global_credits < MPT_MAX_REQUESTS(mpt))? mpt->mpt_global_credits : MPT_MAX_REQUESTS(mpt); handler.reply_handler = mpt_scsi_reply_handler; error = mpt_register_handler(mpt, MPT_HANDLER_REPLY, handler, &scsi_io_handler_id); - if (error != 0) + if (error != 0) { goto cleanup; + } + handler.reply_handler = mpt_scsi_tmf_reply_handler; error = mpt_register_handler(mpt, MPT_HANDLER_REPLY, handler, &scsi_tmf_handler_id); - if (error != 0) + if (error != 0) { goto cleanup; + } + + /* + * We keep two requests reserved for ELS replies/responses + * if we're fibre channel and target mode. + */ + if (mpt->is_fc && (mpt->role & MPT_ROLE_TARGET) != 0) { + request_t *req; + int i; + + handler.reply_handler = mpt_fc_els_reply_handler; + error = mpt_register_handler(mpt, MPT_HANDLER_REPLY, handler, + &fc_els_handler_id); + if (error != 0) { + goto cleanup; + } + + /* + * Feed the chip some ELS buffer resources + */ + for (i = 0; i < MPT_MAX_ELS; i++) { + req = mpt_get_request(mpt, FALSE); + if (req == NULL) { + break; + } + mpt_fc_add_els(mpt, req); + } + if (i == 0) { + mpt_prt(mpt, "Unable to add ELS buffer resources\n"); + goto cleanup; + } + maxq -= i; + } + + /* + * If we're in target mode, register a reply + * handler for it and add some commands. + */ + if ((mpt->role & MPT_ROLE_TARGET) != 0) { + handler.reply_handler = mpt_scsi_tgt_reply_handler; + error = mpt_register_handler(mpt, MPT_HANDLER_REPLY, handler, + &mpt->scsi_tgt_handler_id); + if (error != 0) { + goto cleanup; + } + + /* + * Add some target command resources + */ + mpt_add_target_commands(mpt); + } /* * We keep one request reserved for timeout TMF requests. @@ -252,9 +321,9 @@ mpt_cam_attach(struct mpt_softc *mpt) * Only register a second bus for RAID physical * devices if the controller supports RAID. */ - if (mpt->ioc_page2 == NULL - || mpt->ioc_page2->MaxPhysDisks == 0) + if (mpt->ioc_page2 == NULL || mpt->ioc_page2->MaxPhysDisks == 0) { return (0); + } /* * Create a "bus" to export all hidden disks to CAM. @@ -283,15 +352,412 @@ mpt_cam_attach(struct mpt_softc *mpt) error = ENOMEM; goto cleanup; } - - CAMLOCK_2_MPTLOCK(mpt); return (0); cleanup: - CAMLOCK_2_MPTLOCK(mpt); mpt_cam_detach(mpt); return (error); } +/* + * Read FC configuration information + */ +static int +mpt_read_config_info_fc(struct mpt_softc *mpt) +{ + char *topology = NULL; + int rv, speed = 0; + + rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_FC_PORT, 0, + 0, &mpt->mpt_fcport_page0.Header, FALSE, 5000); + if (rv) { + return (-1); + } + mpt_lprt(mpt, MPT_PRT_DEBUG, "FC Port Page 0 Header: %x %x %x %x\n", + mpt->mpt_fcport_page0.Header.PageVersion, + mpt->mpt_fcport_page0.Header.PageLength, + mpt->mpt_fcport_page0.Header.PageNumber, + mpt->mpt_fcport_page0.Header.PageType); + + + rv = mpt_read_cur_cfg_page(mpt, 0, &mpt->mpt_fcport_page0.Header, + sizeof(mpt->mpt_fcport_page0), FALSE, 5000); + if (rv) { + mpt_prt(mpt, "failed to read FC Port Page 0\n"); + return (-1); + } + + speed = mpt->mpt_fcport_page0.CurrentSpeed; + + switch (mpt->mpt_fcport_page0.Flags & + MPI_FCPORTPAGE0_FLAGS_ATTACH_TYPE_MASK) { + case MPI_FCPORTPAGE0_FLAGS_ATTACH_NO_INIT: + speed = 0; + topology = ""; + break; + case MPI_FCPORTPAGE0_FLAGS_ATTACH_POINT_TO_POINT: + topology = "N-Port"; + break; + case MPI_FCPORTPAGE0_FLAGS_ATTACH_PRIVATE_LOOP: + topology = "NL-Port"; + break; + case MPI_FCPORTPAGE0_FLAGS_ATTACH_FABRIC_DIRECT: + topology = "F-Port"; + break; + case MPI_FCPORTPAGE0_FLAGS_ATTACH_PUBLIC_LOOP: + topology = "FL-Port"; + break; + default: + speed = 0; + topology = "?"; + break; + } + + mpt_lprt(mpt, MPT_PRT_INFO, + "FC Port Page 0: Topology <%s> WWNN 0x%08x%08x WWPN 0x%08x%08x " + "Speed %u-Gbit\n", topology, + mpt->mpt_fcport_page0.WWNN.High, + mpt->mpt_fcport_page0.WWNN.Low, + mpt->mpt_fcport_page0.WWPN.High, + mpt->mpt_fcport_page0.WWPN.Low, + speed); + + return (0); +} + +/* + * Set FC configuration information. + */ +static int +mpt_set_initial_config_fc(struct mpt_softc *mpt) +{ +#if 0 + CONFIG_PAGE_FC_PORT_1 fc; + U32 fl; + int r, doit = 0; + + r = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_FC_PORT, 1, 0, + &fc.Header, FALSE, 5000); + if (r) { + return (mpt_fc_reset_link(mpt, 1)); + } + + r = mpt_read_cfg_page(mpt, MPI_CONFIG_ACTION_PAGE_READ_CURRENT, 0, + &fc.Header, sizeof (fc), FALSE, 5000); + if (r) { + return (mpt_fc_reset_link(mpt, 1)); + } + + fl = le32toh(fc.Flags); + if ((fl & MPI_FCPORTPAGE1_FLAGS_TARGET_MODE_OXID) == 0) { + fl |= MPI_FCPORTPAGE1_FLAGS_TARGET_MODE_OXID; + doit = 1; + } + if ((fl & MPI_FCPORTPAGE1_FLAGS_PROT_FCP_INIT) && + (mpt->role & MPT_ROLE_INITIATOR) == 0) { + fl &= ~MPI_FCPORTPAGE1_FLAGS_PROT_FCP_INIT; + doit = 1; + } + if ((fl & MPI_FCPORTPAGE1_FLAGS_PROT_FCP_TARG) && + (mpt->role & MPT_ROLE_TARGET) == 0) { + fl &= ~MPI_FCPORTPAGE1_FLAGS_PROT_FCP_TARG; + doit = 1; + } + if (doit) { + const char *cc; + + mpt_lprt(mpt, MPT_PRT_INFO, + "FC Port Page 1: New Flags %x \n", fl); + fc.Flags = htole32(fl); + r = mpt_write_cfg_page(mpt, + MPI_CONFIG_ACTION_PAGE_WRITE_CURRENT, 0, &fc.Header, + sizeof(fc), FALSE, 5000); + if (r != 0) { + cc = "FC PORT PAGE1 UPDATE: FAILED\n"; + } else { + cc = "FC PORT PAGE1 UPDATED: SYSTEM NEEDS RESET\n"; + } + mpt_prt(mpt, cc); + } +#endif + return (mpt_fc_reset_link(mpt, 1)); +} + +/* + * Read SAS configuration information. Nothing to do yet. + */ +static int +mpt_read_config_info_sas(struct mpt_softc *mpt) +{ + return (0); +} + +/* + * Set SAS configuration information. Nothing to do yet. + */ +static int +mpt_set_initial_config_sas(struct mpt_softc *mpt) +{ + return (0); +} + +/* + * Read SCSI configuration information + */ +static int +mpt_read_config_info_spi(struct mpt_softc *mpt) +{ + int rv, i; + + rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 0, + 0, &mpt->mpt_port_page0.Header, + /*sleep_ok*/FALSE, /*timeout_ms*/5000); + if (rv) + return (-1); + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Port Page 0 Header: %x %x %x %x\n", + mpt->mpt_port_page0.Header.PageVersion, + mpt->mpt_port_page0.Header.PageLength, + mpt->mpt_port_page0.Header.PageNumber, + mpt->mpt_port_page0.Header.PageType); + + rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 1, + 0, &mpt->mpt_port_page1.Header, + /*sleep_ok*/FALSE, /*timeout_ms*/5000); + if (rv) + return (-1); + + mpt_lprt(mpt, MPT_PRT_DEBUG, "SPI Port Page 1 Header: %x %x %x %x\n", + mpt->mpt_port_page1.Header.PageVersion, + mpt->mpt_port_page1.Header.PageLength, + mpt->mpt_port_page1.Header.PageNumber, + mpt->mpt_port_page1.Header.PageType); + + rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_PORT, 2, + /*PageAddress*/0, &mpt->mpt_port_page2.Header, + /*sleep_ok*/FALSE, /*timeout_ms*/5000); + if (rv) + return (-1); + + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Port Page 2 Header: %x %x %x %x\n", + mpt->mpt_port_page1.Header.PageVersion, + mpt->mpt_port_page1.Header.PageLength, + mpt->mpt_port_page1.Header.PageNumber, + mpt->mpt_port_page1.Header.PageType); + + for (i = 0; i < 16; i++) { + rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_DEVICE, + 0, i, &mpt->mpt_dev_page0[i].Header, + /*sleep_ok*/FALSE, /*timeout_ms*/5000); + if (rv) + return (-1); + + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Target %d Device Page 0 Header: %x %x %x %x\n", + i, mpt->mpt_dev_page0[i].Header.PageVersion, + mpt->mpt_dev_page0[i].Header.PageLength, + mpt->mpt_dev_page0[i].Header.PageNumber, + mpt->mpt_dev_page0[i].Header.PageType); + + rv = mpt_read_cfg_header(mpt, MPI_CONFIG_PAGETYPE_SCSI_DEVICE, + 1, i, &mpt->mpt_dev_page1[i].Header, + /*sleep_ok*/FALSE, /*timeout_ms*/5000); + if (rv) + return (-1); + + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Target %d Device Page 1 Header: %x %x %x %x\n", + i, mpt->mpt_dev_page1[i].Header.PageVersion, + mpt->mpt_dev_page1[i].Header.PageLength, + mpt->mpt_dev_page1[i].Header.PageNumber, + mpt->mpt_dev_page1[i].Header.PageType); + } + + /* + * At this point, we don't *have* to fail. As long as we have + * valid config header information, we can (barely) lurch + * along. + */ + + rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0, + &mpt->mpt_port_page0.Header, + sizeof(mpt->mpt_port_page0), + /*sleep_ok*/FALSE, /*timeout_ms*/5000); + if (rv) { + mpt_prt(mpt, "failed to read SPI Port Page 0\n"); + } else { + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Port Page 0: Capabilities %x PhysicalInterface %x\n", + mpt->mpt_port_page0.Capabilities, + mpt->mpt_port_page0.PhysicalInterface); + } + + rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0, + &mpt->mpt_port_page1.Header, + sizeof(mpt->mpt_port_page1), + /*sleep_ok*/FALSE, /*timeout_ms*/5000); + if (rv) { + mpt_prt(mpt, "failed to read SPI Port Page 1\n"); + } else { + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Port Page 1: Configuration %x OnBusTimerValue %x\n", + mpt->mpt_port_page1.Configuration, + mpt->mpt_port_page1.OnBusTimerValue); + } + + rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0, + &mpt->mpt_port_page2.Header, + sizeof(mpt->mpt_port_page2), + /*sleep_ok*/FALSE, /*timeout_ms*/5000); + if (rv) { + mpt_prt(mpt, "failed to read SPI Port Page 2\n"); + } else { + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Port Page 2: Flags %x Settings %x\n", + mpt->mpt_port_page2.PortFlags, + mpt->mpt_port_page2.PortSettings); + for (i = 0; i < 16; i++) { + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Port Page 2 Tgt %d: timo %x SF %x Flags %x\n", + i, mpt->mpt_port_page2.DeviceSettings[i].Timeout, + mpt->mpt_port_page2.DeviceSettings[i].SyncFactor, + mpt->mpt_port_page2.DeviceSettings[i].DeviceFlags); + } + } + + for (i = 0; i < 16; i++) { + rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i, + &mpt->mpt_dev_page0[i].Header, + sizeof(*mpt->mpt_dev_page0), + /*sleep_ok*/FALSE, + /*timeout_ms*/5000); + if (rv) { + mpt_prt(mpt, + "cannot read SPI Tgt %d Device Page 0\n", i); + continue; + } + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Tgt %d Page 0: NParms %x Information %x", + i, mpt->mpt_dev_page0[i].NegotiatedParameters, + mpt->mpt_dev_page0[i].Information); + + rv = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i, + &mpt->mpt_dev_page1[i].Header, + sizeof(*mpt->mpt_dev_page1), + /*sleep_ok*/FALSE, + /*timeout_ms*/5000); + if (rv) { + mpt_prt(mpt, + "cannot read SPI Tgt %d Device Page 1\n", i); + continue; + } + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Tgt %d Page 1: RParms %x Configuration %x\n", + i, mpt->mpt_dev_page1[i].RequestedParameters, + mpt->mpt_dev_page1[i].Configuration); + } + return (0); +} + +/* + * Validate SPI configuration information. + * + * In particular, validate SPI Port Page 1. + */ +static int +mpt_set_initial_config_spi(struct mpt_softc *mpt) +{ + int i, pp1val = ((1 << mpt->mpt_ini_id) << 16) | mpt->mpt_ini_id; + int error; + + mpt->mpt_disc_enable = 0xff; + mpt->mpt_tag_enable = 0; + + if (mpt->mpt_port_page1.Configuration != pp1val) { + CONFIG_PAGE_SCSI_PORT_1 tmp; + + mpt_prt(mpt, + "SPI Port Page 1 Config value bad (%x)- should be %x\n", + mpt->mpt_port_page1.Configuration, pp1val); + tmp = mpt->mpt_port_page1; + tmp.Configuration = pp1val; + error = mpt_write_cur_cfg_page(mpt, /*PageAddress*/0, + &tmp.Header, sizeof(tmp), + /*sleep_ok*/FALSE, + /*timeout_ms*/5000); + if (error) + return (-1); + error = mpt_read_cur_cfg_page(mpt, /*PageAddress*/0, + &tmp.Header, sizeof(tmp), + /*sleep_ok*/FALSE, + /*timeout_ms*/5000); + if (error) + return (-1); + if (tmp.Configuration != pp1val) { + mpt_prt(mpt, + "failed to reset SPI Port Page 1 Config value\n"); + return (-1); + } + mpt->mpt_port_page1 = tmp; + } + + for (i = 0; i < 16; i++) { + CONFIG_PAGE_SCSI_DEVICE_1 tmp; + tmp = mpt->mpt_dev_page1[i]; + tmp.RequestedParameters = 0; + tmp.Configuration = 0; + mpt_lprt(mpt, MPT_PRT_DEBUG, + "Set Tgt %d SPI DevicePage 1 values to %x 0 %x\n", + i, tmp.RequestedParameters, tmp.Configuration); + error = mpt_write_cur_cfg_page(mpt, /*PageAddress*/i, + &tmp.Header, sizeof(tmp), + /*sleep_ok*/FALSE, + /*timeout_ms*/5000); + if (error) + return (-1); + error = mpt_read_cur_cfg_page(mpt, /*PageAddress*/i, + &tmp.Header, sizeof(tmp), + /*sleep_ok*/FALSE, + /*timeout_ms*/5000); + if (error) + return (-1); + mpt->mpt_dev_page1[i] = tmp; + mpt_lprt(mpt, MPT_PRT_DEBUG, + "SPI Tgt %d Page 1: RParm %x Configuration %x\n", i, + mpt->mpt_dev_page1[i].RequestedParameters, + mpt->mpt_dev_page1[i].Configuration); + } + return (0); +} + +int +mpt_cam_enable(struct mpt_softc *mpt) +{ + if (mpt->is_fc) { + if (mpt_read_config_info_fc(mpt)) { + return (EIO); + } + if (mpt_set_initial_config_fc(mpt)) { + return (EIO); + } + } else if (mpt->is_sas) { + if (mpt_read_config_info_sas(mpt)) { + return (EIO); + } + if (mpt_set_initial_config_sas(mpt)) { + return (EIO); + } + } else { + if (mpt_read_config_info_spi(mpt)) { + return (EIO); + } + if (mpt_set_initial_config_spi(mpt)) { + return (EIO); + } + } + return (0); +} + void mpt_cam_detach(struct mpt_softc *mpt) { @@ -305,6 +771,12 @@ mpt_cam_detach(struct mpt_softc *mpt) handler.reply_handler = mpt_scsi_tmf_reply_handler; mpt_deregister_handler(mpt, MPT_HANDLER_REPLY, handler, scsi_tmf_handler_id); + handler.reply_handler = mpt_fc_els_reply_handler; + mpt_deregister_handler(mpt, MPT_HANDLER_REPLY, handler, + fc_els_handler_id); + handler.reply_handler = mpt_scsi_tgt_reply_handler; + mpt_deregister_handler(mpt, MPT_HANDLER_REPLY, handler, + mpt->scsi_tgt_handler_id); if (mpt->tmf_req != NULL) { mpt_free_request(mpt, mpt->tmf_req); @@ -326,8 +798,7 @@ mpt_cam_detach(struct mpt_softc *mpt) } } -/* This routine is used after a system crash to dump core onto the - * swap device. +/* This routine is used after a system crash to dump core onto the swap device. */ static void mpt_poll(struct cam_sim *sim) @@ -361,7 +832,7 @@ mpt_timeout(void *arg) MPT_LOCK(mpt); req = ccb->ccb_h.ccb_req_ptr; - mpt_prt(mpt, "Request %p:serno Timed out.\n", req, req->serno); + mpt_prt(mpt, "Request %p Timed out.\n", req); if ((req->state & REQ_STATE_QUEUED) == REQ_STATE_QUEUED) { TAILQ_REMOVE(&mpt->request_pending_list, req, links); TAILQ_INSERT_TAIL(&mpt->request_timeout_list, req, links); @@ -379,7 +850,7 @@ mpt_timeout(void *arg) * aborted the transaction. */ static void -mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) +mpt_execute_req_a64(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) { request_t *req, *trq; char *mpt_off; @@ -387,8 +858,8 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) struct mpt_softc *mpt; int seg, first_lim; uint32_t flags, nxt_off; - bus_dmasync_op_t op; - MSG_SCSI_IO_REQUEST *mpt_req; + void *sglp; + MSG_REQUEST_HEADER *hdrp; SGE_SIMPLE64 *se; SGE_CHAIN64 *ce; @@ -397,22 +868,31 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) mpt = ccb->ccb_h.ccb_mpt_ptr; req = ccb->ccb_h.ccb_req_ptr; - mpt_req = req->req_vbuf; + + hdrp = req->req_vbuf; mpt_off = req->req_vbuf; + if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) { + sglp = &((PTR_MSG_SCSI_IO_REQUEST)hdrp)->SGL; + } else /* if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) */ { + sglp = &((PTR_MSG_TARGET_ASSIST_REQUEST)hdrp)->SGL; + } + + if (error == 0 && ((uint32_t)nseg) >= mpt->max_seg_cnt) { error = EFBIG; } bad: if (error != 0) { - if (error != EFBIG && error != ENOMEM) + if (error != EFBIG && error != ENOMEM) { mpt_prt(mpt, "mpt_execute_req: err %d\n", error); - if (ccb->ccb_h.status == CAM_REQ_INPROG) { - xpt_freeze_devq(ccb->ccb_h.path, 1); - ccb->ccb_h.status = CAM_DEV_QFRZN; + } + if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INPROG) { + cam_status status; + mpt_freeze_ccb(ccb); if (error == EFBIG) { - ccb->ccb_h.status |= CAM_REQ_TOO_BIG; + status = CAM_REQ_TOO_BIG; } else if (error == ENOMEM) { if (mpt->outofbeer == 0) { mpt->outofbeer = 1; @@ -420,11 +900,21 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) mpt_lprt(mpt, MPT_PRT_DEBUG, "FREEZEQ\n"); } - ccb->ccb_h.status |= CAM_REQUEUE_REQ; - } else - ccb->ccb_h.status |= CAM_REQ_CMP_ERR; + status = CAM_REQUEUE_REQ; + } else { + status = CAM_REQ_CMP_ERR; + } + mpt_set_ccb_status(ccb, status); + } + if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) { + request_t *cmd_req = + MPT_TAG_2_REQ(mpt, ccb->csio.tag_id); + MPT_TGT_STATE(mpt, cmd_req)->state = TGT_STATE_IN_CAM; + MPT_TGT_STATE(mpt, cmd_req)->ccb = NULL; + MPT_TGT_STATE(mpt, cmd_req)->req = NULL; } ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__)); xpt_done(ccb); CAMLOCK_2_MPTLOCK(mpt); mpt_free_request(mpt, req); @@ -438,29 +928,45 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) */ if (mpt->verbose >= MPT_PRT_DEBUG) { - int tidx = ((char *)&mpt_req->SGL) - mpt_off; + int tidx = ((char *)sglp) - mpt_off; memset(&mpt_off[tidx], 0xff, MPT_REQUEST_AREA - tidx); } if (nseg == 0) { - SGE_SIMPLE32 *se1 = (SGE_SIMPLE32 *) &mpt_req->SGL; + SGE_SIMPLE32 *se1 = (SGE_SIMPLE32 *) sglp; MPI_pSGE_SET_FLAGS(se1, (MPI_SGE_FLAGS_LAST_ELEMENT | MPI_SGE_FLAGS_END_OF_BUFFER | MPI_SGE_FLAGS_SIMPLE_ELEMENT | MPI_SGE_FLAGS_END_OF_LIST)); goto out; } - mpt_req->DataLength = ccb->csio.dxfer_len; - flags = MPI_SGE_FLAGS_SIMPLE_ELEMENT | MPI_SGE_FLAGS_64_BIT_ADDRESSING; - if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT) - flags |= MPI_SGE_FLAGS_HOST_TO_IOC; - if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { - op = BUS_DMASYNC_PREREAD; + flags = MPI_SGE_FLAGS_SIMPLE_ELEMENT | MPI_SGE_FLAGS_64_BIT_ADDRESSING; + if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) { + if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT) { + flags |= MPI_SGE_FLAGS_HOST_TO_IOC; + } } else { - op = BUS_DMASYNC_PREWRITE; + if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { + flags |= MPI_SGE_FLAGS_HOST_TO_IOC; + } } + if (!(ccb->ccb_h.flags & (CAM_SG_LIST_PHYS|CAM_DATA_PHYS))) { + bus_dmasync_op_t op; + if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) { + if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { + op = BUS_DMASYNC_PREREAD; + } else { + op = BUS_DMASYNC_PREWRITE; + } + } else { + if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { + op = BUS_DMASYNC_PREWRITE; + } else { + op = BUS_DMASYNC_PREREAD; + } + } bus_dmamap_sync(mpt->buffer_dmat, req->dmap, op); } @@ -483,7 +989,7 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) first_lim = MPT_NSGL_FIRST(mpt) - 1; } - se = (SGE_SIMPLE64 *) &mpt_req->SGL; + se = (SGE_SIMPLE64 *) sglp; for (seg = 0; seg < first_lim; seg++, se++, dm_segs++) { uint32_t tf; @@ -511,7 +1017,7 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) /* * Tell the IOC where to find the first chain element. */ - mpt_req->ChainOffset = ((char *)se - (char *)mpt_req) >> 2; + hdrp->ChainOffset = ((char *)se - (char *)hdrp) >> 2; nxt_off = MPT_RQSL(mpt); trq = req; @@ -650,7 +1156,6 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) } trq = nrq; mpt_off = trq->req_vbuf; - mpt_req = trq->req_vbuf; if (mpt->verbose >= MPT_PRT_DEBUG) { memset(mpt_off, 0xff, MPT_REQUEST_AREA); } @@ -662,18 +1167,30 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) /* * Last time we need to check if this CCB needs to be aborted. */ - if (ccb->ccb_h.status != CAM_REQ_INPROG) { - if (nseg && (ccb->ccb_h.flags & CAM_SG_LIST_PHYS) == 0) + if ((ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_INPROG) { + if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) { + request_t *cmd_req = + MPT_TAG_2_REQ(mpt, ccb->csio.tag_id); + MPT_TGT_STATE(mpt, cmd_req)->state = TGT_STATE_IN_CAM; + MPT_TGT_STATE(mpt, cmd_req)->ccb = NULL; + MPT_TGT_STATE(mpt, cmd_req)->req = NULL; + } + mpt_prt(mpt, + "mpt_execute_req_a64: I/O cancelled (status 0x%x)\n", + ccb->ccb_h.status & CAM_STATUS_MASK); + if (nseg && (ccb->ccb_h.flags & CAM_SG_LIST_PHYS) == 0) { bus_dmamap_unload(mpt->buffer_dmat, req->dmap); + } + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__)); + xpt_done(ccb); CAMLOCK_2_MPTLOCK(mpt); mpt_free_request(mpt, req); MPTLOCK_2_CAMLOCK(mpt); - xpt_done(ccb); return; } ccb->ccb_h.status |= CAM_SIM_QUEUED; - CAMLOCK_2_MPTLOCK(mpt); if (ccb->ccb_h.timeout != CAM_TIME_INFINITY) { ccb->ccb_h.timeout_ch = timeout(mpt_timeout, (caddr_t)ccb, @@ -683,12 +1200,390 @@ mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) } if (mpt->verbose >= MPT_PRT_DEBUG) { int nc = 0; - mpt_print_scsi_io_request(req->req_vbuf); + mpt_print_request(req->req_vbuf); for (trq = req->chain; trq; trq = trq->chain) { printf(" Additional Chain Area %d\n", nc++); mpt_dump_sgl(trq->req_vbuf, 0); } } + if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) { + request_t *cmd_req = MPT_TAG_2_REQ(mpt, ccb->csio.tag_id); + mpt_tgt_state_t *tgt = MPT_TGT_STATE(mpt, cmd_req); +#ifdef WE_TRUST_AUTO_GOOD_STATUS + if ((ccb->ccb_h.flags & CAM_SEND_STATUS) && + csio->scsi_status == SCSI_STATUS_OK && tgt->resid == 0) { + tgt->state = TGT_STATE_MOVING_DATA_AND_STATUS; + } else { + tgt->state = TGT_STATE_MOVING_DATA; + } +#else + tgt->state = TGT_STATE_MOVING_DATA; +#endif + } + CAMLOCK_2_MPTLOCK(mpt); + mpt_send_cmd(mpt, req); + MPTLOCK_2_CAMLOCK(mpt); +} + +static void +mpt_execute_req(void *arg, bus_dma_segment_t *dm_segs, int nseg, int error) +{ + request_t *req, *trq; + char *mpt_off; + union ccb *ccb; + struct mpt_softc *mpt; + int seg, first_lim; + uint32_t flags, nxt_off; + void *sglp; + MSG_REQUEST_HEADER *hdrp; + SGE_SIMPLE32 *se; + SGE_CHAIN32 *ce; + + req = (request_t *)arg; + ccb = req->ccb; + + mpt = ccb->ccb_h.ccb_mpt_ptr; + req = ccb->ccb_h.ccb_req_ptr; + + hdrp = req->req_vbuf; + mpt_off = req->req_vbuf; + + + if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) { + sglp = &((PTR_MSG_SCSI_IO_REQUEST)hdrp)->SGL; + } else /* if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) */ { + sglp = &((PTR_MSG_TARGET_ASSIST_REQUEST)hdrp)->SGL; + } + + + if (error == 0 && ((uint32_t)nseg) >= mpt->max_seg_cnt) { + error = EFBIG; + mpt_prt(mpt, "segment count %d too large (max %u)\n", + nseg, mpt->max_seg_cnt); + } + +bad: + if (error != 0) { + if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) { + request_t *cmd_req = + MPT_TAG_2_REQ(mpt, ccb->csio.tag_id); + MPT_TGT_STATE(mpt, cmd_req)->state = TGT_STATE_IN_CAM; + MPT_TGT_STATE(mpt, cmd_req)->ccb = NULL; + MPT_TGT_STATE(mpt, cmd_req)->req = NULL; + } + if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INPROG) { + cam_status status; + mpt_freeze_ccb(ccb); + if (error == EFBIG) { + status = CAM_REQ_TOO_BIG; + } else if (error == ENOMEM) { + if (mpt->outofbeer == 0) { + mpt->outofbeer = 1; + xpt_freeze_simq(mpt->sim, 1); + mpt_lprt(mpt, MPT_PRT_DEBUG, + "FREEZEQ\n"); + } + status = CAM_REQUEUE_REQ; + } else { + status = CAM_REQ_CMP_ERR; + } + mpt_set_ccb_status(ccb, status); + } + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__)); + xpt_done(ccb); + CAMLOCK_2_MPTLOCK(mpt); + mpt_free_request(mpt, req); + MPTLOCK_2_CAMLOCK(mpt); + return; + } + + /* + * No data to transfer? + * Just make a single simple SGL with zero length. + */ + + if (mpt->verbose >= MPT_PRT_DEBUG) { + int tidx = ((char *)sglp) - mpt_off; + memset(&mpt_off[tidx], 0xff, MPT_REQUEST_AREA - tidx); + } + + if (nseg == 0) { + SGE_SIMPLE32 *se1 = (SGE_SIMPLE32 *) sglp; + MPI_pSGE_SET_FLAGS(se1, + (MPI_SGE_FLAGS_LAST_ELEMENT | MPI_SGE_FLAGS_END_OF_BUFFER | + MPI_SGE_FLAGS_SIMPLE_ELEMENT | MPI_SGE_FLAGS_END_OF_LIST)); + goto out; + } + + + flags = MPI_SGE_FLAGS_SIMPLE_ELEMENT; + if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) { + if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_OUT) { + flags |= MPI_SGE_FLAGS_HOST_TO_IOC; + } + } else { + if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { + flags |= MPI_SGE_FLAGS_HOST_TO_IOC; + } + } + + if (!(ccb->ccb_h.flags & (CAM_SG_LIST_PHYS|CAM_DATA_PHYS))) { + bus_dmasync_op_t op; + if (hdrp->Function == MPI_FUNCTION_SCSI_IO_REQUEST) { + if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { + op = BUS_DMASYNC_PREREAD; + } else { + op = BUS_DMASYNC_PREWRITE; + } + } else { + if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { + op = BUS_DMASYNC_PREWRITE; + } else { + op = BUS_DMASYNC_PREREAD; + } + } + bus_dmamap_sync(mpt->buffer_dmat, req->dmap, op); + } + + /* + * Okay, fill in what we can at the end of the command frame. + * If we have up to MPT_NSGL_FIRST, we can fit them all into + * the command frame. + * + * Otherwise, we fill up through MPT_NSGL_FIRST less one + * SIMPLE32 pointers and start doing CHAIN32 entries after + * that. + */ + + if (nseg < MPT_NSGL_FIRST(mpt)) { + first_lim = nseg; + } else { + /* + * Leave room for CHAIN element + */ + first_lim = MPT_NSGL_FIRST(mpt) - 1; + } + + se = (SGE_SIMPLE32 *) sglp; + for (seg = 0; seg < first_lim; seg++, se++, dm_segs++) { + uint32_t tf; + + bzero(se, sizeof (*se)); + se->Address = dm_segs->ds_addr; + MPI_pSGE_SET_LENGTH(se, dm_segs->ds_len); + tf = flags; + if (seg == first_lim - 1) { + tf |= MPI_SGE_FLAGS_LAST_ELEMENT; + } + if (seg == nseg - 1) { + tf |= MPI_SGE_FLAGS_END_OF_LIST | + MPI_SGE_FLAGS_END_OF_BUFFER; + } + MPI_pSGE_SET_FLAGS(se, tf); + } + + if (seg == nseg) { + goto out; + } + + /* + * Tell the IOC where to find the first chain element. + */ + hdrp->ChainOffset = ((char *)se - (char *)hdrp) >> 2; + nxt_off = MPT_RQSL(mpt); + trq = req; + + /* + * Make up the rest of the data segments out of a chain element + * (contiained in the current request frame) which points to + * SIMPLE32 elements in the next request frame, possibly ending + * with *another* chain element (if there's more). + */ + while (seg < nseg) { + int this_seg_lim; + uint32_t tf, cur_off; + bus_addr_t chain_list_addr; + + /* + * Point to the chain descriptor. Note that the chain + * descriptor is at the end of the *previous* list (whether + * chain or simple). + */ + ce = (SGE_CHAIN32 *) se; + + /* + * Before we change our current pointer, make sure we won't + * overflow the request area with this frame. Note that we + * test against 'greater than' here as it's okay in this case + * to have next offset be just outside the request area. + */ + if ((nxt_off + MPT_RQSL(mpt)) > MPT_REQUEST_AREA) { + nxt_off = MPT_REQUEST_AREA; + goto next_chain; + } + + /* + * Set our SGE element pointer to the beginning of the chain + * list and update our next chain list offset. + */ + se = (SGE_SIMPLE32 *) &mpt_off[nxt_off]; + cur_off = nxt_off; + nxt_off += MPT_RQSL(mpt); + + /* + * Now initialized the chain descriptor. + */ + bzero(ce, sizeof (SGE_CHAIN32)); + + /* + * Get the physical address of the chain list. + */ + chain_list_addr = trq->req_pbuf; + chain_list_addr += cur_off; + ce->Address = chain_list_addr; + ce->Flags = MPI_SGE_FLAGS_CHAIN_ELEMENT; + + /* + * If we have more than a frame's worth of segments left, + * set up the chain list to have the last element be another + * chain descriptor. + */ + if ((nseg - seg) > MPT_NSGL(mpt)) { + this_seg_lim = seg + MPT_NSGL(mpt) - 1; + /* + * The length of the chain is the length in bytes of the + * number of segments plus the next chain element. + * + * The next chain descriptor offset is the length, + * in words, of the number of segments. + */ + ce->Length = (this_seg_lim - seg) * + sizeof (SGE_SIMPLE32); + ce->NextChainOffset = ce->Length >> 2; + ce->Length += sizeof (SGE_CHAIN32); + } else { + this_seg_lim = nseg; + ce->Length = (this_seg_lim - seg) * + sizeof (SGE_SIMPLE32); + } + + /* + * Fill in the chain list SGE elements with our segment data. + * + * If we're the last element in this chain list, set the last + * element flag. If we're the completely last element period, + * set the end of list and end of buffer flags. + */ + while (seg < this_seg_lim) { + bzero(se, sizeof (*se)); + se->Address = dm_segs->ds_addr; + MPI_pSGE_SET_LENGTH(se, dm_segs->ds_len); + tf = flags; + if (seg == this_seg_lim - 1) { + tf |= MPI_SGE_FLAGS_LAST_ELEMENT; + } + if (seg == nseg - 1) { + tf |= MPI_SGE_FLAGS_END_OF_LIST | + MPI_SGE_FLAGS_END_OF_BUFFER; + } + MPI_pSGE_SET_FLAGS(se, tf); + se++; + seg++; + dm_segs++; + } + + next_chain: + /* + * If we have more segments to do and we've used up all of + * the space in a request area, go allocate another one + * and chain to that. + */ + if (seg < nseg && nxt_off >= MPT_REQUEST_AREA) { + request_t *nrq = mpt_get_request(mpt, FALSE); + + if (nrq == NULL) { + error = ENOMEM; + goto bad; + } + + /* + * Append the new request area on the tail of our list. + */ + if ((trq = req->chain) == NULL) { + req->chain = nrq; + } else { + while (trq->chain != NULL) { + trq = trq->chain; + } + trq->chain = nrq; + } + trq = nrq; + mpt_off = trq->req_vbuf; + if (mpt->verbose >= MPT_PRT_DEBUG) { + memset(mpt_off, 0xff, MPT_REQUEST_AREA); + } + nxt_off = 0; + } + } +out: + + /* + * Last time we need to check if this CCB needs to be aborted. + */ + if ((ccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_INPROG) { + if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) { + request_t *cmd_req = + MPT_TAG_2_REQ(mpt, ccb->csio.tag_id); + MPT_TGT_STATE(mpt, cmd_req)->state = TGT_STATE_IN_CAM; + MPT_TGT_STATE(mpt, cmd_req)->ccb = NULL; + MPT_TGT_STATE(mpt, cmd_req)->req = NULL; + } + mpt_prt(mpt, "mpt_execute_req: I/O cancelled (status 0x%x)\n", + ccb->ccb_h.status & CAM_STATUS_MASK); + if (nseg && (ccb->ccb_h.flags & CAM_SG_LIST_PHYS) == 0) { + bus_dmamap_unload(mpt->buffer_dmat, req->dmap); + } + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__)); + xpt_done(ccb); + CAMLOCK_2_MPTLOCK(mpt); + mpt_free_request(mpt, req); + MPTLOCK_2_CAMLOCK(mpt); + return; + } + + ccb->ccb_h.status |= CAM_SIM_QUEUED; + if (ccb->ccb_h.timeout != CAM_TIME_INFINITY) { + ccb->ccb_h.timeout_ch = + timeout(mpt_timeout, (caddr_t)ccb, + (ccb->ccb_h.timeout * hz) / 1000); + } else { + callout_handle_init(&ccb->ccb_h.timeout_ch); + } + if (mpt->verbose >= MPT_PRT_DEBUG) { + int nc = 0; + mpt_print_request(req->req_vbuf); + for (trq = req->chain; trq; trq = trq->chain) { + printf(" Additional Chain Area %d\n", nc++); + mpt_dump_sgl(trq->req_vbuf, 0); + } + } + if (hdrp->Function == MPI_FUNCTION_TARGET_ASSIST) { + request_t *cmd_req = MPT_TAG_2_REQ(mpt, ccb->csio.tag_id); + mpt_tgt_state_t *tgt = MPT_TGT_STATE(mpt, cmd_req); +#ifdef WE_TRUST_AUTO_GOOD_STATUS + if ((ccb->ccb_h.flags & CAM_SEND_STATUS) && + csio->scsi_status == SCSI_STATUS_OK && tgt->resid == 0) { + tgt->state = TGT_STATE_MOVING_DATA_AND_STATUS; + } else { + tgt->state = TGT_STATE_MOVING_DATA; + } +#else + tgt->state = TGT_STATE_MOVING_DATA; +#endif + } + CAMLOCK_2_MPTLOCK(mpt); mpt_send_cmd(mpt, req); MPTLOCK_2_CAMLOCK(mpt); } @@ -701,6 +1596,7 @@ mpt_start(struct cam_sim *sim, union ccb *ccb) MSG_SCSI_IO_REQUEST *mpt_req; struct ccb_scsiio *csio = &ccb->csio; struct ccb_hdr *ccbh = &ccb->ccb_h; + bus_dmamap_callback_t *cb; int raid_passthru; /* Get the pointer for the physical addapter */ @@ -715,14 +1611,20 @@ mpt_start(struct cam_sim *sim, union ccb *ccb) xpt_freeze_simq(mpt->sim, 1); mpt_lprt(mpt, MPT_PRT_DEBUG, "FREEZEQ\n"); } + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + mpt_set_ccb_status(ccb, CAM_REQUEUE_REQ); MPTLOCK_2_CAMLOCK(mpt); - ccb->ccb_h.status = CAM_REQUEUE_REQ; xpt_done(ccb); return; } - MPTLOCK_2_CAMLOCK(mpt); + if (sizeof (bus_addr_t) > 4) { + cb = mpt_execute_req_a64; + } else { + cb = mpt_execute_req; + } + #if 0 COWWWWW if (raid_passthru) { @@ -743,10 +1645,11 @@ mpt_start(struct cam_sim *sim, union ccb *ccb) bzero(mpt_req, sizeof *mpt_req); mpt_req->Function = MPI_FUNCTION_SCSI_IO_REQUEST; - if (raid_passthru) + if (raid_passthru) { mpt_req->Function = MPI_FUNCTION_RAID_SCSI_IO_PASSTHROUGH; + } - mpt_req->Bus = mpt->bus; + mpt_req->Bus = 0; /* we don't have multiport devices yet */ mpt_req->SenseBufferLength = (csio->sense_len < MPT_SENSE_SIZE) ? @@ -760,10 +1663,14 @@ mpt_start(struct cam_sim *sim, union ccb *ccb) /* Which physical device to do the I/O on */ mpt_req->TargetID = ccb->ccb_h.target_id; - /* - * XXX Assumes Single level, Single byte, CAM LUN type. - */ - mpt_req->LUN[1] = ccb->ccb_h.target_lun; + + /* We assume a single level LUN type */ + if (ccb->ccb_h.target_lun >= 256) { + mpt_req->LUN[0] = 0x40 | ((ccb->ccb_h.target_lun >> 8) & 0x3f); + mpt_req->LUN[1] = ccb->ccb_h.target_lun & 0xff; + } else { + mpt_req->LUN[1] = ccb->ccb_h.target_lun; + } /* Set the direction of the transfer */ if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) @@ -829,10 +1736,11 @@ mpt_start(struct cam_sim *sim, union ccb *ccb) * one or more physical address ranges. */ int error; - + int s = splsoftvm(); error = bus_dmamap_load(mpt->buffer_dmat, req->dmap, csio->data_ptr, csio->dxfer_len, - mpt_execute_req, req, 0); + cb, req, 0); + splx(s); if (error == EINPROGRESS) { /* * So as to maintain ordering, @@ -852,7 +1760,7 @@ mpt_start(struct cam_sim *sim, union ccb *ccb) seg.ds_addr = (bus_addr_t)(vm_offset_t)csio->data_ptr; seg.ds_len = csio->dxfer_len; - mpt_execute_req(req, &seg, 1, 0); + (*cb)(req, &seg, 1, 0); } } else { /* @@ -863,15 +1771,15 @@ mpt_start(struct cam_sim *sim, union ccb *ccb) */ struct bus_dma_segment *segs; if ((ccbh->flags & CAM_SG_LIST_PHYS) == 0) { - mpt_execute_req(req, NULL, 0, EFAULT); + (*cb)(req, NULL, 0, EFAULT); } else { /* Just use the segments provided */ segs = (struct bus_dma_segment *)csio->data_ptr; - mpt_execute_req(req, segs, csio->sglist_cnt, 0); + (*cb)(req, segs, csio->sglist_cnt, 0); } } } else { - mpt_execute_req(req, NULL, 0, 0); + (*cb)(req, NULL, 0, 0); } } @@ -916,12 +1824,35 @@ mpt_bus_reset(struct mpt_softc *mpt, int sleep_ok) return (0); } +static int +mpt_fc_reset_link(struct mpt_softc *mpt, int dowait) +{ + int r = 0; + request_t *req; + PTR_MSG_FC_PRIMITIVE_SEND_REQUEST fc; + + req = mpt_get_request(mpt, FALSE); + if (req == NULL) { + return (ENOMEM); + } + fc = req->req_vbuf; + memset(fc, 0, sizeof(*fc)); + fc->SendFlags = MPI_FC_PRIM_SEND_FLAGS_RESET_LINK; + fc->Function = MPI_FUNCTION_FC_PRIMITIVE_SEND; + fc->MsgContext = htole32(req->index | fc_els_handler_id); + mpt_send_cmd(mpt, req); + if (dowait) { + r = mpt_wait_req(mpt, req, REQ_STATE_DONE, + REQ_STATE_DONE, FALSE, 60 * 1000); + mpt_free_request(mpt, req); + } + return (r); +} + static int mpt_cam_event(struct mpt_softc *mpt, request_t *req, MSG_EVENT_NOTIFY_REPLY *msg) { - mpt_lprt(mpt, MPT_PRT_ALWAYS, "mpt_cam_event: 0x%x\n", - msg->Event & 0xFF); switch(msg->Event & 0xFF) { case MPI_EVENT_UNIT_ATTENTION: mpt_prt(mpt, "Bus: 0x%02x TargetID: 0x%02x\n", @@ -937,7 +1868,7 @@ mpt_cam_event(struct mpt_softc *mpt, request_t *req, case MPI_EVENT_EXT_BUS_RESET: /* Someone else generated a bus reset */ - mpt_prt(mpt, "Ext Bus Reset\n"); + mpt_prt(mpt, "External Bus Reset Detected\n"); /* * These replies don't return EventData like the MPI * spec says they do @@ -972,26 +1903,26 @@ mpt_cam_event(struct mpt_softc *mpt, request_t *req, switch ((msg->Data[0] >> 8) & 0xff) { case 0xF7: if ((msg->Data[0] & 0xff) == 0xF7) { - printf("Device needs AL_PA\n"); + mpt_prt(mpt, "Device needs AL_PA\n"); } else { - printf("Device %02x doesn't like " + mpt_prt(mpt, "Device %02x doesn't like " "FC performance\n", msg->Data[0] & 0xFF); } break; case 0xF8: if ((msg->Data[0] & 0xff) == 0xF7) { - printf("Device had loop failure at its " - "receiver prior to acquiring " - "AL_PA\n"); + mpt_prt(mpt, "Device had loop failure " + "at its receiver prior to acquiring" + " AL_PA\n"); } else { - printf("Device %02x detected loop " - "failure at its receiver\n", + mpt_prt(mpt, "Device %02x detected loop" + " failure at its receiver\n", msg->Data[0] & 0xFF); } break; default: - printf("Device %02x requests that device " + mpt_prt(mpt, "Device %02x requests that device " "%02x reset itself\n", msg->Data[0] & 0xFF, (msg->Data[0] >> 8) & 0xFF); @@ -1038,6 +1969,8 @@ mpt_cam_event(struct mpt_softc *mpt, request_t *req, "mpt_cam_event: MPI_EVENT_SAS_DEVICE_STATUS_CHANGE\n"); break; default: + mpt_lprt(mpt, MPT_PRT_WARN, "mpt_cam_event: 0x%x\n", + msg->Event & 0xFF); return (/*handled*/0); } return (/*handled*/1); @@ -1055,7 +1988,7 @@ mpt_cam_event(struct mpt_softc *mpt, request_t *req, */ static int mpt_scsi_reply_handler(struct mpt_softc *mpt, request_t *req, - MSG_DEFAULT_REPLY *reply_frame) + uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame) { MSG_SCSI_IO_REQUEST *scsi_req; union ccb *ccb; @@ -1084,8 +2017,7 @@ mpt_scsi_reply_handler(struct mpt_softc *mpt, request_t *req, if (reply_frame == NULL) { /* - * Context only reply, completion - * without error status. + * Context only reply, completion without error status. */ ccb->csio.resid = 0; mpt_set_ccb_status(ccb, CAM_REQ_CMP); @@ -1099,13 +2031,9 @@ mpt_scsi_reply_handler(struct mpt_softc *mpt, request_t *req, mpt->outofbeer = 0; mpt_lprt(mpt, MPT_PRT_DEBUG, "THAWQ\n"); } - ccb->ccb_h.status &= ~CAM_SIM_QUEUED; - MPTLOCK_2_CAMLOCK(mpt); - if (scsi_req->Function == MPI_FUNCTION_RAID_SCSI_IO_PASSTHROUGH - && scsi_req->CDB[0] == INQUIRY - && (scsi_req->CDB[1] & SI_EVPD) == 0) { + if (scsi_req->Function == MPI_FUNCTION_RAID_SCSI_IO_PASSTHROUGH && + scsi_req->CDB[0] == INQUIRY && (scsi_req->CDB[1] & SI_EVPD) == 0) { struct scsi_inquiry_data *inq; - /* * Fake out the device type so that only the * pass-thru device will attach. @@ -1114,13 +2042,16 @@ mpt_scsi_reply_handler(struct mpt_softc *mpt, request_t *req, inq->device &= ~0x1F; inq->device |= T_NODEVICE; } + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__)); + MPTLOCK_2_CAMLOCK(mpt); xpt_done(ccb); CAMLOCK_2_MPTLOCK(mpt); - if ((req->state & REQ_STATE_TIMEDOUT) == 0) + if ((req->state & REQ_STATE_TIMEDOUT) == 0) { TAILQ_REMOVE(&mpt->request_pending_list, req, links); - else + } else { TAILQ_REMOVE(&mpt->request_timeout_list, req, links); - + } if ((req->state & REQ_STATE_NEED_WAKEUP) == 0) { mpt_free_request(mpt, req); return (/*free_reply*/TRUE); @@ -1133,7 +2064,7 @@ mpt_scsi_reply_handler(struct mpt_softc *mpt, request_t *req, static int mpt_scsi_tmf_reply_handler(struct mpt_softc *mpt, request_t *req, - MSG_DEFAULT_REPLY *reply_frame) + uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame) { MSG_SCSI_TASK_MGMT_REPLY *tmf_reply; uint16_t status; @@ -1158,6 +2089,884 @@ mpt_scsi_tmf_reply_handler(struct mpt_softc *mpt, request_t *req, return (/*free_reply*/TRUE); } + +/* + * XXX: Move to definitions file + */ +#define ELS 0x22 +#define FC4LS 0x32 +#define ABTS 0x81 +#define BA_ACC 0x84 + +#define LS_RJT 0x01 +#define LS_ACC 0x02 +#define PLOGI 0x03 +#define LOGO 0x05 +#define SRR 0x14 +#define PRLI 0x20 +#define PRLO 0x21 +#define ADISC 0x52 +#define RSCN 0x61 + +static void +mpt_fc_els_send_response(struct mpt_softc *mpt, request_t *req, + PTR_MSG_LINK_SERVICE_BUFFER_POST_REPLY rp, U8 length) +{ + MSG_LINK_SERVICE_RSP_REQUEST tmp; + PTR_MSG_LINK_SERVICE_RSP_REQUEST rsp; + + /* + * We are going to reuse the ELS request to send this response back. + */ + rsp = &tmp; + memset(rsp, 0, sizeof(*rsp)); + +#ifdef USE_IMMEDIATE_LINK_DATA + /* + * Apparently the IMMEDIATE stuff doesn't seem to work. + */ + rsp->RspFlags = LINK_SERVICE_RSP_FLAGS_IMMEDIATE; +#endif + rsp->RspLength = length; + rsp->Function = MPI_FUNCTION_FC_LINK_SRVC_RSP; + rsp->MsgContext = htole32(req->index | fc_els_handler_id); + + /* + * Copy over information from the original reply frame to + * it's correct place in the response. + */ + memcpy((U8 *)rsp + 0x0c, (U8 *)rp + 0x1c, 24); + + /* + * And now copy back the temporary area to the original frame. + */ + memcpy(req->req_vbuf, rsp, sizeof (MSG_LINK_SERVICE_RSP_REQUEST)); + rsp = req->req_vbuf; + +#ifdef USE_IMMEDIATE_LINK_DATA + memcpy((U8 *)&rsp->SGL, &((U8 *)req->req_vbuf)[MPT_RQSL(mpt)], length); +#else +{ + PTR_SGE_SIMPLE32 se = (PTR_SGE_SIMPLE32) &rsp->SGL; + bus_addr_t paddr = req->req_pbuf; + paddr += MPT_RQSL(mpt); + + se->FlagsLength = + MPI_SGE_FLAGS_HOST_TO_IOC | + MPI_SGE_FLAGS_SIMPLE_ELEMENT | + MPI_SGE_FLAGS_LAST_ELEMENT | + MPI_SGE_FLAGS_END_OF_LIST | + MPI_SGE_FLAGS_END_OF_BUFFER; + se->FlagsLength <<= MPI_SGE_FLAGS_SHIFT; + se->FlagsLength |= (length); + se->Address = (uint32_t) paddr; +} +#endif + + /* + * Send it on... + */ + mpt_send_cmd(mpt, req); +} + +static int +mpt_fc_els_reply_handler(struct mpt_softc *mpt, request_t *req, + uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame) +{ + PTR_MSG_LINK_SERVICE_BUFFER_POST_REPLY rp = + (PTR_MSG_LINK_SERVICE_BUFFER_POST_REPLY) reply_frame; + U8 rctl; + U8 type; + U8 cmd; + U16 status = le16toh(reply_frame->IOCStatus); + U32 *elsbuf; + int do_refresh = TRUE; + + mpt_lprt(mpt, MPT_PRT_DEBUG, "FC_ELS Complete: req %p:%u, reply %p\n", + req, req->serno, reply_frame); + + if (status != MPI_IOCSTATUS_SUCCESS) { + mpt_prt(mpt, "ELS REPLY STATUS 0x%x for Function %x\n", + status, reply_frame->Function); + if (status == MPI_IOCSTATUS_INVALID_STATE) { + /* + * XXX: to get around shutdown issue + */ + mpt->disabled = 1; + return (TRUE); + } + return (TRUE); + } + + /* + * If the function of a link service response, we recycle the + * response to be a refresh for a new link service request. + */ + if (rp->Function == MPI_FUNCTION_FC_LINK_SRVC_RSP) { + mpt_fc_add_els(mpt, req); + return (TRUE); + } + + if (rp->Function == MPI_FUNCTION_FC_PRIMITIVE_SEND) { + req->state &= ~REQ_STATE_QUEUED; + req->state |= REQ_STATE_DONE; + if ((req->state & REQ_STATE_NEED_WAKEUP) == 0) { + mpt_lprt(mpt, MPT_PRT_DEBUG, + "Async Primitive Send Complete\n"); + mpt_free_request(mpt, req); + } else { + mpt_lprt(mpt, MPT_PRT_DEBUG, + "Sync Primitive Send Complete\n"); + wakeup(req); + } + return (TRUE); + } + + if (rp->Function != MPI_FUNCTION_FC_LINK_SRVC_BUF_POST) { + mpt_prt(mpt, "unexpected ELS_REPLY: Function 0x%x Flags %x " + "Length %d Message Flags %x\n", rp->Function, rp->Flags, + rp->MsgLength, rp->MsgFlags); + return (TRUE); + } + + if (rp->MsgLength <= 5) { + /* + * This is just a ack of an original ELS buffer post + */ + mpt_lprt(mpt, MPT_PRT_DEBUG, + "Recv'd ACK of FC_ELS buf post %p:%u\n", req, req->serno); + return (TRUE); + } + + + rctl = (le32toh(rp->Rctl_Did) & MPI_FC_RCTL_MASK) >> MPI_FC_RCTL_SHIFT; + type = (le32toh(rp->Type_Fctl) & MPI_FC_TYPE_MASK) >> MPI_FC_TYPE_SHIFT; + + elsbuf = &((U32 *)req->req_vbuf)[MPT_RQSL(mpt)/sizeof (U32)]; + cmd = be32toh(elsbuf[0]) >> 24; + + if (rp->Flags & MPI_LS_BUF_POST_REPLY_FLAG_NO_RSP_NEEDED) { + mpt_lprt(mpt, MPT_PRT_ALWAYS, "ELS_REPLY: response unneeded\n"); + return (TRUE); + } + + + if (rctl == ELS && type == 1) { + switch (cmd) { + case PRLI: + /* + * Send back a PRLI ACC + */ + mpt_prt(mpt, "PRLI from 0x%08x%08x\n", + le32toh(rp->Wwn.PortNameHigh), + le32toh(rp->Wwn.PortNameLow)); + elsbuf[0] = htobe32(0x02100014); + elsbuf[1] |= htobe32(0x00000100); + elsbuf[4] = htobe32(0x00000002); + if (mpt->role & MPT_ROLE_TARGET) + elsbuf[4] |= htobe32(0x00000010); + if (mpt->role & MPT_ROLE_INITIATOR) + elsbuf[4] |= htobe32(0x00000020); + mpt_fc_els_send_response(mpt, req, rp, 20); + do_refresh = FALSE; + break; + case PRLO: + memset(elsbuf, 0, 5 * (sizeof (U32))); + elsbuf[0] = htobe32(0x02100014); + elsbuf[1] = htobe32(0x08000100); + mpt_prt(mpt, "PRLO from 0x%08x%08x\n", + le32toh(rp->Wwn.PortNameHigh), + le32toh(rp->Wwn.PortNameLow)); + mpt_fc_els_send_response(mpt, req, rp, 20); + do_refresh = FALSE; + break; + default: + mpt_prt(mpt, "ELS TYPE 1 COMMAND: %x\n", cmd); + break; + } + } else if (rctl == ABTS && type == 0) { + uint16_t rx_id = le16toh(rp->Rxid); + uint16_t ox_id = le16toh(rp->Oxid); + request_t *tgt_req = NULL; + + mpt_prt(mpt, + "ELS: ABTS OX_ID 0x%x RX_ID 0x%x from 0x%08x%08x\n", + ox_id, rx_id, le32toh(rp->Wwn.PortNameHigh), + le32toh(rp->Wwn.PortNameLow)); + if (rx_id >= mpt->mpt_max_tgtcmds) { + mpt_prt(mpt, "Bad RX_ID 0x%x\n", rx_id); + } else if (mpt->tgt_cmd_ptrs == NULL) { + mpt_prt(mpt, "No TGT CMD PTRS\n"); + } else { + tgt_req = mpt->tgt_cmd_ptrs[rx_id]; + } + if (tgt_req) { + mpt_tgt_state_t *tgt = MPT_TGT_STATE(mpt, tgt_req); + uint8_t *vbuf; + union ccb *ccb = tgt->ccb; + cam_status cs; + uint32_t ct_id; + + vbuf = tgt_req->req_vbuf; + vbuf += MPT_RQSL(mpt); + + /* + * Check to make sure we have the correct command + * The reply descriptor in the target state should + * should contain an IoIndex that should match the + * RX_ID. + * + * It'd be nice to have OX_ID to crosscheck with + * as well. + */ + ct_id = GET_IO_INDEX(tgt->reply_desc); + + if (ct_id != rx_id) { + mpt_lprt(mpt, MPT_PRT_ERROR, "ABORT Mismatch: " + "RX_ID received=0x%x; RX_ID in cmd=0x%x\n", + rx_id, ct_id); + goto skip; + } + + ccb = tgt->ccb; + if (ccb) { + mpt_prt(mpt, + "CCB (%p): lun %u flags %x status %x\n", + ccb, ccb->ccb_h.target_lun, + ccb->ccb_h.flags, ccb->ccb_h.status); + } + mpt_prt(mpt, "target state 0x%x resid %u xfrd %u rpwrd " + "%x nxfers %x flags %x\n", tgt->state, + tgt->resid, tgt->bytes_xfered, tgt->reply_desc, + tgt->nxfers, tgt->flags); + skip: + cs = mpt_abort_target_cmd(mpt, tgt_req); + if (cs != CAM_REQ_INPROG) { + mpt_prt(mpt, "unable to do TargetAbort (%x)\n", + cs); + } + } else { + mpt_prt(mpt, "no back pointer for RX_ID 0x%x\n", rx_id); + } + memset(elsbuf, 0, 5 * (sizeof (U32))); + elsbuf[0] = htobe32(0); + elsbuf[1] = htobe32((ox_id << 16) | rx_id); + elsbuf[2] = htobe32(0x000ffff); + /* + * Dork with the reply frame so that the reponse to it + * will be correct. + */ + rp->Rctl_Did += ((BA_ACC - ABTS) << MPI_FC_RCTL_SHIFT); + mpt_fc_els_send_response(mpt, req, rp, 12); + do_refresh = FALSE; + } else { + mpt_prt(mpt, "ELS: RCTL %x TYPE %x CMD %x\n", rctl, type, cmd); + } + if (do_refresh == TRUE) { + mpt_fc_add_els(mpt, req); + } + return (TRUE); +} + +/* + * WE_TRUST_AUTO_GOOD_STATUS- I've found that setting + * TARGET_STATUS_SEND_FLAGS_AUTO_GOOD_STATUS leads the + * FC929 to set bogus FC_RSP fields (nonzero residuals + * but w/o RESID fields set). This causes QLogic initiators + * to think maybe that a frame was lost. + * + * WE_CAN_USE_AUTO_REPOST- we can't use AUTO_REPOST because + * we use allocated requests to do TARGET_ASSIST and we + * need to know when to release them. + */ + +static void +mpt_scsi_tgt_status(struct mpt_softc *mpt, union ccb *ccb, request_t *cmd_req, + uint8_t status, uint8_t const *sense_data) +{ + uint8_t *cmd_vbuf; + mpt_tgt_state_t *tgt; + PTR_MSG_TARGET_STATUS_SEND_REQUEST tp; + request_t *req; + bus_addr_t paddr; + int resplen = 0; + + cmd_vbuf = cmd_req->req_vbuf; + cmd_vbuf += MPT_RQSL(mpt); + tgt = MPT_TGT_STATE(mpt, cmd_req); + + req = mpt_get_request(mpt, FALSE); + if (req == NULL) { + if (ccb) { + ccb->ccb_h.status &= ~CAM_STATUS_MASK; + mpt_set_ccb_status(ccb, CAM_REQUEUE_REQ); + MPTLOCK_2_CAMLOCK(mpt); + xpt_done(ccb); + CAMLOCK_2_MPTLOCK(mpt); + } else { + /* + * XXX: put in deferred processing if we cannot allocate + */ + mpt_prt(mpt, + "XXXX could not allocate status req- dropping\n"); + } + return; + } + req->ccb = ccb; + if (ccb) { + ccb->ccb_h.ccb_mpt_ptr = mpt; + ccb->ccb_h.ccb_req_ptr = req; + } + + /* + * Record the currently active ccb, if any, and the + * request for it in our target state area. + */ + tgt->ccb = ccb; + tgt->req = req; + tgt->state = TGT_STATE_SENDING_STATUS; + + tp = req->req_vbuf; + paddr = req->req_pbuf; + paddr += MPT_RQSL(mpt); + + memset(tp, 0, sizeof (*tp)); + tp->Function = MPI_FUNCTION_TARGET_STATUS_SEND; + if (mpt->is_fc) { + PTR_MPI_TARGET_FCP_CMD_BUFFER fc = + (PTR_MPI_TARGET_FCP_CMD_BUFFER) cmd_vbuf; + uint8_t *sts_vbuf; + uint32_t *rsp; + + sts_vbuf = req->req_vbuf; + sts_vbuf += MPT_RQSL(mpt); + rsp = (uint32_t *) sts_vbuf; + memcpy(tp->LUN, fc->FcpLun, sizeof (tp->LUN)); + + /* + * The MPI_TARGET_FCP_RSP_BUFFER define is unfortunate. + * It has to be big-endian in memory and is organized + * in 32 bit words, which are much easier to deal with + * as words which are swizzled as needed. + * + * All we're filling here is the FC_RSP payload. + * We may just have the chip synthesize it if + * we have no residual and an OK status. + * + */ + memset(rsp, 0, sizeof (MPI_TARGET_FCP_RSP_BUFFER)); + + rsp[2] = status; + if (tgt->resid) { + rsp[2] |= 0x800; + rsp[3] = htobe32(tgt->resid); +#ifdef WE_TRUST_AUTO_GOOD_STATUS + resplen = sizeof (MPI_TARGET_FCP_RSP_BUFFER); +#endif + } + if (status == SCSI_STATUS_CHECK_COND) { + int i; + + rsp[2] |= 0x200; + rsp[4] = htobe32(MPT_SENSE_SIZE); + memcpy(&rsp[8], sense_data, MPT_SENSE_SIZE); + for (i = 8; i < (8 + (MPT_SENSE_SIZE >> 2)); i++) { + rsp[i] = htobe32(rsp[i]); + } +#ifdef WE_TRUST_AUTO_GOOD_STATUS + resplen = sizeof (MPI_TARGET_FCP_RSP_BUFFER); +#endif + } +#ifndef WE_TRUST_AUTO_GOOD_STATUS + resplen = sizeof (MPI_TARGET_FCP_RSP_BUFFER); +#endif + rsp[2] = htobe32(rsp[2]); + } else if (mpt->is_sas) { + PTR_MPI_TARGET_SSP_CMD_BUFFER ssp = + (PTR_MPI_TARGET_SSP_CMD_BUFFER) cmd_vbuf; + memcpy(tp->LUN, ssp->LogicalUnitNumber, sizeof (tp->LUN)); + } else { + PTR_MPI_TARGET_SCSI_SPI_CMD_BUFFER sp = + (PTR_MPI_TARGET_SCSI_SPI_CMD_BUFFER) cmd_vbuf; + tp->StatusCode = status; + tp->QueueTag = htole16(sp->Tag); + memcpy(tp->LUN, sp->LogicalUnitNumber, sizeof (tp->LUN)); + } + + tp->ReplyWord = htole32(tgt->reply_desc); + tp->MsgContext = htole32(req->index | mpt->scsi_tgt_handler_id); + +#ifdef WE_CAN_USE_AUTO_REPOST + tp->MsgFlags = TARGET_STATUS_SEND_FLAGS_REPOST_CMD_BUFFER; +#endif + if (status == SCSI_STATUS_OK && resplen == 0) { + tp->MsgFlags |= TARGET_STATUS_SEND_FLAGS_AUTO_GOOD_STATUS; + } else { + tp->StatusDataSGE.u.Address32 = (uint32_t) paddr; + tp->StatusDataSGE.FlagsLength = + MPI_SGE_FLAGS_HOST_TO_IOC | + MPI_SGE_FLAGS_SIMPLE_ELEMENT | + MPI_SGE_FLAGS_LAST_ELEMENT | + MPI_SGE_FLAGS_END_OF_LIST | + MPI_SGE_FLAGS_END_OF_BUFFER; + tp->StatusDataSGE.FlagsLength <<= MPI_SGE_FLAGS_SHIFT; + tp->StatusDataSGE.FlagsLength |= resplen; + } + + mpt_lprt(mpt, MPT_PRT_DEBUG, + "STATUS_CCB %p (wit%s sense) tag %x req %p:%u resid %u\n", + ccb, sense_data?"h" : "hout", ccb? ccb->csio.tag_id : -1, req, + req->serno, tgt->resid); + if (ccb) { + ccb->ccb_h.status = CAM_SIM_QUEUED | CAM_REQ_INPROG; + ccb->ccb_h.timeout_ch = timeout(mpt_timeout, (caddr_t)ccb, hz); + } + mpt_send_cmd(mpt, req); +} + +static void +mpt_scsi_tgt_tsk_mgmt(struct mpt_softc *mpt, request_t *req, mpt_task_mgmt_t fc, + tgt_resource_t *trtp, int init_id) +{ + struct ccb_immed_notify *inot; + mpt_tgt_state_t *tgt; + + tgt = MPT_TGT_STATE(mpt, req); + inot = (struct ccb_immed_notify *) STAILQ_FIRST(&trtp->inots); + if (inot == NULL) { + mpt_lprt(mpt, MPT_PRT_WARN, "no INOTSs- sending back BSY\n"); + mpt_scsi_tgt_status(mpt, NULL, req, SCSI_STATUS_BUSY, NULL); + return; + } + STAILQ_REMOVE_HEAD(&trtp->inots, sim_links.stqe); + mpt_lprt(mpt, MPT_PRT_DEBUG1, + "Get FREE INOT %p lun %d\n", inot, inot->ccb_h.target_lun); + + memset(&inot->sense_data, 0, sizeof (inot->sense_data)); + inot->sense_len = 0; + memset(inot->message_args, 0, sizeof (inot->message_args)); + inot->initiator_id = init_id; /* XXX */ + + /* + * This is a somewhat grotesque attempt to map from task management + * to old style SCSI messages. God help us all. + */ + switch (fc) { + case MPT_ABORT_TASK_SET: + inot->message_args[0] = MSG_ABORT_TAG; + break; + case MPT_CLEAR_TASK_SET: + inot->message_args[0] = MSG_CLEAR_TASK_SET; + break; + case MPT_TARGET_RESET: + inot->message_args[0] = MSG_TARGET_RESET; + break; + case MPT_CLEAR_ACA: + inot->message_args[0] = MSG_CLEAR_ACA; + break; + case MPT_TERMINATE_TASK: + inot->message_args[0] = MSG_ABORT_TAG; + break; + default: + inot->message_args[0] = MSG_NOOP; + break; + } + tgt->ccb = (union ccb *) inot; + inot->ccb_h.status = CAM_MESSAGE_RECV|CAM_DEV_QFRZN; + MPTLOCK_2_CAMLOCK(mpt); + xpt_done((union ccb *)inot); + CAMLOCK_2_MPTLOCK(mpt); +} + +static void +mpt_scsi_tgt_atio(struct mpt_softc *mpt, request_t *req, uint32_t reply_desc) +{ + struct ccb_accept_tio *atiop; + lun_id_t lun; + int tag_action = 0; + mpt_tgt_state_t *tgt; + tgt_resource_t *trtp; + U8 *lunptr; + U8 *vbuf; + U16 itag; + U16 ioindex; + mpt_task_mgmt_t fct = MPT_NIL_TMT_VALUE; + uint8_t *cdbp; + + /* + * First, DMA sync the received command- which is in the *request* + * phys area. + * XXX: We could optimize this for a range + */ + bus_dmamap_sync(mpt->request_dmat, mpt->request_dmap, + BUS_DMASYNC_POSTREAD); + + /* + * Stash info for the current command where we can get at it later. + */ + vbuf = req->req_vbuf; + vbuf += MPT_RQSL(mpt); + + /* + * Get our state pointer set up. + */ + tgt = MPT_TGT_STATE(mpt, req); + KASSERT(tgt->state == TGT_STATE_LOADED, + ("bad target state %x in mpt_scsi_tgt_atio for req %p\n", + tgt->state, req)); + memset(tgt, 0, sizeof (mpt_tgt_state_t)); + tgt->state = TGT_STATE_IN_CAM; + tgt->reply_desc = reply_desc; + ioindex = GET_IO_INDEX(reply_desc); + + if (mpt->is_fc) { + PTR_MPI_TARGET_FCP_CMD_BUFFER fc; + fc = (PTR_MPI_TARGET_FCP_CMD_BUFFER) vbuf; + if (fc->FcpCntl[2]) { + /* + * Task Management Request + */ + switch (fc->FcpCntl[2]) { + case 0x2: + fct = MPT_ABORT_TASK_SET; + break; + case 0x4: + fct = MPT_CLEAR_TASK_SET; + break; + case 0x20: + fct = MPT_TARGET_RESET; + break; + case 0x40: + fct = MPT_CLEAR_ACA; + break; + case 0x80: + fct = MPT_TERMINATE_TASK; + break; + default: + mpt_prt(mpt, "CORRUPTED TASK MGMT BITS: 0x%x\n", + fc->FcpCntl[2]); + mpt_scsi_tgt_status(mpt, 0, req, + SCSI_STATUS_OK, 0); + return; + } + return; + } + switch (fc->FcpCntl[1]) { + case 0: + tag_action = MSG_SIMPLE_Q_TAG; + break; + case 1: + tag_action = MSG_HEAD_OF_Q_TAG; + break; + case 2: + tag_action = MSG_ORDERED_Q_TAG; + break; + default: + /* + * Bah. Ignore Untagged Queing and ACA + */ + tag_action = MSG_SIMPLE_Q_TAG; + break; + } + tgt->resid = be32toh(fc->FcpDl); + cdbp = fc->FcpCdb; + lunptr = fc->FcpLun; + itag = be16toh(fc->OptionalOxid); + } else if (mpt->is_sas) { + PTR_MPI_TARGET_SSP_CMD_BUFFER ssp; + ssp = (PTR_MPI_TARGET_SSP_CMD_BUFFER) vbuf; + cdbp = ssp->CDB; + lunptr = ssp->LogicalUnitNumber; + itag = ssp->InitiatorTag; + } else { + PTR_MPI_TARGET_SCSI_SPI_CMD_BUFFER sp; + sp = (PTR_MPI_TARGET_SCSI_SPI_CMD_BUFFER) vbuf; + cdbp = sp->CDB; + lunptr = sp->LogicalUnitNumber; + itag = sp->Tag; + } + + /* + * Generate a simple lun + */ + switch (lunptr[0] & 0xc0) { + case 0x40: + lun = ((lunptr[0] & 0x3f) << 8) | lunptr[1]; + break; + case 0: + lun = lunptr[1]; + break; + default: + mpt_lprt(mpt, MPT_PRT_ERROR, "cannot handle this type lun\n"); + lun = 0xffff; + break; + } + + /* + * Deal with non-enabled or bad luns here. + */ + if (lun >= MPT_MAX_LUNS || mpt->tenabled == 0 || + mpt->trt[lun].enabled == 0) { + if (mpt->twildcard) { + trtp = &mpt->trt_wildcard; + } else { + const uint8_t sp[MPT_SENSE_SIZE] = { + 0xf0, 0, 0x5, 0, 0, 0, 0, 8, 0, 0, 0, 0, 0x25 + }; + mpt_scsi_tgt_status(mpt, NULL, req, + SCSI_STATUS_CHECK_COND, sp); + return; + } + } else { + trtp = &mpt->trt[lun]; + } + + if (fct != MPT_NIL_TMT_VALUE) { + /* undo any tgt residual settings */ + tgt->resid = 0; + mpt_scsi_tgt_tsk_mgmt(mpt, req, fct, trtp, + GET_INITIATOR_INDEX(reply_desc)); + return; + } + + atiop = (struct ccb_accept_tio *) STAILQ_FIRST(&trtp->atios); + if (atiop == NULL) { + mpt_lprt(mpt, MPT_PRT_WARN, + "no ATIOs for lun %u- sending back %s\n", lun, + mpt->tenabled? "QUEUE FULL" : "BUSY"); + mpt_scsi_tgt_status(mpt, NULL, req, + mpt->tenabled? SCSI_STATUS_QUEUE_FULL : SCSI_STATUS_BUSY, + NULL); + return; + } + STAILQ_REMOVE_HEAD(&trtp->atios, sim_links.stqe); + mpt_lprt(mpt, MPT_PRT_DEBUG1, + "Get FREE ATIO %p lun %d\n", atiop, atiop->ccb_h.target_lun); + atiop->ccb_h.ccb_mpt_ptr = mpt; + atiop->ccb_h.status = CAM_CDB_RECVD; + atiop->ccb_h.target_lun = lun; + atiop->sense_len = 0; + atiop->init_id = GET_INITIATOR_INDEX(reply_desc); + atiop->cdb_len = mpt_cdblen(cdbp[0], 16); + memcpy(atiop->cdb_io.cdb_bytes, cdbp, atiop->cdb_len); + + /* + * The tag we construct here allows us to find the + * original request that the command came in with. + * + * This way we don't have to depend on anything but the + * tag to find things when CCBs show back up from CAM. + */ + atiop->tag_id = MPT_MAKE_TAGID(mpt, req, ioindex); + if (tag_action) { + atiop->tag_action = tag_action; + atiop->ccb_h.flags = CAM_TAG_ACTION_VALID; + } + if (mpt->verbose >= MPT_PRT_DEBUG) { + int i; + mpt_prt(mpt, "START_CCB %p for lun %u CDB=<", atiop, + atiop->ccb_h.target_lun); + for (i = 0; i < atiop->cdb_len; i++) { + mpt_prtc(mpt, "%02x%c", cdbp[i] & 0xff, + (i == (atiop->cdb_len - 1))? '>' : ' '); + } + mpt_prtc(mpt, " itag %x tag %x rdesc %x dl=%u\n", + itag, atiop->tag_id, tgt->reply_desc, tgt->resid); + } + tgt->ccb = (union ccb *) atiop; + + MPTLOCK_2_CAMLOCK(mpt); + xpt_done((union ccb *)atiop); + CAMLOCK_2_MPTLOCK(mpt); +} + +static int +mpt_scsi_tgt_reply_handler(struct mpt_softc *mpt, request_t *req, + uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame) +{ + int dbg; + union ccb *ccb; + U16 status; + + if (reply_frame == NULL) { + /* + * Figure out if this is a new command or a target assist + * completing. + */ + mpt_tgt_state_t *tgt = MPT_TGT_STATE(mpt, req); + char serno[8]; + + if (tgt->req) { + snprintf(serno, 8, "%u", tgt->req->serno); + } else { + strncpy(serno, "??", 8); + } + + switch(tgt->state) { + case TGT_STATE_LOADED: + mpt_scsi_tgt_atio(mpt, req, reply_desc); + break; + case TGT_STATE_MOVING_DATA: + { + uint8_t *sp = NULL, sense[MPT_SENSE_SIZE]; + + ccb = tgt->ccb; + tgt->ccb = NULL; + tgt->nxfers++; + untimeout(mpt_timeout, ccb, ccb->ccb_h.timeout_ch); + mpt_lprt(mpt, MPT_PRT_DEBUG, + "TARGET_ASSIST %p (req %p:%s) done tag 0x%x\n", + ccb, tgt->req, serno, ccb->csio.tag_id); + /* + * Free the Target Assist Request + */ + KASSERT(tgt->req && tgt->req->ccb == ccb, + ("tgt->req %p:%s tgt->req->ccb %p", tgt->req, + serno, tgt->req? tgt->req->ccb : NULL)); + mpt_free_request(mpt, tgt->req); + tgt->req = NULL; + /* + * Do we need to send status now? That is, are + * we done with all our data transfers? + */ + if ((ccb->ccb_h.flags & CAM_SEND_STATUS) == 0) { + mpt_set_ccb_status(ccb, CAM_REQ_CMP); + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + KASSERT(ccb->ccb_h.status, + ("zero ccb sts at %d\n", __LINE__)); + tgt->state = TGT_STATE_IN_CAM; + MPTLOCK_2_CAMLOCK(mpt); + xpt_done(ccb); + CAMLOCK_2_MPTLOCK(mpt); + break; + } + if (ccb->ccb_h.flags & CAM_SEND_SENSE) { + sp = sense; + memcpy(sp, &ccb->csio.sense_data, + min(ccb->csio.sense_len, MPT_SENSE_SIZE)); + } + mpt_scsi_tgt_status(mpt, ccb, req, + ccb->csio.scsi_status, sp); + break; + } + case TGT_STATE_SENDING_STATUS: + case TGT_STATE_MOVING_DATA_AND_STATUS: + { + int ioindex; + ccb = tgt->ccb; + + if (ccb) { + tgt->ccb = NULL; + tgt->nxfers++; + untimeout(mpt_timeout, ccb, + ccb->ccb_h.timeout_ch); + if (ccb->ccb_h.flags & CAM_SEND_SENSE) { + ccb->ccb_h.status |= CAM_SENT_SENSE; + } + mpt_lprt(mpt, MPT_PRT_DEBUG, + "TARGET_STATUS tag %x sts %x flgs %x req " + "%p\n", ccb->csio.tag_id, ccb->ccb_h.status, + ccb->ccb_h.flags, tgt->req); + /* + * Free the Target Send Status Request + */ + KASSERT(tgt->req && tgt->req->ccb == ccb, + ("tgt->req %p:%s tgt->req->ccb %p", + tgt->req, serno, + tgt->req? tgt->req->ccb : NULL)); + /* + * Notify CAM that we're done + */ + mpt_set_ccb_status(ccb, CAM_REQ_CMP); + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + KASSERT(ccb->ccb_h.status, + ("ZERO ccb sts at %d\n", __LINE__)); + tgt->ccb = NULL; + } else { + mpt_lprt(mpt, MPT_PRT_DEBUG, + "TARGET_STATUS non-CAM for req %p:%s\n", + tgt->req, serno); + } + mpt_free_request(mpt, tgt->req); + tgt->req = NULL; + + /* + * And re-post the Command Buffer. + */ + ioindex = GET_IO_INDEX(reply_desc); + mpt_post_target_command(mpt, req, ioindex); + + /* + * And post a done for anyone who cares + */ + if (ccb) { + MPTLOCK_2_CAMLOCK(mpt); + xpt_done(ccb); + CAMLOCK_2_MPTLOCK(mpt); + } + break; + } + case TGT_STATE_NIL: /* XXX This Never Happens XXX */ + tgt->state = TGT_STATE_LOADED; + break; + default: + mpt_prt(mpt, "Unknown Target State 0x%x in Context " + "Reply Function\n", tgt->state); + } + return (TRUE); + } + + status = le16toh(reply_frame->IOCStatus); + if (status != MPI_IOCSTATUS_SUCCESS) { + dbg = MPT_PRT_ERROR; + } else { + dbg = MPT_PRT_DEBUG1; + } + + mpt_lprt(mpt, dbg, + "SCSI_TGT REPLY: req=%p:%u reply=%p func=%x IOCstatus 0x%x\n", + req, req->serno, reply_frame, reply_frame->Function, status); + + switch (reply_frame->Function) { + case MPI_FUNCTION_TARGET_CMD_BUFFER_POST: + KASSERT(MPT_TGT_STATE(mpt, + req)->state == TGT_STATE_NIL, + ("bad state %x on reply to buffer post\n", + MPT_TGT_STATE(mpt, req)->state)); + MPT_TGT_STATE(mpt, req)->state = TGT_STATE_LOADED; + break; + case MPI_FUNCTION_TARGET_ASSIST: + mpt_prt(mpt, + "TARGET_ASSIST err for request %p:%u (%x): status 0x%x\n", + req, req->serno, req->index, status); + mpt_free_request(mpt, req); + break; + case MPI_FUNCTION_TARGET_STATUS_SEND: + mpt_prt(mpt, + "TARGET_STATUS_SEND error for request %p:%u(%x): status " + "0x%x\n", req, req->serno, req->index, status); + mpt_free_request(mpt, req); + break; + case MPI_FUNCTION_TARGET_MODE_ABORT: + { + PTR_MSG_TARGET_MODE_ABORT_REPLY abtrp = + (PTR_MSG_TARGET_MODE_ABORT_REPLY) reply_frame; + PTR_MSG_TARGET_MODE_ABORT abtp = + (PTR_MSG_TARGET_MODE_ABORT) req->req_vbuf; + uint32_t cc = GET_IO_INDEX(le32toh(abtp->ReplyWord)); + mpt_prt(mpt, "ABORT RX_ID 0x%x Complete; status 0x%x cnt %u\n", + cc, le16toh(abtrp->IOCStatus), le32toh(abtrp->AbortCount)); + mpt_free_request(mpt, req); + break; + } + default: + mpt_prt(mpt, "Unknown Target Address Reply Function code: " + "0x%x\n", reply_frame->Function); + break; + } + return (TRUE); +} + /* * Clean up all SCSI Initiator personality state in response * to a controller reset. @@ -1354,12 +3163,12 @@ mpt_action(struct cam_sim *sim, union ccb *ccb) raid_passthru = (sim == mpt->phydisk_sim); tgt = ccb->ccb_h.target_id; - if (raid_passthru - && ccb->ccb_h.func_code != XPT_PATH_INQ - && ccb->ccb_h.func_code != XPT_RESET_BUS) { + if (raid_passthru && ccb->ccb_h.func_code != XPT_PATH_INQ && + ccb->ccb_h.func_code != XPT_RESET_BUS) { CAMLOCK_2_MPTLOCK(mpt); if (mpt_map_physdisk(mpt, ccb, &tgt) != 0) { - ccb->ccb_h.status = CAM_DEV_NOT_THERE; + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + mpt_set_ccb_status(ccb, CAM_DEV_NOT_THERE); MPTLOCK_2_CAMLOCK(mpt); xpt_done(ccb); return; @@ -1376,7 +3185,8 @@ mpt_action(struct cam_sim *sim, union ccb *ccb) */ if ((ccb->ccb_h.flags & CAM_CDB_POINTER) != 0) { if ((ccb->ccb_h.flags & CAM_CDB_PHYS) != 0) { - ccb->ccb_h.status = CAM_REQ_INVALID; + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + mpt_set_ccb_status(ccb, CAM_REQ_INVALID); xpt_done(ccb); break; } @@ -1385,7 +3195,8 @@ mpt_action(struct cam_sim *sim, union ccb *ccb) /* XXX Unless we implement the new 32byte message type */ if (ccb->csio.cdb_len > sizeof (((PTR_MSG_SCSI_IO_REQUEST)0)->CDB)) { - ccb->ccb_h.status = CAM_REQ_INVALID; + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + mpt_set_ccb_status(ccb, CAM_REQ_INVALID); xpt_done(ccb); return; } @@ -1405,17 +3216,35 @@ mpt_action(struct cam_sim *sim, union ccb *ccb) * will fall back to a hard reset should a bus * reset attempt fail. */ + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; mpt_set_ccb_status(ccb, CAM_REQ_CMP); xpt_done(ccb); break; case XPT_ABORT: - /* - * XXX: Need to implement - */ - ccb->ccb_h.status = CAM_UA_ABORT; + { + union ccb *accb = ccb->cab.abort_ccb; + CAMLOCK_2_MPTLOCK(mpt); + switch (accb->ccb_h.func_code) { + case XPT_ACCEPT_TARGET_IO: + case XPT_IMMED_NOTIFY: + ccb->ccb_h.status = mpt_abort_target_ccb(mpt, ccb); + break; + case XPT_CONT_TARGET_IO: + mpt_prt(mpt, "cannot abort active CTIOs yet\n"); + ccb->ccb_h.status = CAM_UA_ABORT; + break; + case XPT_SCSI_IO: + ccb->ccb_h.status = CAM_UA_ABORT; + break; + default: + ccb->ccb_h.status = CAM_REQ_INVALID; + break; + } + MPTLOCK_2_CAMLOCK(mpt); xpt_done(ccb); break; + } #ifdef CAM_NEW_TRAN_CODE #define IS_CURRENT_SETTINGS(c) (c->type == CTS_TYPE_CURRENT_SETTINGS) @@ -1440,7 +3269,8 @@ mpt_action(struct cam_sim *sim, union ccb *ccb) cts = &ccb->cts; if (!IS_CURRENT_SETTINGS(cts)) { mpt_prt(mpt, "Attempt to set User settings\n"); - ccb->ccb_h.status = CAM_REQ_INVALID; + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + mpt_set_ccb_status(ccb, CAM_REQ_INVALID); xpt_done(ccb); break; } @@ -1521,8 +3351,10 @@ mpt_action(struct cam_sim *sim, union ccb *ccb) } if (dval & DP_WIDTH) { if (mpt_setwidth(mpt, tgt, dval & DP_WIDE)) { -mpt_prt(mpt, "Set width Failed!\n"); - ccb->ccb_h.status = CAM_REQ_CMP_ERR; + mpt_prt(mpt, "Set width Failed!\n"); + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + mpt_set_ccb_status(ccb, + CAM_REQ_CMP_ERR); MPTLOCK_2_CAMLOCK(mpt); xpt_done(ccb); break; @@ -1530,8 +3362,10 @@ mpt_prt(mpt, "Set width Failed!\n"); } if (dval & DP_SYNC) { if (mpt_setsync(mpt, tgt, period, offset)) { -mpt_prt(mpt, "Set sync Failed!\n"); - ccb->ccb_h.status = CAM_REQ_CMP_ERR; + mpt_prt(mpt, "Set sync Failed!\n"); + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + mpt_set_ccb_status(ccb, + CAM_REQ_CMP_ERR); MPTLOCK_2_CAMLOCK(mpt); xpt_done(ccb); break; @@ -1542,7 +3376,8 @@ mpt_prt(mpt, "Set sync Failed!\n"); "SET tgt %d flags %x period %x off %x\n", tgt, dval, period, offset); } - ccb->ccb_h.status = CAM_REQ_CMP; + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + mpt_set_ccb_status(ccb, CAM_REQ_CMP); xpt_done(ccb); break; @@ -1722,7 +3557,8 @@ mpt_prt(mpt, "Set sync Failed!\n"); ? "ACTIVE" : "NVRAM", tgt, dval, pval, oval); } - ccb->ccb_h.status = CAM_REQ_CMP; + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + mpt_set_ccb_status(ccb, CAM_REQ_CMP); xpt_done(ccb); break; @@ -1732,12 +3568,13 @@ mpt_prt(mpt, "Set sync Failed!\n"); ccg = &ccb->ccg; if (ccg->block_size == 0) { - ccb->ccb_h.status = CAM_REQ_INVALID; + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + mpt_set_ccb_status(ccb, CAM_REQ_INVALID); xpt_done(ccb); break; } - mpt_calc_geometry(ccg, /*extended*/1); + KASSERT(ccb->ccb_h.status, ("zero ccb sts at %d\n", __LINE__)); xpt_done(ccb); break; } @@ -1766,16 +3603,16 @@ mpt_prt(mpt, "Set sync Failed!\n"); PI_SDTR_ABLE|PI_TAG_ABLE|PI_WIDE_16; } } else if (mpt->is_fc) { -/* XXX SHOULD BE BASED UPON IOC FACTS XXX */ + /* XXX SHOULD BE BASED UPON IOC FACTS XXX XXX */ cpi->max_target = 255; cpi->hba_misc = PIM_NOBUSRESET; - cpi->initiator_id = cpi->max_target + 1; + cpi->initiator_id = mpt->mpt_ini_id; cpi->base_transfer_speed = 100000; cpi->hba_inquiry = PI_TAG_ABLE; } else if (mpt->is_sas) { cpi->max_target = 63; /* XXX */ cpi->hba_misc = PIM_NOBUSRESET; - cpi->initiator_id = cpi->max_target; + cpi->initiator_id = mpt->mpt_ini_id; cpi->base_transfer_speed = 300000; cpi->hba_inquiry = PI_TAG_ABLE; } else { @@ -1786,6 +3623,15 @@ mpt_prt(mpt, "Set sync Failed!\n"); cpi->max_target = 15; } + if ((mpt->role & MPT_ROLE_INITIATOR) == 0) { + cpi->hba_misc |= PIM_NOINITIATOR; + } + if ((mpt->role & MPT_ROLE_TARGET) != 0) { + cpi->target_sprt = + PIT_PROCESSOR | PIT_DISCONNECT | PIT_TERM_IO; + } else { + cpi->target_sprt = 0; + } strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN); strncpy(cpi->hba_vid, "LSI", HBA_IDLEN); strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN); @@ -1794,6 +3640,73 @@ mpt_prt(mpt, "Set sync Failed!\n"); xpt_done(ccb); break; } + case XPT_EN_LUN: /* Enable LUN as a target */ + { + int result; + + CAMLOCK_2_MPTLOCK(mpt); + if (ccb->cel.enable) + result = mpt_enable_lun(mpt, + ccb->ccb_h.target_id, ccb->ccb_h.target_lun); + else + result = mpt_disable_lun(mpt, + ccb->ccb_h.target_id, ccb->ccb_h.target_lun); + MPTLOCK_2_CAMLOCK(mpt); + if (result == 0) { + mpt_set_ccb_status(ccb, CAM_REQ_CMP); + } else { + mpt_set_ccb_status(ccb, CAM_REQ_CMP_ERR); + } + xpt_done(ccb); + break; + } + case XPT_NOTIFY_ACK: /* recycle notify ack */ + case XPT_IMMED_NOTIFY: /* Add Immediate Notify Resource */ + case XPT_ACCEPT_TARGET_IO: /* Add Accept Target IO Resource */ + { + tgt_resource_t *trtp; + lun_id_t lun = ccb->ccb_h.target_lun; + ccb->ccb_h.sim_priv.entries[0].field = 0; + ccb->ccb_h.sim_priv.entries[1].ptr = mpt; + ccb->ccb_h.flags = 0; + + if (lun == CAM_LUN_WILDCARD) { + if (ccb->ccb_h.target_id != CAM_TARGET_WILDCARD) { + mpt_set_ccb_status(ccb, CAM_REQ_INVALID); + xpt_done(ccb); + break; + } + trtp = &mpt->trt_wildcard; + } else if (lun >= MPT_MAX_LUNS) { + mpt_set_ccb_status(ccb, CAM_REQ_INVALID); + xpt_done(ccb); + break; + } else { + trtp = &mpt->trt[lun]; + } + CAMLOCK_2_MPTLOCK(mpt); + if (ccb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) { + mpt_lprt(mpt, MPT_PRT_DEBUG1, + "Put FREE ATIO %p lun %d\n", ccb, lun); + STAILQ_INSERT_TAIL(&trtp->atios, &ccb->ccb_h, + sim_links.stqe); + } else if (ccb->ccb_h.func_code == XPT_IMMED_NOTIFY) { + mpt_lprt(mpt, MPT_PRT_DEBUG1, + "Put FREE INOT lun %d\n", lun); + STAILQ_INSERT_TAIL(&trtp->inots, &ccb->ccb_h, + sim_links.stqe); + } else { + mpt_lprt(mpt, MPT_PRT_ALWAYS, "Got Notify ACK\n"); + } + mpt_set_ccb_status(ccb, CAM_REQ_INPROG); + MPTLOCK_2_CAMLOCK(mpt); + break; + } + case XPT_CONT_TARGET_IO: + CAMLOCK_2_MPTLOCK(mpt); + mpt_target_start_io(mpt, ccb); + MPTLOCK_2_CAMLOCK(mpt); + break; default: ccb->ccb_h.status = CAM_REQ_INVALID; xpt_done(ccb); @@ -1892,6 +3805,10 @@ mpt_calc_geometry(struct ccb_calc_geometry *ccg, int extended) uint32_t size_mb; uint32_t secs_per_cylinder; + if (ccg->block_size == 0) { + ccg->ccb_h.status = CAM_REQ_INVALID; + return; + } size_mb = ccg->volume_size / ((1024L * 1024L) / ccg->block_size); if (size_mb > 1024 && extended) { ccg->heads = 255; @@ -1918,16 +3835,10 @@ mpt_spawn_recovery_thread(struct mpt_softc *mpt) return (error); } -/* - * Lock is not held on entry. - */ static void mpt_terminate_recovery_thread(struct mpt_softc *mpt) { - - MPT_LOCK(mpt); if (mpt->recovery_thread == NULL) { - MPT_UNLOCK(mpt); return; } mpt->shutdwn_recovery = 1; @@ -1937,7 +3848,6 @@ mpt_terminate_recovery_thread(struct mpt_softc *mpt) * for this interlock just for added safety. */ mpt_sleep(mpt, &mpt->recovery_thread, PUSER, "thtrm", 0); - MPT_UNLOCK(mpt); } static void @@ -1956,12 +3866,10 @@ mpt_recovery_thread(void *arg) && mpt->shutdwn_recovery == 0) mpt_sleep(mpt, mpt, PUSER, "idle", 0); - if (mpt->shutdwn_recovery != 0) + if (mpt->shutdwn_recovery != 0) { break; - - MPT_UNLOCK(mpt); + } mpt_recover_commands(mpt); - MPT_LOCK(mpt); } mpt->recovery_thread = NULL; wakeup(&mpt->recovery_thread); @@ -2016,12 +3924,443 @@ mpt_scsi_send_tmf(struct mpt_softc *mpt, u_int type, if (mpt->verbose > MPT_PRT_DEBUG) mpt_print_request(tmf_req); - error = mpt_send_handshake_cmd(mpt, sizeof(*tmf_req), tmf_req); - if (error != 0) + if (mpt->is_fc || mpt->is_sas) { + mpt_send_cmd(mpt, mpt->tmf_req); + error = MPT_OK; + } else { + error = mpt_send_handshake_cmd(mpt, sizeof(*tmf_req), tmf_req); + } + if (error != MPT_OK) { mpt_reset(mpt, /*reinit*/TRUE); + } return (error); } +static void +mpt_fc_add_els(struct mpt_softc *mpt, request_t *req) +{ + MSG_LINK_SERVICE_BUFFER_POST_REQUEST *fc; + PTR_SGE_TRANSACTION32 tep; + PTR_SGE_SIMPLE32 se; + bus_addr_t paddr; + + paddr = req->req_pbuf; + paddr += MPT_RQSL(mpt); + + fc = req->req_vbuf; + memset(fc, 0, MPT_REQUEST_AREA); + fc->BufferCount = 1; + fc->Function = MPI_FUNCTION_FC_LINK_SRVC_BUF_POST; + fc->MsgContext = htole32(req->index | fc_els_handler_id); + + /* + * Okay, set up ELS buffer pointers. ELS buffer pointers + * consist of a TE SGL element (with details length of zero) + * followe by a SIMPLE SGL element which holds the address + * of the buffer. + */ + + tep = (PTR_SGE_TRANSACTION32) &fc->SGL; + + tep->ContextSize = 4; + tep->Flags = 0; + tep->TransactionContext[0] = htole32(req->index | fc_els_handler_id); + + se = (PTR_SGE_SIMPLE32) &tep->TransactionDetails[0]; + se->FlagsLength = + MPI_SGE_FLAGS_HOST_TO_IOC | + MPI_SGE_FLAGS_SIMPLE_ELEMENT | + MPI_SGE_FLAGS_LAST_ELEMENT | + MPI_SGE_FLAGS_END_OF_LIST | + MPI_SGE_FLAGS_END_OF_BUFFER; + se->FlagsLength <<= MPI_SGE_FLAGS_SHIFT; + se->FlagsLength |= (MPT_NRFM(mpt) - MPT_RQSL(mpt)); + se->Address = (uint32_t) paddr; + mpt_check_doorbell(mpt); + mpt_send_cmd(mpt, req); +} + +static void +mpt_post_target_command(struct mpt_softc *mpt, request_t *req, int ioindex) +{ + PTR_MSG_TARGET_CMD_BUFFER_POST_REQUEST fc; + PTR_CMD_BUFFER_DESCRIPTOR cb; + bus_addr_t paddr; + + paddr = req->req_pbuf; + paddr += MPT_RQSL(mpt); + memset(req->req_vbuf, 0, MPT_REQUEST_AREA); + + fc = req->req_vbuf; + fc->BufferCount = 1; + fc->Function = MPI_FUNCTION_TARGET_CMD_BUFFER_POST; + fc->MsgContext = htole32(req->index | mpt->scsi_tgt_handler_id); + + cb = &fc->Buffer[0]; + cb->IoIndex = htole16(ioindex); + cb->u.PhysicalAddress32 = (U32) paddr; + + mpt_check_doorbell(mpt); + mpt_send_cmd(mpt, req); +} + +static void +mpt_add_target_commands(struct mpt_softc *mpt) +{ + int i, max; + + if (mpt->tgt_cmd_ptrs) { + return; + } + + max = MPT_MAX_REQUESTS(mpt) >> 1; + if (max > mpt->mpt_max_tgtcmds) { + max = mpt->mpt_max_tgtcmds; + } + mpt->tgt_cmd_ptrs = + malloc(max * sizeof (void *), M_DEVBUF, M_NOWAIT | M_ZERO); + if (mpt->tgt_cmd_ptrs == NULL) { + mpt_prt(mpt, "could not allocate cmdptrs\n"); + return; + } + mpt->tgt_cmds_allocated = max; + + for (i = 0; i < max; i++) { + request_t *req; + + req = mpt_get_request(mpt, FALSE); + if (req == NULL) { + break; + } + mpt->tgt_cmd_ptrs[i] = req; + mpt_post_target_command(mpt, req, i); + } + + if (i == 0) { + mpt_lprt(mpt, MPT_PRT_ERROR, "could not add any target bufs\n"); + free(mpt->tgt_cmd_ptrs, M_DEVBUF); + mpt->tgt_cmd_ptrs = NULL; + mpt->tgt_cmds_allocated = 0; + } else if (i < max) { + mpt_lprt(mpt, MPT_PRT_WARN, "added %d of %d target bufs\n", + i, max); + } +} + +static int +mpt_enable_lun(struct mpt_softc *mpt, target_id_t tgt, lun_id_t lun) +{ + if (tgt == CAM_TARGET_WILDCARD && lun == CAM_LUN_WILDCARD) { + mpt->twildcard = 1; + } else if (lun >= MPT_MAX_LUNS) { + return (EINVAL); + } else if (tgt != CAM_TARGET_WILDCARD && tgt != 0) { + return (EINVAL); + } + if (mpt->tenabled == 0) { +#if 0 + if (mpt->is_fc) { + (void) mpt_fc_reset_link(mpt, 0); + } +#endif + mpt->tenabled = 1; + } + if (lun == CAM_LUN_WILDCARD) { + mpt->trt_wildcard.enabled = 1; + } else { + mpt->trt[lun].enabled = 1; + } + return (0); +} + +static int +mpt_disable_lun(struct mpt_softc *mpt, target_id_t tgt, lun_id_t lun) +{ + int i; + if (tgt == CAM_TARGET_WILDCARD && lun == CAM_LUN_WILDCARD) { + mpt->twildcard = 0; + } else if (lun >= MPT_MAX_LUNS) { + return (EINVAL); + } else if (tgt != CAM_TARGET_WILDCARD && tgt != 0) { + return (EINVAL); + } + if (lun == CAM_LUN_WILDCARD) { + mpt->trt_wildcard.enabled = 0; + } else { + mpt->trt[lun].enabled = 0; + } + for (i = 0; i < MPT_MAX_LUNS; i++) { + if (mpt->trt[lun].enabled) { + break; + } + } + if (i == MPT_MAX_LUNS && mpt->twildcard == 0) { + mpt->tenabled = 0; +#if 0 + if (mpt->is_fc) { + (void) mpt_fc_reset_link(mpt, 0); + } +#endif + } + return (0); +} + +/* + * Called with MPT lock held + */ +static void +mpt_target_start_io(struct mpt_softc *mpt, union ccb *ccb) +{ + struct ccb_scsiio *csio = &ccb->csio; + request_t *cmd_req = MPT_TAG_2_REQ(mpt, csio->tag_id); + mpt_tgt_state_t *tgt = MPT_TGT_STATE(mpt, cmd_req); + + + if (tgt->state != TGT_STATE_IN_CAM) { + mpt_prt(mpt, "tag 0x%08x in state %x when starting I/O\n", + csio->tag_id, tgt->state); + mpt_set_ccb_status(ccb, CAM_REQUEUE_REQ); + MPTLOCK_2_CAMLOCK(mpt); + xpt_done(ccb); + CAMLOCK_2_MPTLOCK(mpt); + return; + } + + if (csio->dxfer_len) { + bus_dmamap_callback_t *cb; + PTR_MSG_TARGET_ASSIST_REQUEST ta; + request_t *req; + + KASSERT((ccb->ccb_h.flags & CAM_DIR_MASK) != CAM_DIR_NONE, + ("dxfer_len %u but direction is NONE\n", csio->dxfer_len)); + + req = mpt_get_request(mpt, FALSE); + if (req == NULL) { + mpt_set_ccb_status(ccb, CAM_REQUEUE_REQ); + MPTLOCK_2_CAMLOCK(mpt); + xpt_done(ccb); + CAMLOCK_2_MPTLOCK(mpt); + return; + } + + ccb->ccb_h.status = CAM_SIM_QUEUED | CAM_REQ_INPROG; + if (sizeof (bus_addr_t) > 4) { + cb = mpt_execute_req_a64; + } else { + cb = mpt_execute_req; + } + + req->ccb = ccb; + ccb->ccb_h.ccb_req_ptr = req; + + /* + * Record the currently active ccb and the + * request for it in our target state area. + */ + tgt->ccb = ccb; + tgt->req = req; + + memset(req->req_vbuf, 0, MPT_RQSL(mpt)); + ta = req->req_vbuf; + + if (mpt->is_fc) { + ; + } else if (mpt->is_sas == 0) { + PTR_MPI_TARGET_SSP_CMD_BUFFER ssp = + cmd_req->req_vbuf; + ta->QueueTag = ssp->InitiatorTag; + } else { + PTR_MPI_TARGET_SCSI_SPI_CMD_BUFFER sp = + cmd_req->req_vbuf; + ta->QueueTag = sp->Tag; + } + ta->Function = MPI_FUNCTION_TARGET_ASSIST; + ta->MsgContext = htole32(req->index | mpt->scsi_tgt_handler_id); + ta->ReplyWord = htole32(tgt->reply_desc); + if (csio->ccb_h.target_lun > 256) { + ta->LUN[0] = + 0x40 | ((csio->ccb_h.target_lun >> 8) & 0x3f); + ta->LUN[1] = csio->ccb_h.target_lun & 0xff; + } else { + ta->LUN[1] = csio->ccb_h.target_lun; + } + + ta->RelativeOffset = tgt->bytes_xfered; + ta->DataLength = ccb->csio.dxfer_len; + if (ta->DataLength > tgt->resid) { + ta->DataLength = tgt->resid; + } + + /* + * XXX Should be done after data transfer completes? + */ + tgt->resid -= csio->dxfer_len; + tgt->bytes_xfered += csio->dxfer_len; + + if ((ccb->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { + ta->TargetAssistFlags |= + TARGET_ASSIST_FLAGS_DATA_DIRECTION; + } + +#ifdef WE_TRUST_AUTO_GOOD_STATUS + if ((ccb->ccb_h.flags & CAM_SEND_STATUS) && + csio->scsi_status == SCSI_STATUS_OK && tgt->resid == 0) { + ta->TargetAssistFlags |= + TARGET_ASSIST_FLAGS_AUTO_STATUS; + } +#endif + tgt->state = TGT_STATE_SETTING_UP_FOR_DATA; + + mpt_lprt(mpt, MPT_PRT_DEBUG, + "DATA_CCB %p tag %x %u bytes %u resid flg %x req %p:%u " + "nxtstate=%d\n", csio, csio->tag_id, csio->dxfer_len, + tgt->resid, ccb->ccb_h.flags, req, req->serno, tgt->state); + + MPTLOCK_2_CAMLOCK(mpt); + if ((ccb->ccb_h.flags & CAM_SCATTER_VALID) == 0) { + if ((ccb->ccb_h.flags & CAM_DATA_PHYS) == 0) { + int error; + int s = splsoftvm(); + error = bus_dmamap_load(mpt->buffer_dmat, + req->dmap, csio->data_ptr, csio->dxfer_len, + cb, req, 0); + splx(s); + if (error == EINPROGRESS) { + xpt_freeze_simq(mpt->sim, 1); + ccb->ccb_h.status |= CAM_RELEASE_SIMQ; + } + } else { + /* + * We have been given a pointer to single + * physical buffer. + */ + struct bus_dma_segment seg; + seg.ds_addr = (bus_addr_t) + (vm_offset_t)csio->data_ptr; + seg.ds_len = csio->dxfer_len; + (*cb)(req, &seg, 1, 0); + } + } else { + /* + * We have been given a list of addresses. + * This case could be easily supported but they are not + * currently generated by the CAM subsystem so there + * is no point in wasting the time right now. + */ + struct bus_dma_segment *sgs; + if ((ccb->ccb_h.flags & CAM_SG_LIST_PHYS) == 0) { + (*cb)(req, NULL, 0, EFAULT); + } else { + /* Just use the segments provided */ + sgs = (struct bus_dma_segment *)csio->data_ptr; + (*cb)(req, sgs, csio->sglist_cnt, 0); + } + } + CAMLOCK_2_MPTLOCK(mpt); + } else { + uint8_t *sp = NULL, sense[MPT_SENSE_SIZE]; + + /* + * XXX: I don't know why this seems to happen, but + * XXX: completing the CCB seems to make things happy. + * XXX: This seems to happen if the initiator requests + * XXX: enough data that we have to do multiple CTIOs. + */ + if ((ccb->ccb_h.flags & CAM_SEND_STATUS) == 0) { + mpt_lprt(mpt, MPT_PRT_DEBUG, + "Meaningless STATUS CCB (%p): flags %x status %x " + "resid %d bytes_xfered %u\n", ccb, ccb->ccb_h.flags, + ccb->ccb_h.status, tgt->resid, tgt->bytes_xfered); + mpt_set_ccb_status(ccb, CAM_REQ_CMP); + ccb->ccb_h.status &= ~CAM_SIM_QUEUED; + tgt->flags |= BOGUS_JO; + MPTLOCK_2_CAMLOCK(mpt); + xpt_done(ccb); + CAMLOCK_2_MPTLOCK(mpt); + return; + } + if (ccb->ccb_h.flags & CAM_SEND_SENSE) { + sp = sense; + memcpy(sp, &csio->sense_data, + min(csio->sense_len, MPT_SENSE_SIZE)); + } + mpt_scsi_tgt_status(mpt, ccb, cmd_req, csio->scsi_status, sp); + } +} + +/* + * Abort queued up CCBs + */ +static cam_status +mpt_abort_target_ccb(struct mpt_softc *mpt, union ccb *ccb) +{ + struct mpt_hdr_stailq *lp; + struct ccb_hdr *srch; + int found = 0; + union ccb *accb = ccb->cab.abort_ccb; + tgt_resource_t *trtp; + + mpt_lprt(mpt, MPT_PRT_DEBUG, "aborting ccb %p\n", accb); + + if (ccb->ccb_h.target_lun == CAM_LUN_WILDCARD) { + trtp = &mpt->trt_wildcard; + } else { + trtp = &mpt->trt[ccb->ccb_h.target_lun]; + } + + if (accb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) { + lp = &trtp->atios; + } else if (accb->ccb_h.func_code == XPT_IMMED_NOTIFY) { + lp = &trtp->inots; + } else { + return (CAM_REQ_INVALID); + } + + STAILQ_FOREACH(srch, lp, sim_links.stqe) { + if (srch == &accb->ccb_h) { + found = 1; + STAILQ_REMOVE(lp, srch, ccb_hdr, sim_links.stqe); + break; + } + } + if (found) { + accb->ccb_h.status = CAM_REQ_ABORTED; + xpt_done(accb); + return (CAM_REQ_CMP); + } + mpt_prt(mpt, "mpt_abort_tgt_ccb: CCB %p not found\n", ccb); + return (CAM_PATH_INVALID); +} + +/* + * Ask the MPT to abort the current target command + */ +static cam_status +mpt_abort_target_cmd(struct mpt_softc *mpt, request_t *cmd_req) +{ + int error; + request_t *req; + PTR_MSG_TARGET_MODE_ABORT abtp; + + req = mpt_get_request(mpt, FALSE); + if (req == NULL) { + return (CAM_RESRC_UNAVAIL); + } + abtp = req->req_vbuf; + memset(abtp, 0, sizeof (*abtp)); + + abtp->MsgContext = htole32(req->index | mpt->scsi_tgt_handler_id); + abtp->AbortType = TARGET_MODE_ABORT_TYPE_EXACT_IO; + abtp->Function = MPI_FUNCTION_TARGET_MODE_ABORT; + abtp->ReplyWord = htole32(MPT_TGT_STATE(mpt, cmd_req)->reply_desc); + if (mpt->is_fc || mpt->is_sas) { + mpt_send_cmd(mpt, req); + } else { + error = mpt_send_handshake_cmd(mpt, sizeof(*req), req); + } + return (CAM_REQ_INPROG); +} + /* * When a command times out, it is placed on the requeust_timeout_list * and we wake our recovery thread. The MPT-Fusion architecture supports @@ -2037,13 +4376,11 @@ mpt_recover_commands(struct mpt_softc *mpt) union ccb *ccb; int error; - MPT_LOCK(mpt); if (TAILQ_EMPTY(&mpt->request_timeout_list) != 0) { /* * No work to do- leave. */ mpt_prt(mpt, "mpt_recover_commands: no requests.\n"); - MPT_UNLOCK(mpt); return; } @@ -2064,7 +4401,6 @@ mpt_recover_commands(struct mpt_softc *mpt) mpt_prt(mpt, "Timedout requests already complete. " "Interrupts may not be functioning.\n"); mpt_enable_ints(mpt); - MPT_UNLOCK(mpt); return; } @@ -2075,20 +4411,18 @@ mpt_recover_commands(struct mpt_softc *mpt) */ while ((req = TAILQ_FIRST(&mpt->request_timeout_list)) != NULL) { u_int status; - u_int32_t serno = req->serno; - mpt_prt(mpt, "Attempting to Abort Req %p:%u\n", req, serno); + mpt_prt(mpt, "Attempting to Abort Req %p\n", req); + ccb = req->ccb; mpt_set_ccb_status(ccb, CAM_CMD_TIMEOUT); error = mpt_scsi_send_tmf(mpt, MPI_SCSITASKMGMT_TASKTYPE_ABORT_TASK, - /*MsgFlags*/0, mpt->bus, ccb->ccb_h.target_id, + /*MsgFlags*/0, /*Bus*/0, ccb->ccb_h.target_id, ccb->ccb_h.target_lun, htole32(req->index | scsi_io_handler_id), /*sleep_ok*/TRUE); if (error != 0) { - mpt_prt(mpt, "Abort Req %p:%u failed to start TMF\n", - req, serno); /* * mpt_scsi_send_tmf hard resets on failure, so no * need to do so here. Our queue should be emptied @@ -2107,8 +4441,8 @@ mpt_recover_commands(struct mpt_softc *mpt) * If we've errored out and the transaction is still * pending, reset the controller. */ - mpt_prt(mpt, "Abort Req %p:%d timed-out. " - "Resetting controller\n", req, serno); + mpt_prt(mpt, "mpt_recover_commands: Abort timed-out. " + "Resetting controller\n"); mpt_reset(mpt, /*reinit*/TRUE); continue; } @@ -2122,9 +4456,8 @@ mpt_recover_commands(struct mpt_softc *mpt) continue; mpt_lprt(mpt, MPT_PRT_DEBUG, - "Abort Req %p: %u Failed " - "with status 0x%x\n. Resetting bus.", - req, serno, status); + "mpt_recover_commands: Abort Failed " + "with status 0x%x\n. Resetting bus", status); /* * If the abort attempt fails for any reason, reset the bus. @@ -2133,6 +4466,4 @@ mpt_recover_commands(struct mpt_softc *mpt) */ mpt_bus_reset(mpt, /*sleep_ok*/TRUE); } - - MPT_UNLOCK(mpt); } diff --git a/sys/dev/mpt/mpt_debug.c b/sys/dev/mpt/mpt_debug.c index 270840afbc73..b9b6bc19b9a7 100644 --- a/sys/dev/mpt/mpt_debug.c +++ b/sys/dev/mpt/mpt_debug.c @@ -68,6 +68,7 @@ __FBSDID("$FreeBSD$"); #include #include #include +#include #include @@ -574,6 +575,35 @@ mpt_print_scsi_tmf_request(MSG_SCSI_TASK_MGMT *msg) printf("\tTaskMsgContext 0x%08x\n", msg->TaskMsgContext); } + +static void +mpt_print_scsi_target_assist_request(PTR_MSG_TARGET_ASSIST_REQUEST msg) +{ + mpt_print_request_hdr((MSG_REQUEST_HEADER *)msg); + printf("\tStatusCode 0x%02x\n", msg->StatusCode); + printf("\tTargetAssist 0x%02x\n", msg->TargetAssistFlags); + printf("\tQueueTag 0x%04x\n", msg->QueueTag); + printf("\tReplyWord 0x%08x\n", msg->ReplyWord); + printf("\tLun 0x%02x\n", msg->LUN[1]); + printf("\tRelativeOff 0x%08x\n", msg->RelativeOffset); + printf("\tDataLength 0x%08x\n", msg->DataLength); + mpt_dump_sgl(msg->SGL, 0); +} + +static void +mpt_print_scsi_target_status_send_request(MSG_TARGET_STATUS_SEND_REQUEST *msg) +{ + SGE_IO_UNION x; + mpt_print_request_hdr((MSG_REQUEST_HEADER *)msg); + printf("\tStatusCode 0x%02x\n", msg->StatusCode); + printf("\tStatusFlags 0x%02x\n", msg->StatusFlags); + printf("\tQueueTag 0x%04x\n", msg->QueueTag); + printf("\tReplyWord 0x%08x\n", msg->ReplyWord); + printf("\tLun 0x%02x\n", msg->LUN[1]); + x.u.Simple = msg->StatusDataSGE; + mpt_dump_sgl(&x, 0); +} + void mpt_print_request(void *vreq) { @@ -585,6 +615,15 @@ mpt_print_request(void *vreq) break; case MPI_FUNCTION_SCSI_TASK_MGMT: mpt_print_scsi_tmf_request((MSG_SCSI_TASK_MGMT *)req); + break; + case MPI_FUNCTION_TARGET_ASSIST: + mpt_print_scsi_target_assist_request( + (PTR_MSG_TARGET_ASSIST_REQUEST)req); + break; + case MPI_FUNCTION_TARGET_STATUS_SEND: + mpt_print_scsi_target_status_send_request( + (MSG_TARGET_STATUS_SEND_REQUEST *)req); + break; default: mpt_print_request_hdr(req); break; @@ -754,6 +793,31 @@ mpt_dump_sgl(SGE_IO_UNION *su, int offset) } while ((flags & MPI_SGE_FLAGS_END_OF_LIST) == 0 && nxtaddr < lim); } +#if __FreeBSD_version < 500000 +void +mpt_lprt(struct mpt_softc *mpt, int level, const char *fmt, ...) +{ + va_list ap; + if (level <= mpt->verbose) { + printf("%s: ", device_get_nameunit(mpt->dev)); + va_start(ap, fmt); + vprintf(fmt, ap); + va_end(ap); + } +} + +void +mpt_lprtc(struct mpt_softc *mpt, int level, const char *fmt, ...) +{ + va_list ap; + if (level <= mpt->verbose) { + va_start(ap, fmt); + vprintf(fmt, ap); + va_end(ap); + } +} +#endif + void mpt_prt(struct mpt_softc *mpt, const char *fmt, ...) { diff --git a/sys/dev/mpt/mpt_pci.c b/sys/dev/mpt/mpt_pci.c index a995618741ff..e29b421642a8 100644 --- a/sys/dev/mpt/mpt_pci.c +++ b/sys/dev/mpt/mpt_pci.c @@ -242,7 +242,7 @@ mpt_pci_probe(device_t dev) return (0); } -#ifdef RELENG_4 +#if __FreeBSD_version < 500000 static void mpt_set_options(struct mpt_softc *mpt) { @@ -261,7 +261,24 @@ mpt_set_options(struct mpt_softc *mpt) mpt->verbose = MPT_PRT_DEBUG; } } - + bitmap = 0; + if (getenv_int("mpt_target", &bitmap)) { + if (bitmap & (1 << mpt->unit)) { + mpt->role = MPT_ROLE_TARGET; + } + } + bitmap = 0; + if (getenv_int("mpt_none", &bitmap)) { + if (bitmap & (1 << mpt->unit)) { + mpt->role = MPT_ROLE_NONE; + } + } + bitmap = 0; + if (getenv_int("mpt_initiator", &bitmap)) { + if (bitmap & (1 << mpt->unit)) { + mpt->role = MPT_ROLE_INITIATOR; + } + } } #else static void @@ -279,6 +296,12 @@ mpt_set_options(struct mpt_softc *mpt) device_get_unit(mpt->dev), "debug", &tval) == 0 && tval != 0) { mpt->verbose += tval; } + tval = 0; + if (resource_int_value(device_get_name(mpt->dev), + device_get_unit(mpt->dev), "role", &tval) == 0 && tval != 0 && + tval <= 3) { + mpt->role = tval; + } } #endif @@ -354,6 +377,7 @@ mpt_pci_attach(device_t dev) mpt->raid_mwce_setting = MPT_RAID_MWCE_DEFAULT; mpt->raid_queue_depth = MPT_RAID_QUEUE_DEPTH_DEFAULT; mpt->verbose = MPT_PRT_NONE; + mpt->role = MPT_ROLE_DEFAULT; mpt_set_options(mpt); if (mpt->verbose == MPT_PRT_NONE) { mpt->verbose = MPT_PRT_WARN; @@ -474,16 +498,44 @@ mpt_pci_attach(device_t dev) */ pci_disable_io(dev, SYS_RES_IOPORT); + switch (mpt->role) { + case MPT_ROLE_TARGET: + break; + case MPT_ROLE_INITIATOR: + break; + case MPT_ROLE_TARGET|MPT_ROLE_INITIATOR: + mpt->disabled = 1; + mpt_prt(mpt, "dual roles unsupported\n"); + goto bad; + case MPT_ROLE_NONE: + device_printf(dev, "role of NONE same as disable\n"); + mpt->disabled = 1; + goto bad; + } + /* Initialize the hardware */ if (mpt->disabled == 0) { - MPT_LOCK(mpt); if (mpt_attach(mpt) != 0) { MPT_UNLOCK(mpt); goto bad; } + MPT_UNLOCK(mpt); + } else { + mpt_prt(mpt, "device disabled at user request\n"); + goto bad; } + mpt->eh = EVENTHANDLER_REGISTER(shutdown_post_sync, mpt_pci_shutdown, + dev, SHUTDOWN_PRI_DEFAULT); + + if (mpt->eh == NULL) { + mpt_prt(mpt, "shutdown event registration failed\n"); + MPT_LOCK(mpt); + (void) mpt_detach(mpt); + MPT_UNLOCK(mpt); + goto bad; + } return (0); bad: @@ -535,24 +587,23 @@ mpt_pci_detach(device_t dev) struct mpt_softc *mpt; mpt = (struct mpt_softc*)device_get_softc(dev); - mpt_prt(mpt, "mpt_detach\n"); if (mpt) { + MPT_LOCK(mpt); mpt_disable_ints(mpt); mpt_detach(mpt); mpt_reset(mpt, /*reinit*/FALSE); mpt_dma_mem_free(mpt); mpt_free_bus_resources(mpt); - if (mpt->raid_volumes != NULL - && mpt->ioc_page2 != NULL) { + if (mpt->raid_volumes != NULL && mpt->ioc_page2 != NULL) { int i; - for (i = 0; i < mpt->ioc_page2->MaxVolumes; i++) { struct mpt_raid_volume *mpt_vol; mpt_vol = &mpt->raid_volumes[i]; - if (mpt_vol->config_page) + if (mpt_vol->config_page) { free(mpt_vol->config_page, M_DEVBUF); + } } } if (mpt->ioc_page2 != NULL) @@ -565,6 +616,7 @@ mpt_pci_detach(device_t dev) free(mpt->raid_disks, M_DEVBUF); if (mpt->eh != NULL) EVENTHANDLER_DEREGISTER(shutdown_final, mpt->eh); + MPT_UNLOCK(mpt); } return(0); } @@ -572,7 +624,6 @@ mpt_pci_detach(device_t dev) /* * Disable the hardware - * XXX - Called too early by New Bus!!! ??? */ static int mpt_pci_shutdown(device_t dev) @@ -580,8 +631,13 @@ mpt_pci_shutdown(device_t dev) struct mpt_softc *mpt; mpt = (struct mpt_softc *)device_get_softc(dev); - if (mpt) - return (mpt_shutdown(mpt)); + if (mpt) { + int r; + MPT_LOCK(mpt); + r = mpt_shutdown(mpt); + MPT_UNLOCK(mpt); + return (r); + } return(0); } diff --git a/sys/dev/mpt/mpt_raid.c b/sys/dev/mpt/mpt_raid.c index 447423ba2677..da25979e3eb5 100644 --- a/sys/dev/mpt/mpt_raid.c +++ b/sys/dev/mpt/mpt_raid.c @@ -46,6 +46,10 @@ __FBSDID("$FreeBSD$"); #include #include +#if __FreeBSD_version < 500000 +#include +#define GIANT_REQUIRED +#endif #include #include @@ -241,9 +245,9 @@ mpt_raid_async(void *callback_arg, u_int32_t code, int mpt_raid_probe(struct mpt_softc *mpt) { - if (mpt->ioc_page2 == NULL - || mpt->ioc_page2->MaxPhysDisks == 0) + if (mpt->ioc_page2 == NULL || mpt->ioc_page2->MaxPhysDisks == 0) { return (ENODEV); + } return (0); } @@ -446,8 +450,10 @@ mpt_raid_event(struct mpt_softc *mpt, request_t *req, if (mpt_disk != NULL) mpt_disk_prt(mpt, mpt_disk, ""); else - mpt_prt(mpt, "Volume(%d:%d:%d: "); - mpt_prtc(mpt, "ASC 0x%x, ASCQ 0x%x\n", + mpt_prt(mpt, "Volume(%d:%d:%d: ", + raid_event->VolumeBus, raid_event->VolumeID, + raid_event->PhysDiskNum); + mpt_prtc(mpt, "ASC 0x%x, ASCQ 0x%x)\n", raid_event->ASC, raid_event->ASCQ); } @@ -460,19 +466,19 @@ mpt_raid_shutdown(struct mpt_softc *mpt) { struct mpt_raid_volume *mpt_vol; - if (mpt->raid_mwce_setting != MPT_RAID_MWCE_REBUILD_ONLY) + if (mpt->raid_mwce_setting != MPT_RAID_MWCE_REBUILD_ONLY) { return; + } mpt->raid_mwce_setting = MPT_RAID_MWCE_OFF; RAID_VOL_FOREACH(mpt, mpt_vol) { - mpt_verify_mwce(mpt, mpt_vol); } } static int mpt_raid_reply_handler(struct mpt_softc *mpt, request_t *req, - MSG_DEFAULT_REPLY *reply_frame) + uint32_t reply_desc, MSG_DEFAULT_REPLY *reply_frame) { int free_req; @@ -607,16 +613,11 @@ mpt_spawn_raid_thread(struct mpt_softc *mpt) return (error); } -/* - * Lock is not held on entry. - */ static void mpt_terminate_raid_thread(struct mpt_softc *mpt) { - MPT_LOCK(mpt); if (mpt->raid_thread == NULL) { - MPT_UNLOCK(mpt); return; } mpt->shutdwn_raid = 1; @@ -626,7 +627,6 @@ mpt_terminate_raid_thread(struct mpt_softc *mpt) * for this interlock just for added safety. */ mpt_sleep(mpt, &mpt->raid_thread, PUSER, "thtrm", 0); - MPT_UNLOCK(mpt); } static void @@ -845,10 +845,9 @@ mpt_verify_mwce(struct mpt_softc *mpt, struct mpt_raid_volume *mpt_vol) */ switch (mpt->raid_mwce_setting) { case MPT_RAID_MWCE_REBUILD_ONLY: - if ((resyncing && mwce) - || (!resyncing && !mwce)) + if ((resyncing && mwce) || (!resyncing && !mwce)) { return; - + } mpt_vol->flags ^= MPT_RVF_WCE_CHANGED; if ((mpt_vol->flags & MPT_RVF_WCE_CHANGED) == 0) { /* @@ -903,7 +902,6 @@ mpt_verify_mwce(struct mpt_softc *mpt, struct mpt_raid_volume *mpt_vol) vol_pg->VolumeSettings.Settings ^= MPI_RAIDVOL0_SETTING_WRITE_CACHING_ENABLE; } - mpt_free_request(mpt, req); } @@ -1228,18 +1226,21 @@ mpt_refresh_raid_data(struct mpt_softc *mpt) int i; u_int nonopt_volumes; - if (mpt->ioc_page2 == NULL || mpt->ioc_page3 == NULL) + if (mpt->ioc_page2 == NULL || mpt->ioc_page3 == NULL) { return; + } /* - * Mark all items as unreferrened by the configuration. + * Mark all items as unreferenced by the configuration. * This allows us to find, report, and discard stale * entries. */ - for (i = 0; i < mpt->ioc_page2->MaxPhysDisks; i++) + for (i = 0; i < mpt->ioc_page2->MaxPhysDisks; i++) { mpt->raid_disks[i].flags &= ~MPT_RDF_REFERENCED; - for (i = 0; i < mpt->ioc_page2->MaxVolumes; i++) + } + for (i = 0; i < mpt->ioc_page2->MaxVolumes; i++) { mpt->raid_volumes[i].flags &= ~MPT_RVF_REFERENCED; + } /* * Get Physical Disk information. @@ -1386,9 +1387,15 @@ mpt_refresh_raid_data(struct mpt_softc *mpt) mpt_vol_prt(mpt, mpt_vol, "%s Priority Re-Sync\n", prio ? "High" : "Low"); } +#if __FreeBSD_version >= 500000 mpt_vol_prt(mpt, mpt_vol, "%ju of %ju " "blocks remaining\n", (uintmax_t)left, (uintmax_t)total); +#else + mpt_vol_prt(mpt, mpt_vol, "%llu of %llu " + "blocks remaining\n", (uint64_t)left, + (uint64_t)total); +#endif /* Periodically report on sync progress. */ mpt_schedule_raid_refresh(mpt); @@ -1484,8 +1491,9 @@ mpt_raid_set_vol_resync_rate(struct mpt_softc *mpt, u_int rate) MPT_LOCK(mpt); mpt->raid_resync_rate = rate; RAID_VOL_FOREACH(mpt, mpt_vol) { - if ((mpt_vol->flags & MPT_RVF_ACTIVE) == 0) + if ((mpt_vol->flags & MPT_RVF_ACTIVE) == 0) { continue; + } mpt_verify_resync_rate(mpt, mpt_vol); } MPT_UNLOCK(mpt); @@ -1497,8 +1505,7 @@ mpt_raid_set_vol_queue_depth(struct mpt_softc *mpt, u_int vol_queue_depth) { struct mpt_raid_volume *mpt_vol; - if (vol_queue_depth > 255 - || vol_queue_depth < 1) + if (vol_queue_depth > 255 || vol_queue_depth < 1) return (EINVAL); MPT_LOCK(mpt); @@ -1662,6 +1669,7 @@ mpt_raid_sysctl_vol_queue_depth(SYSCTL_HANDLER_ARGS) static void mpt_raid_sysctl_attach(struct mpt_softc *mpt) { +#if __FreeBSD_version >= 500000 struct sysctl_ctx_list *ctx = device_get_sysctl_ctx(mpt->dev); struct sysctl_oid *tree = device_get_sysctl_tree(mpt->dev); @@ -1683,4 +1691,5 @@ mpt_raid_sysctl_attach(struct mpt_softc *mpt) "nonoptimal_volumes", CTLFLAG_RD, &mpt->raid_nonopt_volumes, 0, "number of nonoptimal volumes"); +#endif } diff --git a/sys/dev/mpt/mpt_reg.h b/sys/dev/mpt/mpt_reg.h index 23a34e07b54e..fc6d1f3d598d 100644 --- a/sys/dev/mpt/mpt_reg.h +++ b/sys/dev/mpt/mpt_reg.h @@ -125,7 +125,7 @@ enum _MPT_INTR_REQ_BITS { #define MPT_DB_INTR(v) (((v) & MPT_INTR_DB_READY) != 0) #define MPT_REPLY_INTR(v) (((v) & MPT_INTR_REPLY_READY) != 0) -/* Function Maps for INTERRUPT make register */ +/* Function Maps for INTERRUPT mask register */ enum _MPT_INTR_MASK_BITS { MPT_INTR_REPLY_MASK = 0x00000008, MPT_INTR_DB_MASK = 0x00000001