/* $FreeBSD$ */ /* * Platform (FreeBSD) dependent common attachment code for Qlogic adapters. * * Copyright (c) 1997, 1998, 1999, 2000, 2001 by Matthew Jacob * * Redistribution and use in source and binary forms, with or without * modification, are permitted provided that the following conditions * are met: * 1. Redistributions of source code must retain the above copyright * notice immediately at the beginning of the file, without modification, * this list of conditions, and the following disclaimer. * 2. The name of the author may not be used to endorse or promote products * derived from this software without specific prior written permission. * * THIS SOFTWARE IS PROVIDED BY THE AUTHOR AND CONTRIBUTORS ``AS IS'' AND * ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED TO, THE * IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR PURPOSE * ARE DISCLAIMED. IN NO EVENT SHALL THE AUTHOR OR CONTRIBUTORS BE LIABLE FOR * ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR CONSEQUENTIAL * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION) * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT, STRICT * LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING IN ANY WAY * OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE POSSIBILITY OF * SUCH DAMAGE. */ #include #include #include #include /* for use by isp_prt below */ #include #include #include static d_ioctl_t ispioctl; static void isp_intr_enable(void *); static void isp_cam_async(void *, u_int32_t, struct cam_path *, void *); static void isp_poll(struct cam_sim *); #if 0 static void isp_relsim(void *); #endif static timeout_t isp_watchdog; static void isp_kthread(void *); static void isp_action(struct cam_sim *, union ccb *); #define ISP_CDEV_MAJOR 248 static struct cdevsw isp_cdevsw = { /* open */ nullopen, /* close */ nullclose, /* read */ noread, /* write */ nowrite, /* ioctl */ ispioctl, /* poll */ nopoll, /* mmap */ nommap, /* strategy */ nostrategy, /* name */ "isp", /* maj */ ISP_CDEV_MAJOR, /* dump */ nodump, /* psize */ nopsize, /* flags */ D_TAPE, }; static struct ispsoftc *isplist = NULL; void isp_attach(struct ispsoftc *isp) { int primary, secondary; struct ccb_setasync csa; struct cam_devq *devq; struct cam_sim *sim; struct cam_path *path; /* * Establish (in case of 12X0) which bus is the primary. */ primary = 0; secondary = 1; /* * Create the device queue for our SIM(s). */ devq = cam_simq_alloc(isp->isp_maxcmds); if (devq == NULL) { return; } /* * Construct our SIM entry. */ ISPLOCK_2_CAMLOCK(isp); sim = cam_sim_alloc(isp_action, isp_poll, "isp", isp, device_get_unit(isp->isp_dev), 1, isp->isp_maxcmds, devq); if (sim == NULL) { cam_simq_free(devq); CAMLOCK_2_ISPLOCK(isp); return; } CAMLOCK_2_ISPLOCK(isp); isp->isp_osinfo.ehook.ich_func = isp_intr_enable; isp->isp_osinfo.ehook.ich_arg = isp; ISPLOCK_2_CAMLOCK(isp); if (config_intrhook_establish(&isp->isp_osinfo.ehook) != 0) { cam_sim_free(sim, TRUE); CAMLOCK_2_ISPLOCK(isp); isp_prt(isp, ISP_LOGERR, "could not establish interrupt enable hook"); return; } if (xpt_bus_register(sim, primary) != CAM_SUCCESS) { cam_sim_free(sim, TRUE); CAMLOCK_2_ISPLOCK(isp); return; } if (xpt_create_path(&path, NULL, cam_sim_path(sim), CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) { xpt_bus_deregister(cam_sim_path(sim)); cam_sim_free(sim, TRUE); config_intrhook_disestablish(&isp->isp_osinfo.ehook); CAMLOCK_2_ISPLOCK(isp); return; } xpt_setup_ccb(&csa.ccb_h, path, 5); csa.ccb_h.func_code = XPT_SASYNC_CB; csa.event_enable = AC_LOST_DEVICE; csa.callback = isp_cam_async; csa.callback_arg = sim; xpt_action((union ccb *)&csa); CAMLOCK_2_ISPLOCK(isp); isp->isp_sim = sim; isp->isp_path = path; /* * Create a kernel thread for fibre channel instances. We * don't have dual channel FC cards. */ if (IS_FC(isp)) { ISPLOCK_2_CAMLOCK(isp); /* XXX: LOCK VIOLATION */ cv_init(&isp->isp_osinfo.kthread_cv, "isp_kthread_cv"); if (kthread_create(isp_kthread, isp, &isp->isp_osinfo.kproc, RFHIGHPID, "%s: fc_thrd", device_get_nameunit(isp->isp_dev))) { xpt_bus_deregister(cam_sim_path(sim)); cam_sim_free(sim, TRUE); config_intrhook_disestablish(&isp->isp_osinfo.ehook); CAMLOCK_2_ISPLOCK(isp); isp_prt(isp, ISP_LOGERR, "could not create kthread"); return; } CAMLOCK_2_ISPLOCK(isp); } /* * If we have a second channel, construct SIM entry for that. */ if (IS_DUALBUS(isp)) { ISPLOCK_2_CAMLOCK(isp); sim = cam_sim_alloc(isp_action, isp_poll, "isp", isp, device_get_unit(isp->isp_dev), 1, isp->isp_maxcmds, devq); if (sim == NULL) { xpt_bus_deregister(cam_sim_path(isp->isp_sim)); xpt_free_path(isp->isp_path); cam_simq_free(devq); config_intrhook_disestablish(&isp->isp_osinfo.ehook); return; } if (xpt_bus_register(sim, secondary) != CAM_SUCCESS) { xpt_bus_deregister(cam_sim_path(isp->isp_sim)); xpt_free_path(isp->isp_path); cam_sim_free(sim, TRUE); config_intrhook_disestablish(&isp->isp_osinfo.ehook); CAMLOCK_2_ISPLOCK(isp); return; } if (xpt_create_path(&path, NULL, cam_sim_path(sim), CAM_TARGET_WILDCARD, CAM_LUN_WILDCARD) != CAM_REQ_CMP) { xpt_bus_deregister(cam_sim_path(isp->isp_sim)); xpt_free_path(isp->isp_path); xpt_bus_deregister(cam_sim_path(sim)); cam_sim_free(sim, TRUE); config_intrhook_disestablish(&isp->isp_osinfo.ehook); CAMLOCK_2_ISPLOCK(isp); return; } xpt_setup_ccb(&csa.ccb_h, path, 5); csa.ccb_h.func_code = XPT_SASYNC_CB; csa.event_enable = AC_LOST_DEVICE; csa.callback = isp_cam_async; csa.callback_arg = sim; xpt_action((union ccb *)&csa); CAMLOCK_2_ISPLOCK(isp); isp->isp_sim2 = sim; isp->isp_path2 = path; } #ifdef ISP_TARGET_MODE cv_init(&isp->isp_osinfo.tgtcv0[0], "isp_tgcv0a"); cv_init(&isp->isp_osinfo.tgtcv0[1], "isp_tgcv0b"); cv_init(&isp->isp_osinfo.tgtcv1[0], "isp_tgcv1a"); cv_init(&isp->isp_osinfo.tgtcv1[1], "isp_tgcv1b"); #endif /* * Create device nodes */ (void) make_dev(&isp_cdevsw, device_get_unit(isp->isp_dev), UID_ROOT, GID_OPERATOR, 0600, "%s", device_get_nameunit(isp->isp_dev)); if (isp->isp_role != ISP_ROLE_NONE) { isp->isp_state = ISP_RUNSTATE; ENABLE_INTS(isp); } if (isplist == NULL) { isplist = isp; } else { struct ispsoftc *tmp = isplist; while (tmp->isp_osinfo.next) { tmp = tmp->isp_osinfo.next; } tmp->isp_osinfo.next = isp; } } static int ispioctl(dev_t dev, u_long cmd, caddr_t addr, int flags, struct thread *td) { struct ispsoftc *isp; int retval = ENOTTY; isp = isplist; while (isp) { if (minor(dev) == device_get_unit(isp->isp_dev)) { break; } isp = isp->isp_osinfo.next; } if (isp == NULL) return (ENXIO); switch (cmd) { case ISP_SDBLEV: { int olddblev = isp->isp_dblev; isp->isp_dblev = *(int *)addr; *(int *)addr = olddblev; retval = 0; break; } case ISP_RESETHBA: ISP_LOCK(isp); isp_reinit(isp); ISP_UNLOCK(isp); retval = 0; break; case ISP_FC_RESCAN: if (IS_FC(isp)) { ISP_LOCK(isp); if (isp_fc_runstate(isp, 5 * 1000000)) { retval = EIO; } else { retval = 0; } ISP_UNLOCK(isp); } break; case ISP_FC_LIP: if (IS_FC(isp)) { ISP_LOCK(isp); if (isp_control(isp, ISPCTL_SEND_LIP, 0)) { retval = EIO; } else { retval = 0; } ISP_UNLOCK(isp); } break; case ISP_FC_GETDINFO: { struct isp_fc_device *ifc = (struct isp_fc_device *) addr; struct lportdb *lp; if (ifc->loopid < 0 || ifc->loopid >= MAX_FC_TARG) { retval = EINVAL; break; } ISP_LOCK(isp); lp = &FCPARAM(isp)->portdb[ifc->loopid]; if (lp->valid) { ifc->loopid = lp->loopid; ifc->portid = lp->portid; ifc->node_wwn = lp->node_wwn; ifc->port_wwn = lp->port_wwn; retval = 0; } else { retval = ENODEV; } ISP_UNLOCK(isp); break; } case ISP_GET_STATS: { isp_stats_t *sp = (isp_stats_t *) addr; MEMZERO(sp, sizeof (*sp)); sp->isp_stat_version = ISP_STATS_VERSION; sp->isp_type = isp->isp_type; sp->isp_revision = isp->isp_revision; ISP_LOCK(isp); sp->isp_stats[ISP_INTCNT] = isp->isp_intcnt; sp->isp_stats[ISP_INTBOGUS] = isp->isp_intbogus; sp->isp_stats[ISP_INTMBOXC] = isp->isp_intmboxc; sp->isp_stats[ISP_INGOASYNC] = isp->isp_intoasync; sp->isp_stats[ISP_RSLTCCMPLT] = isp->isp_rsltccmplt; sp->isp_stats[ISP_FPHCCMCPLT] = isp->isp_fphccmplt; sp->isp_stats[ISP_RSCCHIWAT] = isp->isp_rscchiwater; sp->isp_stats[ISP_FPCCHIWAT] = isp->isp_fpcchiwater; ISP_UNLOCK(isp); retval = 0; break; } case ISP_CLR_STATS: ISP_LOCK(isp); isp->isp_intcnt = 0; isp->isp_intbogus = 0; isp->isp_intmboxc = 0; isp->isp_intoasync = 0; isp->isp_rsltccmplt = 0; isp->isp_fphccmplt = 0; isp->isp_rscchiwater = 0; isp->isp_fpcchiwater = 0; ISP_UNLOCK(isp); retval = 0; break; default: break; } return (retval); } static void isp_intr_enable(void *arg) { struct ispsoftc *isp = arg; if (isp->isp_role != ISP_ROLE_NONE) { ENABLE_INTS(isp); isp->isp_osinfo.intsok = 1; } /* Release our hook so that the boot can continue. */ config_intrhook_disestablish(&isp->isp_osinfo.ehook); } /* * Put the target mode functions here, because some are inlines */ #ifdef ISP_TARGET_MODE static __inline int is_lun_enabled(struct ispsoftc *, int, lun_id_t); static __inline int are_any_luns_enabled(struct ispsoftc *, int); static __inline tstate_t *get_lun_statep(struct ispsoftc *, int, lun_id_t); static __inline void rls_lun_statep(struct ispsoftc *, tstate_t *); static __inline int isp_psema_sig_rqe(struct ispsoftc *, int); static __inline int isp_cv_wait_timed_rqe(struct ispsoftc *, int, int); static __inline void isp_cv_signal_rqe(struct ispsoftc *, int, int); static __inline void isp_vsema_rqe(struct ispsoftc *, int); static __inline atio_private_data_t *isp_get_atpd(struct ispsoftc *, int); static cam_status create_lun_state(struct ispsoftc *, int, struct cam_path *, tstate_t **); static void destroy_lun_state(struct ispsoftc *, tstate_t *); static void isp_en_lun(struct ispsoftc *, union ccb *); static cam_status isp_abort_tgt_ccb(struct ispsoftc *, union ccb *); static timeout_t isp_refire_putback_atio; static void isp_complete_ctio(union ccb *); static void isp_target_putback_atio(union ccb *); static cam_status isp_target_start_ctio(struct ispsoftc *, union ccb *); static int isp_handle_platform_atio(struct ispsoftc *, at_entry_t *); static int isp_handle_platform_atio2(struct ispsoftc *, at2_entry_t *); static int isp_handle_platform_ctio(struct ispsoftc *, void *); static __inline int is_lun_enabled(struct ispsoftc *isp, int bus, lun_id_t lun) { tstate_t *tptr; tptr = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(isp, bus, lun)]; if (tptr == NULL) { return (0); } do { if (tptr->lun == (lun_id_t) lun && tptr->bus == bus) { return (1); } } while ((tptr = tptr->next) != NULL); return (0); } static __inline int are_any_luns_enabled(struct ispsoftc *isp, int port) { int lo, hi; if (IS_DUALBUS(isp)) { lo = (port * (LUN_HASH_SIZE >> 1)); hi = lo + (LUN_HASH_SIZE >> 1); } else { lo = 0; hi = LUN_HASH_SIZE; } for (lo = 0; lo < hi; lo++) { if (isp->isp_osinfo.lun_hash[lo]) { return (1); } } return (0); } static __inline tstate_t * get_lun_statep(struct ispsoftc *isp, int bus, lun_id_t lun) { tstate_t *tptr = NULL; if (lun == CAM_LUN_WILDCARD) { if (isp->isp_osinfo.tmflags[bus] & TM_WILDCARD_ENABLED) { tptr = &isp->isp_osinfo.tsdflt[bus]; tptr->hold++; return (tptr); } } else { tptr = isp->isp_osinfo.lun_hash[LUN_HASH_FUNC(isp, bus, lun)]; if (tptr == NULL) { return (NULL); } } do { if (tptr->lun == lun && tptr->bus == bus) { tptr->hold++; return (tptr); } } while ((tptr = tptr->next) != NULL); return (tptr); } static __inline void rls_lun_statep(struct ispsoftc *isp, tstate_t *tptr) { if (tptr->hold) tptr->hold--; } static __inline int isp_psema_sig_rqe(struct ispsoftc *isp, int bus) { while (isp->isp_osinfo.tmflags[bus] & TM_BUSY) { isp->isp_osinfo.tmflags[bus] |= TM_WANTED; if (cv_wait_sig(&isp->isp_osinfo.tgtcv0[bus], &isp->isp_lock)) { return (-1); } isp->isp_osinfo.tmflags[bus] |= TM_BUSY; } return (0); } static __inline int isp_cv_wait_timed_rqe(struct ispsoftc *isp, int bus, int timo) { if (cv_timedwait(&isp->isp_osinfo.tgtcv1[bus], &isp->isp_lock, timo)) { return (-1); } return (0); } static __inline void isp_cv_signal_rqe(struct ispsoftc *isp, int bus, int status) { isp->isp_osinfo.rstatus[bus] = status; cv_signal(&isp->isp_osinfo.tgtcv1[bus]); } static __inline void isp_vsema_rqe(struct ispsoftc *isp, int bus) { if (isp->isp_osinfo.tmflags[bus] & TM_WANTED) { isp->isp_osinfo.tmflags[bus] &= ~TM_WANTED; cv_signal(&isp->isp_osinfo.tgtcv0[bus]); } isp->isp_osinfo.tmflags[bus] &= ~TM_BUSY; } static __inline atio_private_data_t * isp_get_atpd(struct ispsoftc *isp, int tag) { atio_private_data_t *atp; for (atp = isp->isp_osinfo.atpdp; atp < &isp->isp_osinfo.atpdp[ATPDPSIZE]; atp++) { if (atp->tag == tag) return (atp); } return (NULL); } static cam_status create_lun_state(struct ispsoftc *isp, int bus, struct cam_path *path, tstate_t **rslt) { cam_status status; lun_id_t lun; int hfx; tstate_t *tptr, *new; lun = xpt_path_lun_id(path); if (lun < 0) { return (CAM_LUN_INVALID); } if (is_lun_enabled(isp, bus, lun)) { return (CAM_LUN_ALRDY_ENA); } new = (tstate_t *) malloc(sizeof (tstate_t), M_DEVBUF, M_NOWAIT|M_ZERO); if (new == NULL) { return (CAM_RESRC_UNAVAIL); } status = xpt_create_path(&new->owner, NULL, xpt_path_path_id(path), xpt_path_target_id(path), xpt_path_lun_id(path)); if (status != CAM_REQ_CMP) { free(new, M_DEVBUF); return (status); } new->bus = bus; new->lun = lun; SLIST_INIT(&new->atios); SLIST_INIT(&new->inots); new->hold = 1; hfx = LUN_HASH_FUNC(isp, new->bus, new->lun); tptr = isp->isp_osinfo.lun_hash[hfx]; if (tptr == NULL) { isp->isp_osinfo.lun_hash[hfx] = new; } else { while (tptr->next) tptr = tptr->next; tptr->next = new; } *rslt = new; return (CAM_REQ_CMP); } static __inline void destroy_lun_state(struct ispsoftc *isp, tstate_t *tptr) { int hfx; tstate_t *lw, *pw; hfx = LUN_HASH_FUNC(isp, tptr->bus, tptr->lun); if (tptr->hold) { return; } pw = isp->isp_osinfo.lun_hash[hfx]; if (pw == NULL) { return; } else if (pw->lun == tptr->lun && pw->bus == tptr->bus) { isp->isp_osinfo.lun_hash[hfx] = pw->next; } else { lw = pw; pw = lw->next; while (pw) { if (pw->lun == tptr->lun && pw->bus == tptr->bus) { lw->next = pw->next; break; } lw = pw; pw = pw->next; } if (pw == NULL) { return; } } free(tptr, M_DEVBUF); } /* * we enter with our locks held. */ static void isp_en_lun(struct ispsoftc *isp, union ccb *ccb) { const char lfmt[] = "Lun now %sabled for target mode on channel %d"; struct ccb_en_lun *cel = &ccb->cel; tstate_t *tptr; u_int16_t rstat; int bus, cmd, av, wildcard; lun_id_t lun; target_id_t tgt; bus = XS_CHANNEL(ccb) & 0x1; tgt = ccb->ccb_h.target_id; lun = ccb->ccb_h.target_lun; /* * Do some sanity checking first. */ if ((lun != CAM_LUN_WILDCARD) && (lun < 0 || lun >= (lun_id_t) isp->isp_maxluns)) { ccb->ccb_h.status = CAM_LUN_INVALID; return; } if (IS_SCSI(isp)) { sdparam *sdp = isp->isp_param; sdp += bus; if (tgt != CAM_TARGET_WILDCARD && tgt != sdp->isp_initiator_id) { ccb->ccb_h.status = CAM_TID_INVALID; return; } } else { if (tgt != CAM_TARGET_WILDCARD && tgt != FCPARAM(isp)->isp_iid) { ccb->ccb_h.status = CAM_TID_INVALID; return; } /* * This is as a good a place as any to check f/w capabilities. */ if ((FCPARAM(isp)->isp_fwattr & ISP_FW_ATTR_TMODE) == 0) { isp_prt(isp, ISP_LOGERR, "firmware does not support target mode"); ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; return; } /* * XXX: We *could* handle non-SCCLUN f/w, but we'd have to * XXX: dorks with our already fragile enable/disable code. */ if ((FCPARAM(isp)->isp_fwattr & ISP_FW_ATTR_SCCLUN) == 0) { isp_prt(isp, ISP_LOGERR, "firmware not SCCLUN capable"); } } if (tgt == CAM_TARGET_WILDCARD) { if (lun == CAM_LUN_WILDCARD) { wildcard = 1; } else { ccb->ccb_h.status = CAM_LUN_INVALID; return; } } else { wildcard = 0; } /* * Next check to see whether this is a target/lun wildcard action. * * If so, we know that we can accept commands for luns that haven't * been enabled yet and send them upstream. Otherwise, we have to * handle them locally (if we see them at all). */ if (wildcard) { tptr = &isp->isp_osinfo.tsdflt[bus]; if (cel->enable) { if (isp->isp_osinfo.tmflags[bus] & TM_WILDCARD_ENABLED) { ccb->ccb_h.status = CAM_LUN_ALRDY_ENA; return; } ccb->ccb_h.status = xpt_create_path(&tptr->owner, NULL, xpt_path_path_id(ccb->ccb_h.path), xpt_path_target_id(ccb->ccb_h.path), xpt_path_lun_id(ccb->ccb_h.path)); if (ccb->ccb_h.status != CAM_REQ_CMP) { return; } SLIST_INIT(&tptr->atios); SLIST_INIT(&tptr->inots); isp->isp_osinfo.tmflags[bus] |= TM_WILDCARD_ENABLED; } else { if ((isp->isp_osinfo.tmflags[bus] & TM_WILDCARD_ENABLED) == 0) { ccb->ccb_h.status = CAM_REQ_CMP; return; } if (tptr->hold) { ccb->ccb_h.status = CAM_SCSI_BUSY; return; } xpt_free_path(tptr->owner); isp->isp_osinfo.tmflags[bus] &= ~TM_WILDCARD_ENABLED; } } /* * Now check to see whether this bus needs to be * enabled/disabled with respect to target mode. */ av = bus << 31; if (cel->enable && !(isp->isp_osinfo.tmflags[bus] & TM_TMODE_ENABLED)) { av |= ENABLE_TARGET_FLAG; av = isp_control(isp, ISPCTL_TOGGLE_TMODE, &av); if (av) { ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; if (wildcard) { isp->isp_osinfo.tmflags[bus] &= ~TM_WILDCARD_ENABLED; xpt_free_path(tptr->owner); } return; } isp->isp_osinfo.tmflags[bus] |= TM_TMODE_ENABLED; isp_prt(isp, ISP_LOGINFO, "Target Mode enabled on channel %d", bus); } else if (cel->enable == 0 && (isp->isp_osinfo.tmflags[bus] & TM_TMODE_ENABLED) && wildcard) { if (are_any_luns_enabled(isp, bus)) { ccb->ccb_h.status = CAM_SCSI_BUSY; return; } av = isp_control(isp, ISPCTL_TOGGLE_TMODE, &av); if (av) { ccb->ccb_h.status = CAM_FUNC_NOTAVAIL; return; } isp->isp_osinfo.tmflags[bus] &= ~TM_TMODE_ENABLED; isp_prt(isp, ISP_LOGINFO, "Target Mode disabled on channel %d", bus); } if (wildcard) { ccb->ccb_h.status = CAM_REQ_CMP; return; } if (cel->enable) { ccb->ccb_h.status = create_lun_state(isp, bus, ccb->ccb_h.path, &tptr); if (ccb->ccb_h.status != CAM_REQ_CMP) { return; } } else { tptr = get_lun_statep(isp, bus, lun); if (tptr == NULL) { ccb->ccb_h.status = CAM_LUN_INVALID; return; } } if (isp_psema_sig_rqe(isp, bus)) { rls_lun_statep(isp, tptr); if (cel->enable) destroy_lun_state(isp, tptr); ccb->ccb_h.status = CAM_REQ_CMP_ERR; return; } if (cel->enable) { u_int32_t seq = isp->isp_osinfo.rollinfo++; int c, n, ulun = lun; cmd = RQSTYPE_ENABLE_LUN; c = DFLT_CMND_CNT; n = DFLT_INOT_CNT; if (IS_FC(isp) && lun != 0) { cmd = RQSTYPE_MODIFY_LUN; n = 0; /* * For SCC firmware, we only deal with setting * (enabling or modifying) lun 0. */ ulun = 0; } rstat = LUN_ERR; if (isp_lun_cmd(isp, cmd, bus, tgt, ulun, c, n, seq)) { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGWARN, "isp_lun_cmd failed"); goto out; } if (isp_cv_wait_timed_rqe(isp, bus, 30 * hz)) { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGERR, "wait for ENABLE/MODIFY LUN timed out"); goto out; } rstat = isp->isp_osinfo.rstatus[bus]; if (rstat != LUN_OK) { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGERR, "ENABLE/MODIFY LUN returned 0x%x", rstat); goto out; } } else { int c, n, ulun = lun; u_int32_t seq; rstat = LUN_ERR; seq = isp->isp_osinfo.rollinfo++; cmd = -RQSTYPE_MODIFY_LUN; c = DFLT_CMND_CNT; n = DFLT_INOT_CNT; if (IS_FC(isp) && lun != 0) { n = 0; /* * For SCC firmware, we only deal with setting * (enabling or modifying) lun 0. */ ulun = 0; } if (isp_lun_cmd(isp, cmd, bus, tgt, ulun, c, n, seq)) { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGERR, "isp_lun_cmd failed"); goto out; } if (isp_cv_wait_timed_rqe(isp, bus, 30 * hz)) { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGERR, "wait for MODIFY LUN timed out"); goto out; } rstat = isp->isp_osinfo.rstatus[bus]; if (rstat != LUN_OK) { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGERR, "MODIFY LUN returned 0x%x", rstat); goto out; } if (IS_FC(isp) && lun) { goto out; } seq = isp->isp_osinfo.rollinfo++; rstat = LUN_ERR; cmd = -RQSTYPE_ENABLE_LUN; if (isp_lun_cmd(isp, cmd, bus, tgt, lun, 0, 0, seq)) { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGERR, "isp_lun_cmd failed"); goto out; } if (isp_cv_wait_timed_rqe(isp, bus, 30 * hz)) { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGERR, "wait for DISABLE LUN timed out"); goto out; } rstat = isp->isp_osinfo.rstatus[bus]; if (rstat != LUN_OK) { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGWARN, "DISABLE LUN returned 0x%x", rstat); goto out; } if (are_any_luns_enabled(isp, bus) == 0) { av = isp_control(isp, ISPCTL_TOGGLE_TMODE, &av); if (av) { isp_prt(isp, ISP_LOGWARN, "disable target mode on channel %d failed", bus); goto out; } isp->isp_osinfo.tmflags[bus] &= ~TM_TMODE_ENABLED; xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGINFO, "Target Mode disabled on channel %d", bus); } } out: isp_vsema_rqe(isp, bus); if (rstat != LUN_OK) { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGWARN, "lun %sable failed", (cel->enable) ? "en" : "dis"); ccb->ccb_h.status = CAM_REQ_CMP_ERR; rls_lun_statep(isp, tptr); if (cel->enable) destroy_lun_state(isp, tptr); } else { xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGINFO, lfmt, (cel->enable) ? "en" : "dis", bus); rls_lun_statep(isp, tptr); if (cel->enable == 0) { destroy_lun_state(isp, tptr); } ccb->ccb_h.status = CAM_REQ_CMP; } } static cam_status isp_abort_tgt_ccb(struct ispsoftc *isp, union ccb *ccb) { tstate_t *tptr; struct ccb_hdr_slist *lp; struct ccb_hdr *curelm; int found; union ccb *accb = ccb->cab.abort_ccb; if (accb->ccb_h.target_id != CAM_TARGET_WILDCARD) { if (IS_FC(isp) && (accb->ccb_h.target_id != ((fcparam *) isp->isp_param)->isp_loopid)) { return (CAM_PATH_INVALID); } else if (IS_SCSI(isp) && (accb->ccb_h.target_id != ((sdparam *) isp->isp_param)->isp_initiator_id)) { return (CAM_PATH_INVALID); } } tptr = get_lun_statep(isp, XS_CHANNEL(ccb), accb->ccb_h.target_lun); if (tptr == NULL) { return (CAM_PATH_INVALID); } if (accb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) { lp = &tptr->atios; } else if (accb->ccb_h.func_code == XPT_IMMED_NOTIFY) { lp = &tptr->inots; } else { rls_lun_statep(isp, tptr); return (CAM_UA_ABORT); } curelm = SLIST_FIRST(lp); found = 0; if (curelm == &accb->ccb_h) { found = 1; SLIST_REMOVE_HEAD(lp, sim_links.sle); } else { while(curelm != NULL) { struct ccb_hdr *nextelm; nextelm = SLIST_NEXT(curelm, sim_links.sle); if (nextelm == &accb->ccb_h) { found = 1; SLIST_NEXT(curelm, sim_links.sle) = SLIST_NEXT(nextelm, sim_links.sle); break; } curelm = nextelm; } } rls_lun_statep(isp, tptr); if (found) { accb->ccb_h.status = CAM_REQ_ABORTED; return (CAM_REQ_CMP); } return(CAM_PATH_INVALID); } static cam_status isp_target_start_ctio(struct ispsoftc *isp, union ccb *ccb) { void *qe; struct ccb_scsiio *cso = &ccb->csio; u_int16_t *hp, save_handle; u_int16_t nxti, optr; u_int8_t local[QENTRY_LEN]; if (isp_getrqentry(isp, &nxti, &optr, &qe)) { xpt_print_path(ccb->ccb_h.path); printf("Request Queue Overflow in isp_target_start_ctio\n"); return (CAM_RESRC_UNAVAIL); } bzero(local, QENTRY_LEN); /* * We're either moving data or completing a command here. */ if (IS_FC(isp)) { atio_private_data_t *atp; ct2_entry_t *cto = (ct2_entry_t *) local; cto->ct_header.rqs_entry_type = RQSTYPE_CTIO2; cto->ct_header.rqs_entry_count = 1; cto->ct_iid = cso->init_id; if ((FCPARAM(isp)->isp_fwattr & ISP_FW_ATTR_SCCLUN) == 0) { cto->ct_lun = ccb->ccb_h.target_lun; } atp = isp_get_atpd(isp, cso->tag_id); if (atp == NULL) { panic("cannot find private data adjunct for tag %x", cso->tag_id); } cto->ct_rxid = cso->tag_id; if (cso->dxfer_len == 0) { cto->ct_flags |= CT2_FLAG_MODE1 | CT2_NO_DATA; if (ccb->ccb_h.flags & CAM_SEND_STATUS) { cto->ct_flags |= CT2_SENDSTATUS; cto->rsp.m1.ct_scsi_status = cso->scsi_status; cto->ct_resid = atp->orig_datalen - atp->bytes_xfered; } if ((ccb->ccb_h.flags & CAM_SEND_SENSE) != 0) { int m = min(cso->sense_len, MAXRESPLEN); bcopy(&cso->sense_data, cto->rsp.m1.ct_resp, m); cto->rsp.m1.ct_senselen = m; cto->rsp.m1.ct_scsi_status |= CT2_SNSLEN_VALID; } } else { cto->ct_flags |= CT2_FLAG_MODE0; if ((cso->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { cto->ct_flags |= CT2_DATA_IN; } else { cto->ct_flags |= CT2_DATA_OUT; } if ((ccb->ccb_h.flags & CAM_SEND_STATUS) != 0) { cto->ct_flags |= CT2_SENDSTATUS; cto->rsp.m0.ct_scsi_status = cso->scsi_status; cto->ct_resid = atp->orig_datalen - (atp->bytes_xfered + cso->dxfer_len); } else { atp->last_xframt = cso->dxfer_len; } /* * If we're sending data and status back together, * we can't also send back sense data as well. */ ccb->ccb_h.flags &= ~CAM_SEND_SENSE; } if (cto->ct_flags & CT2_SENDSTATUS) { isp_prt(isp, ISP_LOGTDEBUG0, "CTIO2[%x] STATUS %x origd %u curd %u resid %u", cto->ct_rxid, cso->scsi_status, atp->orig_datalen, cso->dxfer_len, cto->ct_resid); cto->ct_flags |= CT2_CCINCR; } cto->ct_timeout = 10; hp = &cto->ct_syshandle; } else { ct_entry_t *cto = (ct_entry_t *) local; cto->ct_header.rqs_entry_type = RQSTYPE_CTIO; cto->ct_header.rqs_entry_count = 1; cto->ct_iid = cso->init_id; cto->ct_iid |= XS_CHANNEL(ccb) << 7; cto->ct_tgt = ccb->ccb_h.target_id; cto->ct_lun = ccb->ccb_h.target_lun; cto->ct_fwhandle = AT_GET_HANDLE(cso->tag_id); if (AT_HAS_TAG(cso->tag_id)) { cto->ct_tag_val = (u_int8_t) AT_GET_TAG(cso->tag_id); cto->ct_flags |= CT_TQAE; } if (ccb->ccb_h.flags & CAM_DIS_DISCONNECT) { cto->ct_flags |= CT_NODISC; } if (cso->dxfer_len == 0) { cto->ct_flags |= CT_NO_DATA; } else if ((cso->ccb_h.flags & CAM_DIR_MASK) == CAM_DIR_IN) { cto->ct_flags |= CT_DATA_IN; } else { cto->ct_flags |= CT_DATA_OUT; } if (ccb->ccb_h.flags & CAM_SEND_STATUS) { cto->ct_flags |= CT_SENDSTATUS|CT_CCINCR; cto->ct_scsi_status = cso->scsi_status; cto->ct_resid = cso->resid; isp_prt(isp, ISP_LOGTDEBUG0, "CTIO[%x] SCSI STATUS 0x%x resid %d tag_id %x", cto->ct_fwhandle, cso->scsi_status, cso->resid, cso->tag_id); } ccb->ccb_h.flags &= ~CAM_SEND_SENSE; cto->ct_timeout = 10; hp = &cto->ct_syshandle; } if (isp_save_xs(isp, (XS_T *)ccb, hp)) { xpt_print_path(ccb->ccb_h.path); printf("No XFLIST pointers for isp_target_start_ctio\n"); return (CAM_RESRC_UNAVAIL); } /* * Call the dma setup routines for this entry (and any subsequent * CTIOs) if there's data to move, and then tell the f/w it's got * new things to play with. As with isp_start's usage of DMA setup, * any swizzling is done in the machine dependent layer. Because * of this, we put the request onto the queue area first in native * format. */ save_handle = *hp; switch (ISP_DMASETUP(isp, cso, (ispreq_t *) local, &nxti, optr)) { case CMD_QUEUED: ISP_ADD_REQUEST(isp, nxti); return (CAM_REQ_INPROG); case CMD_EAGAIN: ccb->ccb_h.status = CAM_RESRC_UNAVAIL; isp_destroy_handle(isp, save_handle); return (CAM_RESRC_UNAVAIL); default: isp_destroy_handle(isp, save_handle); return (XS_ERR(ccb)); } } static void isp_refire_putback_atio(void *arg) { int s = splcam(); isp_target_putback_atio(arg); splx(s); } static void isp_target_putback_atio(union ccb *ccb) { struct ispsoftc *isp; struct ccb_scsiio *cso; u_int16_t nxti, optr; void *qe; isp = XS_ISP(ccb); if (isp_getrqentry(isp, &nxti, &optr, &qe)) { (void) timeout(isp_refire_putback_atio, ccb, 10); isp_prt(isp, ISP_LOGWARN, "isp_target_putback_atio: Request Queue Overflow"); return; } bzero(qe, QENTRY_LEN); cso = &ccb->csio; if (IS_FC(isp)) { at2_entry_t local, *at = &local; MEMZERO(at, sizeof (at2_entry_t)); at->at_header.rqs_entry_type = RQSTYPE_ATIO2; at->at_header.rqs_entry_count = 1; if ((FCPARAM(isp)->isp_fwattr & ISP_FW_ATTR_SCCLUN) != 0) { at->at_scclun = (uint16_t) ccb->ccb_h.target_lun; } else { at->at_lun = (uint8_t) ccb->ccb_h.target_lun; } at->at_status = CT_OK; at->at_rxid = cso->tag_id; isp_put_atio2(isp, at, qe); } else { at_entry_t local, *at = &local; MEMZERO(at, sizeof (at_entry_t)); at->at_header.rqs_entry_type = RQSTYPE_ATIO; at->at_header.rqs_entry_count = 1; at->at_iid = cso->init_id; at->at_iid |= XS_CHANNEL(ccb) << 7; at->at_tgt = cso->ccb_h.target_id; at->at_lun = cso->ccb_h.target_lun; at->at_status = CT_OK; at->at_tag_val = AT_GET_TAG(cso->tag_id); at->at_handle = AT_GET_HANDLE(cso->tag_id); isp_put_atio(isp, at, qe); } ISP_TDQE(isp, "isp_target_putback_atio", (int) optr, qe); ISP_ADD_REQUEST(isp, nxti); isp_complete_ctio(ccb); } static void isp_complete_ctio(union ccb *ccb) { struct ispsoftc *isp = XS_ISP(ccb); if ((ccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_INPROG) { ccb->ccb_h.status |= CAM_REQ_CMP; } ccb->ccb_h.status &= ~CAM_SIM_QUEUED; if (isp->isp_osinfo.simqfrozen & SIMQFRZ_RESOURCE) { isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_RESOURCE; if (isp->isp_osinfo.simqfrozen == 0) { if ((ccb->ccb_h.status & CAM_DEV_QFRZN) == 0) { isp_prt(isp, ISP_LOGDEBUG2, "ctio->relsimq"); ccb->ccb_h.status |= CAM_RELEASE_SIMQ; } else { isp_prt(isp, ISP_LOGWARN, "ctio->devqfrozen"); } } else { isp_prt(isp, ISP_LOGWARN, "ctio->simqfrozen(%x)", isp->isp_osinfo.simqfrozen); } } xpt_done(ccb); } /* * Handle ATIO stuff that the generic code can't. * This means handling CDBs. */ static int isp_handle_platform_atio(struct ispsoftc *isp, at_entry_t *aep) { tstate_t *tptr; int status, bus, iswildcard; struct ccb_accept_tio *atiop; /* * The firmware status (except for the QLTM_SVALID bit) * indicates why this ATIO was sent to us. * * If QLTM_SVALID is set, the firware has recommended Sense Data. * * If the DISCONNECTS DISABLED bit is set in the flags field, * we're still connected on the SCSI bus. */ status = aep->at_status; if ((status & ~QLTM_SVALID) == AT_PHASE_ERROR) { /* * Bus Phase Sequence error. We should have sense data * suggested by the f/w. I'm not sure quite yet what * to do about this for CAM. */ isp_prt(isp, ISP_LOGWARN, "PHASE ERROR"); isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); return (0); } if ((status & ~QLTM_SVALID) != AT_CDB) { isp_prt(isp, ISP_LOGWARN, "bad atio (0x%x) leaked to platform", status); isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); return (0); } bus = GET_BUS_VAL(aep->at_iid); tptr = get_lun_statep(isp, bus, aep->at_lun); if (tptr == NULL) { tptr = get_lun_statep(isp, bus, CAM_LUN_WILDCARD); iswildcard = 1; } else { iswildcard = 0; } if (tptr == NULL) { /* * Because we can't autofeed sense data back with * a command for parallel SCSI, we can't give back * a CHECK CONDITION. We'll give back a BUSY status * instead. This works out okay because the only * time we should, in fact, get this, is in the * case that somebody configured us without the * blackhole driver, so they get what they deserve. */ isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); return (0); } atiop = (struct ccb_accept_tio *) SLIST_FIRST(&tptr->atios); if (atiop == NULL) { /* * Because we can't autofeed sense data back with * a command for parallel SCSI, we can't give back * a CHECK CONDITION. We'll give back a QUEUE FULL status * instead. This works out okay because the only time we * should, in fact, get this, is in the case that we've * run out of ATIOS. */ xpt_print_path(tptr->owner); isp_prt(isp, ISP_LOGWARN, "no ATIOS for lun %d from initiator %d on channel %d", aep->at_lun, GET_IID_VAL(aep->at_iid), bus); if (aep->at_flags & AT_TQAE) isp_endcmd(isp, aep, SCSI_STATUS_QUEUE_FULL, 0); else isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); rls_lun_statep(isp, tptr); return (0); } SLIST_REMOVE_HEAD(&tptr->atios, sim_links.sle); if (iswildcard) { atiop->ccb_h.target_id = aep->at_tgt; atiop->ccb_h.target_lun = aep->at_lun; } if (aep->at_flags & AT_NODISC) { atiop->ccb_h.flags = CAM_DIS_DISCONNECT; } else { atiop->ccb_h.flags = 0; } if (status & QLTM_SVALID) { size_t amt = imin(QLTM_SENSELEN, sizeof (atiop->sense_data)); atiop->sense_len = amt; MEMCPY(&atiop->sense_data, aep->at_sense, amt); } else { atiop->sense_len = 0; } atiop->init_id = GET_IID_VAL(aep->at_iid); atiop->cdb_len = aep->at_cdblen; MEMCPY(atiop->cdb_io.cdb_bytes, aep->at_cdb, aep->at_cdblen); atiop->ccb_h.status = CAM_CDB_RECVD; /* * Construct a tag 'id' based upon tag value (which may be 0..255) * and the handle (which we have to preserve). */ AT_MAKE_TAGID(atiop->tag_id, aep); if (aep->at_flags & AT_TQAE) { atiop->tag_action = aep->at_tag_type; atiop->ccb_h.status |= CAM_TAG_ACTION_VALID; } xpt_done((union ccb*)atiop); isp_prt(isp, ISP_LOGTDEBUG0, "ATIO[%x] CDB=0x%x bus %d iid%d->lun%d tag 0x%x ttype 0x%x %s", aep->at_handle, aep->at_cdb[0] & 0xff, GET_BUS_VAL(aep->at_iid), GET_IID_VAL(aep->at_iid), aep->at_lun, aep->at_tag_val & 0xff, aep->at_tag_type, (aep->at_flags & AT_NODISC)? "nondisc" : "disconnecting"); rls_lun_statep(isp, tptr); return (0); } static int isp_handle_platform_atio2(struct ispsoftc *isp, at2_entry_t *aep) { lun_id_t lun; tstate_t *tptr; struct ccb_accept_tio *atiop; atio_private_data_t *atp; /* * The firmware status (except for the QLTM_SVALID bit) * indicates why this ATIO was sent to us. * * If QLTM_SVALID is set, the firware has recommended Sense Data. */ if ((aep->at_status & ~QLTM_SVALID) != AT_CDB) { isp_prt(isp, ISP_LOGWARN, "bogus atio (0x%x) leaked to platform", aep->at_status); isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); return (0); } if ((FCPARAM(isp)->isp_fwattr & ISP_FW_ATTR_SCCLUN) != 0) { lun = aep->at_scclun; } else { lun = aep->at_lun; } tptr = get_lun_statep(isp, 0, lun); if (tptr == NULL) { tptr = get_lun_statep(isp, 0, CAM_LUN_WILDCARD); } if (tptr == NULL) { /* * What we'd like to know is whether or not we have a listener * upstream that really hasn't configured yet. If we do, then * we can give a more sensible reply here. If not, then we can * reject this out of hand. * * Choices for what to send were * * Not Ready, Unit Not Self-Configured Yet * (0x2,0x3e,0x00) * * for the former and * * Illegal Request, Logical Unit Not Supported * (0x5,0x25,0x00) * * for the latter. * * We used to decide whether there was at least one listener * based upon whether the black hole driver was configured. * However, recent config(8) changes have made this hard to do * at this time. * */ u_int32_t ccode = SCSI_STATUS_BUSY; /* * Because we can't autofeed sense data back with * a command for parallel SCSI, we can't give back * a CHECK CONDITION. We'll give back a BUSY status * instead. This works out okay because the only * time we should, in fact, get this, is in the * case that somebody configured us without the * blackhole driver, so they get what they deserve. */ isp_endcmd(isp, aep, ccode, 0); return (0); } atp = isp_get_atpd(isp, 0); atiop = (struct ccb_accept_tio *) SLIST_FIRST(&tptr->atios); if (atiop == NULL || atp == NULL) { /* * Because we can't autofeed sense data back with * a command for parallel SCSI, we can't give back * a CHECK CONDITION. We'll give back a QUEUE FULL status * instead. This works out okay because the only time we * should, in fact, get this, is in the case that we've * run out of ATIOS. */ xpt_print_path(tptr->owner); isp_prt(isp, ISP_LOGWARN, "no ATIOS for lun %d from initiator %d", lun, aep->at_iid); rls_lun_statep(isp, tptr); if (aep->at_flags & AT_TQAE) isp_endcmd(isp, aep, SCSI_STATUS_QUEUE_FULL, 0); else isp_endcmd(isp, aep, SCSI_STATUS_BUSY, 0); return (0); } SLIST_REMOVE_HEAD(&tptr->atios, sim_links.sle); if (tptr == &isp->isp_osinfo.tsdflt[0]) { atiop->ccb_h.target_id = ((fcparam *)isp->isp_param)->isp_loopid; atiop->ccb_h.target_lun = lun; } /* * We don't get 'suggested' sense data as we do with SCSI cards. */ atiop->sense_len = 0; atiop->init_id = aep->at_iid; atiop->cdb_len = ATIO2_CDBLEN; MEMCPY(atiop->cdb_io.cdb_bytes, aep->at_cdb, ATIO2_CDBLEN); atiop->ccb_h.status = CAM_CDB_RECVD; atiop->tag_id = aep->at_rxid; switch (aep->at_taskflags & ATIO2_TC_ATTR_MASK) { case ATIO2_TC_ATTR_SIMPLEQ: atiop->tag_action = MSG_SIMPLE_Q_TAG; break; case ATIO2_TC_ATTR_HEADOFQ: atiop->tag_action = MSG_HEAD_OF_Q_TAG; break; case ATIO2_TC_ATTR_ORDERED: atiop->tag_action = MSG_ORDERED_Q_TAG; break; case ATIO2_TC_ATTR_ACAQ: /* ?? */ case ATIO2_TC_ATTR_UNTAGGED: default: atiop->tag_action = 0; break; } if (atiop->tag_action != 0) { atiop->ccb_h.status |= CAM_TAG_ACTION_VALID; } atp->tag = atiop->tag_id; atp->orig_datalen = aep->at_datalen; atp->last_xframt = 0; atp->bytes_xfered = 0; xpt_done((union ccb*)atiop); isp_prt(isp, ISP_LOGTDEBUG0, "ATIO2[%x] CDB=0x%x iid%d->lun%d tattr 0x%x datalen %u", aep->at_rxid, aep->at_cdb[0] & 0xff, aep->at_iid, lun, aep->at_taskflags, aep->at_datalen); rls_lun_statep(isp, tptr); return (0); } static int isp_handle_platform_ctio(struct ispsoftc *isp, void *arg) { union ccb *ccb; int sentstatus, ok, notify_cam, resid = 0; u_int16_t tval; /* * CTIO and CTIO2 are close enough.... */ ccb = (union ccb *) isp_find_xs(isp, ((ct_entry_t *)arg)->ct_syshandle); KASSERT((ccb != NULL), ("null ccb in isp_handle_platform_ctio")); isp_destroy_handle(isp, ((ct_entry_t *)arg)->ct_syshandle); if (IS_FC(isp)) { ct2_entry_t *ct = arg; sentstatus = ct->ct_flags & CT2_SENDSTATUS; ok = (ct->ct_status & ~QLTM_SVALID) == CT_OK; if (ok && sentstatus && (ccb->ccb_h.flags & CAM_SEND_SENSE)) { ccb->ccb_h.status |= CAM_SENT_SENSE; } notify_cam = ct->ct_header.rqs_seqno & 0x1; if ((ct->ct_flags & CT2_DATAMASK) != CT2_NO_DATA) { atio_private_data_t *atp = isp_get_atpd(isp, ct->ct_rxid); if (atp == NULL) { panic("cannot find adjunct after I/O"); } resid = ct->ct_resid; atp->bytes_xfered += (atp->last_xframt - resid); atp->last_xframt = 0; if (sentstatus) { atp->tag = 0; } } isp_prt(isp, ISP_LOGTDEBUG0, "CTIO2[%x] sts 0x%x flg 0x%x sns %d resid %d %s", ct->ct_rxid, ct->ct_status, ct->ct_flags, (ccb->ccb_h.status & CAM_SENT_SENSE) != 0, resid, sentstatus? "FIN" : "MID"); tval = ct->ct_rxid; } else { ct_entry_t *ct = arg; sentstatus = ct->ct_flags & CT_SENDSTATUS; ok = (ct->ct_status & ~QLTM_SVALID) == CT_OK; /* * We *ought* to be able to get back to the original ATIO * here, but for some reason this gets lost. It's just as * well because it's squirrelled away as part of periph * private data. * * We can live without it as long as we continue to use * the auto-replenish feature for CTIOs. */ notify_cam = ct->ct_header.rqs_seqno & 0x1; if (ct->ct_status & QLTM_SVALID) { char *sp = (char *)ct; sp += CTIO_SENSE_OFFSET; ccb->csio.sense_len = min(sizeof (ccb->csio.sense_data), QLTM_SENSELEN); MEMCPY(&ccb->csio.sense_data, sp, ccb->csio.sense_len); ccb->ccb_h.status |= CAM_AUTOSNS_VALID; } if ((ct->ct_flags & CT_DATAMASK) != CT_NO_DATA) { resid = ct->ct_resid; } isp_prt(isp, ISP_LOGTDEBUG0, "CTIO[%x] tag %x iid %d lun %d sts %x flg %x resid %d %s", ct->ct_fwhandle, ct->ct_tag_val, ct->ct_iid, ct->ct_lun, ct->ct_status, ct->ct_flags, resid, sentstatus? "FIN" : "MID"); tval = ct->ct_fwhandle; } ccb->csio.resid += resid; /* * We're here either because intermediate data transfers are done * and/or the final status CTIO (which may have joined with a * Data Transfer) is done. * * In any case, for this platform, the upper layers figure out * what to do next, so all we do here is collect status and * pass information along. Any DMA handles have already been * freed. */ if (notify_cam == 0) { isp_prt(isp, ISP_LOGTDEBUG0, " INTER CTIO[0x%x] done", tval); return (0); } isp_prt(isp, ISP_LOGTDEBUG0, "%s CTIO[0x%x] done", (sentstatus)? " FINAL " : "MIDTERM ", tval); if (!ok) { isp_target_putback_atio(ccb); } else { isp_complete_ctio(ccb); } return (0); } #endif static void isp_cam_async(void *cbarg, u_int32_t code, struct cam_path *path, void *arg) { struct cam_sim *sim; struct ispsoftc *isp; sim = (struct cam_sim *)cbarg; isp = (struct ispsoftc *) cam_sim_softc(sim); switch (code) { case AC_LOST_DEVICE: if (IS_SCSI(isp)) { u_int16_t oflags, nflags; sdparam *sdp = isp->isp_param; int tgt; tgt = xpt_path_target_id(path); ISP_LOCK(isp); sdp += cam_sim_bus(sim); nflags = sdp->isp_devparam[tgt].nvrm_flags; #ifndef ISP_TARGET_MODE nflags &= DPARM_SAFE_DFLT; if (isp->isp_loaded_fw) { nflags |= DPARM_NARROW | DPARM_ASYNC; } #else nflags = DPARM_DEFAULT; #endif oflags = sdp->isp_devparam[tgt].goal_flags; sdp->isp_devparam[tgt].goal_flags = nflags; sdp->isp_devparam[tgt].dev_update = 1; isp->isp_update |= (1 << cam_sim_bus(sim)); (void) isp_control(isp, ISPCTL_UPDATE_PARAMS, NULL); sdp->isp_devparam[tgt].goal_flags = oflags; ISP_UNLOCK(isp); } break; default: isp_prt(isp, ISP_LOGWARN, "isp_cam_async: Code 0x%x", code); break; } } static void isp_poll(struct cam_sim *sim) { struct ispsoftc *isp = cam_sim_softc(sim); u_int16_t isr, sema, mbox; ISP_LOCK(isp); if (ISP_READ_ISR(isp, &isr, &sema, &mbox)) { isp_intr(isp, isr, sema, mbox); } ISP_UNLOCK(isp); } #if 0 static void isp_relsim(void *arg) { struct ispsoftc *isp = arg; ISP_LOCK(isp); if (isp->isp_osinfo.simqfrozen & SIMQFRZ_TIMED) { int wasfrozen = isp->isp_osinfo.simqfrozen & SIMQFRZ_TIMED; isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_TIMED; if (wasfrozen && isp->isp_osinfo.simqfrozen == 0) { xpt_release_simq(isp->isp_sim, 1); isp_prt(isp, ISP_LOGDEBUG2, "timed relsimq"); } } ISP_UNLOCK(isp); } #endif static void isp_watchdog(void *arg) { XS_T *xs = arg; struct ispsoftc *isp = XS_ISP(xs); u_int32_t handle; /* * We've decided this command is dead. Make sure we're not trying * to kill a command that's already dead by getting it's handle and * and seeing whether it's still alive. */ ISP_LOCK(isp); handle = isp_find_handle(isp, xs); if (handle) { u_int16_t isr, sema, mbox; if (XS_CMD_DONE_P(xs)) { isp_prt(isp, ISP_LOGDEBUG1, "watchdog found done cmd (handle 0x%x)", handle); ISP_UNLOCK(isp); return; } if (XS_CMD_WDOG_P(xs)) { isp_prt(isp, ISP_LOGDEBUG2, "recursive watchdog (handle 0x%x)", handle); ISP_UNLOCK(isp); return; } XS_CMD_S_WDOG(xs); if (ISP_READ_ISR(isp, &isr, &sema, &mbox)) { isp_intr(isp, isr, sema, mbox); } if (XS_CMD_DONE_P(xs)) { isp_prt(isp, ISP_LOGDEBUG2, "watchdog cleanup for handle 0x%x", handle); xpt_done((union ccb *) xs); } else if (XS_CMD_GRACE_P(xs)) { /* * Make sure the command is *really* dead before we * release the handle (and DMA resources) for reuse. */ (void) isp_control(isp, ISPCTL_ABORT_CMD, arg); /* * After this point, the comamnd is really dead. */ if (XS_XFRLEN(xs)) { ISP_DMAFREE(isp, xs, handle); } isp_destroy_handle(isp, handle); xpt_print_path(xs->ccb_h.path); isp_prt(isp, ISP_LOGWARN, "watchdog timeout for handle 0x%x", handle); XS_SETERR(xs, CAM_CMD_TIMEOUT); XS_CMD_C_WDOG(xs); isp_done(xs); } else { u_int16_t nxti, optr; ispreq_t local, *mp= &local, *qe; XS_CMD_C_WDOG(xs); xs->ccb_h.timeout_ch = timeout(isp_watchdog, xs, hz); if (isp_getrqentry(isp, &nxti, &optr, (void **) &qe)) { ISP_UNLOCK(isp); return; } XS_CMD_S_GRACE(xs); MEMZERO((void *) mp, sizeof (*mp)); mp->req_header.rqs_entry_count = 1; mp->req_header.rqs_entry_type = RQSTYPE_MARKER; mp->req_modifier = SYNC_ALL; mp->req_target = XS_CHANNEL(xs) << 7; isp_put_request(isp, mp, qe); ISP_ADD_REQUEST(isp, nxti); } } else { isp_prt(isp, ISP_LOGDEBUG2, "watchdog with no command"); } ISP_UNLOCK(isp); } static int isp_ktmature = 0; static void isp_kthread(void *arg) { int wasfrozen; struct ispsoftc *isp = arg; mtx_lock(&isp->isp_lock); for (;;) { isp_prt(isp, ISP_LOGDEBUG0, "kthread checking FC state"); while (isp_fc_runstate(isp, 2 * 1000000) != 0) { if (FCPARAM(isp)->isp_fwstate != FW_READY || FCPARAM(isp)->isp_loopstate < LOOP_PDB_RCVD) { if (FCPARAM(isp)->loop_seen_once == 0 || isp_ktmature == 0) { break; } } msleep(isp_kthread, &isp->isp_lock, PRIBIO, "isp_fcthrd", hz); } /* * Even if we didn't get good loop state we may be * unfreezing the SIMQ so that we can kill off * commands (if we've never seen loop before, e.g.) */ isp_ktmature = 1; wasfrozen = isp->isp_osinfo.simqfrozen & SIMQFRZ_LOOPDOWN; isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_LOOPDOWN; if (wasfrozen && isp->isp_osinfo.simqfrozen == 0) { isp_prt(isp, ISP_LOGDEBUG0, "kthread up release simq"); ISPLOCK_2_CAMLOCK(isp); xpt_release_simq(isp->isp_sim, 1); CAMLOCK_2_ISPLOCK(isp); } cv_wait(&isp->isp_osinfo.kthread_cv, &isp->isp_lock); } } static void isp_action(struct cam_sim *sim, union ccb *ccb) { int bus, tgt, error; struct ispsoftc *isp; struct ccb_trans_settings *cts; CAM_DEBUG(ccb->ccb_h.path, CAM_DEBUG_TRACE, ("isp_action\n")); isp = (struct ispsoftc *)cam_sim_softc(sim); ccb->ccb_h.sim_priv.entries[0].field = 0; ccb->ccb_h.sim_priv.entries[1].ptr = isp; if (isp->isp_state != ISP_RUNSTATE && ccb->ccb_h.func_code == XPT_SCSI_IO) { CAMLOCK_2_ISPLOCK(isp); isp_init(isp); if (isp->isp_state != ISP_INITSTATE) { ISP_UNLOCK(isp); /* * Lie. Say it was a selection timeout. */ ccb->ccb_h.status = CAM_SEL_TIMEOUT | CAM_DEV_QFRZN; xpt_freeze_devq(ccb->ccb_h.path, 1); xpt_done(ccb); return; } isp->isp_state = ISP_RUNSTATE; ISPLOCK_2_CAMLOCK(isp); } isp_prt(isp, ISP_LOGDEBUG2, "isp_action code %x", ccb->ccb_h.func_code); switch (ccb->ccb_h.func_code) { case XPT_SCSI_IO: /* Execute the requested I/O operation */ /* * Do a couple of preliminary checks... */ if ((ccb->ccb_h.flags & CAM_CDB_POINTER) != 0) { if ((ccb->ccb_h.flags & CAM_CDB_PHYS) != 0) { ccb->ccb_h.status = CAM_REQ_INVALID; xpt_done(ccb); break; } } #ifdef DIAGNOSTIC if (ccb->ccb_h.target_id > (ISP_MAX_TARGETS(isp) - 1)) { ccb->ccb_h.status = CAM_PATH_INVALID; } else if (ccb->ccb_h.target_lun > (ISP_MAX_LUNS(isp) - 1)) { ccb->ccb_h.status = CAM_PATH_INVALID; } if (ccb->ccb_h.status == CAM_PATH_INVALID) { isp_prt(isp, ISP_LOGERR, "invalid tgt/lun (%d.%d) in XPT_SCSI_IO", ccb->ccb_h.target_id, ccb->ccb_h.target_lun); xpt_done(ccb); break; } #endif ((struct ccb_scsiio *) ccb)->scsi_status = SCSI_STATUS_OK; CAMLOCK_2_ISPLOCK(isp); error = isp_start((XS_T *) ccb); switch (error) { case CMD_QUEUED: ccb->ccb_h.status |= CAM_SIM_QUEUED; if (ccb->ccb_h.timeout != CAM_TIME_INFINITY) { u_int64_t ticks = (u_int64_t) hz; if (ccb->ccb_h.timeout == CAM_TIME_DEFAULT) ticks = 60 * 1000 * ticks; else ticks = ccb->ccb_h.timeout * hz; ticks = ((ticks + 999) / 1000) + hz + hz; if (ticks >= 0x80000000) { isp_prt(isp, ISP_LOGERR, "timeout overflow"); ticks = 0x80000000; } ccb->ccb_h.timeout_ch = timeout(isp_watchdog, (caddr_t)ccb, (int)ticks); } else { callout_handle_init(&ccb->ccb_h.timeout_ch); } ISPLOCK_2_CAMLOCK(isp); break; case CMD_RQLATER: /* * This can only happen for Fibre Channel */ KASSERT((IS_FC(isp)), ("CMD_RQLATER for FC only")); if (FCPARAM(isp)->loop_seen_once == 0 && isp_ktmature) { ISPLOCK_2_CAMLOCK(isp); XS_SETERR(ccb, CAM_SEL_TIMEOUT); xpt_done(ccb); break; } cv_signal(&isp->isp_osinfo.kthread_cv); if (isp->isp_osinfo.simqfrozen == 0) { isp_prt(isp, ISP_LOGDEBUG2, "RQLATER freeze simq"); isp->isp_osinfo.simqfrozen |= SIMQFRZ_LOOPDOWN; ISPLOCK_2_CAMLOCK(isp); xpt_freeze_simq(sim, 1); } else { ISPLOCK_2_CAMLOCK(isp); } XS_SETERR(ccb, CAM_REQUEUE_REQ); xpt_done(ccb); break; case CMD_EAGAIN: if (isp->isp_osinfo.simqfrozen == 0) { xpt_freeze_simq(sim, 1); isp_prt(isp, ISP_LOGDEBUG2, "EAGAIN freeze simq"); } isp->isp_osinfo.simqfrozen |= SIMQFRZ_RESOURCE; XS_SETERR(ccb, CAM_REQUEUE_REQ); ISPLOCK_2_CAMLOCK(isp); xpt_done(ccb); break; case CMD_COMPLETE: isp_done((struct ccb_scsiio *) ccb); ISPLOCK_2_CAMLOCK(isp); break; default: isp_prt(isp, ISP_LOGERR, "What's this? 0x%x at %d in file %s", error, __LINE__, __FILE__); XS_SETERR(ccb, CAM_REQ_CMP_ERR); xpt_done(ccb); ISPLOCK_2_CAMLOCK(isp); } break; #ifdef ISP_TARGET_MODE case XPT_EN_LUN: /* Enable LUN as a target */ { int iok; CAMLOCK_2_ISPLOCK(isp); iok = isp->isp_osinfo.intsok; isp->isp_osinfo.intsok = 0; isp_en_lun(isp, ccb); isp->isp_osinfo.intsok = iok; ISPLOCK_2_CAMLOCK(isp); xpt_done(ccb); break; } case XPT_NOTIFY_ACK: /* recycle notify ack */ case XPT_IMMED_NOTIFY: /* Add Immediate Notify Resource */ case XPT_ACCEPT_TARGET_IO: /* Add Accept Target IO Resource */ { tstate_t *tptr = get_lun_statep(isp, XS_CHANNEL(ccb), ccb->ccb_h.target_lun); if (tptr == NULL) { ccb->ccb_h.status = CAM_LUN_INVALID; xpt_done(ccb); break; } ccb->ccb_h.sim_priv.entries[0].field = 0; ccb->ccb_h.sim_priv.entries[1].ptr = isp; CAMLOCK_2_ISPLOCK(isp); if (ccb->ccb_h.func_code == XPT_ACCEPT_TARGET_IO) { SLIST_INSERT_HEAD(&tptr->atios, &ccb->ccb_h, sim_links.sle); } else { SLIST_INSERT_HEAD(&tptr->inots, &ccb->ccb_h, sim_links.sle); } rls_lun_statep(isp, tptr); ccb->ccb_h.status = CAM_REQ_INPROG; ISPLOCK_2_CAMLOCK(isp); break; } case XPT_CONT_TARGET_IO: { CAMLOCK_2_ISPLOCK(isp); ccb->ccb_h.status = isp_target_start_ctio(isp, ccb); if (ccb->ccb_h.status != CAM_REQ_INPROG) { if (isp->isp_osinfo.simqfrozen == 0) { xpt_freeze_simq(sim, 1); xpt_print_path(ccb->ccb_h.path); isp_prt(isp, ISP_LOGINFO, "XPT_CONT_TARGET_IO freeze simq"); } isp->isp_osinfo.simqfrozen |= SIMQFRZ_RESOURCE; XS_SETERR(ccb, CAM_REQUEUE_REQ); ISPLOCK_2_CAMLOCK(isp); xpt_done(ccb); } else { ISPLOCK_2_CAMLOCK(isp); ccb->ccb_h.status |= CAM_SIM_QUEUED; } break; } #endif case XPT_RESET_DEV: /* BDR the specified SCSI device */ bus = cam_sim_bus(xpt_path_sim(ccb->ccb_h.path)); tgt = ccb->ccb_h.target_id; tgt |= (bus << 16); CAMLOCK_2_ISPLOCK(isp); error = isp_control(isp, ISPCTL_RESET_DEV, &tgt); ISPLOCK_2_CAMLOCK(isp); if (error) { ccb->ccb_h.status = CAM_REQ_CMP_ERR; } else { ccb->ccb_h.status = CAM_REQ_CMP; } xpt_done(ccb); break; case XPT_ABORT: /* Abort the specified CCB */ { union ccb *accb = ccb->cab.abort_ccb; CAMLOCK_2_ISPLOCK(isp); switch (accb->ccb_h.func_code) { #ifdef ISP_TARGET_MODE case XPT_ACCEPT_TARGET_IO: case XPT_IMMED_NOTIFY: ccb->ccb_h.status = isp_abort_tgt_ccb(isp, ccb); break; case XPT_CONT_TARGET_IO: isp_prt(isp, ISP_LOGERR, "cannot abort CTIOs yet"); ccb->ccb_h.status = CAM_UA_ABORT; break; #endif case XPT_SCSI_IO: error = isp_control(isp, ISPCTL_ABORT_CMD, ccb); if (error) { ccb->ccb_h.status = CAM_UA_ABORT; } else { ccb->ccb_h.status = CAM_REQ_CMP; } break; default: ccb->ccb_h.status = CAM_REQ_INVALID; break; } ISPLOCK_2_CAMLOCK(isp); xpt_done(ccb); break; } #ifdef CAM_NEW_TRAN_CODE #define IS_CURRENT_SETTINGS(c) (c->type == CTS_TYPE_CURRENT_SETTINGS) #else #define IS_CURRENT_SETTINGS(c) (c->flags & CCB_TRANS_CURRENT_SETTINGS) #endif case XPT_SET_TRAN_SETTINGS: /* Nexus Settings */ cts = &ccb->cts; if (!IS_CURRENT_SETTINGS(cts)) { ccb->ccb_h.status = CAM_REQ_INVALID; xpt_done(ccb); break; } tgt = cts->ccb_h.target_id; CAMLOCK_2_ISPLOCK(isp); if (IS_SCSI(isp)) { #ifndef CAM_NEW_TRAN_CODE sdparam *sdp = isp->isp_param; u_int16_t *dptr; bus = cam_sim_bus(xpt_path_sim(cts->ccb_h.path)); sdp += bus; /* * We always update (internally) from goal_flags * so any request to change settings just gets * vectored to that location. */ dptr = &sdp->isp_devparam[tgt].goal_flags; /* * Note that these operations affect the * the goal flags (goal_flags)- not * the current state flags. Then we mark * things so that the next operation to * this HBA will cause the update to occur. */ if (cts->valid & CCB_TRANS_DISC_VALID) { if ((cts->flags & CCB_TRANS_DISC_ENB) != 0) { *dptr |= DPARM_DISC; } else { *dptr &= ~DPARM_DISC; } } if (cts->valid & CCB_TRANS_TQ_VALID) { if ((cts->flags & CCB_TRANS_TAG_ENB) != 0) { *dptr |= DPARM_TQING; } else { *dptr &= ~DPARM_TQING; } } if (cts->valid & CCB_TRANS_BUS_WIDTH_VALID) { switch (cts->bus_width) { case MSG_EXT_WDTR_BUS_16_BIT: *dptr |= DPARM_WIDE; break; default: *dptr &= ~DPARM_WIDE; } } /* * Any SYNC RATE of nonzero and SYNC_OFFSET * of nonzero will cause us to go to the * selected (from NVRAM) maximum value for * this device. At a later point, we'll * allow finer control. */ if ((cts->valid & CCB_TRANS_SYNC_RATE_VALID) && (cts->valid & CCB_TRANS_SYNC_OFFSET_VALID) && (cts->sync_offset > 0)) { *dptr |= DPARM_SYNC; } else { *dptr &= ~DPARM_SYNC; } *dptr |= DPARM_SAFE_DFLT; #else struct ccb_trans_settings_scsi *scsi = &cts->proto_specific.scsi; struct ccb_trans_settings_spi *spi = &cts->xport_specific.spi; sdparam *sdp = isp->isp_param; u_int16_t *dptr; bus = cam_sim_bus(xpt_path_sim(cts->ccb_h.path)); sdp += bus; /* * We always update (internally) from goal_flags * so any request to change settings just gets * vectored to that location. */ dptr = &sdp->isp_devparam[tgt].goal_flags; if ((spi->valid & CTS_SPI_VALID_DISC) != 0) { if ((spi->flags & CTS_SPI_FLAGS_DISC_ENB) != 0) *dptr |= DPARM_DISC; else *dptr &= ~DPARM_DISC; } if ((scsi->valid & CTS_SCSI_VALID_TQ) != 0) { if ((scsi->flags & CTS_SCSI_FLAGS_TAG_ENB) != 0) *dptr |= DPARM_TQING; else *dptr &= ~DPARM_TQING; } if ((spi->valid & CTS_SPI_VALID_BUS_WIDTH) != 0) { if (spi->bus_width == MSG_EXT_WDTR_BUS_16_BIT) *dptr |= DPARM_WIDE; else *dptr &= ~DPARM_WIDE; } /* * XXX: FIX ME */ if ((spi->valid & CTS_SPI_VALID_SYNC_OFFSET) && (spi->valid & CTS_SPI_VALID_SYNC_RATE) && (spi->sync_period && spi->sync_offset)) { *dptr |= DPARM_SYNC; /* * XXX: CHECK FOR LEGALITY */ sdp->isp_devparam[tgt].goal_period = spi->sync_period; sdp->isp_devparam[tgt].goal_offset = spi->sync_offset; } else { *dptr &= ~DPARM_SYNC; } #endif isp_prt(isp, ISP_LOGDEBUG0, "SET bus %d targ %d to flags %x off %x per %x", bus, tgt, sdp->isp_devparam[tgt].goal_flags, sdp->isp_devparam[tgt].goal_offset, sdp->isp_devparam[tgt].goal_period); sdp->isp_devparam[tgt].dev_update = 1; isp->isp_update |= (1 << bus); } ISPLOCK_2_CAMLOCK(isp); ccb->ccb_h.status = CAM_REQ_CMP; xpt_done(ccb); break; case XPT_GET_TRAN_SETTINGS: cts = &ccb->cts; tgt = cts->ccb_h.target_id; CAMLOCK_2_ISPLOCK(isp); if (IS_FC(isp)) { #ifndef CAM_NEW_TRAN_CODE /* * a lot of normal SCSI things don't make sense. */ cts->flags = CCB_TRANS_TAG_ENB | CCB_TRANS_DISC_ENB; cts->valid = CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID; /* * How do you measure the width of a high * speed serial bus? Well, in bytes. * * Offset and period make no sense, though, so we set * (above) a 'base' transfer speed to be gigabit. */ cts->bus_width = MSG_EXT_WDTR_BUS_8_BIT; #else fcparam *fcp = isp->isp_param; struct ccb_trans_settings_fc *fc = &cts->xport_specific.fc; cts->protocol = PROTO_SCSI; cts->protocol_version = SCSI_REV_2; cts->transport = XPORT_FC; cts->transport_version = 0; fc->valid = CTS_FC_VALID_SPEED; if (fcp->isp_gbspeed == 2) fc->bitrate = 200000; else fc->bitrate = 100000; if (tgt > 0 && tgt < MAX_FC_TARG) { struct lportdb *lp = &fcp->portdb[tgt]; fc->wwnn = lp->node_wwn; fc->wwpn = lp->port_wwn; fc->port = lp->portid; fc->valid |= CTS_FC_VALID_WWNN | CTS_FC_VALID_WWPN | CTS_FC_VALID_PORT; } #endif } else { #ifdef CAM_NEW_TRAN_CODE struct ccb_trans_settings_scsi *scsi = &cts->proto_specific.scsi; struct ccb_trans_settings_spi *spi = &cts->xport_specific.spi; #endif sdparam *sdp = isp->isp_param; int bus = cam_sim_bus(xpt_path_sim(cts->ccb_h.path)); u_int16_t dval, pval, oval; sdp += bus; if (IS_CURRENT_SETTINGS(cts)) { sdp->isp_devparam[tgt].dev_refresh = 1; isp->isp_update |= (1 << bus); (void) isp_control(isp, ISPCTL_UPDATE_PARAMS, NULL); dval = sdp->isp_devparam[tgt].actv_flags; oval = sdp->isp_devparam[tgt].actv_offset; pval = sdp->isp_devparam[tgt].actv_period; } else { dval = sdp->isp_devparam[tgt].nvrm_flags; oval = sdp->isp_devparam[tgt].nvrm_offset; pval = sdp->isp_devparam[tgt].nvrm_period; } #ifndef CAM_NEW_TRAN_CODE cts->flags &= ~(CCB_TRANS_DISC_ENB|CCB_TRANS_TAG_ENB); if (dval & DPARM_DISC) { cts->flags |= CCB_TRANS_DISC_ENB; } if (dval & DPARM_TQING) { cts->flags |= CCB_TRANS_TAG_ENB; } if (dval & DPARM_WIDE) { cts->bus_width = MSG_EXT_WDTR_BUS_16_BIT; } else { cts->bus_width = MSG_EXT_WDTR_BUS_8_BIT; } cts->valid = CCB_TRANS_BUS_WIDTH_VALID | CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID; if ((dval & DPARM_SYNC) && oval != 0) { cts->sync_period = pval; cts->sync_offset = oval; cts->valid |= CCB_TRANS_SYNC_RATE_VALID | CCB_TRANS_SYNC_OFFSET_VALID; } #else cts->protocol = PROTO_SCSI; cts->protocol_version = SCSI_REV_2; cts->transport = XPORT_SPI; cts->transport_version = 2; scsi->flags &= ~CTS_SCSI_FLAGS_TAG_ENB; spi->flags &= ~CTS_SPI_FLAGS_DISC_ENB; if (dval & DPARM_DISC) { spi->flags |= CTS_SPI_FLAGS_DISC_ENB; } if (dval & DPARM_TQING) { scsi->flags |= CTS_SCSI_FLAGS_TAG_ENB; } if ((dval & DPARM_SYNC) && oval && pval) { spi->sync_offset = oval; spi->sync_period = pval; spi->valid |= CTS_SPI_VALID_SYNC_OFFSET; spi->valid |= CTS_SPI_VALID_SYNC_RATE; } spi->valid |= CTS_SPI_VALID_BUS_WIDTH; if (dval & DPARM_WIDE) { spi->bus_width = MSG_EXT_WDTR_BUS_16_BIT; } else { spi->bus_width = MSG_EXT_WDTR_BUS_8_BIT; } if (cts->ccb_h.target_lun != CAM_LUN_WILDCARD) { scsi->valid = CTS_SCSI_VALID_TQ; spi->valid |= CTS_SPI_VALID_DISC; } else { scsi->valid = 0; } #endif isp_prt(isp, ISP_LOGDEBUG0, "GET %s bus %d targ %d to flags %x off %x per %x", IS_CURRENT_SETTINGS(cts)? "ACTIVE" : "NVRAM", bus, tgt, dval, oval, pval); } ISPLOCK_2_CAMLOCK(isp); ccb->ccb_h.status = CAM_REQ_CMP; xpt_done(ccb); break; case XPT_CALC_GEOMETRY: { struct ccb_calc_geometry *ccg; u_int32_t secs_per_cylinder; u_int32_t size_mb; ccg = &ccb->ccg; if (ccg->block_size == 0) { isp_prt(isp, ISP_LOGERR, "%d.%d XPT_CALC_GEOMETRY block size 0?", ccg->ccb_h.target_id, ccg->ccb_h.target_lun); ccb->ccb_h.status = CAM_REQ_INVALID; xpt_done(ccb); break; } size_mb = ccg->volume_size /((1024L * 1024L) / ccg->block_size); if (size_mb > 1024) { ccg->heads = 255; ccg->secs_per_track = 63; } else { ccg->heads = 64; ccg->secs_per_track = 32; } secs_per_cylinder = ccg->heads * ccg->secs_per_track; ccg->cylinders = ccg->volume_size / secs_per_cylinder; ccb->ccb_h.status = CAM_REQ_CMP; xpt_done(ccb); break; } case XPT_RESET_BUS: /* Reset the specified bus */ bus = cam_sim_bus(sim); CAMLOCK_2_ISPLOCK(isp); error = isp_control(isp, ISPCTL_RESET_BUS, &bus); ISPLOCK_2_CAMLOCK(isp); if (error) ccb->ccb_h.status = CAM_REQ_CMP_ERR; else { if (cam_sim_bus(sim) && isp->isp_path2 != NULL) xpt_async(AC_BUS_RESET, isp->isp_path2, NULL); else if (isp->isp_path != NULL) xpt_async(AC_BUS_RESET, isp->isp_path, NULL); ccb->ccb_h.status = CAM_REQ_CMP; } xpt_done(ccb); break; case XPT_TERM_IO: /* Terminate the I/O process */ ccb->ccb_h.status = CAM_REQ_INVALID; xpt_done(ccb); break; case XPT_PATH_INQ: /* Path routing inquiry */ { struct ccb_pathinq *cpi = &ccb->cpi; cpi->version_num = 1; #ifdef ISP_TARGET_MODE cpi->target_sprt = PIT_PROCESSOR | PIT_DISCONNECT | PIT_TERM_IO; #else cpi->target_sprt = 0; #endif cpi->hba_eng_cnt = 0; cpi->max_target = ISP_MAX_TARGETS(isp) - 1; cpi->max_lun = ISP_MAX_LUNS(isp) - 1; cpi->bus_id = cam_sim_bus(sim); if (IS_FC(isp)) { cpi->hba_misc = PIM_NOBUSRESET; /* * Because our loop ID can shift from time to time, * make our initiator ID out of range of our bus. */ cpi->initiator_id = cpi->max_target + 1; /* * Set base transfer capabilities for Fibre Channel. * Technically not correct because we don't know * what media we're running on top of- but we'll * look good if we always say 100MB/s. */ if (FCPARAM(isp)->isp_gbspeed == 2) cpi->base_transfer_speed = 200000; else cpi->base_transfer_speed = 100000; cpi->hba_inquiry = PI_TAG_ABLE; #ifdef CAM_NEW_TRAN_CODE cpi->transport = XPORT_FC; cpi->transport_version = 0; /* WHAT'S THIS FOR? */ #endif } else { sdparam *sdp = isp->isp_param; sdp += cam_sim_bus(xpt_path_sim(cpi->ccb_h.path)); cpi->hba_inquiry = PI_SDTR_ABLE|PI_TAG_ABLE|PI_WIDE_16; cpi->hba_misc = 0; cpi->initiator_id = sdp->isp_initiator_id; cpi->base_transfer_speed = 3300; #ifdef CAM_NEW_TRAN_CODE cpi->transport = XPORT_SPI; cpi->transport_version = 2; /* WHAT'S THIS FOR? */ #endif } #ifdef CAM_NEW_TRAN_CODE cpi->protocol = PROTO_SCSI; cpi->protocol_version = SCSI_REV_2; #endif strncpy(cpi->sim_vid, "FreeBSD", SIM_IDLEN); strncpy(cpi->hba_vid, "Qlogic", HBA_IDLEN); strncpy(cpi->dev_name, cam_sim_name(sim), DEV_IDLEN); cpi->unit_number = cam_sim_unit(sim); cpi->ccb_h.status = CAM_REQ_CMP; xpt_done(ccb); break; } default: ccb->ccb_h.status = CAM_REQ_INVALID; xpt_done(ccb); break; } } #define ISPDDB (CAM_DEBUG_INFO|CAM_DEBUG_TRACE|CAM_DEBUG_CDB) void isp_done(struct ccb_scsiio *sccb) { struct ispsoftc *isp = XS_ISP(sccb); if (XS_NOERR(sccb)) XS_SETERR(sccb, CAM_REQ_CMP); if ((sccb->ccb_h.status & CAM_STATUS_MASK) == CAM_REQ_CMP && (sccb->scsi_status != SCSI_STATUS_OK)) { sccb->ccb_h.status &= ~CAM_STATUS_MASK; if ((sccb->scsi_status == SCSI_STATUS_CHECK_COND) && (sccb->ccb_h.status & CAM_AUTOSNS_VALID) == 0) { sccb->ccb_h.status |= CAM_AUTOSENSE_FAIL; } else { sccb->ccb_h.status |= CAM_SCSI_STATUS_ERROR; } } sccb->ccb_h.status &= ~CAM_SIM_QUEUED; if ((sccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) { if ((sccb->ccb_h.status & CAM_DEV_QFRZN) == 0) { sccb->ccb_h.status |= CAM_DEV_QFRZN; xpt_freeze_devq(sccb->ccb_h.path, 1); if (sccb->scsi_status != SCSI_STATUS_OK) isp_prt(isp, ISP_LOGDEBUG2, "freeze devq %d.%d %x %x", sccb->ccb_h.target_id, sccb->ccb_h.target_lun, sccb->ccb_h.status, sccb->scsi_status); } } /* * If we were frozen waiting resources, clear that we were frozen * waiting for resources. If we are no longer frozen, and the devq * isn't frozen, mark the completing CCB to have the XPT layer * release the simq. */ if (isp->isp_osinfo.simqfrozen & SIMQFRZ_RESOURCE) { isp->isp_osinfo.simqfrozen &= ~SIMQFRZ_RESOURCE; if (isp->isp_osinfo.simqfrozen == 0) { if ((sccb->ccb_h.status & CAM_DEV_QFRZN) == 0) { isp_prt(isp, ISP_LOGDEBUG2, "isp_done->relsimq"); sccb->ccb_h.status |= CAM_RELEASE_SIMQ; } else { isp_prt(isp, ISP_LOGDEBUG2, "isp_done->devq frozen"); } } else { isp_prt(isp, ISP_LOGDEBUG2, "isp_done -> simqfrozen = %x", isp->isp_osinfo.simqfrozen); } } if ((CAM_DEBUGGED(sccb->ccb_h.path, ISPDDB)) && (sccb->ccb_h.status & CAM_STATUS_MASK) != CAM_REQ_CMP) { xpt_print_path(sccb->ccb_h.path); isp_prt(isp, ISP_LOGINFO, "cam completion status 0x%x", sccb->ccb_h.status); } XS_CMD_S_DONE(sccb); if (XS_CMD_WDOG_P(sccb) == 0) { untimeout(isp_watchdog, (caddr_t)sccb, sccb->ccb_h.timeout_ch); if (XS_CMD_GRACE_P(sccb)) { isp_prt(isp, ISP_LOGDEBUG2, "finished command on borrowed time"); } XS_CMD_S_CLEAR(sccb); ISPLOCK_2_CAMLOCK(isp); xpt_done((union ccb *) sccb); CAMLOCK_2_ISPLOCK(isp); } } int isp_async(struct ispsoftc *isp, ispasync_t cmd, void *arg) { int bus, rv = 0; switch (cmd) { case ISPASYNC_NEW_TGT_PARAMS: { #ifdef CAM_NEW_TRAN_CODE struct ccb_trans_settings_scsi *scsi; struct ccb_trans_settings_spi *spi; #endif int flags, tgt; sdparam *sdp = isp->isp_param; struct ccb_trans_settings cts; struct cam_path *tmppath; bzero(&cts, sizeof (struct ccb_trans_settings)); tgt = *((int *)arg); bus = (tgt >> 16) & 0xffff; tgt &= 0xffff; sdp += bus; ISPLOCK_2_CAMLOCK(isp); if (xpt_create_path(&tmppath, NULL, cam_sim_path(bus? isp->isp_sim2 : isp->isp_sim), tgt, CAM_LUN_WILDCARD) != CAM_REQ_CMP) { CAMLOCK_2_ISPLOCK(isp); isp_prt(isp, ISP_LOGWARN, "isp_async cannot make temp path for %d.%d", tgt, bus); rv = -1; break; } CAMLOCK_2_ISPLOCK(isp); flags = sdp->isp_devparam[tgt].actv_flags; #ifdef CAM_NEW_TRAN_CODE cts.type = CTS_TYPE_CURRENT_SETTINGS; cts.protocol = PROTO_SCSI; cts.transport = XPORT_SPI; scsi = &cts.proto_specific.scsi; spi = &cts.xport_specific.spi; if (flags & DPARM_TQING) { scsi->valid |= CTS_SCSI_VALID_TQ; scsi->flags |= CTS_SCSI_FLAGS_TAG_ENB; spi->flags |= CTS_SPI_FLAGS_TAG_ENB; } if (flags & DPARM_DISC) { spi->valid |= CTS_SPI_VALID_DISC; spi->flags |= CTS_SPI_FLAGS_DISC_ENB; } spi->flags |= CTS_SPI_VALID_BUS_WIDTH; if (flags & DPARM_WIDE) { spi->bus_width = MSG_EXT_WDTR_BUS_16_BIT; } else { spi->bus_width = MSG_EXT_WDTR_BUS_8_BIT; } if (flags & DPARM_SYNC) { spi->valid |= CTS_SPI_VALID_SYNC_RATE; spi->valid |= CTS_SPI_VALID_SYNC_OFFSET; spi->sync_period = sdp->isp_devparam[tgt].actv_period; spi->sync_offset = sdp->isp_devparam[tgt].actv_offset; } #else cts.flags = CCB_TRANS_CURRENT_SETTINGS; cts.valid = CCB_TRANS_DISC_VALID | CCB_TRANS_TQ_VALID; if (flags & DPARM_DISC) { cts.flags |= CCB_TRANS_DISC_ENB; } if (flags & DPARM_TQING) { cts.flags |= CCB_TRANS_TAG_ENB; } cts.valid |= CCB_TRANS_BUS_WIDTH_VALID; cts.bus_width = (flags & DPARM_WIDE)? MSG_EXT_WDTR_BUS_8_BIT : MSG_EXT_WDTR_BUS_16_BIT; cts.sync_period = sdp->isp_devparam[tgt].actv_period; cts.sync_offset = sdp->isp_devparam[tgt].actv_offset; if (flags & DPARM_SYNC) { cts.valid |= CCB_TRANS_SYNC_RATE_VALID | CCB_TRANS_SYNC_OFFSET_VALID; } #endif isp_prt(isp, ISP_LOGDEBUG2, "NEW_TGT_PARAMS bus %d tgt %d period %x offset %x flags %x", bus, tgt, sdp->isp_devparam[tgt].actv_period, sdp->isp_devparam[tgt].actv_offset, flags); xpt_setup_ccb(&cts.ccb_h, tmppath, 1); ISPLOCK_2_CAMLOCK(isp); xpt_async(AC_TRANSFER_NEG, tmppath, &cts); xpt_free_path(tmppath); CAMLOCK_2_ISPLOCK(isp); break; } case ISPASYNC_BUS_RESET: bus = *((int *)arg); isp_prt(isp, ISP_LOGINFO, "SCSI bus reset on bus %d detected", bus); if (bus > 0 && isp->isp_path2) { ISPLOCK_2_CAMLOCK(isp); xpt_async(AC_BUS_RESET, isp->isp_path2, NULL); CAMLOCK_2_ISPLOCK(isp); } else if (isp->isp_path) { ISPLOCK_2_CAMLOCK(isp); xpt_async(AC_BUS_RESET, isp->isp_path, NULL); CAMLOCK_2_ISPLOCK(isp); } break; case ISPASYNC_LIP: if (isp->isp_path) { if (isp->isp_osinfo.simqfrozen == 0) { isp_prt(isp, ISP_LOGDEBUG0, "LIP freeze simq"); ISPLOCK_2_CAMLOCK(isp); xpt_freeze_simq(isp->isp_sim, 1); CAMLOCK_2_ISPLOCK(isp); } isp->isp_osinfo.simqfrozen |= SIMQFRZ_LOOPDOWN; } isp_prt(isp, ISP_LOGINFO, "LIP Received"); break; case ISPASYNC_LOOP_RESET: if (isp->isp_path) { if (isp->isp_osinfo.simqfrozen == 0) { isp_prt(isp, ISP_LOGDEBUG0, "Loop Reset freeze simq"); ISPLOCK_2_CAMLOCK(isp); xpt_freeze_simq(isp->isp_sim, 1); CAMLOCK_2_ISPLOCK(isp); } isp->isp_osinfo.simqfrozen |= SIMQFRZ_LOOPDOWN; } isp_prt(isp, ISP_LOGINFO, "Loop Reset Received"); break; case ISPASYNC_LOOP_DOWN: if (isp->isp_path) { if (isp->isp_osinfo.simqfrozen == 0) { isp_prt(isp, ISP_LOGDEBUG0, "loop down freeze simq"); ISPLOCK_2_CAMLOCK(isp); xpt_freeze_simq(isp->isp_sim, 1); CAMLOCK_2_ISPLOCK(isp); } isp->isp_osinfo.simqfrozen |= SIMQFRZ_LOOPDOWN; } isp_prt(isp, ISP_LOGINFO, "Loop DOWN"); break; case ISPASYNC_LOOP_UP: /* * Now we just note that Loop has come up. We don't * actually do anything because we're waiting for a * Change Notify before activating the FC cleanup * thread to look at the state of the loop again. */ isp_prt(isp, ISP_LOGINFO, "Loop UP"); break; case ISPASYNC_PROMENADE: { struct cam_path *tmppath; const char *fmt = "Target %d (Loop 0x%x) Port ID 0x%x " "(role %s) %s\n Port WWN 0x%08x%08x\n Node WWN 0x%08x%08x"; static const char *roles[4] = { "(none)", "Target", "Initiator", "Target/Initiator" }; fcparam *fcp = isp->isp_param; int tgt = *((int *) arg); int is_tgt_mask = (SVC3_TGT_ROLE >> SVC3_ROLE_SHIFT); struct lportdb *lp = &fcp->portdb[tgt]; isp_prt(isp, ISP_LOGINFO, fmt, tgt, lp->loopid, lp->portid, roles[lp->roles & 0x3], (lp->valid)? "Arrived" : "Departed", (u_int32_t) (lp->port_wwn >> 32), (u_int32_t) (lp->port_wwn & 0xffffffffLL), (u_int32_t) (lp->node_wwn >> 32), (u_int32_t) (lp->node_wwn & 0xffffffffLL)); ISPLOCK_2_CAMLOCK(isp); if (xpt_create_path(&tmppath, NULL, cam_sim_path(isp->isp_sim), (target_id_t)tgt, CAM_LUN_WILDCARD) != CAM_REQ_CMP) { CAMLOCK_2_ISPLOCK(isp); break; } /* * Policy: only announce targets. */ if (lp->roles & is_tgt_mask) { if (lp->valid) { xpt_async(AC_FOUND_DEVICE, tmppath, NULL); } else { xpt_async(AC_LOST_DEVICE, tmppath, NULL); } } xpt_free_path(tmppath); CAMLOCK_2_ISPLOCK(isp); break; } case ISPASYNC_CHANGE_NOTIFY: if (arg == ISPASYNC_CHANGE_PDB) { isp_prt(isp, ISP_LOGINFO, "Port Database Changed"); } else if (arg == ISPASYNC_CHANGE_SNS) { isp_prt(isp, ISP_LOGINFO, "Name Server Database Changed"); } cv_signal(&isp->isp_osinfo.kthread_cv); break; case ISPASYNC_FABRIC_DEV: { int target, lrange; struct lportdb *lp = NULL; char *pt; sns_ganrsp_t *resp = (sns_ganrsp_t *) arg; u_int32_t portid; u_int64_t wwpn, wwnn; fcparam *fcp = isp->isp_param; portid = (((u_int32_t) resp->snscb_port_id[0]) << 16) | (((u_int32_t) resp->snscb_port_id[1]) << 8) | (((u_int32_t) resp->snscb_port_id[2])); wwpn = (((u_int64_t)resp->snscb_portname[0]) << 56) | (((u_int64_t)resp->snscb_portname[1]) << 48) | (((u_int64_t)resp->snscb_portname[2]) << 40) | (((u_int64_t)resp->snscb_portname[3]) << 32) | (((u_int64_t)resp->snscb_portname[4]) << 24) | (((u_int64_t)resp->snscb_portname[5]) << 16) | (((u_int64_t)resp->snscb_portname[6]) << 8) | (((u_int64_t)resp->snscb_portname[7])); wwnn = (((u_int64_t)resp->snscb_nodename[0]) << 56) | (((u_int64_t)resp->snscb_nodename[1]) << 48) | (((u_int64_t)resp->snscb_nodename[2]) << 40) | (((u_int64_t)resp->snscb_nodename[3]) << 32) | (((u_int64_t)resp->snscb_nodename[4]) << 24) | (((u_int64_t)resp->snscb_nodename[5]) << 16) | (((u_int64_t)resp->snscb_nodename[6]) << 8) | (((u_int64_t)resp->snscb_nodename[7])); if (portid == 0 || wwpn == 0) { break; } switch (resp->snscb_port_type) { case 1: pt = " N_Port"; break; case 2: pt = " NL_Port"; break; case 3: pt = "F/NL_Port"; break; case 0x7f: pt = " Nx_Port"; break; case 0x81: pt = " F_port"; break; case 0x82: pt = " FL_Port"; break; case 0x84: pt = " E_port"; break; default: pt = "?"; break; } isp_prt(isp, ISP_LOGINFO, "%s @ 0x%x, Node 0x%08x%08x Port %08x%08x", pt, portid, ((u_int32_t) (wwnn >> 32)), ((u_int32_t) wwnn), ((u_int32_t) (wwpn >> 32)), ((u_int32_t) wwpn)); /* * We're only interested in SCSI_FCP types (for now) */ if ((resp->snscb_fc4_types[2] & 1) == 0) { break; } if (fcp->isp_topo != TOPO_F_PORT) lrange = FC_SNS_ID+1; else lrange = 0; /* * Is it already in our list? */ for (target = lrange; target < MAX_FC_TARG; target++) { if (target >= FL_PORT_ID && target <= FC_SNS_ID) { continue; } lp = &fcp->portdb[target]; if (lp->port_wwn == wwpn && lp->node_wwn == wwnn) { lp->fabric_dev = 1; break; } } if (target < MAX_FC_TARG) { break; } for (target = lrange; target < MAX_FC_TARG; target++) { if (target >= FL_PORT_ID && target <= FC_SNS_ID) { continue; } lp = &fcp->portdb[target]; if (lp->port_wwn == 0) { break; } } if (target == MAX_FC_TARG) { isp_prt(isp, ISP_LOGWARN, "no more space for fabric devices"); break; } lp->node_wwn = wwnn; lp->port_wwn = wwpn; lp->portid = portid; lp->fabric_dev = 1; break; } #ifdef ISP_TARGET_MODE case ISPASYNC_TARGET_MESSAGE: { tmd_msg_t *mp = arg; isp_prt(isp, ISP_LOGALL, "bus %d iid %d tgt %d lun %d ttype %x tval %x msg[0]=%x", mp->nt_bus, (int) mp->nt_iid, (int) mp->nt_tgt, (int) mp->nt_lun, mp->nt_tagtype, mp->nt_tagval, mp->nt_msg[0]); break; } case ISPASYNC_TARGET_EVENT: { tmd_event_t *ep = arg; isp_prt(isp, ISP_LOGALL, "bus %d event code 0x%x", ep->ev_bus, ep->ev_event); break; } case ISPASYNC_TARGET_ACTION: switch (((isphdr_t *)arg)->rqs_entry_type) { default: isp_prt(isp, ISP_LOGWARN, "event 0x%x for unhandled target action", ((isphdr_t *)arg)->rqs_entry_type); break; case RQSTYPE_ATIO: rv = isp_handle_platform_atio(isp, (at_entry_t *) arg); break; case RQSTYPE_ATIO2: rv = isp_handle_platform_atio2(isp, (at2_entry_t *)arg); break; case RQSTYPE_CTIO2: case RQSTYPE_CTIO: rv = isp_handle_platform_ctio(isp, arg); break; case RQSTYPE_ENABLE_LUN: case RQSTYPE_MODIFY_LUN: if (IS_DUALBUS(isp)) { bus = GET_BUS_VAL(((lun_entry_t *)arg)->le_rsvd); } else { bus = 0; } isp_cv_signal_rqe(isp, bus, ((lun_entry_t *)arg)->le_status); break; } break; #endif case ISPASYNC_FW_CRASH: { u_int16_t mbox1, mbox6; mbox1 = ISP_READ(isp, OUTMAILBOX1); if (IS_DUALBUS(isp)) { mbox6 = ISP_READ(isp, OUTMAILBOX6); } else { mbox6 = 0; } isp_prt(isp, ISP_LOGERR, "Internal Firmware on bus %d Error @ RISC Address 0x%x", mbox6, mbox1); isp_reinit(isp); break; } case ISPASYNC_UNHANDLED_RESPONSE: break; default: isp_prt(isp, ISP_LOGERR, "unknown isp_async event %d", cmd); break; } return (rv); } /* * Locks are held before coming here. */ void isp_uninit(struct ispsoftc *isp) { ISP_WRITE(isp, HCCR, HCCR_CMD_RESET); DISABLE_INTS(isp); } void isp_prt(struct ispsoftc *isp, int level, const char *fmt, ...) { va_list ap; if (level != ISP_LOGALL && (level & isp->isp_dblev) == 0) { return; } printf("%s: ", device_get_nameunit(isp->isp_dev)); va_start(ap, fmt); vprintf(fmt, ap); va_end(ap); printf("\n"); }