1/*
2 * Core routines and tables shareable across OS platforms.
3 *
4 * Copyright (c) 1994-2002 Justin T. Gibbs.
5 * Copyright (c) 2000-2002 Adaptec Inc.
6 * All rights reserved.
7 *
8 * Redistribution and use in source and binary forms, with or without
9 * modification, are permitted provided that the following conditions
10 * are met:
11 * 1. Redistributions of source code must retain the above copyright
12 *    notice, this list of conditions, and the following disclaimer,
13 *    without modification.
14 * 2. Redistributions in binary form must reproduce at minimum a disclaimer
15 *    substantially similar to the "NO WARRANTY" disclaimer below
16 *    ("Disclaimer") and any redistribution must be conditioned upon
17 *    including a substantially similar Disclaimer requirement for further
18 *    binary redistribution.
19 * 3. Neither the names of the above-listed copyright holders nor the names
20 *    of any contributors may be used to endorse or promote products derived
21 *    from this software without specific prior written permission.
22 *
23 * Alternatively, this software may be distributed under the terms of the
24 * GNU General Public License ("GPL") version 2 as published by the Free
25 * Software Foundation.
26 *
27 * NO WARRANTY
28 * THIS SOFTWARE IS PROVIDED BY THE COPYRIGHT HOLDERS AND CONTRIBUTORS
29 * "AS IS" AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT
30 * LIMITED TO, THE IMPLIED WARRANTIES OF MERCHANTIBILITY AND FITNESS FOR
31 * A PARTICULAR PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE COPYRIGHT
32 * HOLDERS OR CONTRIBUTORS BE LIABLE FOR SPECIAL, EXEMPLARY, OR CONSEQUENTIAL
33 * DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF SUBSTITUTE GOODS
34 * OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS INTERRUPTION)
35 * HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN CONTRACT,
36 * STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) ARISING
37 * IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE
38 * POSSIBILITY OF SUCH DAMAGES.
39 *
40 * $Id: //depot/aic7xxx/aic7xxx/aic7xxx.c#155 $
41 */
42
43#ifdef __linux__
44#include "aic7xxx_osm.h"
45#include "aic7xxx_inline.h"
46#include "aicasm/aicasm_insformat.h"
47#else
48#include <dev/aic7xxx/aic7xxx_osm.h>
49#include <dev/aic7xxx/aic7xxx_inline.h>
50#include <dev/aic7xxx/aicasm/aicasm_insformat.h>
51#endif
52
53/***************************** Lookup Tables **********************************/
54static const char *const ahc_chip_names[] = {
55	"NONE",
56	"aic7770",
57	"aic7850",
58	"aic7855",
59	"aic7859",
60	"aic7860",
61	"aic7870",
62	"aic7880",
63	"aic7895",
64	"aic7895C",
65	"aic7890/91",
66	"aic7896/97",
67	"aic7892",
68	"aic7899"
69};
70static const u_int num_chip_names = ARRAY_SIZE(ahc_chip_names);
71
72/*
73 * Hardware error codes.
74 */
75struct ahc_hard_error_entry {
76        uint8_t errno;
77	const char *errmesg;
78};
79
80static const struct ahc_hard_error_entry ahc_hard_errors[] = {
81	{ ILLHADDR,	"Illegal Host Access" },
82	{ ILLSADDR,	"Illegal Sequencer Address referrenced" },
83	{ ILLOPCODE,	"Illegal Opcode in sequencer program" },
84	{ SQPARERR,	"Sequencer Parity Error" },
85	{ DPARERR,	"Data-path Parity Error" },
86	{ MPARERR,	"Scratch or SCB Memory Parity Error" },
87	{ PCIERRSTAT,	"PCI Error detected" },
88	{ CIOPARERR,	"CIOBUS Parity Error" },
89};
90static const u_int num_errors = ARRAY_SIZE(ahc_hard_errors);
91
92static const struct ahc_phase_table_entry ahc_phase_table[] =
93{
94	{ P_DATAOUT,	MSG_NOOP,		"in Data-out phase"	},
95	{ P_DATAIN,	MSG_INITIATOR_DET_ERR,	"in Data-in phase"	},
96	{ P_DATAOUT_DT,	MSG_NOOP,		"in DT Data-out phase"	},
97	{ P_DATAIN_DT,	MSG_INITIATOR_DET_ERR,	"in DT Data-in phase"	},
98	{ P_COMMAND,	MSG_NOOP,		"in Command phase"	},
99	{ P_MESGOUT,	MSG_NOOP,		"in Message-out phase"	},
100	{ P_STATUS,	MSG_INITIATOR_DET_ERR,	"in Status phase"	},
101	{ P_MESGIN,	MSG_PARITY_ERROR,	"in Message-in phase"	},
102	{ P_BUSFREE,	MSG_NOOP,		"while idle"		},
103	{ 0,		MSG_NOOP,		"in unknown phase"	}
104};
105
106/*
107 * In most cases we only wish to itterate over real phases, so
108 * exclude the last element from the count.
109 */
110static const u_int num_phases = ARRAY_SIZE(ahc_phase_table) - 1;
111
112/*
113 * Valid SCSIRATE values.  (p. 3-17)
114 * Provides a mapping of tranfer periods in ns to the proper value to
115 * stick in the scsixfer reg.
116 */
117static const struct ahc_syncrate ahc_syncrates[] =
118{
119      /* ultra2    fast/ultra  period     rate */
120	{ 0x42,      0x000,      9,      "80.0" },
121	{ 0x03,      0x000,     10,      "40.0" },
122	{ 0x04,      0x000,     11,      "33.0" },
123	{ 0x05,      0x100,     12,      "20.0" },
124	{ 0x06,      0x110,     15,      "16.0" },
125	{ 0x07,      0x120,     18,      "13.4" },
126	{ 0x08,      0x000,     25,      "10.0" },
127	{ 0x19,      0x010,     31,      "8.0"  },
128	{ 0x1a,      0x020,     37,      "6.67" },
129	{ 0x1b,      0x030,     43,      "5.7"  },
130	{ 0x1c,      0x040,     50,      "5.0"  },
131	{ 0x00,      0x050,     56,      "4.4"  },
132	{ 0x00,      0x060,     62,      "4.0"  },
133	{ 0x00,      0x070,     68,      "3.6"  },
134	{ 0x00,      0x000,      0,      NULL   }
135};
136
137/* Our Sequencer Program */
138#include "aic7xxx_seq.h"
139
140/**************************** Function Declarations ***************************/
141static void		ahc_force_renegotiation(struct ahc_softc *ahc,
142						struct ahc_devinfo *devinfo);
143static struct ahc_tmode_tstate*
144			ahc_alloc_tstate(struct ahc_softc *ahc,
145					 u_int scsi_id, char channel);
146#ifdef AHC_TARGET_MODE
147static void		ahc_free_tstate(struct ahc_softc *ahc,
148					u_int scsi_id, char channel, int force);
149#endif
150static const struct ahc_syncrate*
151			ahc_devlimited_syncrate(struct ahc_softc *ahc,
152					        struct ahc_initiator_tinfo *,
153						u_int *period,
154						u_int *ppr_options,
155						role_t role);
156static void		ahc_update_pending_scbs(struct ahc_softc *ahc);
157static void		ahc_fetch_devinfo(struct ahc_softc *ahc,
158					  struct ahc_devinfo *devinfo);
159static void		ahc_scb_devinfo(struct ahc_softc *ahc,
160					struct ahc_devinfo *devinfo,
161					struct scb *scb);
162static void		ahc_assert_atn(struct ahc_softc *ahc);
163static void		ahc_setup_initiator_msgout(struct ahc_softc *ahc,
164						   struct ahc_devinfo *devinfo,
165						   struct scb *scb);
166static void		ahc_build_transfer_msg(struct ahc_softc *ahc,
167					       struct ahc_devinfo *devinfo);
168static void		ahc_construct_sdtr(struct ahc_softc *ahc,
169					   struct ahc_devinfo *devinfo,
170					   u_int period, u_int offset);
171static void		ahc_construct_wdtr(struct ahc_softc *ahc,
172					   struct ahc_devinfo *devinfo,
173					   u_int bus_width);
174static void		ahc_construct_ppr(struct ahc_softc *ahc,
175					  struct ahc_devinfo *devinfo,
176					  u_int period, u_int offset,
177					  u_int bus_width, u_int ppr_options);
178static void		ahc_clear_msg_state(struct ahc_softc *ahc);
179static void		ahc_handle_proto_violation(struct ahc_softc *ahc);
180static void		ahc_handle_message_phase(struct ahc_softc *ahc);
181typedef enum {
182	AHCMSG_1B,
183	AHCMSG_2B,
184	AHCMSG_EXT
185} ahc_msgtype;
186static int		ahc_sent_msg(struct ahc_softc *ahc, ahc_msgtype type,
187				     u_int msgval, int full);
188static int		ahc_parse_msg(struct ahc_softc *ahc,
189				      struct ahc_devinfo *devinfo);
190static int		ahc_handle_msg_reject(struct ahc_softc *ahc,
191					      struct ahc_devinfo *devinfo);
192static void		ahc_handle_ign_wide_residue(struct ahc_softc *ahc,
193						struct ahc_devinfo *devinfo);
194static void		ahc_reinitialize_dataptrs(struct ahc_softc *ahc);
195static void		ahc_handle_devreset(struct ahc_softc *ahc,
196					    struct ahc_devinfo *devinfo,
197					    cam_status status, char *message,
198					    int verbose_level);
199#ifdef AHC_TARGET_MODE
200static void		ahc_setup_target_msgin(struct ahc_softc *ahc,
201					       struct ahc_devinfo *devinfo,
202					       struct scb *scb);
203#endif
204
205static bus_dmamap_callback_t	ahc_dmamap_cb;
206static void		ahc_build_free_scb_list(struct ahc_softc *ahc);
207static int		ahc_init_scbdata(struct ahc_softc *ahc);
208static void		ahc_fini_scbdata(struct ahc_softc *ahc);
209static void		ahc_qinfifo_requeue(struct ahc_softc *ahc,
210					    struct scb *prev_scb,
211					    struct scb *scb);
212static int		ahc_qinfifo_count(struct ahc_softc *ahc);
213static u_int		ahc_rem_scb_from_disc_list(struct ahc_softc *ahc,
214						   u_int prev, u_int scbptr);
215static void		ahc_add_curscb_to_free_list(struct ahc_softc *ahc);
216static u_int		ahc_rem_wscb(struct ahc_softc *ahc,
217				     u_int scbpos, u_int prev);
218static void		ahc_reset_current_bus(struct ahc_softc *ahc);
219#ifdef AHC_DUMP_SEQ
220static void		ahc_dumpseq(struct ahc_softc *ahc);
221#endif
222static int		ahc_loadseq(struct ahc_softc *ahc);
223static int		ahc_check_patch(struct ahc_softc *ahc,
224					const struct patch **start_patch,
225					u_int start_instr, u_int *skip_addr);
226static void		ahc_download_instr(struct ahc_softc *ahc,
227					   u_int instrptr, uint8_t *dconsts);
228#ifdef AHC_TARGET_MODE
229static void		ahc_queue_lstate_event(struct ahc_softc *ahc,
230					       struct ahc_tmode_lstate *lstate,
231					       u_int initiator_id,
232					       u_int event_type,
233					       u_int event_arg);
234static void		ahc_update_scsiid(struct ahc_softc *ahc,
235					  u_int targid_mask);
236static int		ahc_handle_target_cmd(struct ahc_softc *ahc,
237					      struct target_cmd *cmd);
238#endif
239
240static u_int		ahc_index_busy_tcl(struct ahc_softc *ahc, u_int tcl);
241static void		ahc_unbusy_tcl(struct ahc_softc *ahc, u_int tcl);
242static void		ahc_busy_tcl(struct ahc_softc *ahc,
243				     u_int tcl, u_int busyid);
244
245/************************** SCB and SCB queue management **********************/
246static void		ahc_run_untagged_queues(struct ahc_softc *ahc);
247static void		ahc_run_untagged_queue(struct ahc_softc *ahc,
248					       struct scb_tailq *queue);
249
250/****************************** Initialization ********************************/
251static void		 ahc_alloc_scbs(struct ahc_softc *ahc);
252static void		 ahc_shutdown(void *arg);
253
254/*************************** Interrupt Services *******************************/
255static void		ahc_clear_intstat(struct ahc_softc *ahc);
256static void		ahc_run_qoutfifo(struct ahc_softc *ahc);
257#ifdef AHC_TARGET_MODE
258static void		ahc_run_tqinfifo(struct ahc_softc *ahc, int paused);
259#endif
260static void		ahc_handle_brkadrint(struct ahc_softc *ahc);
261static void		ahc_handle_seqint(struct ahc_softc *ahc, u_int intstat);
262static void		ahc_handle_scsiint(struct ahc_softc *ahc,
263					   u_int intstat);
264static void		ahc_clear_critical_section(struct ahc_softc *ahc);
265
266/***************************** Error Recovery *********************************/
267static void		ahc_freeze_devq(struct ahc_softc *ahc, struct scb *scb);
268static int		ahc_abort_scbs(struct ahc_softc *ahc, int target,
269				       char channel, int lun, u_int tag,
270				       role_t role, uint32_t status);
271static void		ahc_calc_residual(struct ahc_softc *ahc,
272					  struct scb *scb);
273
274/*********************** Untagged Transaction Routines ************************/
275static inline void	ahc_freeze_untagged_queues(struct ahc_softc *ahc);
276static inline void	ahc_release_untagged_queues(struct ahc_softc *ahc);
277
278/*
279 * Block our completion routine from starting the next untagged
280 * transaction for this target or target lun.
281 */
282static inline void
283ahc_freeze_untagged_queues(struct ahc_softc *ahc)
284{
285	if ((ahc->flags & AHC_SCB_BTT) == 0)
286		ahc->untagged_queue_lock++;
287}
288
289/*
290 * Allow the next untagged transaction for this target or target lun
291 * to be executed.  We use a counting semaphore to allow the lock
292 * to be acquired recursively.  Once the count drops to zero, the
293 * transaction queues will be run.
294 */
295static inline void
296ahc_release_untagged_queues(struct ahc_softc *ahc)
297{
298	if ((ahc->flags & AHC_SCB_BTT) == 0) {
299		ahc->untagged_queue_lock--;
300		if (ahc->untagged_queue_lock == 0)
301			ahc_run_untagged_queues(ahc);
302	}
303}
304
305/************************* Sequencer Execution Control ************************/
306/*
307 * Work around any chip bugs related to halting sequencer execution.
308 * On Ultra2 controllers, we must clear the CIOBUS stretch signal by
309 * reading a register that will set this signal and deassert it.
310 * Without this workaround, if the chip is paused, by an interrupt or
311 * manual pause while accessing scb ram, accesses to certain registers
312 * will hang the system (infinite pci retries).
313 */
314static void
315ahc_pause_bug_fix(struct ahc_softc *ahc)
316{
317	if ((ahc->features & AHC_ULTRA2) != 0)
318		(void)ahc_inb(ahc, CCSCBCTL);
319}
320
321/*
322 * Determine whether the sequencer has halted code execution.
323 * Returns non-zero status if the sequencer is stopped.
324 */
325int
326ahc_is_paused(struct ahc_softc *ahc)
327{
328	return ((ahc_inb(ahc, HCNTRL) & PAUSE) != 0);
329}
330
331/*
332 * Request that the sequencer stop and wait, indefinitely, for it
333 * to stop.  The sequencer will only acknowledge that it is paused
334 * once it has reached an instruction boundary and PAUSEDIS is
335 * cleared in the SEQCTL register.  The sequencer may use PAUSEDIS
336 * for critical sections.
337 */
338void
339ahc_pause(struct ahc_softc *ahc)
340{
341	ahc_outb(ahc, HCNTRL, ahc->pause);
342
343	/*
344	 * Since the sequencer can disable pausing in a critical section, we
345	 * must loop until it actually stops.
346	 */
347	while (ahc_is_paused(ahc) == 0)
348		;
349
350	ahc_pause_bug_fix(ahc);
351}
352
353/*
354 * Allow the sequencer to continue program execution.
355 * We check here to ensure that no additional interrupt
356 * sources that would cause the sequencer to halt have been
357 * asserted.  If, for example, a SCSI bus reset is detected
358 * while we are fielding a different, pausing, interrupt type,
359 * we don't want to release the sequencer before going back
360 * into our interrupt handler and dealing with this new
361 * condition.
362 */
363void
364ahc_unpause(struct ahc_softc *ahc)
365{
366	if ((ahc_inb(ahc, INTSTAT) & (SCSIINT | SEQINT | BRKADRINT)) == 0)
367		ahc_outb(ahc, HCNTRL, ahc->unpause);
368}
369
370/************************** Memory mapping routines ***************************/
371static struct ahc_dma_seg *
372ahc_sg_bus_to_virt(struct scb *scb, uint32_t sg_busaddr)
373{
374	int sg_index;
375
376	sg_index = (sg_busaddr - scb->sg_list_phys)/sizeof(struct ahc_dma_seg);
377	/* sg_list_phys points to entry 1, not 0 */
378	sg_index++;
379
380	return (&scb->sg_list[sg_index]);
381}
382
383static uint32_t
384ahc_sg_virt_to_bus(struct scb *scb, struct ahc_dma_seg *sg)
385{
386	int sg_index;
387
388	/* sg_list_phys points to entry 1, not 0 */
389	sg_index = sg - &scb->sg_list[1];
390
391	return (scb->sg_list_phys + (sg_index * sizeof(*scb->sg_list)));
392}
393
394static uint32_t
395ahc_hscb_busaddr(struct ahc_softc *ahc, u_int index)
396{
397	return (ahc->scb_data->hscb_busaddr
398		+ (sizeof(struct hardware_scb) * index));
399}
400
401static void
402ahc_sync_scb(struct ahc_softc *ahc, struct scb *scb, int op)
403{
404	ahc_dmamap_sync(ahc, ahc->scb_data->hscb_dmat,
405			ahc->scb_data->hscb_dmamap,
406			/*offset*/(scb->hscb - ahc->hscbs) * sizeof(*scb->hscb),
407			/*len*/sizeof(*scb->hscb), op);
408}
409
410void
411ahc_sync_sglist(struct ahc_softc *ahc, struct scb *scb, int op)
412{
413	if (scb->sg_count == 0)
414		return;
415
416	ahc_dmamap_sync(ahc, ahc->scb_data->sg_dmat, scb->sg_map->sg_dmamap,
417			/*offset*/(scb->sg_list - scb->sg_map->sg_vaddr)
418				* sizeof(struct ahc_dma_seg),
419			/*len*/sizeof(struct ahc_dma_seg) * scb->sg_count, op);
420}
421
422#ifdef AHC_TARGET_MODE
423static uint32_t
424ahc_targetcmd_offset(struct ahc_softc *ahc, u_int index)
425{
426	return (((uint8_t *)&ahc->targetcmds[index]) - ahc->qoutfifo);
427}
428#endif
429
430/*********************** Miscellaneous Support Functions ***********************/
431/*
432 * Determine whether the sequencer reported a residual
433 * for this SCB/transaction.
434 */
435static void
436ahc_update_residual(struct ahc_softc *ahc, struct scb *scb)
437{
438	uint32_t sgptr;
439
440	sgptr = ahc_le32toh(scb->hscb->sgptr);
441	if ((sgptr & SG_RESID_VALID) != 0)
442		ahc_calc_residual(ahc, scb);
443}
444
445/*
446 * Return pointers to the transfer negotiation information
447 * for the specified our_id/remote_id pair.
448 */
449struct ahc_initiator_tinfo *
450ahc_fetch_transinfo(struct ahc_softc *ahc, char channel, u_int our_id,
451		    u_int remote_id, struct ahc_tmode_tstate **tstate)
452{
453	/*
454	 * Transfer data structures are stored from the perspective
455	 * of the target role.  Since the parameters for a connection
456	 * in the initiator role to a given target are the same as
457	 * when the roles are reversed, we pretend we are the target.
458	 */
459	if (channel == 'B')
460		our_id += 8;
461	*tstate = ahc->enabled_targets[our_id];
462	return (&(*tstate)->transinfo[remote_id]);
463}
464
465uint16_t
466ahc_inw(struct ahc_softc *ahc, u_int port)
467{
468	uint16_t r = ahc_inb(ahc, port+1) << 8;
469	return r | ahc_inb(ahc, port);
470}
471
472void
473ahc_outw(struct ahc_softc *ahc, u_int port, u_int value)
474{
475	ahc_outb(ahc, port, value & 0xFF);
476	ahc_outb(ahc, port+1, (value >> 8) & 0xFF);
477}
478
479uint32_t
480ahc_inl(struct ahc_softc *ahc, u_int port)
481{
482	return ((ahc_inb(ahc, port))
483	      | (ahc_inb(ahc, port+1) << 8)
484	      | (ahc_inb(ahc, port+2) << 16)
485	      | (ahc_inb(ahc, port+3) << 24));
486}
487
488void
489ahc_outl(struct ahc_softc *ahc, u_int port, uint32_t value)
490{
491	ahc_outb(ahc, port, (value) & 0xFF);
492	ahc_outb(ahc, port+1, ((value) >> 8) & 0xFF);
493	ahc_outb(ahc, port+2, ((value) >> 16) & 0xFF);
494	ahc_outb(ahc, port+3, ((value) >> 24) & 0xFF);
495}
496
497uint64_t
498ahc_inq(struct ahc_softc *ahc, u_int port)
499{
500	return ((ahc_inb(ahc, port))
501	      | (ahc_inb(ahc, port+1) << 8)
502	      | (ahc_inb(ahc, port+2) << 16)
503	      | (ahc_inb(ahc, port+3) << 24)
504	      | (((uint64_t)ahc_inb(ahc, port+4)) << 32)
505	      | (((uint64_t)ahc_inb(ahc, port+5)) << 40)
506	      | (((uint64_t)ahc_inb(ahc, port+6)) << 48)
507	      | (((uint64_t)ahc_inb(ahc, port+7)) << 56));
508}
509
510void
511ahc_outq(struct ahc_softc *ahc, u_int port, uint64_t value)
512{
513	ahc_outb(ahc, port, value & 0xFF);
514	ahc_outb(ahc, port+1, (value >> 8) & 0xFF);
515	ahc_outb(ahc, port+2, (value >> 16) & 0xFF);
516	ahc_outb(ahc, port+3, (value >> 24) & 0xFF);
517	ahc_outb(ahc, port+4, (value >> 32) & 0xFF);
518	ahc_outb(ahc, port+5, (value >> 40) & 0xFF);
519	ahc_outb(ahc, port+6, (value >> 48) & 0xFF);
520	ahc_outb(ahc, port+7, (value >> 56) & 0xFF);
521}
522
523/*
524 * Get a free scb. If there are none, see if we can allocate a new SCB.
525 */
526struct scb *
527ahc_get_scb(struct ahc_softc *ahc)
528{
529	struct scb *scb;
530
531	if ((scb = SLIST_FIRST(&ahc->scb_data->free_scbs)) == NULL) {
532		ahc_alloc_scbs(ahc);
533		scb = SLIST_FIRST(&ahc->scb_data->free_scbs);
534		if (scb == NULL)
535			return (NULL);
536	}
537	SLIST_REMOVE_HEAD(&ahc->scb_data->free_scbs, links.sle);
538	return (scb);
539}
540
541/*
542 * Return an SCB resource to the free list.
543 */
544void
545ahc_free_scb(struct ahc_softc *ahc, struct scb *scb)
546{
547	struct hardware_scb *hscb;
548
549	hscb = scb->hscb;
550	/* Clean up for the next user */
551	ahc->scb_data->scbindex[hscb->tag] = NULL;
552	scb->flags = SCB_FREE;
553	hscb->control = 0;
554
555	SLIST_INSERT_HEAD(&ahc->scb_data->free_scbs, scb, links.sle);
556
557	/* Notify the OSM that a resource is now available. */
558	ahc_platform_scb_free(ahc, scb);
559}
560
561struct scb *
562ahc_lookup_scb(struct ahc_softc *ahc, u_int tag)
563{
564	struct scb* scb;
565
566	scb = ahc->scb_data->scbindex[tag];
567	if (scb != NULL)
568		ahc_sync_scb(ahc, scb,
569			     BUS_DMASYNC_POSTREAD|BUS_DMASYNC_POSTWRITE);
570	return (scb);
571}
572
573static void
574ahc_swap_with_next_hscb(struct ahc_softc *ahc, struct scb *scb)
575{
576	struct hardware_scb *q_hscb;
577	u_int  saved_tag;
578
579	/*
580	 * Our queuing method is a bit tricky.  The card
581	 * knows in advance which HSCB to download, and we
582	 * can't disappoint it.  To achieve this, the next
583	 * SCB to download is saved off in ahc->next_queued_scb.
584	 * When we are called to queue "an arbitrary scb",
585	 * we copy the contents of the incoming HSCB to the one
586	 * the sequencer knows about, swap HSCB pointers and
587	 * finally assign the SCB to the tag indexed location
588	 * in the scb_array.  This makes sure that we can still
589	 * locate the correct SCB by SCB_TAG.
590	 */
591	q_hscb = ahc->next_queued_scb->hscb;
592	saved_tag = q_hscb->tag;
593	memcpy(q_hscb, scb->hscb, sizeof(*scb->hscb));
594	if ((scb->flags & SCB_CDB32_PTR) != 0) {
595		q_hscb->shared_data.cdb_ptr =
596		    ahc_htole32(ahc_hscb_busaddr(ahc, q_hscb->tag)
597			      + offsetof(struct hardware_scb, cdb32));
598	}
599	q_hscb->tag = saved_tag;
600	q_hscb->next = scb->hscb->tag;
601
602	/* Now swap HSCB pointers. */
603	ahc->next_queued_scb->hscb = scb->hscb;
604	scb->hscb = q_hscb;
605
606	/* Now define the mapping from tag to SCB in the scbindex */
607	ahc->scb_data->scbindex[scb->hscb->tag] = scb;
608}
609
610/*
611 * Tell the sequencer about a new transaction to execute.
612 */
613void
614ahc_queue_scb(struct ahc_softc *ahc, struct scb *scb)
615{
616	ahc_swap_with_next_hscb(ahc, scb);
617
618	if (scb->hscb->tag == SCB_LIST_NULL
619	 || scb->hscb->next == SCB_LIST_NULL)
620		panic("Attempt to queue invalid SCB tag %x:%x\n",
621		      scb->hscb->tag, scb->hscb->next);
622
623	/*
624	 * Setup data "oddness".
625	 */
626	scb->hscb->lun &= LID;
627	if (ahc_get_transfer_length(scb) & 0x1)
628		scb->hscb->lun |= SCB_XFERLEN_ODD;
629
630	/*
631	 * Keep a history of SCBs we've downloaded in the qinfifo.
632	 */
633	ahc->qinfifo[ahc->qinfifonext++] = scb->hscb->tag;
634
635	/*
636	 * Make sure our data is consistent from the
637	 * perspective of the adapter.
638	 */
639	ahc_sync_scb(ahc, scb, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
640
641	/* Tell the adapter about the newly queued SCB */
642	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
643		ahc_outb(ahc, HNSCB_QOFF, ahc->qinfifonext);
644	} else {
645		if ((ahc->features & AHC_AUTOPAUSE) == 0)
646			ahc_pause(ahc);
647		ahc_outb(ahc, KERNEL_QINPOS, ahc->qinfifonext);
648		if ((ahc->features & AHC_AUTOPAUSE) == 0)
649			ahc_unpause(ahc);
650	}
651}
652
653struct scsi_sense_data *
654ahc_get_sense_buf(struct ahc_softc *ahc, struct scb *scb)
655{
656	int offset;
657
658	offset = scb - ahc->scb_data->scbarray;
659	return (&ahc->scb_data->sense[offset]);
660}
661
662static uint32_t
663ahc_get_sense_bufaddr(struct ahc_softc *ahc, struct scb *scb)
664{
665	int offset;
666
667	offset = scb - ahc->scb_data->scbarray;
668	return (ahc->scb_data->sense_busaddr
669	      + (offset * sizeof(struct scsi_sense_data)));
670}
671
672/************************** Interrupt Processing ******************************/
673static void
674ahc_sync_qoutfifo(struct ahc_softc *ahc, int op)
675{
676	ahc_dmamap_sync(ahc, ahc->shared_data_dmat, ahc->shared_data_dmamap,
677			/*offset*/0, /*len*/256, op);
678}
679
680static void
681ahc_sync_tqinfifo(struct ahc_softc *ahc, int op)
682{
683#ifdef AHC_TARGET_MODE
684	if ((ahc->flags & AHC_TARGETROLE) != 0) {
685		ahc_dmamap_sync(ahc, ahc->shared_data_dmat,
686				ahc->shared_data_dmamap,
687				ahc_targetcmd_offset(ahc, 0),
688				sizeof(struct target_cmd) * AHC_TMODE_CMDS,
689				op);
690	}
691#endif
692}
693
694/*
695 * See if the firmware has posted any completed commands
696 * into our in-core command complete fifos.
697 */
698#define AHC_RUN_QOUTFIFO 0x1
699#define AHC_RUN_TQINFIFO 0x2
700static u_int
701ahc_check_cmdcmpltqueues(struct ahc_softc *ahc)
702{
703	u_int retval;
704
705	retval = 0;
706	ahc_dmamap_sync(ahc, ahc->shared_data_dmat, ahc->shared_data_dmamap,
707			/*offset*/ahc->qoutfifonext, /*len*/1,
708			BUS_DMASYNC_POSTREAD);
709	if (ahc->qoutfifo[ahc->qoutfifonext] != SCB_LIST_NULL)
710		retval |= AHC_RUN_QOUTFIFO;
711#ifdef AHC_TARGET_MODE
712	if ((ahc->flags & AHC_TARGETROLE) != 0
713	 && (ahc->flags & AHC_TQINFIFO_BLOCKED) == 0) {
714		ahc_dmamap_sync(ahc, ahc->shared_data_dmat,
715				ahc->shared_data_dmamap,
716				ahc_targetcmd_offset(ahc, ahc->tqinfifofnext),
717				/*len*/sizeof(struct target_cmd),
718				BUS_DMASYNC_POSTREAD);
719		if (ahc->targetcmds[ahc->tqinfifonext].cmd_valid != 0)
720			retval |= AHC_RUN_TQINFIFO;
721	}
722#endif
723	return (retval);
724}
725
726/*
727 * Catch an interrupt from the adapter
728 */
729int
730ahc_intr(struct ahc_softc *ahc)
731{
732	u_int	intstat;
733
734	if ((ahc->pause & INTEN) == 0) {
735		/*
736		 * Our interrupt is not enabled on the chip
737		 * and may be disabled for re-entrancy reasons,
738		 * so just return.  This is likely just a shared
739		 * interrupt.
740		 */
741		return (0);
742	}
743	/*
744	 * Instead of directly reading the interrupt status register,
745	 * infer the cause of the interrupt by checking our in-core
746	 * completion queues.  This avoids a costly PCI bus read in
747	 * most cases.
748	 */
749	if ((ahc->flags & (AHC_ALL_INTERRUPTS|AHC_EDGE_INTERRUPT)) == 0
750	 && (ahc_check_cmdcmpltqueues(ahc) != 0))
751		intstat = CMDCMPLT;
752	else {
753		intstat = ahc_inb(ahc, INTSTAT);
754	}
755
756	if ((intstat & INT_PEND) == 0) {
757#if AHC_PCI_CONFIG > 0
758		if (ahc->unsolicited_ints > 500) {
759			ahc->unsolicited_ints = 0;
760			if ((ahc->chip & AHC_PCI) != 0
761			 && (ahc_inb(ahc, ERROR) & PCIERRSTAT) != 0)
762				ahc->bus_intr(ahc);
763		}
764#endif
765		ahc->unsolicited_ints++;
766		return (0);
767	}
768	ahc->unsolicited_ints = 0;
769
770	if (intstat & CMDCMPLT) {
771		ahc_outb(ahc, CLRINT, CLRCMDINT);
772
773		/*
774		 * Ensure that the chip sees that we've cleared
775		 * this interrupt before we walk the output fifo.
776		 * Otherwise, we may, due to posted bus writes,
777		 * clear the interrupt after we finish the scan,
778		 * and after the sequencer has added new entries
779		 * and asserted the interrupt again.
780		 */
781		ahc_flush_device_writes(ahc);
782		ahc_run_qoutfifo(ahc);
783#ifdef AHC_TARGET_MODE
784		if ((ahc->flags & AHC_TARGETROLE) != 0)
785			ahc_run_tqinfifo(ahc, /*paused*/FALSE);
786#endif
787	}
788
789	/*
790	 * Handle statuses that may invalidate our cached
791	 * copy of INTSTAT separately.
792	 */
793	if (intstat == 0xFF && (ahc->features & AHC_REMOVABLE) != 0) {
794		/* Hot eject.  Do nothing */
795	} else if (intstat & BRKADRINT) {
796		ahc_handle_brkadrint(ahc);
797	} else if ((intstat & (SEQINT|SCSIINT)) != 0) {
798
799		ahc_pause_bug_fix(ahc);
800
801		if ((intstat & SEQINT) != 0)
802			ahc_handle_seqint(ahc, intstat);
803
804		if ((intstat & SCSIINT) != 0)
805			ahc_handle_scsiint(ahc, intstat);
806	}
807	return (1);
808}
809
810/************************* Sequencer Execution Control ************************/
811/*
812 * Restart the sequencer program from address zero
813 */
814static void
815ahc_restart(struct ahc_softc *ahc)
816{
817	uint8_t	sblkctl;
818
819	ahc_pause(ahc);
820
821	/* No more pending messages. */
822	ahc_clear_msg_state(ahc);
823
824	ahc_outb(ahc, SCSISIGO, 0);		/* De-assert BSY */
825	ahc_outb(ahc, MSG_OUT, MSG_NOOP);	/* No message to send */
826	ahc_outb(ahc, SXFRCTL1, ahc_inb(ahc, SXFRCTL1) & ~BITBUCKET);
827	ahc_outb(ahc, LASTPHASE, P_BUSFREE);
828	ahc_outb(ahc, SAVED_SCSIID, 0xFF);
829	ahc_outb(ahc, SAVED_LUN, 0xFF);
830
831	/*
832	 * Ensure that the sequencer's idea of TQINPOS
833	 * matches our own.  The sequencer increments TQINPOS
834	 * only after it sees a DMA complete and a reset could
835	 * occur before the increment leaving the kernel to believe
836	 * the command arrived but the sequencer to not.
837	 */
838	ahc_outb(ahc, TQINPOS, ahc->tqinfifonext);
839
840	/* Always allow reselection */
841	ahc_outb(ahc, SCSISEQ,
842		 ahc_inb(ahc, SCSISEQ_TEMPLATE) & (ENSELI|ENRSELI|ENAUTOATNP));
843	if ((ahc->features & AHC_CMD_CHAN) != 0) {
844		/* Ensure that no DMA operations are in progress */
845		ahc_outb(ahc, CCSCBCNT, 0);
846		ahc_outb(ahc, CCSGCTL, 0);
847		ahc_outb(ahc, CCSCBCTL, 0);
848	}
849	/*
850	 * If we were in the process of DMA'ing SCB data into
851	 * an SCB, replace that SCB on the free list.  This prevents
852	 * an SCB leak.
853	 */
854	if ((ahc_inb(ahc, SEQ_FLAGS2) & SCB_DMA) != 0) {
855		ahc_add_curscb_to_free_list(ahc);
856		ahc_outb(ahc, SEQ_FLAGS2,
857			 ahc_inb(ahc, SEQ_FLAGS2) & ~SCB_DMA);
858	}
859
860	/*
861	 * Clear any pending sequencer interrupt.  It is no
862	 * longer relevant since we're resetting the Program
863	 * Counter.
864	 */
865	ahc_outb(ahc, CLRINT, CLRSEQINT);
866
867	ahc_outb(ahc, MWI_RESIDUAL, 0);
868	ahc_outb(ahc, SEQCTL, ahc->seqctl);
869	ahc_outb(ahc, SEQADDR0, 0);
870	ahc_outb(ahc, SEQADDR1, 0);
871
872	/*
873	 * Take the LED out of diagnostic mode on PM resume, too
874	 */
875	sblkctl = ahc_inb(ahc, SBLKCTL);
876	ahc_outb(ahc, SBLKCTL, (sblkctl & ~(DIAGLEDEN|DIAGLEDON)));
877
878	ahc_unpause(ahc);
879}
880
881/************************* Input/Output Queues ********************************/
882static void
883ahc_run_qoutfifo(struct ahc_softc *ahc)
884{
885	struct scb *scb;
886	u_int  scb_index;
887
888	ahc_sync_qoutfifo(ahc, BUS_DMASYNC_POSTREAD);
889	while (ahc->qoutfifo[ahc->qoutfifonext] != SCB_LIST_NULL) {
890
891		scb_index = ahc->qoutfifo[ahc->qoutfifonext];
892		if ((ahc->qoutfifonext & 0x03) == 0x03) {
893			u_int modnext;
894
895			/*
896			 * Clear 32bits of QOUTFIFO at a time
897			 * so that we don't clobber an incoming
898			 * byte DMA to the array on architectures
899			 * that only support 32bit load and store
900			 * operations.
901			 */
902			modnext = ahc->qoutfifonext & ~0x3;
903			*((uint32_t *)(&ahc->qoutfifo[modnext])) = 0xFFFFFFFFUL;
904			ahc_dmamap_sync(ahc, ahc->shared_data_dmat,
905					ahc->shared_data_dmamap,
906					/*offset*/modnext, /*len*/4,
907					BUS_DMASYNC_PREREAD);
908		}
909		ahc->qoutfifonext++;
910
911		scb = ahc_lookup_scb(ahc, scb_index);
912		if (scb == NULL) {
913			printk("%s: WARNING no command for scb %d "
914			       "(cmdcmplt)\nQOUTPOS = %d\n",
915			       ahc_name(ahc), scb_index,
916			       (ahc->qoutfifonext - 1) & 0xFF);
917			continue;
918		}
919
920		/*
921		 * Save off the residual
922		 * if there is one.
923		 */
924		ahc_update_residual(ahc, scb);
925		ahc_done(ahc, scb);
926	}
927}
928
929static void
930ahc_run_untagged_queues(struct ahc_softc *ahc)
931{
932	int i;
933
934	for (i = 0; i < 16; i++)
935		ahc_run_untagged_queue(ahc, &ahc->untagged_queues[i]);
936}
937
938static void
939ahc_run_untagged_queue(struct ahc_softc *ahc, struct scb_tailq *queue)
940{
941	struct scb *scb;
942
943	if (ahc->untagged_queue_lock != 0)
944		return;
945
946	if ((scb = TAILQ_FIRST(queue)) != NULL
947	 && (scb->flags & SCB_ACTIVE) == 0) {
948		scb->flags |= SCB_ACTIVE;
949		ahc_queue_scb(ahc, scb);
950	}
951}
952
953/************************* Interrupt Handling *********************************/
954static void
955ahc_handle_brkadrint(struct ahc_softc *ahc)
956{
957	/*
958	 * We upset the sequencer :-(
959	 * Lookup the error message
960	 */
961	int i;
962	int error;
963
964	error = ahc_inb(ahc, ERROR);
965	for (i = 0; error != 1 && i < num_errors; i++)
966		error >>= 1;
967	printk("%s: brkadrint, %s at seqaddr = 0x%x\n",
968	       ahc_name(ahc), ahc_hard_errors[i].errmesg,
969	       ahc_inb(ahc, SEQADDR0) |
970	       (ahc_inb(ahc, SEQADDR1) << 8));
971
972	ahc_dump_card_state(ahc);
973
974	/* Tell everyone that this HBA is no longer available */
975	ahc_abort_scbs(ahc, CAM_TARGET_WILDCARD, ALL_CHANNELS,
976		       CAM_LUN_WILDCARD, SCB_LIST_NULL, ROLE_UNKNOWN,
977		       CAM_NO_HBA);
978
979	/* Disable all interrupt sources by resetting the controller */
980	ahc_shutdown(ahc);
981}
982
983static void
984ahc_handle_seqint(struct ahc_softc *ahc, u_int intstat)
985{
986	struct scb *scb;
987	struct ahc_devinfo devinfo;
988
989	ahc_fetch_devinfo(ahc, &devinfo);
990
991	/*
992	 * Clear the upper byte that holds SEQINT status
993	 * codes and clear the SEQINT bit. We will unpause
994	 * the sequencer, if appropriate, after servicing
995	 * the request.
996	 */
997	ahc_outb(ahc, CLRINT, CLRSEQINT);
998	switch (intstat & SEQINT_MASK) {
999	case BAD_STATUS:
1000	{
1001		u_int  scb_index;
1002		struct hardware_scb *hscb;
1003
1004		/*
1005		 * Set the default return value to 0 (don't
1006		 * send sense).  The sense code will change
1007		 * this if needed.
1008		 */
1009		ahc_outb(ahc, RETURN_1, 0);
1010
1011		/*
1012		 * The sequencer will notify us when a command
1013		 * has an error that would be of interest to
1014		 * the kernel.  This allows us to leave the sequencer
1015		 * running in the common case of command completes
1016		 * without error.  The sequencer will already have
1017		 * dma'd the SCB back up to us, so we can reference
1018		 * the in kernel copy directly.
1019		 */
1020		scb_index = ahc_inb(ahc, SCB_TAG);
1021		scb = ahc_lookup_scb(ahc, scb_index);
1022		if (scb == NULL) {
1023			ahc_print_devinfo(ahc, &devinfo);
1024			printk("ahc_intr - referenced scb "
1025			       "not valid during seqint 0x%x scb(%d)\n",
1026			       intstat, scb_index);
1027			ahc_dump_card_state(ahc);
1028			panic("for safety");
1029			goto unpause;
1030		}
1031
1032		hscb = scb->hscb;
1033
1034		/* Don't want to clobber the original sense code */
1035		if ((scb->flags & SCB_SENSE) != 0) {
1036			/*
1037			 * Clear the SCB_SENSE Flag and have
1038			 * the sequencer do a normal command
1039			 * complete.
1040			 */
1041			scb->flags &= ~SCB_SENSE;
1042			ahc_set_transaction_status(scb, CAM_AUTOSENSE_FAIL);
1043			break;
1044		}
1045		ahc_set_transaction_status(scb, CAM_SCSI_STATUS_ERROR);
1046		/* Freeze the queue until the client sees the error. */
1047		ahc_freeze_devq(ahc, scb);
1048		ahc_freeze_scb(scb);
1049		ahc_set_scsi_status(scb, hscb->shared_data.status.scsi_status);
1050		switch (hscb->shared_data.status.scsi_status) {
1051		case SCSI_STATUS_OK:
1052			printk("%s: Interrupted for status of 0???\n",
1053			       ahc_name(ahc));
1054			break;
1055		case SCSI_STATUS_CMD_TERMINATED:
1056		case SCSI_STATUS_CHECK_COND:
1057		{
1058			struct ahc_dma_seg *sg;
1059			struct scsi_sense *sc;
1060			struct ahc_initiator_tinfo *targ_info;
1061			struct ahc_tmode_tstate *tstate;
1062			struct ahc_transinfo *tinfo;
1063#ifdef AHC_DEBUG
1064			if (ahc_debug & AHC_SHOW_SENSE) {
1065				ahc_print_path(ahc, scb);
1066				printk("SCB %d: requests Check Status\n",
1067				       scb->hscb->tag);
1068			}
1069#endif
1070
1071			if (ahc_perform_autosense(scb) == 0)
1072				break;
1073
1074			targ_info = ahc_fetch_transinfo(ahc,
1075							devinfo.channel,
1076							devinfo.our_scsiid,
1077							devinfo.target,
1078							&tstate);
1079			tinfo = &targ_info->curr;
1080			sg = scb->sg_list;
1081			sc = (struct scsi_sense *)(&hscb->shared_data.cdb);
1082			/*
1083			 * Save off the residual if there is one.
1084			 */
1085			ahc_update_residual(ahc, scb);
1086#ifdef AHC_DEBUG
1087			if (ahc_debug & AHC_SHOW_SENSE) {
1088				ahc_print_path(ahc, scb);
1089				printk("Sending Sense\n");
1090			}
1091#endif
1092			sg->addr = ahc_get_sense_bufaddr(ahc, scb);
1093			sg->len = ahc_get_sense_bufsize(ahc, scb);
1094			sg->len |= AHC_DMA_LAST_SEG;
1095
1096			/* Fixup byte order */
1097			sg->addr = ahc_htole32(sg->addr);
1098			sg->len = ahc_htole32(sg->len);
1099
1100			sc->opcode = REQUEST_SENSE;
1101			sc->byte2 = 0;
1102			if (tinfo->protocol_version <= SCSI_REV_2
1103			 && SCB_GET_LUN(scb) < 8)
1104				sc->byte2 = SCB_GET_LUN(scb) << 5;
1105			sc->unused[0] = 0;
1106			sc->unused[1] = 0;
1107			sc->length = sg->len;
1108			sc->control = 0;
1109
1110			/*
1111			 * We can't allow the target to disconnect.
1112			 * This will be an untagged transaction and
1113			 * having the target disconnect will make this
1114			 * transaction indestinguishable from outstanding
1115			 * tagged transactions.
1116			 */
1117			hscb->control = 0;
1118
1119			/*
1120			 * This request sense could be because the
1121			 * the device lost power or in some other
1122			 * way has lost our transfer negotiations.
1123			 * Renegotiate if appropriate.  Unit attention
1124			 * errors will be reported before any data
1125			 * phases occur.
1126			 */
1127			if (ahc_get_residual(scb)
1128			 == ahc_get_transfer_length(scb)) {
1129				ahc_update_neg_request(ahc, &devinfo,
1130						       tstate, targ_info,
1131						       AHC_NEG_IF_NON_ASYNC);
1132			}
1133			if (tstate->auto_negotiate & devinfo.target_mask) {
1134				hscb->control |= MK_MESSAGE;
1135				scb->flags &= ~SCB_NEGOTIATE;
1136				scb->flags |= SCB_AUTO_NEGOTIATE;
1137			}
1138			hscb->cdb_len = sizeof(*sc);
1139			hscb->dataptr = sg->addr;
1140			hscb->datacnt = sg->len;
1141			hscb->sgptr = scb->sg_list_phys | SG_FULL_RESID;
1142			hscb->sgptr = ahc_htole32(hscb->sgptr);
1143			scb->sg_count = 1;
1144			scb->flags |= SCB_SENSE;
1145			ahc_qinfifo_requeue_tail(ahc, scb);
1146			ahc_outb(ahc, RETURN_1, SEND_SENSE);
1147			/*
1148			 * Ensure we have enough time to actually
1149			 * retrieve the sense.
1150			 */
1151			ahc_scb_timer_reset(scb, 5 * 1000000);
1152			break;
1153		}
1154		default:
1155			break;
1156		}
1157		break;
1158	}
1159	case NO_MATCH:
1160	{
1161		/* Ensure we don't leave the selection hardware on */
1162		ahc_outb(ahc, SCSISEQ,
1163			 ahc_inb(ahc, SCSISEQ) & (ENSELI|ENRSELI|ENAUTOATNP));
1164
1165		printk("%s:%c:%d: no active SCB for reconnecting "
1166		       "target - issuing BUS DEVICE RESET\n",
1167		       ahc_name(ahc), devinfo.channel, devinfo.target);
1168		printk("SAVED_SCSIID == 0x%x, SAVED_LUN == 0x%x, "
1169		       "ARG_1 == 0x%x ACCUM = 0x%x\n",
1170		       ahc_inb(ahc, SAVED_SCSIID), ahc_inb(ahc, SAVED_LUN),
1171		       ahc_inb(ahc, ARG_1), ahc_inb(ahc, ACCUM));
1172		printk("SEQ_FLAGS == 0x%x, SCBPTR == 0x%x, BTT == 0x%x, "
1173		       "SINDEX == 0x%x\n",
1174		       ahc_inb(ahc, SEQ_FLAGS), ahc_inb(ahc, SCBPTR),
1175		       ahc_index_busy_tcl(ahc,
1176			    BUILD_TCL(ahc_inb(ahc, SAVED_SCSIID),
1177				      ahc_inb(ahc, SAVED_LUN))),
1178		       ahc_inb(ahc, SINDEX));
1179		printk("SCSIID == 0x%x, SCB_SCSIID == 0x%x, SCB_LUN == 0x%x, "
1180		       "SCB_TAG == 0x%x, SCB_CONTROL == 0x%x\n",
1181		       ahc_inb(ahc, SCSIID), ahc_inb(ahc, SCB_SCSIID),
1182		       ahc_inb(ahc, SCB_LUN), ahc_inb(ahc, SCB_TAG),
1183		       ahc_inb(ahc, SCB_CONTROL));
1184		printk("SCSIBUSL == 0x%x, SCSISIGI == 0x%x\n",
1185		       ahc_inb(ahc, SCSIBUSL), ahc_inb(ahc, SCSISIGI));
1186		printk("SXFRCTL0 == 0x%x\n", ahc_inb(ahc, SXFRCTL0));
1187		printk("SEQCTL == 0x%x\n", ahc_inb(ahc, SEQCTL));
1188		ahc_dump_card_state(ahc);
1189		ahc->msgout_buf[0] = MSG_BUS_DEV_RESET;
1190		ahc->msgout_len = 1;
1191		ahc->msgout_index = 0;
1192		ahc->msg_type = MSG_TYPE_INITIATOR_MSGOUT;
1193		ahc_outb(ahc, MSG_OUT, HOST_MSG);
1194		ahc_assert_atn(ahc);
1195		break;
1196	}
1197	case SEND_REJECT:
1198	{
1199		u_int rejbyte = ahc_inb(ahc, ACCUM);
1200		printk("%s:%c:%d: Warning - unknown message received from "
1201		       "target (0x%x).  Rejecting\n",
1202		       ahc_name(ahc), devinfo.channel, devinfo.target, rejbyte);
1203		break;
1204	}
1205	case PROTO_VIOLATION:
1206	{
1207		ahc_handle_proto_violation(ahc);
1208		break;
1209	}
1210	case IGN_WIDE_RES:
1211		ahc_handle_ign_wide_residue(ahc, &devinfo);
1212		break;
1213	case PDATA_REINIT:
1214		ahc_reinitialize_dataptrs(ahc);
1215		break;
1216	case BAD_PHASE:
1217	{
1218		u_int lastphase;
1219
1220		lastphase = ahc_inb(ahc, LASTPHASE);
1221		printk("%s:%c:%d: unknown scsi bus phase %x, "
1222		       "lastphase = 0x%x.  Attempting to continue\n",
1223		       ahc_name(ahc), devinfo.channel, devinfo.target,
1224		       lastphase, ahc_inb(ahc, SCSISIGI));
1225		break;
1226	}
1227	case MISSED_BUSFREE:
1228	{
1229		u_int lastphase;
1230
1231		lastphase = ahc_inb(ahc, LASTPHASE);
1232		printk("%s:%c:%d: Missed busfree. "
1233		       "Lastphase = 0x%x, Curphase = 0x%x\n",
1234		       ahc_name(ahc), devinfo.channel, devinfo.target,
1235		       lastphase, ahc_inb(ahc, SCSISIGI));
1236		ahc_restart(ahc);
1237		return;
1238	}
1239	case HOST_MSG_LOOP:
1240	{
1241		/*
1242		 * The sequencer has encountered a message phase
1243		 * that requires host assistance for completion.
1244		 * While handling the message phase(s), we will be
1245		 * notified by the sequencer after each byte is
1246		 * transferred so we can track bus phase changes.
1247		 *
1248		 * If this is the first time we've seen a HOST_MSG_LOOP
1249		 * interrupt, initialize the state of the host message
1250		 * loop.
1251		 */
1252		if (ahc->msg_type == MSG_TYPE_NONE) {
1253			struct scb *scb;
1254			u_int scb_index;
1255			u_int bus_phase;
1256
1257			bus_phase = ahc_inb(ahc, SCSISIGI) & PHASE_MASK;
1258			if (bus_phase != P_MESGIN
1259			 && bus_phase != P_MESGOUT) {
1260				printk("ahc_intr: HOST_MSG_LOOP bad "
1261				       "phase 0x%x\n",
1262				      bus_phase);
1263				/*
1264				 * Probably transitioned to bus free before
1265				 * we got here.  Just punt the message.
1266				 */
1267				ahc_clear_intstat(ahc);
1268				ahc_restart(ahc);
1269				return;
1270			}
1271
1272			scb_index = ahc_inb(ahc, SCB_TAG);
1273			scb = ahc_lookup_scb(ahc, scb_index);
1274			if (devinfo.role == ROLE_INITIATOR) {
1275				if (bus_phase == P_MESGOUT) {
1276					if (scb == NULL)
1277						panic("HOST_MSG_LOOP with "
1278						      "invalid SCB %x\n",
1279						      scb_index);
1280
1281					ahc_setup_initiator_msgout(ahc,
1282								   &devinfo,
1283								   scb);
1284				} else {
1285					ahc->msg_type =
1286					    MSG_TYPE_INITIATOR_MSGIN;
1287					ahc->msgin_index = 0;
1288				}
1289			}
1290#ifdef AHC_TARGET_MODE
1291			else {
1292				if (bus_phase == P_MESGOUT) {
1293					ahc->msg_type =
1294					    MSG_TYPE_TARGET_MSGOUT;
1295					ahc->msgin_index = 0;
1296				}
1297				else
1298					ahc_setup_target_msgin(ahc,
1299							       &devinfo,
1300							       scb);
1301			}
1302#endif
1303		}
1304
1305		ahc_handle_message_phase(ahc);
1306		break;
1307	}
1308	case PERR_DETECTED:
1309	{
1310		/*
1311		 * If we've cleared the parity error interrupt
1312		 * but the sequencer still believes that SCSIPERR
1313		 * is true, it must be that the parity error is
1314		 * for the currently presented byte on the bus,
1315		 * and we are not in a phase (data-in) where we will
1316		 * eventually ack this byte.  Ack the byte and
1317		 * throw it away in the hope that the target will
1318		 * take us to message out to deliver the appropriate
1319		 * error message.
1320		 */
1321		if ((intstat & SCSIINT) == 0
1322		 && (ahc_inb(ahc, SSTAT1) & SCSIPERR) != 0) {
1323
1324			if ((ahc->features & AHC_DT) == 0) {
1325				u_int curphase;
1326
1327				/*
1328				 * The hardware will only let you ack bytes
1329				 * if the expected phase in SCSISIGO matches
1330				 * the current phase.  Make sure this is
1331				 * currently the case.
1332				 */
1333				curphase = ahc_inb(ahc, SCSISIGI) & PHASE_MASK;
1334				ahc_outb(ahc, LASTPHASE, curphase);
1335				ahc_outb(ahc, SCSISIGO, curphase);
1336			}
1337			if ((ahc_inb(ahc, SCSISIGI) & (CDI|MSGI)) == 0) {
1338				int wait;
1339
1340				/*
1341				 * In a data phase.  Faster to bitbucket
1342				 * the data than to individually ack each
1343				 * byte.  This is also the only strategy
1344				 * that will work with AUTOACK enabled.
1345				 */
1346				ahc_outb(ahc, SXFRCTL1,
1347					 ahc_inb(ahc, SXFRCTL1) | BITBUCKET);
1348				wait = 5000;
1349				while (--wait != 0) {
1350					if ((ahc_inb(ahc, SCSISIGI)
1351					  & (CDI|MSGI)) != 0)
1352						break;
1353					ahc_delay(100);
1354				}
1355				ahc_outb(ahc, SXFRCTL1,
1356					 ahc_inb(ahc, SXFRCTL1) & ~BITBUCKET);
1357				if (wait == 0) {
1358					struct	scb *scb;
1359					u_int	scb_index;
1360
1361					ahc_print_devinfo(ahc, &devinfo);
1362					printk("Unable to clear parity error.  "
1363					       "Resetting bus.\n");
1364					scb_index = ahc_inb(ahc, SCB_TAG);
1365					scb = ahc_lookup_scb(ahc, scb_index);
1366					if (scb != NULL)
1367						ahc_set_transaction_status(scb,
1368						    CAM_UNCOR_PARITY);
1369					ahc_reset_channel(ahc, devinfo.channel,
1370							  /*init reset*/TRUE);
1371				}
1372			} else {
1373				ahc_inb(ahc, SCSIDATL);
1374			}
1375		}
1376		break;
1377	}
1378	case DATA_OVERRUN:
1379	{
1380		/*
1381		 * When the sequencer detects an overrun, it
1382		 * places the controller in "BITBUCKET" mode
1383		 * and allows the target to complete its transfer.
1384		 * Unfortunately, none of the counters get updated
1385		 * when the controller is in this mode, so we have
1386		 * no way of knowing how large the overrun was.
1387		 */
1388		u_int scbindex = ahc_inb(ahc, SCB_TAG);
1389		u_int lastphase = ahc_inb(ahc, LASTPHASE);
1390		u_int i;
1391
1392		scb = ahc_lookup_scb(ahc, scbindex);
1393		for (i = 0; i < num_phases; i++) {
1394			if (lastphase == ahc_phase_table[i].phase)
1395				break;
1396		}
1397		ahc_print_path(ahc, scb);
1398		printk("data overrun detected %s."
1399		       "  Tag == 0x%x.\n",
1400		       ahc_phase_table[i].phasemsg,
1401  		       scb->hscb->tag);
1402		ahc_print_path(ahc, scb);
1403		printk("%s seen Data Phase.  Length = %ld.  NumSGs = %d.\n",
1404		       ahc_inb(ahc, SEQ_FLAGS) & DPHASE ? "Have" : "Haven't",
1405		       ahc_get_transfer_length(scb), scb->sg_count);
1406		if (scb->sg_count > 0) {
1407			for (i = 0; i < scb->sg_count; i++) {
1408
1409				printk("sg[%d] - Addr 0x%x%x : Length %d\n",
1410				       i,
1411				       (ahc_le32toh(scb->sg_list[i].len) >> 24
1412				        & SG_HIGH_ADDR_BITS),
1413				       ahc_le32toh(scb->sg_list[i].addr),
1414				       ahc_le32toh(scb->sg_list[i].len)
1415				       & AHC_SG_LEN_MASK);
1416			}
1417		}
1418		/*
1419		 * Set this and it will take effect when the
1420		 * target does a command complete.
1421		 */
1422		ahc_freeze_devq(ahc, scb);
1423		if ((scb->flags & SCB_SENSE) == 0) {
1424			ahc_set_transaction_status(scb, CAM_DATA_RUN_ERR);
1425		} else {
1426			scb->flags &= ~SCB_SENSE;
1427			ahc_set_transaction_status(scb, CAM_AUTOSENSE_FAIL);
1428		}
1429		ahc_freeze_scb(scb);
1430
1431		if ((ahc->features & AHC_ULTRA2) != 0) {
1432			/*
1433			 * Clear the channel in case we return
1434			 * to data phase later.
1435			 */
1436			ahc_outb(ahc, SXFRCTL0,
1437				 ahc_inb(ahc, SXFRCTL0) | CLRSTCNT|CLRCHN);
1438			ahc_outb(ahc, SXFRCTL0,
1439				 ahc_inb(ahc, SXFRCTL0) | CLRSTCNT|CLRCHN);
1440		}
1441		if ((ahc->flags & AHC_39BIT_ADDRESSING) != 0) {
1442			u_int dscommand1;
1443
1444			/* Ensure HHADDR is 0 for future DMA operations. */
1445			dscommand1 = ahc_inb(ahc, DSCOMMAND1);
1446			ahc_outb(ahc, DSCOMMAND1, dscommand1 | HADDLDSEL0);
1447			ahc_outb(ahc, HADDR, 0);
1448			ahc_outb(ahc, DSCOMMAND1, dscommand1);
1449		}
1450		break;
1451	}
1452	case MKMSG_FAILED:
1453	{
1454		u_int scbindex;
1455
1456		printk("%s:%c:%d:%d: Attempt to issue message failed\n",
1457		       ahc_name(ahc), devinfo.channel, devinfo.target,
1458		       devinfo.lun);
1459		scbindex = ahc_inb(ahc, SCB_TAG);
1460		scb = ahc_lookup_scb(ahc, scbindex);
1461		if (scb != NULL
1462		 && (scb->flags & SCB_RECOVERY_SCB) != 0)
1463			/*
1464			 * Ensure that we didn't put a second instance of this
1465			 * SCB into the QINFIFO.
1466			 */
1467			ahc_search_qinfifo(ahc, SCB_GET_TARGET(ahc, scb),
1468					   SCB_GET_CHANNEL(ahc, scb),
1469					   SCB_GET_LUN(scb), scb->hscb->tag,
1470					   ROLE_INITIATOR, /*status*/0,
1471					   SEARCH_REMOVE);
1472		break;
1473	}
1474	case NO_FREE_SCB:
1475	{
1476		printk("%s: No free or disconnected SCBs\n", ahc_name(ahc));
1477		ahc_dump_card_state(ahc);
1478		panic("for safety");
1479		break;
1480	}
1481	case SCB_MISMATCH:
1482	{
1483		u_int scbptr;
1484
1485		scbptr = ahc_inb(ahc, SCBPTR);
1486		printk("Bogus TAG after DMA.  SCBPTR %d, tag %d, our tag %d\n",
1487		       scbptr, ahc_inb(ahc, ARG_1),
1488		       ahc->scb_data->hscbs[scbptr].tag);
1489		ahc_dump_card_state(ahc);
1490		panic("for safety");
1491		break;
1492	}
1493	case OUT_OF_RANGE:
1494	{
1495		printk("%s: BTT calculation out of range\n", ahc_name(ahc));
1496		printk("SAVED_SCSIID == 0x%x, SAVED_LUN == 0x%x, "
1497		       "ARG_1 == 0x%x ACCUM = 0x%x\n",
1498		       ahc_inb(ahc, SAVED_SCSIID), ahc_inb(ahc, SAVED_LUN),
1499		       ahc_inb(ahc, ARG_1), ahc_inb(ahc, ACCUM));
1500		printk("SEQ_FLAGS == 0x%x, SCBPTR == 0x%x, BTT == 0x%x, "
1501		       "SINDEX == 0x%x\n, A == 0x%x\n",
1502		       ahc_inb(ahc, SEQ_FLAGS), ahc_inb(ahc, SCBPTR),
1503		       ahc_index_busy_tcl(ahc,
1504			    BUILD_TCL(ahc_inb(ahc, SAVED_SCSIID),
1505				      ahc_inb(ahc, SAVED_LUN))),
1506		       ahc_inb(ahc, SINDEX),
1507		       ahc_inb(ahc, ACCUM));
1508		printk("SCSIID == 0x%x, SCB_SCSIID == 0x%x, SCB_LUN == 0x%x, "
1509		       "SCB_TAG == 0x%x, SCB_CONTROL == 0x%x\n",
1510		       ahc_inb(ahc, SCSIID), ahc_inb(ahc, SCB_SCSIID),
1511		       ahc_inb(ahc, SCB_LUN), ahc_inb(ahc, SCB_TAG),
1512		       ahc_inb(ahc, SCB_CONTROL));
1513		printk("SCSIBUSL == 0x%x, SCSISIGI == 0x%x\n",
1514		       ahc_inb(ahc, SCSIBUSL), ahc_inb(ahc, SCSISIGI));
1515		ahc_dump_card_state(ahc);
1516		panic("for safety");
1517		break;
1518	}
1519	default:
1520		printk("ahc_intr: seqint, "
1521		       "intstat == 0x%x, scsisigi = 0x%x\n",
1522		       intstat, ahc_inb(ahc, SCSISIGI));
1523		break;
1524	}
1525unpause:
1526	/*
1527	 *  The sequencer is paused immediately on
1528	 *  a SEQINT, so we should restart it when
1529	 *  we're done.
1530	 */
1531	ahc_unpause(ahc);
1532}
1533
1534static void
1535ahc_handle_scsiint(struct ahc_softc *ahc, u_int intstat)
1536{
1537	u_int	scb_index;
1538	u_int	status0;
1539	u_int	status;
1540	struct	scb *scb;
1541	char	cur_channel;
1542	char	intr_channel;
1543
1544	if ((ahc->features & AHC_TWIN) != 0
1545	 && ((ahc_inb(ahc, SBLKCTL) & SELBUSB) != 0))
1546		cur_channel = 'B';
1547	else
1548		cur_channel = 'A';
1549	intr_channel = cur_channel;
1550
1551	if ((ahc->features & AHC_ULTRA2) != 0)
1552		status0 = ahc_inb(ahc, SSTAT0) & IOERR;
1553	else
1554		status0 = 0;
1555	status = ahc_inb(ahc, SSTAT1) & (SELTO|SCSIRSTI|BUSFREE|SCSIPERR);
1556	if (status == 0 && status0 == 0) {
1557		if ((ahc->features & AHC_TWIN) != 0) {
1558			/* Try the other channel */
1559		 	ahc_outb(ahc, SBLKCTL, ahc_inb(ahc, SBLKCTL) ^ SELBUSB);
1560			status = ahc_inb(ahc, SSTAT1)
1561			       & (SELTO|SCSIRSTI|BUSFREE|SCSIPERR);
1562			intr_channel = (cur_channel == 'A') ? 'B' : 'A';
1563		}
1564		if (status == 0) {
1565			printk("%s: Spurious SCSI interrupt\n", ahc_name(ahc));
1566			ahc_outb(ahc, CLRINT, CLRSCSIINT);
1567			ahc_unpause(ahc);
1568			return;
1569		}
1570	}
1571
1572	/* Make sure the sequencer is in a safe location. */
1573	ahc_clear_critical_section(ahc);
1574
1575	scb_index = ahc_inb(ahc, SCB_TAG);
1576	scb = ahc_lookup_scb(ahc, scb_index);
1577	if (scb != NULL
1578	 && (ahc_inb(ahc, SEQ_FLAGS) & NOT_IDENTIFIED) != 0)
1579		scb = NULL;
1580
1581	if ((ahc->features & AHC_ULTRA2) != 0
1582	 && (status0 & IOERR) != 0) {
1583		int now_lvd;
1584
1585		now_lvd = ahc_inb(ahc, SBLKCTL) & ENAB40;
1586		printk("%s: Transceiver State Has Changed to %s mode\n",
1587		       ahc_name(ahc), now_lvd ? "LVD" : "SE");
1588		ahc_outb(ahc, CLRSINT0, CLRIOERR);
1589		/*
1590		 * When transitioning to SE mode, the reset line
1591		 * glitches, triggering an arbitration bug in some
1592		 * Ultra2 controllers.  This bug is cleared when we
1593		 * assert the reset line.  Since a reset glitch has
1594		 * already occurred with this transition and a
1595		 * transceiver state change is handled just like
1596		 * a bus reset anyway, asserting the reset line
1597		 * ourselves is safe.
1598		 */
1599		ahc_reset_channel(ahc, intr_channel,
1600				 /*Initiate Reset*/now_lvd == 0);
1601	} else if ((status & SCSIRSTI) != 0) {
1602		printk("%s: Someone reset channel %c\n",
1603			ahc_name(ahc), intr_channel);
1604		if (intr_channel != cur_channel)
1605		 	ahc_outb(ahc, SBLKCTL, ahc_inb(ahc, SBLKCTL) ^ SELBUSB);
1606		ahc_reset_channel(ahc, intr_channel, /*Initiate Reset*/FALSE);
1607	} else if ((status & SCSIPERR) != 0) {
1608		/*
1609		 * Determine the bus phase and queue an appropriate message.
1610		 * SCSIPERR is latched true as soon as a parity error
1611		 * occurs.  If the sequencer acked the transfer that
1612		 * caused the parity error and the currently presented
1613		 * transfer on the bus has correct parity, SCSIPERR will
1614		 * be cleared by CLRSCSIPERR.  Use this to determine if
1615		 * we should look at the last phase the sequencer recorded,
1616		 * or the current phase presented on the bus.
1617		 */
1618		struct	ahc_devinfo devinfo;
1619		u_int	mesg_out;
1620		u_int	curphase;
1621		u_int	errorphase;
1622		u_int	lastphase;
1623		u_int	scsirate;
1624		u_int	i;
1625		u_int	sstat2;
1626		int	silent;
1627
1628		lastphase = ahc_inb(ahc, LASTPHASE);
1629		curphase = ahc_inb(ahc, SCSISIGI) & PHASE_MASK;
1630		sstat2 = ahc_inb(ahc, SSTAT2);
1631		ahc_outb(ahc, CLRSINT1, CLRSCSIPERR);
1632		/*
1633		 * For all phases save DATA, the sequencer won't
1634		 * automatically ack a byte that has a parity error
1635		 * in it.  So the only way that the current phase
1636		 * could be 'data-in' is if the parity error is for
1637		 * an already acked byte in the data phase.  During
1638		 * synchronous data-in transfers, we may actually
1639		 * ack bytes before latching the current phase in
1640		 * LASTPHASE, leading to the discrepancy between
1641		 * curphase and lastphase.
1642		 */
1643		if ((ahc_inb(ahc, SSTAT1) & SCSIPERR) != 0
1644		 || curphase == P_DATAIN || curphase == P_DATAIN_DT)
1645			errorphase = curphase;
1646		else
1647			errorphase = lastphase;
1648
1649		for (i = 0; i < num_phases; i++) {
1650			if (errorphase == ahc_phase_table[i].phase)
1651				break;
1652		}
1653		mesg_out = ahc_phase_table[i].mesg_out;
1654		silent = FALSE;
1655		if (scb != NULL) {
1656			if (SCB_IS_SILENT(scb))
1657				silent = TRUE;
1658			else
1659				ahc_print_path(ahc, scb);
1660			scb->flags |= SCB_TRANSMISSION_ERROR;
1661		} else
1662			printk("%s:%c:%d: ", ahc_name(ahc), intr_channel,
1663			       SCSIID_TARGET(ahc, ahc_inb(ahc, SAVED_SCSIID)));
1664		scsirate = ahc_inb(ahc, SCSIRATE);
1665		if (silent == FALSE) {
1666			printk("parity error detected %s. "
1667			       "SEQADDR(0x%x) SCSIRATE(0x%x)\n",
1668			       ahc_phase_table[i].phasemsg,
1669			       ahc_inw(ahc, SEQADDR0),
1670			       scsirate);
1671			if ((ahc->features & AHC_DT) != 0) {
1672				if ((sstat2 & CRCVALERR) != 0)
1673					printk("\tCRC Value Mismatch\n");
1674				if ((sstat2 & CRCENDERR) != 0)
1675					printk("\tNo terminal CRC packet "
1676					       "recevied\n");
1677				if ((sstat2 & CRCREQERR) != 0)
1678					printk("\tIllegal CRC packet "
1679					       "request\n");
1680				if ((sstat2 & DUAL_EDGE_ERR) != 0)
1681					printk("\tUnexpected %sDT Data Phase\n",
1682					       (scsirate & SINGLE_EDGE)
1683					     ? "" : "non-");
1684			}
1685		}
1686
1687		if ((ahc->features & AHC_DT) != 0
1688		 && (sstat2 & DUAL_EDGE_ERR) != 0) {
1689			/*
1690			 * This error applies regardless of
1691			 * data direction, so ignore the value
1692			 * in the phase table.
1693			 */
1694			mesg_out = MSG_INITIATOR_DET_ERR;
1695		}
1696
1697		/*
1698		 * We've set the hardware to assert ATN if we
1699		 * get a parity error on "in" phases, so all we
1700		 * need to do is stuff the message buffer with
1701		 * the appropriate message.  "In" phases have set
1702		 * mesg_out to something other than MSG_NOP.
1703		 */
1704		if (mesg_out != MSG_NOOP) {
1705			if (ahc->msg_type != MSG_TYPE_NONE)
1706				ahc->send_msg_perror = TRUE;
1707			else
1708				ahc_outb(ahc, MSG_OUT, mesg_out);
1709		}
1710		/*
1711		 * Force a renegotiation with this target just in
1712		 * case we are out of sync for some external reason
1713		 * unknown (or unreported) by the target.
1714		 */
1715		ahc_fetch_devinfo(ahc, &devinfo);
1716		ahc_force_renegotiation(ahc, &devinfo);
1717
1718		ahc_outb(ahc, CLRINT, CLRSCSIINT);
1719		ahc_unpause(ahc);
1720	} else if ((status & SELTO) != 0) {
1721		u_int	scbptr;
1722
1723		/* Stop the selection */
1724		ahc_outb(ahc, SCSISEQ, 0);
1725
1726		/* No more pending messages */
1727		ahc_clear_msg_state(ahc);
1728
1729		/* Clear interrupt state */
1730		ahc_outb(ahc, SIMODE1, ahc_inb(ahc, SIMODE1) & ~ENBUSFREE);
1731		ahc_outb(ahc, CLRSINT1, CLRSELTIMEO|CLRBUSFREE|CLRSCSIPERR);
1732
1733		/*
1734		 * Although the driver does not care about the
1735		 * 'Selection in Progress' status bit, the busy
1736		 * LED does.  SELINGO is only cleared by a successful
1737		 * selection, so we must manually clear it to insure
1738		 * the LED turns off just incase no future successful
1739		 * selections occur (e.g. no devices on the bus).
1740		 */
1741		ahc_outb(ahc, CLRSINT0, CLRSELINGO);
1742
1743		scbptr = ahc_inb(ahc, WAITING_SCBH);
1744		ahc_outb(ahc, SCBPTR, scbptr);
1745		scb_index = ahc_inb(ahc, SCB_TAG);
1746
1747		scb = ahc_lookup_scb(ahc, scb_index);
1748		if (scb == NULL) {
1749			printk("%s: ahc_intr - referenced scb not "
1750			       "valid during SELTO scb(%d, %d)\n",
1751			       ahc_name(ahc), scbptr, scb_index);
1752			ahc_dump_card_state(ahc);
1753		} else {
1754			struct ahc_devinfo devinfo;
1755#ifdef AHC_DEBUG
1756			if ((ahc_debug & AHC_SHOW_SELTO) != 0) {
1757				ahc_print_path(ahc, scb);
1758				printk("Saw Selection Timeout for SCB 0x%x\n",
1759				       scb_index);
1760			}
1761#endif
1762			ahc_scb_devinfo(ahc, &devinfo, scb);
1763			ahc_set_transaction_status(scb, CAM_SEL_TIMEOUT);
1764			ahc_freeze_devq(ahc, scb);
1765
1766			/*
1767			 * Cancel any pending transactions on the device
1768			 * now that it seems to be missing.  This will
1769			 * also revert us to async/narrow transfers until
1770			 * we can renegotiate with the device.
1771			 */
1772			ahc_handle_devreset(ahc, &devinfo,
1773					    CAM_SEL_TIMEOUT,
1774					    "Selection Timeout",
1775					    /*verbose_level*/1);
1776		}
1777		ahc_outb(ahc, CLRINT, CLRSCSIINT);
1778		ahc_restart(ahc);
1779	} else if ((status & BUSFREE) != 0
1780		&& (ahc_inb(ahc, SIMODE1) & ENBUSFREE) != 0) {
1781		struct	ahc_devinfo devinfo;
1782		u_int	lastphase;
1783		u_int	saved_scsiid;
1784		u_int	saved_lun;
1785		u_int	target;
1786		u_int	initiator_role_id;
1787		char	channel;
1788		int	printerror;
1789
1790		/*
1791		 * Clear our selection hardware as soon as possible.
1792		 * We may have an entry in the waiting Q for this target,
1793		 * that is affected by this busfree and we don't want to
1794		 * go about selecting the target while we handle the event.
1795		 */
1796		ahc_outb(ahc, SCSISEQ,
1797			 ahc_inb(ahc, SCSISEQ) & (ENSELI|ENRSELI|ENAUTOATNP));
1798
1799		/*
1800		 * Disable busfree interrupts and clear the busfree
1801		 * interrupt status.  We do this here so that several
1802		 * bus transactions occur prior to clearing the SCSIINT
1803		 * latch.  It can take a bit for the clearing to take effect.
1804		 */
1805		ahc_outb(ahc, SIMODE1, ahc_inb(ahc, SIMODE1) & ~ENBUSFREE);
1806		ahc_outb(ahc, CLRSINT1, CLRBUSFREE|CLRSCSIPERR);
1807
1808		/*
1809		 * Look at what phase we were last in.
1810		 * If its message out, chances are pretty good
1811		 * that the busfree was in response to one of
1812		 * our abort requests.
1813		 */
1814		lastphase = ahc_inb(ahc, LASTPHASE);
1815		saved_scsiid = ahc_inb(ahc, SAVED_SCSIID);
1816		saved_lun = ahc_inb(ahc, SAVED_LUN);
1817		target = SCSIID_TARGET(ahc, saved_scsiid);
1818		initiator_role_id = SCSIID_OUR_ID(saved_scsiid);
1819		channel = SCSIID_CHANNEL(ahc, saved_scsiid);
1820		ahc_compile_devinfo(&devinfo, initiator_role_id,
1821				    target, saved_lun, channel, ROLE_INITIATOR);
1822		printerror = 1;
1823
1824		if (lastphase == P_MESGOUT) {
1825			u_int tag;
1826
1827			tag = SCB_LIST_NULL;
1828			if (ahc_sent_msg(ahc, AHCMSG_1B, MSG_ABORT_TAG, TRUE)
1829			 || ahc_sent_msg(ahc, AHCMSG_1B, MSG_ABORT, TRUE)) {
1830				if (ahc->msgout_buf[ahc->msgout_index - 1]
1831				 == MSG_ABORT_TAG)
1832					tag = scb->hscb->tag;
1833				ahc_print_path(ahc, scb);
1834				printk("SCB %d - Abort%s Completed.\n",
1835				       scb->hscb->tag, tag == SCB_LIST_NULL ?
1836				       "" : " Tag");
1837				ahc_abort_scbs(ahc, target, channel,
1838					       saved_lun, tag,
1839					       ROLE_INITIATOR,
1840					       CAM_REQ_ABORTED);
1841				printerror = 0;
1842			} else if (ahc_sent_msg(ahc, AHCMSG_1B,
1843						MSG_BUS_DEV_RESET, TRUE)) {
1844#ifdef __FreeBSD__
1845				/*
1846				 * Don't mark the user's request for this BDR
1847				 * as completing with CAM_BDR_SENT.  CAM3
1848				 * specifies CAM_REQ_CMP.
1849				 */
1850				if (scb != NULL
1851				 && scb->io_ctx->ccb_h.func_code== XPT_RESET_DEV
1852				 && ahc_match_scb(ahc, scb, target, channel,
1853						  CAM_LUN_WILDCARD,
1854						  SCB_LIST_NULL,
1855						  ROLE_INITIATOR)) {
1856					ahc_set_transaction_status(scb, CAM_REQ_CMP);
1857				}
1858#endif
1859				ahc_compile_devinfo(&devinfo,
1860						    initiator_role_id,
1861						    target,
1862						    CAM_LUN_WILDCARD,
1863						    channel,
1864						    ROLE_INITIATOR);
1865				ahc_handle_devreset(ahc, &devinfo,
1866						    CAM_BDR_SENT,
1867						    "Bus Device Reset",
1868						    /*verbose_level*/0);
1869				printerror = 0;
1870			} else if (ahc_sent_msg(ahc, AHCMSG_EXT,
1871						MSG_EXT_PPR, FALSE)) {
1872				struct ahc_initiator_tinfo *tinfo;
1873				struct ahc_tmode_tstate *tstate;
1874
1875				/*
1876				 * PPR Rejected.  Try non-ppr negotiation
1877				 * and retry command.
1878				 */
1879				tinfo = ahc_fetch_transinfo(ahc,
1880							    devinfo.channel,
1881							    devinfo.our_scsiid,
1882							    devinfo.target,
1883							    &tstate);
1884				tinfo->curr.transport_version = 2;
1885				tinfo->goal.transport_version = 2;
1886				tinfo->goal.ppr_options = 0;
1887				ahc_qinfifo_requeue_tail(ahc, scb);
1888				printerror = 0;
1889			} else if (ahc_sent_msg(ahc, AHCMSG_EXT,
1890						MSG_EXT_WDTR, FALSE)) {
1891				/*
1892				 * Negotiation Rejected.  Go-narrow and
1893				 * retry command.
1894				 */
1895				ahc_set_width(ahc, &devinfo,
1896					      MSG_EXT_WDTR_BUS_8_BIT,
1897					      AHC_TRANS_CUR|AHC_TRANS_GOAL,
1898					      /*paused*/TRUE);
1899				ahc_qinfifo_requeue_tail(ahc, scb);
1900				printerror = 0;
1901			} else if (ahc_sent_msg(ahc, AHCMSG_EXT,
1902						MSG_EXT_SDTR, FALSE)) {
1903				/*
1904				 * Negotiation Rejected.  Go-async and
1905				 * retry command.
1906				 */
1907				ahc_set_syncrate(ahc, &devinfo,
1908						/*syncrate*/NULL,
1909						/*period*/0, /*offset*/0,
1910						/*ppr_options*/0,
1911						AHC_TRANS_CUR|AHC_TRANS_GOAL,
1912						/*paused*/TRUE);
1913				ahc_qinfifo_requeue_tail(ahc, scb);
1914				printerror = 0;
1915			}
1916		}
1917		if (printerror != 0) {
1918			u_int i;
1919
1920			if (scb != NULL) {
1921				u_int tag;
1922
1923				if ((scb->hscb->control & TAG_ENB) != 0)
1924					tag = scb->hscb->tag;
1925				else
1926					tag = SCB_LIST_NULL;
1927				ahc_print_path(ahc, scb);
1928				ahc_abort_scbs(ahc, target, channel,
1929					       SCB_GET_LUN(scb), tag,
1930					       ROLE_INITIATOR,
1931					       CAM_UNEXP_BUSFREE);
1932			} else {
1933				/*
1934				 * We had not fully identified this connection,
1935				 * so we cannot abort anything.
1936				 */
1937				printk("%s: ", ahc_name(ahc));
1938			}
1939			for (i = 0; i < num_phases; i++) {
1940				if (lastphase == ahc_phase_table[i].phase)
1941					break;
1942			}
1943			if (lastphase != P_BUSFREE) {
1944				/*
1945				 * Renegotiate with this device at the
1946				 * next opportunity just in case this busfree
1947				 * is due to a negotiation mismatch with the
1948				 * device.
1949				 */
1950				ahc_force_renegotiation(ahc, &devinfo);
1951			}
1952			printk("Unexpected busfree %s\n"
1953			       "SEQADDR == 0x%x\n",
1954			       ahc_phase_table[i].phasemsg,
1955			       ahc_inb(ahc, SEQADDR0)
1956				| (ahc_inb(ahc, SEQADDR1) << 8));
1957		}
1958		ahc_outb(ahc, CLRINT, CLRSCSIINT);
1959		ahc_restart(ahc);
1960	} else {
1961		printk("%s: Missing case in ahc_handle_scsiint. status = %x\n",
1962		       ahc_name(ahc), status);
1963		ahc_outb(ahc, CLRINT, CLRSCSIINT);
1964	}
1965}
1966
1967/*
1968 * Force renegotiation to occur the next time we initiate
1969 * a command to the current device.
1970 */
1971static void
1972ahc_force_renegotiation(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
1973{
1974	struct	ahc_initiator_tinfo *targ_info;
1975	struct	ahc_tmode_tstate *tstate;
1976
1977	targ_info = ahc_fetch_transinfo(ahc,
1978					devinfo->channel,
1979					devinfo->our_scsiid,
1980					devinfo->target,
1981					&tstate);
1982	ahc_update_neg_request(ahc, devinfo, tstate,
1983			       targ_info, AHC_NEG_IF_NON_ASYNC);
1984}
1985
1986#define AHC_MAX_STEPS 2000
1987static void
1988ahc_clear_critical_section(struct ahc_softc *ahc)
1989{
1990	int	stepping;
1991	int	steps;
1992	u_int	simode0;
1993	u_int	simode1;
1994
1995	if (ahc->num_critical_sections == 0)
1996		return;
1997
1998	stepping = FALSE;
1999	steps = 0;
2000	simode0 = 0;
2001	simode1 = 0;
2002	for (;;) {
2003		struct	cs *cs;
2004		u_int	seqaddr;
2005		u_int	i;
2006
2007		seqaddr = ahc_inb(ahc, SEQADDR0)
2008			| (ahc_inb(ahc, SEQADDR1) << 8);
2009
2010		/*
2011		 * Seqaddr represents the next instruction to execute,
2012		 * so we are really executing the instruction just
2013		 * before it.
2014		 */
2015		if (seqaddr != 0)
2016			seqaddr -= 1;
2017		cs = ahc->critical_sections;
2018		for (i = 0; i < ahc->num_critical_sections; i++, cs++) {
2019
2020			if (cs->begin < seqaddr && cs->end >= seqaddr)
2021				break;
2022		}
2023
2024		if (i == ahc->num_critical_sections)
2025			break;
2026
2027		if (steps > AHC_MAX_STEPS) {
2028			printk("%s: Infinite loop in critical section\n",
2029			       ahc_name(ahc));
2030			ahc_dump_card_state(ahc);
2031			panic("critical section loop");
2032		}
2033
2034		steps++;
2035		if (stepping == FALSE) {
2036
2037			/*
2038			 * Disable all interrupt sources so that the
2039			 * sequencer will not be stuck by a pausing
2040			 * interrupt condition while we attempt to
2041			 * leave a critical section.
2042			 */
2043			simode0 = ahc_inb(ahc, SIMODE0);
2044			ahc_outb(ahc, SIMODE0, 0);
2045			simode1 = ahc_inb(ahc, SIMODE1);
2046			if ((ahc->features & AHC_DT) != 0)
2047				/*
2048				 * On DT class controllers, we
2049				 * use the enhanced busfree logic.
2050				 * Unfortunately we cannot re-enable
2051				 * busfree detection within the
2052				 * current connection, so we must
2053				 * leave it on while single stepping.
2054				 */
2055				ahc_outb(ahc, SIMODE1, simode1 & ENBUSFREE);
2056			else
2057				ahc_outb(ahc, SIMODE1, 0);
2058			ahc_outb(ahc, CLRINT, CLRSCSIINT);
2059			ahc_outb(ahc, SEQCTL, ahc->seqctl | STEP);
2060			stepping = TRUE;
2061		}
2062		if ((ahc->features & AHC_DT) != 0) {
2063			ahc_outb(ahc, CLRSINT1, CLRBUSFREE);
2064			ahc_outb(ahc, CLRINT, CLRSCSIINT);
2065		}
2066		ahc_outb(ahc, HCNTRL, ahc->unpause);
2067		while (!ahc_is_paused(ahc))
2068			ahc_delay(200);
2069	}
2070	if (stepping) {
2071		ahc_outb(ahc, SIMODE0, simode0);
2072		ahc_outb(ahc, SIMODE1, simode1);
2073		ahc_outb(ahc, SEQCTL, ahc->seqctl);
2074	}
2075}
2076
2077/*
2078 * Clear any pending interrupt status.
2079 */
2080static void
2081ahc_clear_intstat(struct ahc_softc *ahc)
2082{
2083	/* Clear any interrupt conditions this may have caused */
2084	ahc_outb(ahc, CLRSINT1, CLRSELTIMEO|CLRATNO|CLRSCSIRSTI
2085				|CLRBUSFREE|CLRSCSIPERR|CLRPHASECHG|
2086				CLRREQINIT);
2087	ahc_flush_device_writes(ahc);
2088	ahc_outb(ahc, CLRSINT0, CLRSELDO|CLRSELDI|CLRSELINGO);
2089 	ahc_flush_device_writes(ahc);
2090	ahc_outb(ahc, CLRINT, CLRSCSIINT);
2091	ahc_flush_device_writes(ahc);
2092}
2093
2094/**************************** Debugging Routines ******************************/
2095#ifdef AHC_DEBUG
2096uint32_t ahc_debug = AHC_DEBUG_OPTS;
2097#endif
2098
2099#if 0 /* unused */
2100static void
2101ahc_print_scb(struct scb *scb)
2102{
2103	int i;
2104
2105	struct hardware_scb *hscb = scb->hscb;
2106
2107	printk("scb:%p control:0x%x scsiid:0x%x lun:%d cdb_len:%d\n",
2108	       (void *)scb,
2109	       hscb->control,
2110	       hscb->scsiid,
2111	       hscb->lun,
2112	       hscb->cdb_len);
2113	printk("Shared Data: ");
2114	for (i = 0; i < sizeof(hscb->shared_data.cdb); i++)
2115		printk("%#02x", hscb->shared_data.cdb[i]);
2116	printk("        dataptr:%#x datacnt:%#x sgptr:%#x tag:%#x\n",
2117		ahc_le32toh(hscb->dataptr),
2118		ahc_le32toh(hscb->datacnt),
2119		ahc_le32toh(hscb->sgptr),
2120		hscb->tag);
2121	if (scb->sg_count > 0) {
2122		for (i = 0; i < scb->sg_count; i++) {
2123			printk("sg[%d] - Addr 0x%x%x : Length %d\n",
2124			       i,
2125			       (ahc_le32toh(scb->sg_list[i].len) >> 24
2126			        & SG_HIGH_ADDR_BITS),
2127			       ahc_le32toh(scb->sg_list[i].addr),
2128			       ahc_le32toh(scb->sg_list[i].len));
2129		}
2130	}
2131}
2132#endif
2133
2134/************************* Transfer Negotiation *******************************/
2135/*
2136 * Allocate per target mode instance (ID we respond to as a target)
2137 * transfer negotiation data structures.
2138 */
2139static struct ahc_tmode_tstate *
2140ahc_alloc_tstate(struct ahc_softc *ahc, u_int scsi_id, char channel)
2141{
2142	struct ahc_tmode_tstate *master_tstate;
2143	struct ahc_tmode_tstate *tstate;
2144	int i;
2145
2146	master_tstate = ahc->enabled_targets[ahc->our_id];
2147	if (channel == 'B') {
2148		scsi_id += 8;
2149		master_tstate = ahc->enabled_targets[ahc->our_id_b + 8];
2150	}
2151	if (ahc->enabled_targets[scsi_id] != NULL
2152	 && ahc->enabled_targets[scsi_id] != master_tstate)
2153		panic("%s: ahc_alloc_tstate - Target already allocated",
2154		      ahc_name(ahc));
2155	tstate = kmalloc(sizeof(*tstate), GFP_ATOMIC);
2156	if (tstate == NULL)
2157		return (NULL);
2158
2159	/*
2160	 * If we have allocated a master tstate, copy user settings from
2161	 * the master tstate (taken from SRAM or the EEPROM) for this
2162	 * channel, but reset our current and goal settings to async/narrow
2163	 * until an initiator talks to us.
2164	 */
2165	if (master_tstate != NULL) {
2166		memcpy(tstate, master_tstate, sizeof(*tstate));
2167		memset(tstate->enabled_luns, 0, sizeof(tstate->enabled_luns));
2168		tstate->ultraenb = 0;
2169		for (i = 0; i < AHC_NUM_TARGETS; i++) {
2170			memset(&tstate->transinfo[i].curr, 0,
2171			      sizeof(tstate->transinfo[i].curr));
2172			memset(&tstate->transinfo[i].goal, 0,
2173			      sizeof(tstate->transinfo[i].goal));
2174		}
2175	} else
2176		memset(tstate, 0, sizeof(*tstate));
2177	ahc->enabled_targets[scsi_id] = tstate;
2178	return (tstate);
2179}
2180
2181#ifdef AHC_TARGET_MODE
2182/*
2183 * Free per target mode instance (ID we respond to as a target)
2184 * transfer negotiation data structures.
2185 */
2186static void
2187ahc_free_tstate(struct ahc_softc *ahc, u_int scsi_id, char channel, int force)
2188{
2189	struct ahc_tmode_tstate *tstate;
2190
2191	/*
2192	 * Don't clean up our "master" tstate.
2193	 * It has our default user settings.
2194	 */
2195	if (((channel == 'B' && scsi_id == ahc->our_id_b)
2196	  || (channel == 'A' && scsi_id == ahc->our_id))
2197	 && force == FALSE)
2198		return;
2199
2200	if (channel == 'B')
2201		scsi_id += 8;
2202	tstate = ahc->enabled_targets[scsi_id];
2203	if (tstate != NULL)
2204		kfree(tstate);
2205	ahc->enabled_targets[scsi_id] = NULL;
2206}
2207#endif
2208
2209/*
2210 * Called when we have an active connection to a target on the bus,
2211 * this function finds the nearest syncrate to the input period limited
2212 * by the capabilities of the bus connectivity of and sync settings for
2213 * the target.
2214 */
2215const struct ahc_syncrate *
2216ahc_devlimited_syncrate(struct ahc_softc *ahc,
2217			struct ahc_initiator_tinfo *tinfo,
2218			u_int *period, u_int *ppr_options, role_t role)
2219{
2220	struct	ahc_transinfo *transinfo;
2221	u_int	maxsync;
2222
2223	if ((ahc->features & AHC_ULTRA2) != 0) {
2224		if ((ahc_inb(ahc, SBLKCTL) & ENAB40) != 0
2225		 && (ahc_inb(ahc, SSTAT2) & EXP_ACTIVE) == 0) {
2226			maxsync = AHC_SYNCRATE_DT;
2227		} else {
2228			maxsync = AHC_SYNCRATE_ULTRA;
2229			/* Can't do DT on an SE bus */
2230			*ppr_options &= ~MSG_EXT_PPR_DT_REQ;
2231		}
2232	} else if ((ahc->features & AHC_ULTRA) != 0) {
2233		maxsync = AHC_SYNCRATE_ULTRA;
2234	} else {
2235		maxsync = AHC_SYNCRATE_FAST;
2236	}
2237	/*
2238	 * Never allow a value higher than our current goal
2239	 * period otherwise we may allow a target initiated
2240	 * negotiation to go above the limit as set by the
2241	 * user.  In the case of an initiator initiated
2242	 * sync negotiation, we limit based on the user
2243	 * setting.  This allows the system to still accept
2244	 * incoming negotiations even if target initiated
2245	 * negotiation is not performed.
2246	 */
2247	if (role == ROLE_TARGET)
2248		transinfo = &tinfo->user;
2249	else
2250		transinfo = &tinfo->goal;
2251	*ppr_options &= transinfo->ppr_options;
2252	if (transinfo->width == MSG_EXT_WDTR_BUS_8_BIT) {
2253		maxsync = max(maxsync, (u_int)AHC_SYNCRATE_ULTRA2);
2254		*ppr_options &= ~MSG_EXT_PPR_DT_REQ;
2255	}
2256	if (transinfo->period == 0) {
2257		*period = 0;
2258		*ppr_options = 0;
2259		return (NULL);
2260	}
2261	*period = max(*period, (u_int)transinfo->period);
2262	return (ahc_find_syncrate(ahc, period, ppr_options, maxsync));
2263}
2264
2265/*
2266 * Look up the valid period to SCSIRATE conversion in our table.
2267 * Return the period and offset that should be sent to the target
2268 * if this was the beginning of an SDTR.
2269 */
2270const struct ahc_syncrate *
2271ahc_find_syncrate(struct ahc_softc *ahc, u_int *period,
2272		  u_int *ppr_options, u_int maxsync)
2273{
2274	const struct ahc_syncrate *syncrate;
2275
2276	if ((ahc->features & AHC_DT) == 0)
2277		*ppr_options &= ~MSG_EXT_PPR_DT_REQ;
2278
2279	/* Skip all DT only entries if DT is not available */
2280	if ((*ppr_options & MSG_EXT_PPR_DT_REQ) == 0
2281	 && maxsync < AHC_SYNCRATE_ULTRA2)
2282		maxsync = AHC_SYNCRATE_ULTRA2;
2283
2284	/* Now set the maxsync based on the card capabilities
2285	 * DT is already done above */
2286	if ((ahc->features & (AHC_DT | AHC_ULTRA2)) == 0
2287	    && maxsync < AHC_SYNCRATE_ULTRA)
2288		maxsync = AHC_SYNCRATE_ULTRA;
2289	if ((ahc->features & (AHC_DT | AHC_ULTRA2 | AHC_ULTRA)) == 0
2290	    && maxsync < AHC_SYNCRATE_FAST)
2291		maxsync = AHC_SYNCRATE_FAST;
2292
2293	for (syncrate = &ahc_syncrates[maxsync];
2294	     syncrate->rate != NULL;
2295	     syncrate++) {
2296
2297		/*
2298		 * The Ultra2 table doesn't go as low
2299		 * as for the Fast/Ultra cards.
2300		 */
2301		if ((ahc->features & AHC_ULTRA2) != 0
2302		 && (syncrate->sxfr_u2 == 0))
2303			break;
2304
2305		if (*period <= syncrate->period) {
2306			/*
2307			 * When responding to a target that requests
2308			 * sync, the requested rate may fall between
2309			 * two rates that we can output, but still be
2310			 * a rate that we can receive.  Because of this,
2311			 * we want to respond to the target with
2312			 * the same rate that it sent to us even
2313			 * if the period we use to send data to it
2314			 * is lower.  Only lower the response period
2315			 * if we must.
2316			 */
2317			if (syncrate == &ahc_syncrates[maxsync])
2318				*period = syncrate->period;
2319
2320			/*
2321			 * At some speeds, we only support
2322			 * ST transfers.
2323			 */
2324		 	if ((syncrate->sxfr_u2 & ST_SXFR) != 0)
2325				*ppr_options &= ~MSG_EXT_PPR_DT_REQ;
2326			break;
2327		}
2328	}
2329
2330	if ((*period == 0)
2331	 || (syncrate->rate == NULL)
2332	 || ((ahc->features & AHC_ULTRA2) != 0
2333	  && (syncrate->sxfr_u2 == 0))) {
2334		/* Use asynchronous transfers. */
2335		*period = 0;
2336		syncrate = NULL;
2337		*ppr_options &= ~MSG_EXT_PPR_DT_REQ;
2338	}
2339	return (syncrate);
2340}
2341
2342/*
2343 * Convert from an entry in our syncrate table to the SCSI equivalent
2344 * sync "period" factor.
2345 */
2346u_int
2347ahc_find_period(struct ahc_softc *ahc, u_int scsirate, u_int maxsync)
2348{
2349	const struct ahc_syncrate *syncrate;
2350
2351	if ((ahc->features & AHC_ULTRA2) != 0)
2352		scsirate &= SXFR_ULTRA2;
2353	else
2354		scsirate &= SXFR;
2355
2356	/* now set maxsync based on card capabilities */
2357	if ((ahc->features & AHC_DT) == 0 && maxsync < AHC_SYNCRATE_ULTRA2)
2358		maxsync = AHC_SYNCRATE_ULTRA2;
2359	if ((ahc->features & (AHC_DT | AHC_ULTRA2)) == 0
2360	    && maxsync < AHC_SYNCRATE_ULTRA)
2361		maxsync = AHC_SYNCRATE_ULTRA;
2362	if ((ahc->features & (AHC_DT | AHC_ULTRA2 | AHC_ULTRA)) == 0
2363	    && maxsync < AHC_SYNCRATE_FAST)
2364		maxsync = AHC_SYNCRATE_FAST;
2365
2366
2367	syncrate = &ahc_syncrates[maxsync];
2368	while (syncrate->rate != NULL) {
2369
2370		if ((ahc->features & AHC_ULTRA2) != 0) {
2371			if (syncrate->sxfr_u2 == 0)
2372				break;
2373			else if (scsirate == (syncrate->sxfr_u2 & SXFR_ULTRA2))
2374				return (syncrate->period);
2375		} else if (scsirate == (syncrate->sxfr & SXFR)) {
2376				return (syncrate->period);
2377		}
2378		syncrate++;
2379	}
2380	return (0); /* async */
2381}
2382
2383/*
2384 * Truncate the given synchronous offset to a value the
2385 * current adapter type and syncrate are capable of.
2386 */
2387static void
2388ahc_validate_offset(struct ahc_softc *ahc,
2389		    struct ahc_initiator_tinfo *tinfo,
2390		    const struct ahc_syncrate *syncrate,
2391		    u_int *offset, int wide, role_t role)
2392{
2393	u_int maxoffset;
2394
2395	/* Limit offset to what we can do */
2396	if (syncrate == NULL) {
2397		maxoffset = 0;
2398	} else if ((ahc->features & AHC_ULTRA2) != 0) {
2399		maxoffset = MAX_OFFSET_ULTRA2;
2400	} else {
2401		if (wide)
2402			maxoffset = MAX_OFFSET_16BIT;
2403		else
2404			maxoffset = MAX_OFFSET_8BIT;
2405	}
2406	*offset = min(*offset, maxoffset);
2407	if (tinfo != NULL) {
2408		if (role == ROLE_TARGET)
2409			*offset = min(*offset, (u_int)tinfo->user.offset);
2410		else
2411			*offset = min(*offset, (u_int)tinfo->goal.offset);
2412	}
2413}
2414
2415/*
2416 * Truncate the given transfer width parameter to a value the
2417 * current adapter type is capable of.
2418 */
2419static void
2420ahc_validate_width(struct ahc_softc *ahc, struct ahc_initiator_tinfo *tinfo,
2421		   u_int *bus_width, role_t role)
2422{
2423	switch (*bus_width) {
2424	default:
2425		if (ahc->features & AHC_WIDE) {
2426			/* Respond Wide */
2427			*bus_width = MSG_EXT_WDTR_BUS_16_BIT;
2428			break;
2429		}
2430		/* FALLTHROUGH */
2431	case MSG_EXT_WDTR_BUS_8_BIT:
2432		*bus_width = MSG_EXT_WDTR_BUS_8_BIT;
2433		break;
2434	}
2435	if (tinfo != NULL) {
2436		if (role == ROLE_TARGET)
2437			*bus_width = min((u_int)tinfo->user.width, *bus_width);
2438		else
2439			*bus_width = min((u_int)tinfo->goal.width, *bus_width);
2440	}
2441}
2442
2443/*
2444 * Update the bitmask of targets for which the controller should
2445 * negotiate with at the next convenient opportunity.  This currently
2446 * means the next time we send the initial identify messages for
2447 * a new transaction.
2448 */
2449int
2450ahc_update_neg_request(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
2451		       struct ahc_tmode_tstate *tstate,
2452		       struct ahc_initiator_tinfo *tinfo, ahc_neg_type neg_type)
2453{
2454	u_int auto_negotiate_orig;
2455
2456	auto_negotiate_orig = tstate->auto_negotiate;
2457	if (neg_type == AHC_NEG_ALWAYS) {
2458		/*
2459		 * Force our "current" settings to be
2460		 * unknown so that unless a bus reset
2461		 * occurs the need to renegotiate is
2462		 * recorded persistently.
2463		 */
2464		if ((ahc->features & AHC_WIDE) != 0)
2465			tinfo->curr.width = AHC_WIDTH_UNKNOWN;
2466		tinfo->curr.period = AHC_PERIOD_UNKNOWN;
2467		tinfo->curr.offset = AHC_OFFSET_UNKNOWN;
2468	}
2469	if (tinfo->curr.period != tinfo->goal.period
2470	 || tinfo->curr.width != tinfo->goal.width
2471	 || tinfo->curr.offset != tinfo->goal.offset
2472	 || tinfo->curr.ppr_options != tinfo->goal.ppr_options
2473	 || (neg_type == AHC_NEG_IF_NON_ASYNC
2474	  && (tinfo->goal.offset != 0
2475	   || tinfo->goal.width != MSG_EXT_WDTR_BUS_8_BIT
2476	   || tinfo->goal.ppr_options != 0)))
2477		tstate->auto_negotiate |= devinfo->target_mask;
2478	else
2479		tstate->auto_negotiate &= ~devinfo->target_mask;
2480
2481	return (auto_negotiate_orig != tstate->auto_negotiate);
2482}
2483
2484/*
2485 * Update the user/goal/curr tables of synchronous negotiation
2486 * parameters as well as, in the case of a current or active update,
2487 * any data structures on the host controller.  In the case of an
2488 * active update, the specified target is currently talking to us on
2489 * the bus, so the transfer parameter update must take effect
2490 * immediately.
2491 */
2492void
2493ahc_set_syncrate(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
2494		 const struct ahc_syncrate *syncrate, u_int period,
2495		 u_int offset, u_int ppr_options, u_int type, int paused)
2496{
2497	struct	ahc_initiator_tinfo *tinfo;
2498	struct	ahc_tmode_tstate *tstate;
2499	u_int	old_period;
2500	u_int	old_offset;
2501	u_int	old_ppr;
2502	int	active;
2503	int	update_needed;
2504
2505	active = (type & AHC_TRANS_ACTIVE) == AHC_TRANS_ACTIVE;
2506	update_needed = 0;
2507
2508	if (syncrate == NULL) {
2509		period = 0;
2510		offset = 0;
2511	}
2512
2513	tinfo = ahc_fetch_transinfo(ahc, devinfo->channel, devinfo->our_scsiid,
2514				    devinfo->target, &tstate);
2515
2516	if ((type & AHC_TRANS_USER) != 0) {
2517		tinfo->user.period = period;
2518		tinfo->user.offset = offset;
2519		tinfo->user.ppr_options = ppr_options;
2520	}
2521
2522	if ((type & AHC_TRANS_GOAL) != 0) {
2523		tinfo->goal.period = period;
2524		tinfo->goal.offset = offset;
2525		tinfo->goal.ppr_options = ppr_options;
2526	}
2527
2528	old_period = tinfo->curr.period;
2529	old_offset = tinfo->curr.offset;
2530	old_ppr	   = tinfo->curr.ppr_options;
2531
2532	if ((type & AHC_TRANS_CUR) != 0
2533	 && (old_period != period
2534	  || old_offset != offset
2535	  || old_ppr != ppr_options)) {
2536		u_int	scsirate;
2537
2538		update_needed++;
2539		scsirate = tinfo->scsirate;
2540		if ((ahc->features & AHC_ULTRA2) != 0) {
2541
2542			scsirate &= ~(SXFR_ULTRA2|SINGLE_EDGE|ENABLE_CRC);
2543			if (syncrate != NULL) {
2544				scsirate |= syncrate->sxfr_u2;
2545				if ((ppr_options & MSG_EXT_PPR_DT_REQ) != 0)
2546					scsirate |= ENABLE_CRC;
2547				else
2548					scsirate |= SINGLE_EDGE;
2549			}
2550		} else {
2551
2552			scsirate &= ~(SXFR|SOFS);
2553			/*
2554			 * Ensure Ultra mode is set properly for
2555			 * this target.
2556			 */
2557			tstate->ultraenb &= ~devinfo->target_mask;
2558			if (syncrate != NULL) {
2559				if (syncrate->sxfr & ULTRA_SXFR) {
2560					tstate->ultraenb |=
2561						devinfo->target_mask;
2562				}
2563				scsirate |= syncrate->sxfr & SXFR;
2564				scsirate |= offset & SOFS;
2565			}
2566			if (active) {
2567				u_int sxfrctl0;
2568
2569				sxfrctl0 = ahc_inb(ahc, SXFRCTL0);
2570				sxfrctl0 &= ~FAST20;
2571				if (tstate->ultraenb & devinfo->target_mask)
2572					sxfrctl0 |= FAST20;
2573				ahc_outb(ahc, SXFRCTL0, sxfrctl0);
2574			}
2575		}
2576		if (active) {
2577			ahc_outb(ahc, SCSIRATE, scsirate);
2578			if ((ahc->features & AHC_ULTRA2) != 0)
2579				ahc_outb(ahc, SCSIOFFSET, offset);
2580		}
2581
2582		tinfo->scsirate = scsirate;
2583		tinfo->curr.period = period;
2584		tinfo->curr.offset = offset;
2585		tinfo->curr.ppr_options = ppr_options;
2586
2587		ahc_send_async(ahc, devinfo->channel, devinfo->target,
2588			       CAM_LUN_WILDCARD, AC_TRANSFER_NEG);
2589		if (bootverbose) {
2590			if (offset != 0) {
2591				printk("%s: target %d synchronous at %sMHz%s, "
2592				       "offset = 0x%x\n", ahc_name(ahc),
2593				       devinfo->target, syncrate->rate,
2594				       (ppr_options & MSG_EXT_PPR_DT_REQ)
2595				       ? " DT" : "", offset);
2596			} else {
2597				printk("%s: target %d using "
2598				       "asynchronous transfers\n",
2599				       ahc_name(ahc), devinfo->target);
2600			}
2601		}
2602	}
2603
2604	update_needed += ahc_update_neg_request(ahc, devinfo, tstate,
2605						tinfo, AHC_NEG_TO_GOAL);
2606
2607	if (update_needed)
2608		ahc_update_pending_scbs(ahc);
2609}
2610
2611/*
2612 * Update the user/goal/curr tables of wide negotiation
2613 * parameters as well as, in the case of a current or active update,
2614 * any data structures on the host controller.  In the case of an
2615 * active update, the specified target is currently talking to us on
2616 * the bus, so the transfer parameter update must take effect
2617 * immediately.
2618 */
2619void
2620ahc_set_width(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
2621	      u_int width, u_int type, int paused)
2622{
2623	struct	ahc_initiator_tinfo *tinfo;
2624	struct	ahc_tmode_tstate *tstate;
2625	u_int	oldwidth;
2626	int	active;
2627	int	update_needed;
2628
2629	active = (type & AHC_TRANS_ACTIVE) == AHC_TRANS_ACTIVE;
2630	update_needed = 0;
2631	tinfo = ahc_fetch_transinfo(ahc, devinfo->channel, devinfo->our_scsiid,
2632				    devinfo->target, &tstate);
2633
2634	if ((type & AHC_TRANS_USER) != 0)
2635		tinfo->user.width = width;
2636
2637	if ((type & AHC_TRANS_GOAL) != 0)
2638		tinfo->goal.width = width;
2639
2640	oldwidth = tinfo->curr.width;
2641	if ((type & AHC_TRANS_CUR) != 0 && oldwidth != width) {
2642		u_int	scsirate;
2643
2644		update_needed++;
2645		scsirate =  tinfo->scsirate;
2646		scsirate &= ~WIDEXFER;
2647		if (width == MSG_EXT_WDTR_BUS_16_BIT)
2648			scsirate |= WIDEXFER;
2649
2650		tinfo->scsirate = scsirate;
2651
2652		if (active)
2653			ahc_outb(ahc, SCSIRATE, scsirate);
2654
2655		tinfo->curr.width = width;
2656
2657		ahc_send_async(ahc, devinfo->channel, devinfo->target,
2658			       CAM_LUN_WILDCARD, AC_TRANSFER_NEG);
2659		if (bootverbose) {
2660			printk("%s: target %d using %dbit transfers\n",
2661			       ahc_name(ahc), devinfo->target,
2662			       8 * (0x01 << width));
2663		}
2664	}
2665
2666	update_needed += ahc_update_neg_request(ahc, devinfo, tstate,
2667						tinfo, AHC_NEG_TO_GOAL);
2668	if (update_needed)
2669		ahc_update_pending_scbs(ahc);
2670}
2671
2672/*
2673 * Update the current state of tagged queuing for a given target.
2674 */
2675static void
2676ahc_set_tags(struct ahc_softc *ahc, struct scsi_cmnd *cmd,
2677	     struct ahc_devinfo *devinfo, ahc_queue_alg alg)
2678{
2679	struct scsi_device *sdev = cmd->device;
2680
2681 	ahc_platform_set_tags(ahc, sdev, devinfo, alg);
2682 	ahc_send_async(ahc, devinfo->channel, devinfo->target,
2683 		       devinfo->lun, AC_TRANSFER_NEG);
2684}
2685
2686/*
2687 * When the transfer settings for a connection change, update any
2688 * in-transit SCBs to contain the new data so the hardware will
2689 * be set correctly during future (re)selections.
2690 */
2691static void
2692ahc_update_pending_scbs(struct ahc_softc *ahc)
2693{
2694	struct	scb *pending_scb;
2695	int	pending_scb_count;
2696	int	i;
2697	int	paused;
2698	u_int	saved_scbptr;
2699
2700	/*
2701	 * Traverse the pending SCB list and ensure that all of the
2702	 * SCBs there have the proper settings.
2703	 */
2704	pending_scb_count = 0;
2705	LIST_FOREACH(pending_scb, &ahc->pending_scbs, pending_links) {
2706		struct ahc_devinfo devinfo;
2707		struct hardware_scb *pending_hscb;
2708		struct ahc_initiator_tinfo *tinfo;
2709		struct ahc_tmode_tstate *tstate;
2710
2711		ahc_scb_devinfo(ahc, &devinfo, pending_scb);
2712		tinfo = ahc_fetch_transinfo(ahc, devinfo.channel,
2713					    devinfo.our_scsiid,
2714					    devinfo.target, &tstate);
2715		pending_hscb = pending_scb->hscb;
2716		pending_hscb->control &= ~ULTRAENB;
2717		if ((tstate->ultraenb & devinfo.target_mask) != 0)
2718			pending_hscb->control |= ULTRAENB;
2719		pending_hscb->scsirate = tinfo->scsirate;
2720		pending_hscb->scsioffset = tinfo->curr.offset;
2721		if ((tstate->auto_negotiate & devinfo.target_mask) == 0
2722		 && (pending_scb->flags & SCB_AUTO_NEGOTIATE) != 0) {
2723			pending_scb->flags &= ~SCB_AUTO_NEGOTIATE;
2724			pending_hscb->control &= ~MK_MESSAGE;
2725		}
2726		ahc_sync_scb(ahc, pending_scb,
2727			     BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
2728		pending_scb_count++;
2729	}
2730
2731	if (pending_scb_count == 0)
2732		return;
2733
2734	if (ahc_is_paused(ahc)) {
2735		paused = 1;
2736	} else {
2737		paused = 0;
2738		ahc_pause(ahc);
2739	}
2740
2741	saved_scbptr = ahc_inb(ahc, SCBPTR);
2742	/* Ensure that the hscbs down on the card match the new information */
2743	for (i = 0; i < ahc->scb_data->maxhscbs; i++) {
2744		struct	hardware_scb *pending_hscb;
2745		u_int	control;
2746		u_int	scb_tag;
2747
2748		ahc_outb(ahc, SCBPTR, i);
2749		scb_tag = ahc_inb(ahc, SCB_TAG);
2750		pending_scb = ahc_lookup_scb(ahc, scb_tag);
2751		if (pending_scb == NULL)
2752			continue;
2753
2754		pending_hscb = pending_scb->hscb;
2755		control = ahc_inb(ahc, SCB_CONTROL);
2756		control &= ~(ULTRAENB|MK_MESSAGE);
2757		control |= pending_hscb->control & (ULTRAENB|MK_MESSAGE);
2758		ahc_outb(ahc, SCB_CONTROL, control);
2759		ahc_outb(ahc, SCB_SCSIRATE, pending_hscb->scsirate);
2760		ahc_outb(ahc, SCB_SCSIOFFSET, pending_hscb->scsioffset);
2761	}
2762	ahc_outb(ahc, SCBPTR, saved_scbptr);
2763
2764	if (paused == 0)
2765		ahc_unpause(ahc);
2766}
2767
2768/**************************** Pathing Information *****************************/
2769static void
2770ahc_fetch_devinfo(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
2771{
2772	u_int	saved_scsiid;
2773	role_t	role;
2774	int	our_id;
2775
2776	if (ahc_inb(ahc, SSTAT0) & TARGET)
2777		role = ROLE_TARGET;
2778	else
2779		role = ROLE_INITIATOR;
2780
2781	if (role == ROLE_TARGET
2782	 && (ahc->features & AHC_MULTI_TID) != 0
2783	 && (ahc_inb(ahc, SEQ_FLAGS)
2784 	   & (CMDPHASE_PENDING|TARG_CMD_PENDING|NO_DISCONNECT)) != 0) {
2785		/* We were selected, so pull our id from TARGIDIN */
2786		our_id = ahc_inb(ahc, TARGIDIN) & OID;
2787	} else if ((ahc->features & AHC_ULTRA2) != 0)
2788		our_id = ahc_inb(ahc, SCSIID_ULTRA2) & OID;
2789	else
2790		our_id = ahc_inb(ahc, SCSIID) & OID;
2791
2792	saved_scsiid = ahc_inb(ahc, SAVED_SCSIID);
2793	ahc_compile_devinfo(devinfo,
2794			    our_id,
2795			    SCSIID_TARGET(ahc, saved_scsiid),
2796			    ahc_inb(ahc, SAVED_LUN),
2797			    SCSIID_CHANNEL(ahc, saved_scsiid),
2798			    role);
2799}
2800
2801static const struct ahc_phase_table_entry*
2802ahc_lookup_phase_entry(int phase)
2803{
2804	const struct ahc_phase_table_entry *entry;
2805	const struct ahc_phase_table_entry *last_entry;
2806
2807	/*
2808	 * num_phases doesn't include the default entry which
2809	 * will be returned if the phase doesn't match.
2810	 */
2811	last_entry = &ahc_phase_table[num_phases];
2812	for (entry = ahc_phase_table; entry < last_entry; entry++) {
2813		if (phase == entry->phase)
2814			break;
2815	}
2816	return (entry);
2817}
2818
2819void
2820ahc_compile_devinfo(struct ahc_devinfo *devinfo, u_int our_id, u_int target,
2821		    u_int lun, char channel, role_t role)
2822{
2823	devinfo->our_scsiid = our_id;
2824	devinfo->target = target;
2825	devinfo->lun = lun;
2826	devinfo->target_offset = target;
2827	devinfo->channel = channel;
2828	devinfo->role = role;
2829	if (channel == 'B')
2830		devinfo->target_offset += 8;
2831	devinfo->target_mask = (0x01 << devinfo->target_offset);
2832}
2833
2834void
2835ahc_print_devinfo(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
2836{
2837	printk("%s:%c:%d:%d: ", ahc_name(ahc), devinfo->channel,
2838	       devinfo->target, devinfo->lun);
2839}
2840
2841static void
2842ahc_scb_devinfo(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
2843		struct scb *scb)
2844{
2845	role_t	role;
2846	int	our_id;
2847
2848	our_id = SCSIID_OUR_ID(scb->hscb->scsiid);
2849	role = ROLE_INITIATOR;
2850	if ((scb->flags & SCB_TARGET_SCB) != 0)
2851		role = ROLE_TARGET;
2852	ahc_compile_devinfo(devinfo, our_id, SCB_GET_TARGET(ahc, scb),
2853			    SCB_GET_LUN(scb), SCB_GET_CHANNEL(ahc, scb), role);
2854}
2855
2856
2857/************************ Message Phase Processing ****************************/
2858static void
2859ahc_assert_atn(struct ahc_softc *ahc)
2860{
2861	u_int scsisigo;
2862
2863	scsisigo = ATNO;
2864	if ((ahc->features & AHC_DT) == 0)
2865		scsisigo |= ahc_inb(ahc, SCSISIGI);
2866	ahc_outb(ahc, SCSISIGO, scsisigo);
2867}
2868
2869/*
2870 * When an initiator transaction with the MK_MESSAGE flag either reconnects
2871 * or enters the initial message out phase, we are interrupted.  Fill our
2872 * outgoing message buffer with the appropriate message and beging handing
2873 * the message phase(s) manually.
2874 */
2875static void
2876ahc_setup_initiator_msgout(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
2877			   struct scb *scb)
2878{
2879	/*
2880	 * To facilitate adding multiple messages together,
2881	 * each routine should increment the index and len
2882	 * variables instead of setting them explicitly.
2883	 */
2884	ahc->msgout_index = 0;
2885	ahc->msgout_len = 0;
2886
2887	if ((scb->flags & SCB_DEVICE_RESET) == 0
2888	 && ahc_inb(ahc, MSG_OUT) == MSG_IDENTIFYFLAG) {
2889		u_int identify_msg;
2890
2891		identify_msg = MSG_IDENTIFYFLAG | SCB_GET_LUN(scb);
2892		if ((scb->hscb->control & DISCENB) != 0)
2893			identify_msg |= MSG_IDENTIFY_DISCFLAG;
2894		ahc->msgout_buf[ahc->msgout_index++] = identify_msg;
2895		ahc->msgout_len++;
2896
2897		if ((scb->hscb->control & TAG_ENB) != 0) {
2898			ahc->msgout_buf[ahc->msgout_index++] =
2899			    scb->hscb->control & (TAG_ENB|SCB_TAG_TYPE);
2900			ahc->msgout_buf[ahc->msgout_index++] = scb->hscb->tag;
2901			ahc->msgout_len += 2;
2902		}
2903	}
2904
2905	if (scb->flags & SCB_DEVICE_RESET) {
2906		ahc->msgout_buf[ahc->msgout_index++] = MSG_BUS_DEV_RESET;
2907		ahc->msgout_len++;
2908		ahc_print_path(ahc, scb);
2909		printk("Bus Device Reset Message Sent\n");
2910		/*
2911		 * Clear our selection hardware in advance of
2912		 * the busfree.  We may have an entry in the waiting
2913		 * Q for this target, and we don't want to go about
2914		 * selecting while we handle the busfree and blow it
2915		 * away.
2916		 */
2917		ahc_outb(ahc, SCSISEQ, (ahc_inb(ahc, SCSISEQ) & ~ENSELO));
2918	} else if ((scb->flags & SCB_ABORT) != 0) {
2919		if ((scb->hscb->control & TAG_ENB) != 0)
2920			ahc->msgout_buf[ahc->msgout_index++] = MSG_ABORT_TAG;
2921		else
2922			ahc->msgout_buf[ahc->msgout_index++] = MSG_ABORT;
2923		ahc->msgout_len++;
2924		ahc_print_path(ahc, scb);
2925		printk("Abort%s Message Sent\n",
2926		       (scb->hscb->control & TAG_ENB) != 0 ? " Tag" : "");
2927		/*
2928		 * Clear our selection hardware in advance of
2929		 * the busfree.  We may have an entry in the waiting
2930		 * Q for this target, and we don't want to go about
2931		 * selecting while we handle the busfree and blow it
2932		 * away.
2933		 */
2934		ahc_outb(ahc, SCSISEQ, (ahc_inb(ahc, SCSISEQ) & ~ENSELO));
2935	} else if ((scb->flags & (SCB_AUTO_NEGOTIATE|SCB_NEGOTIATE)) != 0) {
2936		ahc_build_transfer_msg(ahc, devinfo);
2937	} else {
2938		printk("ahc_intr: AWAITING_MSG for an SCB that "
2939		       "does not have a waiting message\n");
2940		printk("SCSIID = %x, target_mask = %x\n", scb->hscb->scsiid,
2941		       devinfo->target_mask);
2942		panic("SCB = %d, SCB Control = %x, MSG_OUT = %x "
2943		      "SCB flags = %x", scb->hscb->tag, scb->hscb->control,
2944		      ahc_inb(ahc, MSG_OUT), scb->flags);
2945	}
2946
2947	/*
2948	 * Clear the MK_MESSAGE flag from the SCB so we aren't
2949	 * asked to send this message again.
2950	 */
2951	ahc_outb(ahc, SCB_CONTROL, ahc_inb(ahc, SCB_CONTROL) & ~MK_MESSAGE);
2952	scb->hscb->control &= ~MK_MESSAGE;
2953	ahc->msgout_index = 0;
2954	ahc->msg_type = MSG_TYPE_INITIATOR_MSGOUT;
2955}
2956
2957/*
2958 * Build an appropriate transfer negotiation message for the
2959 * currently active target.
2960 */
2961static void
2962ahc_build_transfer_msg(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
2963{
2964	/*
2965	 * We need to initiate transfer negotiations.
2966	 * If our current and goal settings are identical,
2967	 * we want to renegotiate due to a check condition.
2968	 */
2969	struct	ahc_initiator_tinfo *tinfo;
2970	struct	ahc_tmode_tstate *tstate;
2971	const struct ahc_syncrate *rate;
2972	int	dowide;
2973	int	dosync;
2974	int	doppr;
2975	u_int	period;
2976	u_int	ppr_options;
2977	u_int	offset;
2978
2979	tinfo = ahc_fetch_transinfo(ahc, devinfo->channel, devinfo->our_scsiid,
2980				    devinfo->target, &tstate);
2981	/*
2982	 * Filter our period based on the current connection.
2983	 * If we can't perform DT transfers on this segment (not in LVD
2984	 * mode for instance), then our decision to issue a PPR message
2985	 * may change.
2986	 */
2987	period = tinfo->goal.period;
2988	offset = tinfo->goal.offset;
2989	ppr_options = tinfo->goal.ppr_options;
2990	/* Target initiated PPR is not allowed in the SCSI spec */
2991	if (devinfo->role == ROLE_TARGET)
2992		ppr_options = 0;
2993	rate = ahc_devlimited_syncrate(ahc, tinfo, &period,
2994				       &ppr_options, devinfo->role);
2995	dowide = tinfo->curr.width != tinfo->goal.width;
2996	dosync = tinfo->curr.offset != offset || tinfo->curr.period != period;
2997	/*
2998	 * Only use PPR if we have options that need it, even if the device
2999	 * claims to support it.  There might be an expander in the way
3000	 * that doesn't.
3001	 */
3002	doppr = ppr_options != 0;
3003
3004	if (!dowide && !dosync && !doppr) {
3005		dowide = tinfo->goal.width != MSG_EXT_WDTR_BUS_8_BIT;
3006		dosync = tinfo->goal.offset != 0;
3007	}
3008
3009	if (!dowide && !dosync && !doppr) {
3010		/*
3011		 * Force async with a WDTR message if we have a wide bus,
3012		 * or just issue an SDTR with a 0 offset.
3013		 */
3014		if ((ahc->features & AHC_WIDE) != 0)
3015			dowide = 1;
3016		else
3017			dosync = 1;
3018
3019		if (bootverbose) {
3020			ahc_print_devinfo(ahc, devinfo);
3021			printk("Ensuring async\n");
3022		}
3023	}
3024
3025	/* Target initiated PPR is not allowed in the SCSI spec */
3026	if (devinfo->role == ROLE_TARGET)
3027		doppr = 0;
3028
3029	/*
3030	 * Both the PPR message and SDTR message require the
3031	 * goal syncrate to be limited to what the target device
3032	 * is capable of handling (based on whether an LVD->SE
3033	 * expander is on the bus), so combine these two cases.
3034	 * Regardless, guarantee that if we are using WDTR and SDTR
3035	 * messages that WDTR comes first.
3036	 */
3037	if (doppr || (dosync && !dowide)) {
3038
3039		offset = tinfo->goal.offset;
3040		ahc_validate_offset(ahc, tinfo, rate, &offset,
3041				    doppr ? tinfo->goal.width
3042					  : tinfo->curr.width,
3043				    devinfo->role);
3044		if (doppr) {
3045			ahc_construct_ppr(ahc, devinfo, period, offset,
3046					  tinfo->goal.width, ppr_options);
3047		} else {
3048			ahc_construct_sdtr(ahc, devinfo, period, offset);
3049		}
3050	} else {
3051		ahc_construct_wdtr(ahc, devinfo, tinfo->goal.width);
3052	}
3053}
3054
3055/*
3056 * Build a synchronous negotiation message in our message
3057 * buffer based on the input parameters.
3058 */
3059static void
3060ahc_construct_sdtr(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
3061		   u_int period, u_int offset)
3062{
3063	if (offset == 0)
3064		period = AHC_ASYNC_XFER_PERIOD;
3065	ahc->msgout_index += spi_populate_sync_msg(
3066			ahc->msgout_buf + ahc->msgout_index, period, offset);
3067	ahc->msgout_len += 5;
3068	if (bootverbose) {
3069		printk("(%s:%c:%d:%d): Sending SDTR period %x, offset %x\n",
3070		       ahc_name(ahc), devinfo->channel, devinfo->target,
3071		       devinfo->lun, period, offset);
3072	}
3073}
3074
3075/*
3076 * Build a wide negotiation message in our message
3077 * buffer based on the input parameters.
3078 */
3079static void
3080ahc_construct_wdtr(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
3081		   u_int bus_width)
3082{
3083	ahc->msgout_index += spi_populate_width_msg(
3084			ahc->msgout_buf + ahc->msgout_index, bus_width);
3085	ahc->msgout_len += 4;
3086	if (bootverbose) {
3087		printk("(%s:%c:%d:%d): Sending WDTR %x\n",
3088		       ahc_name(ahc), devinfo->channel, devinfo->target,
3089		       devinfo->lun, bus_width);
3090	}
3091}
3092
3093/*
3094 * Build a parallel protocol request message in our message
3095 * buffer based on the input parameters.
3096 */
3097static void
3098ahc_construct_ppr(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
3099		  u_int period, u_int offset, u_int bus_width,
3100		  u_int ppr_options)
3101{
3102	if (offset == 0)
3103		period = AHC_ASYNC_XFER_PERIOD;
3104	ahc->msgout_index += spi_populate_ppr_msg(
3105			ahc->msgout_buf + ahc->msgout_index, period, offset,
3106			bus_width, ppr_options);
3107	ahc->msgout_len += 8;
3108	if (bootverbose) {
3109		printk("(%s:%c:%d:%d): Sending PPR bus_width %x, period %x, "
3110		       "offset %x, ppr_options %x\n", ahc_name(ahc),
3111		       devinfo->channel, devinfo->target, devinfo->lun,
3112		       bus_width, period, offset, ppr_options);
3113	}
3114}
3115
3116/*
3117 * Clear any active message state.
3118 */
3119static void
3120ahc_clear_msg_state(struct ahc_softc *ahc)
3121{
3122	ahc->msgout_len = 0;
3123	ahc->msgin_index = 0;
3124	ahc->msg_type = MSG_TYPE_NONE;
3125	if ((ahc_inb(ahc, SCSISIGI) & ATNI) != 0) {
3126		/*
3127		 * The target didn't care to respond to our
3128		 * message request, so clear ATN.
3129		 */
3130		ahc_outb(ahc, CLRSINT1, CLRATNO);
3131	}
3132	ahc_outb(ahc, MSG_OUT, MSG_NOOP);
3133	ahc_outb(ahc, SEQ_FLAGS2,
3134		 ahc_inb(ahc, SEQ_FLAGS2) & ~TARGET_MSG_PENDING);
3135}
3136
3137static void
3138ahc_handle_proto_violation(struct ahc_softc *ahc)
3139{
3140	struct	ahc_devinfo devinfo;
3141	struct	scb *scb;
3142	u_int	scbid;
3143	u_int	seq_flags;
3144	u_int	curphase;
3145	u_int	lastphase;
3146	int	found;
3147
3148	ahc_fetch_devinfo(ahc, &devinfo);
3149	scbid = ahc_inb(ahc, SCB_TAG);
3150	scb = ahc_lookup_scb(ahc, scbid);
3151	seq_flags = ahc_inb(ahc, SEQ_FLAGS);
3152	curphase = ahc_inb(ahc, SCSISIGI) & PHASE_MASK;
3153	lastphase = ahc_inb(ahc, LASTPHASE);
3154	if ((seq_flags & NOT_IDENTIFIED) != 0) {
3155
3156		/*
3157		 * The reconnecting target either did not send an
3158		 * identify message, or did, but we didn't find an SCB
3159		 * to match.
3160		 */
3161		ahc_print_devinfo(ahc, &devinfo);
3162		printk("Target did not send an IDENTIFY message. "
3163		       "LASTPHASE = 0x%x.\n", lastphase);
3164		scb = NULL;
3165	} else if (scb == NULL) {
3166		/*
3167		 * We don't seem to have an SCB active for this
3168		 * transaction.  Print an error and reset the bus.
3169		 */
3170		ahc_print_devinfo(ahc, &devinfo);
3171		printk("No SCB found during protocol violation\n");
3172		goto proto_violation_reset;
3173	} else {
3174		ahc_set_transaction_status(scb, CAM_SEQUENCE_FAIL);
3175		if ((seq_flags & NO_CDB_SENT) != 0) {
3176			ahc_print_path(ahc, scb);
3177			printk("No or incomplete CDB sent to device.\n");
3178		} else if ((ahc_inb(ahc, SCB_CONTROL) & STATUS_RCVD) == 0) {
3179			/*
3180			 * The target never bothered to provide status to
3181			 * us prior to completing the command.  Since we don't
3182			 * know the disposition of this command, we must attempt
3183			 * to abort it.  Assert ATN and prepare to send an abort
3184			 * message.
3185			 */
3186			ahc_print_path(ahc, scb);
3187			printk("Completed command without status.\n");
3188		} else {
3189			ahc_print_path(ahc, scb);
3190			printk("Unknown protocol violation.\n");
3191			ahc_dump_card_state(ahc);
3192		}
3193	}
3194	if ((lastphase & ~P_DATAIN_DT) == 0
3195	 || lastphase == P_COMMAND) {
3196proto_violation_reset:
3197		/*
3198		 * Target either went directly to data/command
3199		 * phase or didn't respond to our ATN.
3200		 * The only safe thing to do is to blow
3201		 * it away with a bus reset.
3202		 */
3203		found = ahc_reset_channel(ahc, 'A', TRUE);
3204		printk("%s: Issued Channel %c Bus Reset. "
3205		       "%d SCBs aborted\n", ahc_name(ahc), 'A', found);
3206	} else {
3207		/*
3208		 * Leave the selection hardware off in case
3209		 * this abort attempt will affect yet to
3210		 * be sent commands.
3211		 */
3212		ahc_outb(ahc, SCSISEQ,
3213			 ahc_inb(ahc, SCSISEQ) & ~ENSELO);
3214		ahc_assert_atn(ahc);
3215		ahc_outb(ahc, MSG_OUT, HOST_MSG);
3216		if (scb == NULL) {
3217			ahc_print_devinfo(ahc, &devinfo);
3218			ahc->msgout_buf[0] = MSG_ABORT_TASK;
3219			ahc->msgout_len = 1;
3220			ahc->msgout_index = 0;
3221			ahc->msg_type = MSG_TYPE_INITIATOR_MSGOUT;
3222		} else {
3223			ahc_print_path(ahc, scb);
3224			scb->flags |= SCB_ABORT;
3225		}
3226		printk("Protocol violation %s.  Attempting to abort.\n",
3227		       ahc_lookup_phase_entry(curphase)->phasemsg);
3228	}
3229}
3230
3231/*
3232 * Manual message loop handler.
3233 */
3234static void
3235ahc_handle_message_phase(struct ahc_softc *ahc)
3236{
3237	struct	ahc_devinfo devinfo;
3238	u_int	bus_phase;
3239	int	end_session;
3240
3241	ahc_fetch_devinfo(ahc, &devinfo);
3242	end_session = FALSE;
3243	bus_phase = ahc_inb(ahc, SCSISIGI) & PHASE_MASK;
3244
3245reswitch:
3246	switch (ahc->msg_type) {
3247	case MSG_TYPE_INITIATOR_MSGOUT:
3248	{
3249		int lastbyte;
3250		int phasemis;
3251		int msgdone;
3252
3253		if (ahc->msgout_len == 0)
3254			panic("HOST_MSG_LOOP interrupt with no active message");
3255
3256#ifdef AHC_DEBUG
3257		if ((ahc_debug & AHC_SHOW_MESSAGES) != 0) {
3258			ahc_print_devinfo(ahc, &devinfo);
3259			printk("INITIATOR_MSG_OUT");
3260		}
3261#endif
3262		phasemis = bus_phase != P_MESGOUT;
3263		if (phasemis) {
3264#ifdef AHC_DEBUG
3265			if ((ahc_debug & AHC_SHOW_MESSAGES) != 0) {
3266				printk(" PHASEMIS %s\n",
3267				       ahc_lookup_phase_entry(bus_phase)
3268							     ->phasemsg);
3269			}
3270#endif
3271			if (bus_phase == P_MESGIN) {
3272				/*
3273				 * Change gears and see if
3274				 * this messages is of interest to
3275				 * us or should be passed back to
3276				 * the sequencer.
3277				 */
3278				ahc_outb(ahc, CLRSINT1, CLRATNO);
3279				ahc->send_msg_perror = FALSE;
3280				ahc->msg_type = MSG_TYPE_INITIATOR_MSGIN;
3281				ahc->msgin_index = 0;
3282				goto reswitch;
3283			}
3284			end_session = TRUE;
3285			break;
3286		}
3287
3288		if (ahc->send_msg_perror) {
3289			ahc_outb(ahc, CLRSINT1, CLRATNO);
3290			ahc_outb(ahc, CLRSINT1, CLRREQINIT);
3291#ifdef AHC_DEBUG
3292			if ((ahc_debug & AHC_SHOW_MESSAGES) != 0)
3293				printk(" byte 0x%x\n", ahc->send_msg_perror);
3294#endif
3295			ahc_outb(ahc, SCSIDATL, MSG_PARITY_ERROR);
3296			break;
3297		}
3298
3299		msgdone	= ahc->msgout_index == ahc->msgout_len;
3300		if (msgdone) {
3301			/*
3302			 * The target has requested a retry.
3303			 * Re-assert ATN, reset our message index to
3304			 * 0, and try again.
3305			 */
3306			ahc->msgout_index = 0;
3307			ahc_assert_atn(ahc);
3308		}
3309
3310		lastbyte = ahc->msgout_index == (ahc->msgout_len - 1);
3311		if (lastbyte) {
3312			/* Last byte is signified by dropping ATN */
3313			ahc_outb(ahc, CLRSINT1, CLRATNO);
3314		}
3315
3316		/*
3317		 * Clear our interrupt status and present
3318		 * the next byte on the bus.
3319		 */
3320		ahc_outb(ahc, CLRSINT1, CLRREQINIT);
3321#ifdef AHC_DEBUG
3322		if ((ahc_debug & AHC_SHOW_MESSAGES) != 0)
3323			printk(" byte 0x%x\n",
3324			       ahc->msgout_buf[ahc->msgout_index]);
3325#endif
3326		ahc_outb(ahc, SCSIDATL, ahc->msgout_buf[ahc->msgout_index++]);
3327		break;
3328	}
3329	case MSG_TYPE_INITIATOR_MSGIN:
3330	{
3331		int phasemis;
3332		int message_done;
3333
3334#ifdef AHC_DEBUG
3335		if ((ahc_debug & AHC_SHOW_MESSAGES) != 0) {
3336			ahc_print_devinfo(ahc, &devinfo);
3337			printk("INITIATOR_MSG_IN");
3338		}
3339#endif
3340		phasemis = bus_phase != P_MESGIN;
3341		if (phasemis) {
3342#ifdef AHC_DEBUG
3343			if ((ahc_debug & AHC_SHOW_MESSAGES) != 0) {
3344				printk(" PHASEMIS %s\n",
3345				       ahc_lookup_phase_entry(bus_phase)
3346							     ->phasemsg);
3347			}
3348#endif
3349			ahc->msgin_index = 0;
3350			if (bus_phase == P_MESGOUT
3351			 && (ahc->send_msg_perror == TRUE
3352			  || (ahc->msgout_len != 0
3353			   && ahc->msgout_index == 0))) {
3354				ahc->msg_type = MSG_TYPE_INITIATOR_MSGOUT;
3355				goto reswitch;
3356			}
3357			end_session = TRUE;
3358			break;
3359		}
3360
3361		/* Pull the byte in without acking it */
3362		ahc->msgin_buf[ahc->msgin_index] = ahc_inb(ahc, SCSIBUSL);
3363#ifdef AHC_DEBUG
3364		if ((ahc_debug & AHC_SHOW_MESSAGES) != 0)
3365			printk(" byte 0x%x\n",
3366			       ahc->msgin_buf[ahc->msgin_index]);
3367#endif
3368
3369		message_done = ahc_parse_msg(ahc, &devinfo);
3370
3371		if (message_done) {
3372			/*
3373			 * Clear our incoming message buffer in case there
3374			 * is another message following this one.
3375			 */
3376			ahc->msgin_index = 0;
3377
3378			/*
3379			 * If this message illicited a response,
3380			 * assert ATN so the target takes us to the
3381			 * message out phase.
3382			 */
3383			if (ahc->msgout_len != 0) {
3384#ifdef AHC_DEBUG
3385				if ((ahc_debug & AHC_SHOW_MESSAGES) != 0) {
3386					ahc_print_devinfo(ahc, &devinfo);
3387					printk("Asserting ATN for response\n");
3388				}
3389#endif
3390				ahc_assert_atn(ahc);
3391			}
3392		} else
3393			ahc->msgin_index++;
3394
3395		if (message_done == MSGLOOP_TERMINATED) {
3396			end_session = TRUE;
3397		} else {
3398			/* Ack the byte */
3399			ahc_outb(ahc, CLRSINT1, CLRREQINIT);
3400			ahc_inb(ahc, SCSIDATL);
3401		}
3402		break;
3403	}
3404	case MSG_TYPE_TARGET_MSGIN:
3405	{
3406		int msgdone;
3407		int msgout_request;
3408
3409		if (ahc->msgout_len == 0)
3410			panic("Target MSGIN with no active message");
3411
3412		/*
3413		 * If we interrupted a mesgout session, the initiator
3414		 * will not know this until our first REQ.  So, we
3415		 * only honor mesgout requests after we've sent our
3416		 * first byte.
3417		 */
3418		if ((ahc_inb(ahc, SCSISIGI) & ATNI) != 0
3419		 && ahc->msgout_index > 0)
3420			msgout_request = TRUE;
3421		else
3422			msgout_request = FALSE;
3423
3424		if (msgout_request) {
3425
3426			/*
3427			 * Change gears and see if
3428			 * this messages is of interest to
3429			 * us or should be passed back to
3430			 * the sequencer.
3431			 */
3432			ahc->msg_type = MSG_TYPE_TARGET_MSGOUT;
3433			ahc_outb(ahc, SCSISIGO, P_MESGOUT | BSYO);
3434			ahc->msgin_index = 0;
3435			/* Dummy read to REQ for first byte */
3436			ahc_inb(ahc, SCSIDATL);
3437			ahc_outb(ahc, SXFRCTL0,
3438				 ahc_inb(ahc, SXFRCTL0) | SPIOEN);
3439			break;
3440		}
3441
3442		msgdone = ahc->msgout_index == ahc->msgout_len;
3443		if (msgdone) {
3444			ahc_outb(ahc, SXFRCTL0,
3445				 ahc_inb(ahc, SXFRCTL0) & ~SPIOEN);
3446			end_session = TRUE;
3447			break;
3448		}
3449
3450		/*
3451		 * Present the next byte on the bus.
3452		 */
3453		ahc_outb(ahc, SXFRCTL0, ahc_inb(ahc, SXFRCTL0) | SPIOEN);
3454		ahc_outb(ahc, SCSIDATL, ahc->msgout_buf[ahc->msgout_index++]);
3455		break;
3456	}
3457	case MSG_TYPE_TARGET_MSGOUT:
3458	{
3459		int lastbyte;
3460		int msgdone;
3461
3462		/*
3463		 * The initiator signals that this is
3464		 * the last byte by dropping ATN.
3465		 */
3466		lastbyte = (ahc_inb(ahc, SCSISIGI) & ATNI) == 0;
3467
3468		/*
3469		 * Read the latched byte, but turn off SPIOEN first
3470		 * so that we don't inadvertently cause a REQ for the
3471		 * next byte.
3472		 */
3473		ahc_outb(ahc, SXFRCTL0, ahc_inb(ahc, SXFRCTL0) & ~SPIOEN);
3474		ahc->msgin_buf[ahc->msgin_index] = ahc_inb(ahc, SCSIDATL);
3475		msgdone = ahc_parse_msg(ahc, &devinfo);
3476		if (msgdone == MSGLOOP_TERMINATED) {
3477			/*
3478			 * The message is *really* done in that it caused
3479			 * us to go to bus free.  The sequencer has already
3480			 * been reset at this point, so pull the ejection
3481			 * handle.
3482			 */
3483			return;
3484		}
3485
3486		ahc->msgin_index++;
3487
3488		/*
3489		 * XXX Read spec about initiator dropping ATN too soon
3490		 *     and use msgdone to detect it.
3491		 */
3492		if (msgdone == MSGLOOP_MSGCOMPLETE) {
3493			ahc->msgin_index = 0;
3494
3495			/*
3496			 * If this message illicited a response, transition
3497			 * to the Message in phase and send it.
3498			 */
3499			if (ahc->msgout_len != 0) {
3500				ahc_outb(ahc, SCSISIGO, P_MESGIN | BSYO);
3501				ahc_outb(ahc, SXFRCTL0,
3502					 ahc_inb(ahc, SXFRCTL0) | SPIOEN);
3503				ahc->msg_type = MSG_TYPE_TARGET_MSGIN;
3504				ahc->msgin_index = 0;
3505				break;
3506			}
3507		}
3508
3509		if (lastbyte)
3510			end_session = TRUE;
3511		else {
3512			/* Ask for the next byte. */
3513			ahc_outb(ahc, SXFRCTL0,
3514				 ahc_inb(ahc, SXFRCTL0) | SPIOEN);
3515		}
3516
3517		break;
3518	}
3519	default:
3520		panic("Unknown REQINIT message type");
3521	}
3522
3523	if (end_session) {
3524		ahc_clear_msg_state(ahc);
3525		ahc_outb(ahc, RETURN_1, EXIT_MSG_LOOP);
3526	} else
3527		ahc_outb(ahc, RETURN_1, CONT_MSG_LOOP);
3528}
3529
3530/*
3531 * See if we sent a particular extended message to the target.
3532 * If "full" is true, return true only if the target saw the full
3533 * message.  If "full" is false, return true if the target saw at
3534 * least the first byte of the message.
3535 */
3536static int
3537ahc_sent_msg(struct ahc_softc *ahc, ahc_msgtype type, u_int msgval, int full)
3538{
3539	int found;
3540	u_int index;
3541
3542	found = FALSE;
3543	index = 0;
3544
3545	while (index < ahc->msgout_len) {
3546		if (ahc->msgout_buf[index] == MSG_EXTENDED) {
3547			u_int end_index;
3548
3549			end_index = index + 1 + ahc->msgout_buf[index + 1];
3550			if (ahc->msgout_buf[index+2] == msgval
3551			 && type == AHCMSG_EXT) {
3552
3553				if (full) {
3554					if (ahc->msgout_index > end_index)
3555						found = TRUE;
3556				} else if (ahc->msgout_index > index)
3557					found = TRUE;
3558			}
3559			index = end_index;
3560		} else if (ahc->msgout_buf[index] >= MSG_SIMPLE_TASK
3561			&& ahc->msgout_buf[index] <= MSG_IGN_WIDE_RESIDUE) {
3562
3563			/* Skip tag type and tag id or residue param*/
3564			index += 2;
3565		} else {
3566			/* Single byte message */
3567			if (type == AHCMSG_1B
3568			 && ahc->msgout_buf[index] == msgval
3569			 && ahc->msgout_index > index)
3570				found = TRUE;
3571			index++;
3572		}
3573
3574		if (found)
3575			break;
3576	}
3577	return (found);
3578}
3579
3580/*
3581 * Wait for a complete incoming message, parse it, and respond accordingly.
3582 */
3583static int
3584ahc_parse_msg(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
3585{
3586	struct	ahc_initiator_tinfo *tinfo;
3587	struct	ahc_tmode_tstate *tstate;
3588	int	reject;
3589	int	done;
3590	int	response;
3591	u_int	targ_scsirate;
3592
3593	done = MSGLOOP_IN_PROG;
3594	response = FALSE;
3595	reject = FALSE;
3596	tinfo = ahc_fetch_transinfo(ahc, devinfo->channel, devinfo->our_scsiid,
3597				    devinfo->target, &tstate);
3598	targ_scsirate = tinfo->scsirate;
3599
3600	/*
3601	 * Parse as much of the message as is available,
3602	 * rejecting it if we don't support it.  When
3603	 * the entire message is available and has been
3604	 * handled, return MSGLOOP_MSGCOMPLETE, indicating
3605	 * that we have parsed an entire message.
3606	 *
3607	 * In the case of extended messages, we accept the length
3608	 * byte outright and perform more checking once we know the
3609	 * extended message type.
3610	 */
3611	switch (ahc->msgin_buf[0]) {
3612	case MSG_DISCONNECT:
3613	case MSG_SAVEDATAPOINTER:
3614	case MSG_CMDCOMPLETE:
3615	case MSG_RESTOREPOINTERS:
3616	case MSG_IGN_WIDE_RESIDUE:
3617		/*
3618		 * End our message loop as these are messages
3619		 * the sequencer handles on its own.
3620		 */
3621		done = MSGLOOP_TERMINATED;
3622		break;
3623	case MSG_MESSAGE_REJECT:
3624		response = ahc_handle_msg_reject(ahc, devinfo);
3625		/* FALLTHROUGH */
3626	case MSG_NOOP:
3627		done = MSGLOOP_MSGCOMPLETE;
3628		break;
3629	case MSG_EXTENDED:
3630	{
3631		/* Wait for enough of the message to begin validation */
3632		if (ahc->msgin_index < 2)
3633			break;
3634		switch (ahc->msgin_buf[2]) {
3635		case MSG_EXT_SDTR:
3636		{
3637			const struct ahc_syncrate *syncrate;
3638			u_int	 period;
3639			u_int	 ppr_options;
3640			u_int	 offset;
3641			u_int	 saved_offset;
3642
3643			if (ahc->msgin_buf[1] != MSG_EXT_SDTR_LEN) {
3644				reject = TRUE;
3645				break;
3646			}
3647
3648			/*
3649			 * Wait until we have both args before validating
3650			 * and acting on this message.
3651			 *
3652			 * Add one to MSG_EXT_SDTR_LEN to account for
3653			 * the extended message preamble.
3654			 */
3655			if (ahc->msgin_index < (MSG_EXT_SDTR_LEN + 1))
3656				break;
3657
3658			period = ahc->msgin_buf[3];
3659			ppr_options = 0;
3660			saved_offset = offset = ahc->msgin_buf[4];
3661			syncrate = ahc_devlimited_syncrate(ahc, tinfo, &period,
3662							   &ppr_options,
3663							   devinfo->role);
3664			ahc_validate_offset(ahc, tinfo, syncrate, &offset,
3665					    targ_scsirate & WIDEXFER,
3666					    devinfo->role);
3667			if (bootverbose) {
3668				printk("(%s:%c:%d:%d): Received "
3669				       "SDTR period %x, offset %x\n\t"
3670				       "Filtered to period %x, offset %x\n",
3671				       ahc_name(ahc), devinfo->channel,
3672				       devinfo->target, devinfo->lun,
3673				       ahc->msgin_buf[3], saved_offset,
3674				       period, offset);
3675			}
3676			ahc_set_syncrate(ahc, devinfo,
3677					 syncrate, period,
3678					 offset, ppr_options,
3679					 AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
3680					 /*paused*/TRUE);
3681
3682			/*
3683			 * See if we initiated Sync Negotiation
3684			 * and didn't have to fall down to async
3685			 * transfers.
3686			 */
3687			if (ahc_sent_msg(ahc, AHCMSG_EXT, MSG_EXT_SDTR, TRUE)) {
3688				/* We started it */
3689				if (saved_offset != offset) {
3690					/* Went too low - force async */
3691					reject = TRUE;
3692				}
3693			} else {
3694				/*
3695				 * Send our own SDTR in reply
3696				 */
3697				if (bootverbose
3698				 && devinfo->role == ROLE_INITIATOR) {
3699					printk("(%s:%c:%d:%d): Target "
3700					       "Initiated SDTR\n",
3701					       ahc_name(ahc), devinfo->channel,
3702					       devinfo->target, devinfo->lun);
3703				}
3704				ahc->msgout_index = 0;
3705				ahc->msgout_len = 0;
3706				ahc_construct_sdtr(ahc, devinfo,
3707						   period, offset);
3708				ahc->msgout_index = 0;
3709				response = TRUE;
3710			}
3711			done = MSGLOOP_MSGCOMPLETE;
3712			break;
3713		}
3714		case MSG_EXT_WDTR:
3715		{
3716			u_int bus_width;
3717			u_int saved_width;
3718			u_int sending_reply;
3719
3720			sending_reply = FALSE;
3721			if (ahc->msgin_buf[1] != MSG_EXT_WDTR_LEN) {
3722				reject = TRUE;
3723				break;
3724			}
3725
3726			/*
3727			 * Wait until we have our arg before validating
3728			 * and acting on this message.
3729			 *
3730			 * Add one to MSG_EXT_WDTR_LEN to account for
3731			 * the extended message preamble.
3732			 */
3733			if (ahc->msgin_index < (MSG_EXT_WDTR_LEN + 1))
3734				break;
3735
3736			bus_width = ahc->msgin_buf[3];
3737			saved_width = bus_width;
3738			ahc_validate_width(ahc, tinfo, &bus_width,
3739					   devinfo->role);
3740			if (bootverbose) {
3741				printk("(%s:%c:%d:%d): Received WDTR "
3742				       "%x filtered to %x\n",
3743				       ahc_name(ahc), devinfo->channel,
3744				       devinfo->target, devinfo->lun,
3745				       saved_width, bus_width);
3746			}
3747
3748			if (ahc_sent_msg(ahc, AHCMSG_EXT, MSG_EXT_WDTR, TRUE)) {
3749				/*
3750				 * Don't send a WDTR back to the
3751				 * target, since we asked first.
3752				 * If the width went higher than our
3753				 * request, reject it.
3754				 */
3755				if (saved_width > bus_width) {
3756					reject = TRUE;
3757					printk("(%s:%c:%d:%d): requested %dBit "
3758					       "transfers.  Rejecting...\n",
3759					       ahc_name(ahc), devinfo->channel,
3760					       devinfo->target, devinfo->lun,
3761					       8 * (0x01 << bus_width));
3762					bus_width = 0;
3763				}
3764			} else {
3765				/*
3766				 * Send our own WDTR in reply
3767				 */
3768				if (bootverbose
3769				 && devinfo->role == ROLE_INITIATOR) {
3770					printk("(%s:%c:%d:%d): Target "
3771					       "Initiated WDTR\n",
3772					       ahc_name(ahc), devinfo->channel,
3773					       devinfo->target, devinfo->lun);
3774				}
3775				ahc->msgout_index = 0;
3776				ahc->msgout_len = 0;
3777				ahc_construct_wdtr(ahc, devinfo, bus_width);
3778				ahc->msgout_index = 0;
3779				response = TRUE;
3780				sending_reply = TRUE;
3781			}
3782			/*
3783			 * After a wide message, we are async, but
3784			 * some devices don't seem to honor this portion
3785			 * of the spec.  Force a renegotiation of the
3786			 * sync component of our transfer agreement even
3787			 * if our goal is async.  By updating our width
3788			 * after forcing the negotiation, we avoid
3789			 * renegotiating for width.
3790			 */
3791			ahc_update_neg_request(ahc, devinfo, tstate,
3792					       tinfo, AHC_NEG_ALWAYS);
3793			ahc_set_width(ahc, devinfo, bus_width,
3794				      AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
3795				      /*paused*/TRUE);
3796			if (sending_reply == FALSE && reject == FALSE) {
3797
3798				/*
3799				 * We will always have an SDTR to send.
3800				 */
3801				ahc->msgout_index = 0;
3802				ahc->msgout_len = 0;
3803				ahc_build_transfer_msg(ahc, devinfo);
3804				ahc->msgout_index = 0;
3805				response = TRUE;
3806			}
3807			done = MSGLOOP_MSGCOMPLETE;
3808			break;
3809		}
3810		case MSG_EXT_PPR:
3811		{
3812			const struct ahc_syncrate *syncrate;
3813			u_int	period;
3814			u_int	offset;
3815			u_int	bus_width;
3816			u_int	ppr_options;
3817			u_int	saved_width;
3818			u_int	saved_offset;
3819			u_int	saved_ppr_options;
3820
3821			if (ahc->msgin_buf[1] != MSG_EXT_PPR_LEN) {
3822				reject = TRUE;
3823				break;
3824			}
3825
3826			/*
3827			 * Wait until we have all args before validating
3828			 * and acting on this message.
3829			 *
3830			 * Add one to MSG_EXT_PPR_LEN to account for
3831			 * the extended message preamble.
3832			 */
3833			if (ahc->msgin_index < (MSG_EXT_PPR_LEN + 1))
3834				break;
3835
3836			period = ahc->msgin_buf[3];
3837			offset = ahc->msgin_buf[5];
3838			bus_width = ahc->msgin_buf[6];
3839			saved_width = bus_width;
3840			ppr_options = ahc->msgin_buf[7];
3841			/*
3842			 * According to the spec, a DT only
3843			 * period factor with no DT option
3844			 * set implies async.
3845			 */
3846			if ((ppr_options & MSG_EXT_PPR_DT_REQ) == 0
3847			 && period == 9)
3848				offset = 0;
3849			saved_ppr_options = ppr_options;
3850			saved_offset = offset;
3851
3852			/*
3853			 * Mask out any options we don't support
3854			 * on any controller.  Transfer options are
3855			 * only available if we are negotiating wide.
3856			 */
3857			ppr_options &= MSG_EXT_PPR_DT_REQ;
3858			if (bus_width == 0)
3859				ppr_options = 0;
3860
3861			ahc_validate_width(ahc, tinfo, &bus_width,
3862					   devinfo->role);
3863			syncrate = ahc_devlimited_syncrate(ahc, tinfo, &period,
3864							   &ppr_options,
3865							   devinfo->role);
3866			ahc_validate_offset(ahc, tinfo, syncrate,
3867					    &offset, bus_width,
3868					    devinfo->role);
3869
3870			if (ahc_sent_msg(ahc, AHCMSG_EXT, MSG_EXT_PPR, TRUE)) {
3871				/*
3872				 * If we are unable to do any of the
3873				 * requested options (we went too low),
3874				 * then we'll have to reject the message.
3875				 */
3876				if (saved_width > bus_width
3877				 || saved_offset != offset
3878				 || saved_ppr_options != ppr_options) {
3879					reject = TRUE;
3880					period = 0;
3881					offset = 0;
3882					bus_width = 0;
3883					ppr_options = 0;
3884					syncrate = NULL;
3885				}
3886			} else {
3887				if (devinfo->role != ROLE_TARGET)
3888					printk("(%s:%c:%d:%d): Target "
3889					       "Initiated PPR\n",
3890					       ahc_name(ahc), devinfo->channel,
3891					       devinfo->target, devinfo->lun);
3892				else
3893					printk("(%s:%c:%d:%d): Initiator "
3894					       "Initiated PPR\n",
3895					       ahc_name(ahc), devinfo->channel,
3896					       devinfo->target, devinfo->lun);
3897				ahc->msgout_index = 0;
3898				ahc->msgout_len = 0;
3899				ahc_construct_ppr(ahc, devinfo, period, offset,
3900						  bus_width, ppr_options);
3901				ahc->msgout_index = 0;
3902				response = TRUE;
3903			}
3904			if (bootverbose) {
3905				printk("(%s:%c:%d:%d): Received PPR width %x, "
3906				       "period %x, offset %x,options %x\n"
3907				       "\tFiltered to width %x, period %x, "
3908				       "offset %x, options %x\n",
3909				       ahc_name(ahc), devinfo->channel,
3910				       devinfo->target, devinfo->lun,
3911				       saved_width, ahc->msgin_buf[3],
3912				       saved_offset, saved_ppr_options,
3913				       bus_width, period, offset, ppr_options);
3914			}
3915			ahc_set_width(ahc, devinfo, bus_width,
3916				      AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
3917				      /*paused*/TRUE);
3918			ahc_set_syncrate(ahc, devinfo,
3919					 syncrate, period,
3920					 offset, ppr_options,
3921					 AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
3922					 /*paused*/TRUE);
3923			done = MSGLOOP_MSGCOMPLETE;
3924			break;
3925		}
3926		default:
3927			/* Unknown extended message.  Reject it. */
3928			reject = TRUE;
3929			break;
3930		}
3931		break;
3932	}
3933#ifdef AHC_TARGET_MODE
3934	case MSG_BUS_DEV_RESET:
3935		ahc_handle_devreset(ahc, devinfo,
3936				    CAM_BDR_SENT,
3937				    "Bus Device Reset Received",
3938				    /*verbose_level*/0);
3939		ahc_restart(ahc);
3940		done = MSGLOOP_TERMINATED;
3941		break;
3942	case MSG_ABORT_TAG:
3943	case MSG_ABORT:
3944	case MSG_CLEAR_QUEUE:
3945	{
3946		int tag;
3947
3948		/* Target mode messages */
3949		if (devinfo->role != ROLE_TARGET) {
3950			reject = TRUE;
3951			break;
3952		}
3953		tag = SCB_LIST_NULL;
3954		if (ahc->msgin_buf[0] == MSG_ABORT_TAG)
3955			tag = ahc_inb(ahc, INITIATOR_TAG);
3956		ahc_abort_scbs(ahc, devinfo->target, devinfo->channel,
3957			       devinfo->lun, tag, ROLE_TARGET,
3958			       CAM_REQ_ABORTED);
3959
3960		tstate = ahc->enabled_targets[devinfo->our_scsiid];
3961		if (tstate != NULL) {
3962			struct ahc_tmode_lstate* lstate;
3963
3964			lstate = tstate->enabled_luns[devinfo->lun];
3965			if (lstate != NULL) {
3966				ahc_queue_lstate_event(ahc, lstate,
3967						       devinfo->our_scsiid,
3968						       ahc->msgin_buf[0],
3969						       /*arg*/tag);
3970				ahc_send_lstate_events(ahc, lstate);
3971			}
3972		}
3973		ahc_restart(ahc);
3974		done = MSGLOOP_TERMINATED;
3975		break;
3976	}
3977#endif
3978	case MSG_TERM_IO_PROC:
3979	default:
3980		reject = TRUE;
3981		break;
3982	}
3983
3984	if (reject) {
3985		/*
3986		 * Setup to reject the message.
3987		 */
3988		ahc->msgout_index = 0;
3989		ahc->msgout_len = 1;
3990		ahc->msgout_buf[0] = MSG_MESSAGE_REJECT;
3991		done = MSGLOOP_MSGCOMPLETE;
3992		response = TRUE;
3993	}
3994
3995	if (done != MSGLOOP_IN_PROG && !response)
3996		/* Clear the outgoing message buffer */
3997		ahc->msgout_len = 0;
3998
3999	return (done);
4000}
4001
4002/*
4003 * Process a message reject message.
4004 */
4005static int
4006ahc_handle_msg_reject(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
4007{
4008	/*
4009	 * What we care about here is if we had an
4010	 * outstanding SDTR or WDTR message for this
4011	 * target.  If we did, this is a signal that
4012	 * the target is refusing negotiation.
4013	 */
4014	struct scb *scb;
4015	struct ahc_initiator_tinfo *tinfo;
4016	struct ahc_tmode_tstate *tstate;
4017	u_int scb_index;
4018	u_int last_msg;
4019	int   response = 0;
4020
4021	scb_index = ahc_inb(ahc, SCB_TAG);
4022	scb = ahc_lookup_scb(ahc, scb_index);
4023	tinfo = ahc_fetch_transinfo(ahc, devinfo->channel,
4024				    devinfo->our_scsiid,
4025				    devinfo->target, &tstate);
4026	/* Might be necessary */
4027	last_msg = ahc_inb(ahc, LAST_MSG);
4028
4029	if (ahc_sent_msg(ahc, AHCMSG_EXT, MSG_EXT_PPR, /*full*/FALSE)) {
4030		/*
4031		 * Target does not support the PPR message.
4032		 * Attempt to negotiate SPI-2 style.
4033		 */
4034		if (bootverbose) {
4035			printk("(%s:%c:%d:%d): PPR Rejected. "
4036			       "Trying WDTR/SDTR\n",
4037			       ahc_name(ahc), devinfo->channel,
4038			       devinfo->target, devinfo->lun);
4039		}
4040		tinfo->goal.ppr_options = 0;
4041		tinfo->curr.transport_version = 2;
4042		tinfo->goal.transport_version = 2;
4043		ahc->msgout_index = 0;
4044		ahc->msgout_len = 0;
4045		ahc_build_transfer_msg(ahc, devinfo);
4046		ahc->msgout_index = 0;
4047		response = 1;
4048	} else if (ahc_sent_msg(ahc, AHCMSG_EXT, MSG_EXT_WDTR, /*full*/FALSE)) {
4049
4050		/* note 8bit xfers */
4051		printk("(%s:%c:%d:%d): refuses WIDE negotiation.  Using "
4052		       "8bit transfers\n", ahc_name(ahc),
4053		       devinfo->channel, devinfo->target, devinfo->lun);
4054		ahc_set_width(ahc, devinfo, MSG_EXT_WDTR_BUS_8_BIT,
4055			      AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
4056			      /*paused*/TRUE);
4057		/*
4058		 * No need to clear the sync rate.  If the target
4059		 * did not accept the command, our syncrate is
4060		 * unaffected.  If the target started the negotiation,
4061		 * but rejected our response, we already cleared the
4062		 * sync rate before sending our WDTR.
4063		 */
4064		if (tinfo->goal.offset != tinfo->curr.offset) {
4065
4066			/* Start the sync negotiation */
4067			ahc->msgout_index = 0;
4068			ahc->msgout_len = 0;
4069			ahc_build_transfer_msg(ahc, devinfo);
4070			ahc->msgout_index = 0;
4071			response = 1;
4072		}
4073	} else if (ahc_sent_msg(ahc, AHCMSG_EXT, MSG_EXT_SDTR, /*full*/FALSE)) {
4074		/* note asynch xfers and clear flag */
4075		ahc_set_syncrate(ahc, devinfo, /*syncrate*/NULL, /*period*/0,
4076				 /*offset*/0, /*ppr_options*/0,
4077				 AHC_TRANS_ACTIVE|AHC_TRANS_GOAL,
4078				 /*paused*/TRUE);
4079		printk("(%s:%c:%d:%d): refuses synchronous negotiation. "
4080		       "Using asynchronous transfers\n",
4081		       ahc_name(ahc), devinfo->channel,
4082		       devinfo->target, devinfo->lun);
4083	} else if ((scb->hscb->control & MSG_SIMPLE_TASK) != 0) {
4084		int tag_type;
4085		int mask;
4086
4087		tag_type = (scb->hscb->control & MSG_SIMPLE_TASK);
4088
4089		if (tag_type == MSG_SIMPLE_TASK) {
4090			printk("(%s:%c:%d:%d): refuses tagged commands.  "
4091			       "Performing non-tagged I/O\n", ahc_name(ahc),
4092			       devinfo->channel, devinfo->target, devinfo->lun);
4093			ahc_set_tags(ahc, scb->io_ctx, devinfo, AHC_QUEUE_NONE);
4094			mask = ~0x23;
4095		} else {
4096			printk("(%s:%c:%d:%d): refuses %s tagged commands.  "
4097			       "Performing simple queue tagged I/O only\n",
4098			       ahc_name(ahc), devinfo->channel, devinfo->target,
4099			       devinfo->lun, tag_type == MSG_ORDERED_TASK
4100			       ? "ordered" : "head of queue");
4101			ahc_set_tags(ahc, scb->io_ctx, devinfo, AHC_QUEUE_BASIC);
4102			mask = ~0x03;
4103		}
4104
4105		/*
4106		 * Resend the identify for this CCB as the target
4107		 * may believe that the selection is invalid otherwise.
4108		 */
4109		ahc_outb(ahc, SCB_CONTROL,
4110			 ahc_inb(ahc, SCB_CONTROL) & mask);
4111	 	scb->hscb->control &= mask;
4112		ahc_set_transaction_tag(scb, /*enabled*/FALSE,
4113					/*type*/MSG_SIMPLE_TASK);
4114		ahc_outb(ahc, MSG_OUT, MSG_IDENTIFYFLAG);
4115		ahc_assert_atn(ahc);
4116
4117		/*
4118		 * This transaction is now at the head of
4119		 * the untagged queue for this target.
4120		 */
4121		if ((ahc->flags & AHC_SCB_BTT) == 0) {
4122			struct scb_tailq *untagged_q;
4123
4124			untagged_q =
4125			    &(ahc->untagged_queues[devinfo->target_offset]);
4126			TAILQ_INSERT_HEAD(untagged_q, scb, links.tqe);
4127			scb->flags |= SCB_UNTAGGEDQ;
4128		}
4129		ahc_busy_tcl(ahc, BUILD_TCL(scb->hscb->scsiid, devinfo->lun),
4130			     scb->hscb->tag);
4131
4132		/*
4133		 * Requeue all tagged commands for this target
4134		 * currently in our possession so they can be
4135		 * converted to untagged commands.
4136		 */
4137		ahc_search_qinfifo(ahc, SCB_GET_TARGET(ahc, scb),
4138				   SCB_GET_CHANNEL(ahc, scb),
4139				   SCB_GET_LUN(scb), /*tag*/SCB_LIST_NULL,
4140				   ROLE_INITIATOR, CAM_REQUEUE_REQ,
4141				   SEARCH_COMPLETE);
4142	} else {
4143		/*
4144		 * Otherwise, we ignore it.
4145		 */
4146		printk("%s:%c:%d: Message reject for %x -- ignored\n",
4147		       ahc_name(ahc), devinfo->channel, devinfo->target,
4148		       last_msg);
4149	}
4150	return (response);
4151}
4152
4153/*
4154 * Process an ingnore wide residue message.
4155 */
4156static void
4157ahc_handle_ign_wide_residue(struct ahc_softc *ahc, struct ahc_devinfo *devinfo)
4158{
4159	u_int scb_index;
4160	struct scb *scb;
4161
4162	scb_index = ahc_inb(ahc, SCB_TAG);
4163	scb = ahc_lookup_scb(ahc, scb_index);
4164	/*
4165	 * XXX Actually check data direction in the sequencer?
4166	 * Perhaps add datadir to some spare bits in the hscb?
4167	 */
4168	if ((ahc_inb(ahc, SEQ_FLAGS) & DPHASE) == 0
4169	 || ahc_get_transfer_dir(scb) != CAM_DIR_IN) {
4170		/*
4171		 * Ignore the message if we haven't
4172		 * seen an appropriate data phase yet.
4173		 */
4174	} else {
4175		/*
4176		 * If the residual occurred on the last
4177		 * transfer and the transfer request was
4178		 * expected to end on an odd count, do
4179		 * nothing.  Otherwise, subtract a byte
4180		 * and update the residual count accordingly.
4181		 */
4182		uint32_t sgptr;
4183
4184		sgptr = ahc_inb(ahc, SCB_RESIDUAL_SGPTR);
4185		if ((sgptr & SG_LIST_NULL) != 0
4186		 && (ahc_inb(ahc, SCB_LUN) & SCB_XFERLEN_ODD) != 0) {
4187			/*
4188			 * If the residual occurred on the last
4189			 * transfer and the transfer request was
4190			 * expected to end on an odd count, do
4191			 * nothing.
4192			 */
4193		} else {
4194			struct ahc_dma_seg *sg;
4195			uint32_t data_cnt;
4196			uint32_t data_addr;
4197			uint32_t sglen;
4198
4199			/* Pull in all of the sgptr */
4200			sgptr = ahc_inl(ahc, SCB_RESIDUAL_SGPTR);
4201			data_cnt = ahc_inl(ahc, SCB_RESIDUAL_DATACNT);
4202
4203			if ((sgptr & SG_LIST_NULL) != 0) {
4204				/*
4205				 * The residual data count is not updated
4206				 * for the command run to completion case.
4207				 * Explicitly zero the count.
4208				 */
4209				data_cnt &= ~AHC_SG_LEN_MASK;
4210			}
4211
4212			data_addr = ahc_inl(ahc, SHADDR);
4213
4214			data_cnt += 1;
4215			data_addr -= 1;
4216			sgptr &= SG_PTR_MASK;
4217
4218			sg = ahc_sg_bus_to_virt(scb, sgptr);
4219
4220			/*
4221			 * The residual sg ptr points to the next S/G
4222			 * to load so we must go back one.
4223			 */
4224			sg--;
4225			sglen = ahc_le32toh(sg->len) & AHC_SG_LEN_MASK;
4226			if (sg != scb->sg_list
4227			 && sglen < (data_cnt & AHC_SG_LEN_MASK)) {
4228
4229				sg--;
4230				sglen = ahc_le32toh(sg->len);
4231				/*
4232				 * Preserve High Address and SG_LIST bits
4233				 * while setting the count to 1.
4234				 */
4235				data_cnt = 1 | (sglen & (~AHC_SG_LEN_MASK));
4236				data_addr = ahc_le32toh(sg->addr)
4237					  + (sglen & AHC_SG_LEN_MASK) - 1;
4238
4239				/*
4240				 * Increment sg so it points to the
4241				 * "next" sg.
4242				 */
4243				sg++;
4244				sgptr = ahc_sg_virt_to_bus(scb, sg);
4245			}
4246			ahc_outl(ahc, SCB_RESIDUAL_SGPTR, sgptr);
4247			ahc_outl(ahc, SCB_RESIDUAL_DATACNT, data_cnt);
4248			/*
4249			 * Toggle the "oddness" of the transfer length
4250			 * to handle this mid-transfer ignore wide
4251			 * residue.  This ensures that the oddness is
4252			 * correct for subsequent data transfers.
4253			 */
4254			ahc_outb(ahc, SCB_LUN,
4255				 ahc_inb(ahc, SCB_LUN) ^ SCB_XFERLEN_ODD);
4256		}
4257	}
4258}
4259
4260
4261/*
4262 * Reinitialize the data pointers for the active transfer
4263 * based on its current residual.
4264 */
4265static void
4266ahc_reinitialize_dataptrs(struct ahc_softc *ahc)
4267{
4268	struct	 scb *scb;
4269	struct	 ahc_dma_seg *sg;
4270	u_int	 scb_index;
4271	uint32_t sgptr;
4272	uint32_t resid;
4273	uint32_t dataptr;
4274
4275	scb_index = ahc_inb(ahc, SCB_TAG);
4276	scb = ahc_lookup_scb(ahc, scb_index);
4277	sgptr = (ahc_inb(ahc, SCB_RESIDUAL_SGPTR + 3) << 24)
4278	      | (ahc_inb(ahc, SCB_RESIDUAL_SGPTR + 2) << 16)
4279	      | (ahc_inb(ahc, SCB_RESIDUAL_SGPTR + 1) << 8)
4280	      |	ahc_inb(ahc, SCB_RESIDUAL_SGPTR);
4281
4282	sgptr &= SG_PTR_MASK;
4283	sg = ahc_sg_bus_to_virt(scb, sgptr);
4284
4285	/* The residual sg_ptr always points to the next sg */
4286	sg--;
4287
4288	resid = (ahc_inb(ahc, SCB_RESIDUAL_DATACNT + 2) << 16)
4289	      | (ahc_inb(ahc, SCB_RESIDUAL_DATACNT + 1) << 8)
4290	      | ahc_inb(ahc, SCB_RESIDUAL_DATACNT);
4291
4292	dataptr = ahc_le32toh(sg->addr)
4293		+ (ahc_le32toh(sg->len) & AHC_SG_LEN_MASK)
4294		- resid;
4295	if ((ahc->flags & AHC_39BIT_ADDRESSING) != 0) {
4296		u_int dscommand1;
4297
4298		dscommand1 = ahc_inb(ahc, DSCOMMAND1);
4299		ahc_outb(ahc, DSCOMMAND1, dscommand1 | HADDLDSEL0);
4300		ahc_outb(ahc, HADDR,
4301			 (ahc_le32toh(sg->len) >> 24) & SG_HIGH_ADDR_BITS);
4302		ahc_outb(ahc, DSCOMMAND1, dscommand1);
4303	}
4304	ahc_outb(ahc, HADDR + 3, dataptr >> 24);
4305	ahc_outb(ahc, HADDR + 2, dataptr >> 16);
4306	ahc_outb(ahc, HADDR + 1, dataptr >> 8);
4307	ahc_outb(ahc, HADDR, dataptr);
4308	ahc_outb(ahc, HCNT + 2, resid >> 16);
4309	ahc_outb(ahc, HCNT + 1, resid >> 8);
4310	ahc_outb(ahc, HCNT, resid);
4311	if ((ahc->features & AHC_ULTRA2) == 0) {
4312		ahc_outb(ahc, STCNT + 2, resid >> 16);
4313		ahc_outb(ahc, STCNT + 1, resid >> 8);
4314		ahc_outb(ahc, STCNT, resid);
4315	}
4316}
4317
4318/*
4319 * Handle the effects of issuing a bus device reset message.
4320 */
4321static void
4322ahc_handle_devreset(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
4323		    cam_status status, char *message, int verbose_level)
4324{
4325#ifdef AHC_TARGET_MODE
4326	struct ahc_tmode_tstate* tstate;
4327	u_int lun;
4328#endif
4329	int found;
4330
4331	found = ahc_abort_scbs(ahc, devinfo->target, devinfo->channel,
4332			       CAM_LUN_WILDCARD, SCB_LIST_NULL, devinfo->role,
4333			       status);
4334
4335#ifdef AHC_TARGET_MODE
4336	/*
4337	 * Send an immediate notify ccb to all target mord peripheral
4338	 * drivers affected by this action.
4339	 */
4340	tstate = ahc->enabled_targets[devinfo->our_scsiid];
4341	if (tstate != NULL) {
4342		for (lun = 0; lun < AHC_NUM_LUNS; lun++) {
4343			struct ahc_tmode_lstate* lstate;
4344
4345			lstate = tstate->enabled_luns[lun];
4346			if (lstate == NULL)
4347				continue;
4348
4349			ahc_queue_lstate_event(ahc, lstate, devinfo->our_scsiid,
4350					       MSG_BUS_DEV_RESET, /*arg*/0);
4351			ahc_send_lstate_events(ahc, lstate);
4352		}
4353	}
4354#endif
4355
4356	/*
4357	 * Go back to async/narrow transfers and renegotiate.
4358	 */
4359	ahc_set_width(ahc, devinfo, MSG_EXT_WDTR_BUS_8_BIT,
4360		      AHC_TRANS_CUR, /*paused*/TRUE);
4361	ahc_set_syncrate(ahc, devinfo, /*syncrate*/NULL,
4362			 /*period*/0, /*offset*/0, /*ppr_options*/0,
4363			 AHC_TRANS_CUR, /*paused*/TRUE);
4364
4365	if (status != CAM_SEL_TIMEOUT)
4366		ahc_send_async(ahc, devinfo->channel, devinfo->target,
4367			       CAM_LUN_WILDCARD, AC_SENT_BDR);
4368
4369	if (message != NULL
4370	 && (verbose_level <= bootverbose))
4371		printk("%s: %s on %c:%d. %d SCBs aborted\n", ahc_name(ahc),
4372		       message, devinfo->channel, devinfo->target, found);
4373}
4374
4375#ifdef AHC_TARGET_MODE
4376static void
4377ahc_setup_target_msgin(struct ahc_softc *ahc, struct ahc_devinfo *devinfo,
4378		       struct scb *scb)
4379{
4380
4381	/*
4382	 * To facilitate adding multiple messages together,
4383	 * each routine should increment the index and len
4384	 * variables instead of setting them explicitly.
4385	 */
4386	ahc->msgout_index = 0;
4387	ahc->msgout_len = 0;
4388
4389	if (scb != NULL && (scb->flags & SCB_AUTO_NEGOTIATE) != 0)
4390		ahc_build_transfer_msg(ahc, devinfo);
4391	else
4392		panic("ahc_intr: AWAITING target message with no message");
4393
4394	ahc->msgout_index = 0;
4395	ahc->msg_type = MSG_TYPE_TARGET_MSGIN;
4396}
4397#endif
4398/**************************** Initialization **********************************/
4399/*
4400 * Allocate a controller structure for a new device
4401 * and perform initial initializion.
4402 */
4403struct ahc_softc *
4404ahc_alloc(void *platform_arg, char *name)
4405{
4406	struct  ahc_softc *ahc;
4407	int	i;
4408
4409#ifndef	__FreeBSD__
4410	ahc = kmalloc(sizeof(*ahc), GFP_ATOMIC);
4411	if (!ahc) {
4412		printk("aic7xxx: cannot malloc softc!\n");
4413		kfree(name);
4414		return NULL;
4415	}
4416#else
4417	ahc = device_get_softc((device_t)platform_arg);
4418#endif
4419	memset(ahc, 0, sizeof(*ahc));
4420	ahc->seep_config = kmalloc(sizeof(*ahc->seep_config), GFP_ATOMIC);
4421	if (ahc->seep_config == NULL) {
4422#ifndef	__FreeBSD__
4423		kfree(ahc);
4424#endif
4425		kfree(name);
4426		return (NULL);
4427	}
4428	LIST_INIT(&ahc->pending_scbs);
4429	/* We don't know our unit number until the OSM sets it */
4430	ahc->name = name;
4431	ahc->unit = -1;
4432	ahc->description = NULL;
4433	ahc->channel = 'A';
4434	ahc->channel_b = 'B';
4435	ahc->chip = AHC_NONE;
4436	ahc->features = AHC_FENONE;
4437	ahc->bugs = AHC_BUGNONE;
4438	ahc->flags = AHC_FNONE;
4439	/*
4440	 * Default to all error reporting enabled with the
4441	 * sequencer operating at its fastest speed.
4442	 * The bus attach code may modify this.
4443	 */
4444	ahc->seqctl = FASTMODE;
4445
4446	for (i = 0; i < AHC_NUM_TARGETS; i++)
4447		TAILQ_INIT(&ahc->untagged_queues[i]);
4448	if (ahc_platform_alloc(ahc, platform_arg) != 0) {
4449		ahc_free(ahc);
4450		ahc = NULL;
4451	}
4452	return (ahc);
4453}
4454
4455int
4456ahc_softc_init(struct ahc_softc *ahc)
4457{
4458
4459	/* The IRQMS bit is only valid on VL and EISA chips */
4460	if ((ahc->chip & AHC_PCI) == 0)
4461		ahc->unpause = ahc_inb(ahc, HCNTRL) & IRQMS;
4462	else
4463		ahc->unpause = 0;
4464	ahc->pause = ahc->unpause | PAUSE;
4465	/* XXX The shared scb data stuff should be deprecated */
4466	if (ahc->scb_data == NULL) {
4467		ahc->scb_data = kmalloc(sizeof(*ahc->scb_data), GFP_ATOMIC);
4468		if (ahc->scb_data == NULL)
4469			return (ENOMEM);
4470		memset(ahc->scb_data, 0, sizeof(*ahc->scb_data));
4471	}
4472
4473	return (0);
4474}
4475
4476void
4477ahc_set_unit(struct ahc_softc *ahc, int unit)
4478{
4479	ahc->unit = unit;
4480}
4481
4482void
4483ahc_set_name(struct ahc_softc *ahc, char *name)
4484{
4485	if (ahc->name != NULL)
4486		kfree(ahc->name);
4487	ahc->name = name;
4488}
4489
4490void
4491ahc_free(struct ahc_softc *ahc)
4492{
4493	int i;
4494
4495	switch (ahc->init_level) {
4496	default:
4497	case 5:
4498		ahc_shutdown(ahc);
4499		/* FALLTHROUGH */
4500	case 4:
4501		ahc_dmamap_unload(ahc, ahc->shared_data_dmat,
4502				  ahc->shared_data_dmamap);
4503		/* FALLTHROUGH */
4504	case 3:
4505		ahc_dmamem_free(ahc, ahc->shared_data_dmat, ahc->qoutfifo,
4506				ahc->shared_data_dmamap);
4507		ahc_dmamap_destroy(ahc, ahc->shared_data_dmat,
4508				   ahc->shared_data_dmamap);
4509		/* FALLTHROUGH */
4510	case 2:
4511		ahc_dma_tag_destroy(ahc, ahc->shared_data_dmat);
4512	case 1:
4513#ifndef __linux__
4514		ahc_dma_tag_destroy(ahc, ahc->buffer_dmat);
4515#endif
4516		break;
4517	case 0:
4518		break;
4519	}
4520
4521#ifndef __linux__
4522	ahc_dma_tag_destroy(ahc, ahc->parent_dmat);
4523#endif
4524	ahc_platform_free(ahc);
4525	ahc_fini_scbdata(ahc);
4526	for (i = 0; i < AHC_NUM_TARGETS; i++) {
4527		struct ahc_tmode_tstate *tstate;
4528
4529		tstate = ahc->enabled_targets[i];
4530		if (tstate != NULL) {
4531#ifdef AHC_TARGET_MODE
4532			int j;
4533
4534			for (j = 0; j < AHC_NUM_LUNS; j++) {
4535				struct ahc_tmode_lstate *lstate;
4536
4537				lstate = tstate->enabled_luns[j];
4538				if (lstate != NULL) {
4539					xpt_free_path(lstate->path);
4540					kfree(lstate);
4541				}
4542			}
4543#endif
4544			kfree(tstate);
4545		}
4546	}
4547#ifdef AHC_TARGET_MODE
4548	if (ahc->black_hole != NULL) {
4549		xpt_free_path(ahc->black_hole->path);
4550		kfree(ahc->black_hole);
4551	}
4552#endif
4553	if (ahc->name != NULL)
4554		kfree(ahc->name);
4555	if (ahc->seep_config != NULL)
4556		kfree(ahc->seep_config);
4557#ifndef __FreeBSD__
4558	kfree(ahc);
4559#endif
4560	return;
4561}
4562
4563static void
4564ahc_shutdown(void *arg)
4565{
4566	struct	ahc_softc *ahc;
4567	int	i;
4568
4569	ahc = (struct ahc_softc *)arg;
4570
4571	/* This will reset most registers to 0, but not all */
4572	ahc_reset(ahc, /*reinit*/FALSE);
4573	ahc_outb(ahc, SCSISEQ, 0);
4574	ahc_outb(ahc, SXFRCTL0, 0);
4575	ahc_outb(ahc, DSPCISTATUS, 0);
4576
4577	for (i = TARG_SCSIRATE; i < SCSICONF; i++)
4578		ahc_outb(ahc, i, 0);
4579}
4580
4581/*
4582 * Reset the controller and record some information about it
4583 * that is only available just after a reset.  If "reinit" is
4584 * non-zero, this reset occurred after initial configuration
4585 * and the caller requests that the chip be fully reinitialized
4586 * to a runable state.  Chip interrupts are *not* enabled after
4587 * a reinitialization.  The caller must enable interrupts via
4588 * ahc_intr_enable().
4589 */
4590int
4591ahc_reset(struct ahc_softc *ahc, int reinit)
4592{
4593	u_int	sblkctl;
4594	u_int	sxfrctl1_a, sxfrctl1_b;
4595	int	error;
4596	int	wait;
4597
4598	/*
4599	 * Preserve the value of the SXFRCTL1 register for all channels.
4600	 * It contains settings that affect termination and we don't want
4601	 * to disturb the integrity of the bus.
4602	 */
4603	ahc_pause(ahc);
4604	sxfrctl1_b = 0;
4605	if ((ahc->chip & AHC_CHIPID_MASK) == AHC_AIC7770) {
4606		u_int sblkctl;
4607
4608		/*
4609		 * Save channel B's settings in case this chip
4610		 * is setup for TWIN channel operation.
4611		 */
4612		sblkctl = ahc_inb(ahc, SBLKCTL);
4613		ahc_outb(ahc, SBLKCTL, sblkctl | SELBUSB);
4614		sxfrctl1_b = ahc_inb(ahc, SXFRCTL1);
4615		ahc_outb(ahc, SBLKCTL, sblkctl & ~SELBUSB);
4616	}
4617	sxfrctl1_a = ahc_inb(ahc, SXFRCTL1);
4618
4619	ahc_outb(ahc, HCNTRL, CHIPRST | ahc->pause);
4620
4621	/*
4622	 * Ensure that the reset has finished.  We delay 1000us
4623	 * prior to reading the register to make sure the chip
4624	 * has sufficiently completed its reset to handle register
4625	 * accesses.
4626	 */
4627	wait = 1000;
4628	do {
4629		ahc_delay(1000);
4630	} while (--wait && !(ahc_inb(ahc, HCNTRL) & CHIPRSTACK));
4631
4632	if (wait == 0) {
4633		printk("%s: WARNING - Failed chip reset!  "
4634		       "Trying to initialize anyway.\n", ahc_name(ahc));
4635	}
4636	ahc_outb(ahc, HCNTRL, ahc->pause);
4637
4638	/* Determine channel configuration */
4639	sblkctl = ahc_inb(ahc, SBLKCTL) & (SELBUSB|SELWIDE);
4640	/* No Twin Channel PCI cards */
4641	if ((ahc->chip & AHC_PCI) != 0)
4642		sblkctl &= ~SELBUSB;
4643	switch (sblkctl) {
4644	case 0:
4645		/* Single Narrow Channel */
4646		break;
4647	case 2:
4648		/* Wide Channel */
4649		ahc->features |= AHC_WIDE;
4650		break;
4651	case 8:
4652		/* Twin Channel */
4653		ahc->features |= AHC_TWIN;
4654		break;
4655	default:
4656		printk(" Unsupported adapter type.  Ignoring\n");
4657		return(-1);
4658	}
4659
4660	/*
4661	 * Reload sxfrctl1.
4662	 *
4663	 * We must always initialize STPWEN to 1 before we
4664	 * restore the saved values.  STPWEN is initialized
4665	 * to a tri-state condition which can only be cleared
4666	 * by turning it on.
4667	 */
4668	if ((ahc->features & AHC_TWIN) != 0) {
4669		u_int sblkctl;
4670
4671		sblkctl = ahc_inb(ahc, SBLKCTL);
4672		ahc_outb(ahc, SBLKCTL, sblkctl | SELBUSB);
4673		ahc_outb(ahc, SXFRCTL1, sxfrctl1_b);
4674		ahc_outb(ahc, SBLKCTL, sblkctl & ~SELBUSB);
4675	}
4676	ahc_outb(ahc, SXFRCTL1, sxfrctl1_a);
4677
4678	error = 0;
4679	if (reinit != 0)
4680		/*
4681		 * If a recovery action has forced a chip reset,
4682		 * re-initialize the chip to our liking.
4683		 */
4684		error = ahc->bus_chip_init(ahc);
4685#ifdef AHC_DUMP_SEQ
4686	else
4687		ahc_dumpseq(ahc);
4688#endif
4689
4690	return (error);
4691}
4692
4693/*
4694 * Determine the number of SCBs available on the controller
4695 */
4696int
4697ahc_probe_scbs(struct ahc_softc *ahc) {
4698	int i;
4699
4700	for (i = 0; i < AHC_SCB_MAX; i++) {
4701
4702		ahc_outb(ahc, SCBPTR, i);
4703		ahc_outb(ahc, SCB_BASE, i);
4704		if (ahc_inb(ahc, SCB_BASE) != i)
4705			break;
4706		ahc_outb(ahc, SCBPTR, 0);
4707		if (ahc_inb(ahc, SCB_BASE) != 0)
4708			break;
4709	}
4710	return (i);
4711}
4712
4713static void
4714ahc_dmamap_cb(void *arg, bus_dma_segment_t *segs, int nseg, int error)
4715{
4716	dma_addr_t *baddr;
4717
4718	baddr = (dma_addr_t *)arg;
4719	*baddr = segs->ds_addr;
4720}
4721
4722static void
4723ahc_build_free_scb_list(struct ahc_softc *ahc)
4724{
4725	int scbsize;
4726	int i;
4727
4728	scbsize = 32;
4729	if ((ahc->flags & AHC_LSCBS_ENABLED) != 0)
4730		scbsize = 64;
4731
4732	for (i = 0; i < ahc->scb_data->maxhscbs; i++) {
4733		int j;
4734
4735		ahc_outb(ahc, SCBPTR, i);
4736
4737		/*
4738		 * Touch all SCB bytes to avoid parity errors
4739		 * should one of our debugging routines read
4740		 * an otherwise uninitiatlized byte.
4741		 */
4742		for (j = 0; j < scbsize; j++)
4743			ahc_outb(ahc, SCB_BASE+j, 0xFF);
4744
4745		/* Clear the control byte. */
4746		ahc_outb(ahc, SCB_CONTROL, 0);
4747
4748		/* Set the next pointer */
4749		if ((ahc->flags & AHC_PAGESCBS) != 0)
4750			ahc_outb(ahc, SCB_NEXT, i+1);
4751		else
4752			ahc_outb(ahc, SCB_NEXT, SCB_LIST_NULL);
4753
4754		/* Make the tag number, SCSIID, and lun invalid */
4755		ahc_outb(ahc, SCB_TAG, SCB_LIST_NULL);
4756		ahc_outb(ahc, SCB_SCSIID, 0xFF);
4757		ahc_outb(ahc, SCB_LUN, 0xFF);
4758	}
4759
4760	if ((ahc->flags & AHC_PAGESCBS) != 0) {
4761		/* SCB 0 heads the free list. */
4762		ahc_outb(ahc, FREE_SCBH, 0);
4763	} else {
4764		/* No free list. */
4765		ahc_outb(ahc, FREE_SCBH, SCB_LIST_NULL);
4766	}
4767
4768	/* Make sure that the last SCB terminates the free list */
4769	ahc_outb(ahc, SCBPTR, i-1);
4770	ahc_outb(ahc, SCB_NEXT, SCB_LIST_NULL);
4771}
4772
4773static int
4774ahc_init_scbdata(struct ahc_softc *ahc)
4775{
4776	struct scb_data *scb_data;
4777
4778	scb_data = ahc->scb_data;
4779	SLIST_INIT(&scb_data->free_scbs);
4780	SLIST_INIT(&scb_data->sg_maps);
4781
4782	/* Allocate SCB resources */
4783	scb_data->scbarray = kmalloc(sizeof(struct scb) * AHC_SCB_MAX_ALLOC, GFP_ATOMIC);
4784	if (scb_data->scbarray == NULL)
4785		return (ENOMEM);
4786	memset(scb_data->scbarray, 0, sizeof(struct scb) * AHC_SCB_MAX_ALLOC);
4787
4788	/* Determine the number of hardware SCBs and initialize them */
4789
4790	scb_data->maxhscbs = ahc_probe_scbs(ahc);
4791	if (ahc->scb_data->maxhscbs == 0) {
4792		printk("%s: No SCB space found\n", ahc_name(ahc));
4793		return (ENXIO);
4794	}
4795
4796	/*
4797	 * Create our DMA tags.  These tags define the kinds of device
4798	 * accessible memory allocations and memory mappings we will
4799	 * need to perform during normal operation.
4800	 *
4801	 * Unless we need to further restrict the allocation, we rely
4802	 * on the restrictions of the parent dmat, hence the common
4803	 * use of MAXADDR and MAXSIZE.
4804	 */
4805
4806	/* DMA tag for our hardware scb structures */
4807	if (ahc_dma_tag_create(ahc, ahc->parent_dmat, /*alignment*/1,
4808			       /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1,
4809			       /*lowaddr*/BUS_SPACE_MAXADDR_32BIT,
4810			       /*highaddr*/BUS_SPACE_MAXADDR,
4811			       /*filter*/NULL, /*filterarg*/NULL,
4812			       AHC_SCB_MAX_ALLOC * sizeof(struct hardware_scb),
4813			       /*nsegments*/1,
4814			       /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
4815			       /*flags*/0, &scb_data->hscb_dmat) != 0) {
4816		goto error_exit;
4817	}
4818
4819	scb_data->init_level++;
4820
4821	/* Allocation for our hscbs */
4822	if (ahc_dmamem_alloc(ahc, scb_data->hscb_dmat,
4823			     (void **)&scb_data->hscbs,
4824			     BUS_DMA_NOWAIT, &scb_data->hscb_dmamap) != 0) {
4825		goto error_exit;
4826	}
4827
4828	scb_data->init_level++;
4829
4830	/* And permanently map them */
4831	ahc_dmamap_load(ahc, scb_data->hscb_dmat, scb_data->hscb_dmamap,
4832			scb_data->hscbs,
4833			AHC_SCB_MAX_ALLOC * sizeof(struct hardware_scb),
4834			ahc_dmamap_cb, &scb_data->hscb_busaddr, /*flags*/0);
4835
4836	scb_data->init_level++;
4837
4838	/* DMA tag for our sense buffers */
4839	if (ahc_dma_tag_create(ahc, ahc->parent_dmat, /*alignment*/1,
4840			       /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1,
4841			       /*lowaddr*/BUS_SPACE_MAXADDR_32BIT,
4842			       /*highaddr*/BUS_SPACE_MAXADDR,
4843			       /*filter*/NULL, /*filterarg*/NULL,
4844			       AHC_SCB_MAX_ALLOC * sizeof(struct scsi_sense_data),
4845			       /*nsegments*/1,
4846			       /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
4847			       /*flags*/0, &scb_data->sense_dmat) != 0) {
4848		goto error_exit;
4849	}
4850
4851	scb_data->init_level++;
4852
4853	/* Allocate them */
4854	if (ahc_dmamem_alloc(ahc, scb_data->sense_dmat,
4855			     (void **)&scb_data->sense,
4856			     BUS_DMA_NOWAIT, &scb_data->sense_dmamap) != 0) {
4857		goto error_exit;
4858	}
4859
4860	scb_data->init_level++;
4861
4862	/* And permanently map them */
4863	ahc_dmamap_load(ahc, scb_data->sense_dmat, scb_data->sense_dmamap,
4864			scb_data->sense,
4865			AHC_SCB_MAX_ALLOC * sizeof(struct scsi_sense_data),
4866			ahc_dmamap_cb, &scb_data->sense_busaddr, /*flags*/0);
4867
4868	scb_data->init_level++;
4869
4870	/* DMA tag for our S/G structures.  We allocate in page sized chunks */
4871	if (ahc_dma_tag_create(ahc, ahc->parent_dmat, /*alignment*/8,
4872			       /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1,
4873			       /*lowaddr*/BUS_SPACE_MAXADDR_32BIT,
4874			       /*highaddr*/BUS_SPACE_MAXADDR,
4875			       /*filter*/NULL, /*filterarg*/NULL,
4876			       PAGE_SIZE, /*nsegments*/1,
4877			       /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
4878			       /*flags*/0, &scb_data->sg_dmat) != 0) {
4879		goto error_exit;
4880	}
4881
4882	scb_data->init_level++;
4883
4884	/* Perform initial CCB allocation */
4885	memset(scb_data->hscbs, 0,
4886	       AHC_SCB_MAX_ALLOC * sizeof(struct hardware_scb));
4887	ahc_alloc_scbs(ahc);
4888
4889	if (scb_data->numscbs == 0) {
4890		printk("%s: ahc_init_scbdata - "
4891		       "Unable to allocate initial scbs\n",
4892		       ahc_name(ahc));
4893		goto error_exit;
4894	}
4895
4896	/*
4897	 * Reserve the next queued SCB.
4898	 */
4899	ahc->next_queued_scb = ahc_get_scb(ahc);
4900
4901	/*
4902	 * Note that we were successful
4903	 */
4904	return (0);
4905
4906error_exit:
4907
4908	return (ENOMEM);
4909}
4910
4911static void
4912ahc_fini_scbdata(struct ahc_softc *ahc)
4913{
4914	struct scb_data *scb_data;
4915
4916	scb_data = ahc->scb_data;
4917	if (scb_data == NULL)
4918		return;
4919
4920	switch (scb_data->init_level) {
4921	default:
4922	case 7:
4923	{
4924		struct sg_map_node *sg_map;
4925
4926		while ((sg_map = SLIST_FIRST(&scb_data->sg_maps))!= NULL) {
4927			SLIST_REMOVE_HEAD(&scb_data->sg_maps, links);
4928			ahc_dmamap_unload(ahc, scb_data->sg_dmat,
4929					  sg_map->sg_dmamap);
4930			ahc_dmamem_free(ahc, scb_data->sg_dmat,
4931					sg_map->sg_vaddr,
4932					sg_map->sg_dmamap);
4933			kfree(sg_map);
4934		}
4935		ahc_dma_tag_destroy(ahc, scb_data->sg_dmat);
4936	}
4937	case 6:
4938		ahc_dmamap_unload(ahc, scb_data->sense_dmat,
4939				  scb_data->sense_dmamap);
4940	case 5:
4941		ahc_dmamem_free(ahc, scb_data->sense_dmat, scb_data->sense,
4942				scb_data->sense_dmamap);
4943		ahc_dmamap_destroy(ahc, scb_data->sense_dmat,
4944				   scb_data->sense_dmamap);
4945	case 4:
4946		ahc_dma_tag_destroy(ahc, scb_data->sense_dmat);
4947	case 3:
4948		ahc_dmamap_unload(ahc, scb_data->hscb_dmat,
4949				  scb_data->hscb_dmamap);
4950	case 2:
4951		ahc_dmamem_free(ahc, scb_data->hscb_dmat, scb_data->hscbs,
4952				scb_data->hscb_dmamap);
4953		ahc_dmamap_destroy(ahc, scb_data->hscb_dmat,
4954				   scb_data->hscb_dmamap);
4955	case 1:
4956		ahc_dma_tag_destroy(ahc, scb_data->hscb_dmat);
4957		break;
4958	case 0:
4959		break;
4960	}
4961	if (scb_data->scbarray != NULL)
4962		kfree(scb_data->scbarray);
4963}
4964
4965static void
4966ahc_alloc_scbs(struct ahc_softc *ahc)
4967{
4968	struct scb_data *scb_data;
4969	struct scb *next_scb;
4970	struct sg_map_node *sg_map;
4971	dma_addr_t physaddr;
4972	struct ahc_dma_seg *segs;
4973	int newcount;
4974	int i;
4975
4976	scb_data = ahc->scb_data;
4977	if (scb_data->numscbs >= AHC_SCB_MAX_ALLOC)
4978		/* Can't allocate any more */
4979		return;
4980
4981	next_scb = &scb_data->scbarray[scb_data->numscbs];
4982
4983	sg_map = kmalloc(sizeof(*sg_map), GFP_ATOMIC);
4984
4985	if (sg_map == NULL)
4986		return;
4987
4988	/* Allocate S/G space for the next batch of SCBS */
4989	if (ahc_dmamem_alloc(ahc, scb_data->sg_dmat,
4990			     (void **)&sg_map->sg_vaddr,
4991			     BUS_DMA_NOWAIT, &sg_map->sg_dmamap) != 0) {
4992		kfree(sg_map);
4993		return;
4994	}
4995
4996	SLIST_INSERT_HEAD(&scb_data->sg_maps, sg_map, links);
4997
4998	ahc_dmamap_load(ahc, scb_data->sg_dmat, sg_map->sg_dmamap,
4999			sg_map->sg_vaddr, PAGE_SIZE, ahc_dmamap_cb,
5000			&sg_map->sg_physaddr, /*flags*/0);
5001
5002	segs = sg_map->sg_vaddr;
5003	physaddr = sg_map->sg_physaddr;
5004
5005	newcount = (PAGE_SIZE / (AHC_NSEG * sizeof(struct ahc_dma_seg)));
5006	newcount = min(newcount, (AHC_SCB_MAX_ALLOC - scb_data->numscbs));
5007	for (i = 0; i < newcount; i++) {
5008		struct scb_platform_data *pdata;
5009#ifndef __linux__
5010		int error;
5011#endif
5012		pdata = kmalloc(sizeof(*pdata), GFP_ATOMIC);
5013		if (pdata == NULL)
5014			break;
5015		next_scb->platform_data = pdata;
5016		next_scb->sg_map = sg_map;
5017		next_scb->sg_list = segs;
5018		/*
5019		 * The sequencer always starts with the second entry.
5020		 * The first entry is embedded in the scb.
5021		 */
5022		next_scb->sg_list_phys = physaddr + sizeof(struct ahc_dma_seg);
5023		next_scb->ahc_softc = ahc;
5024		next_scb->flags = SCB_FREE;
5025#ifndef __linux__
5026		error = ahc_dmamap_create(ahc, ahc->buffer_dmat, /*flags*/0,
5027					  &next_scb->dmamap);
5028		if (error != 0)
5029			break;
5030#endif
5031		next_scb->hscb = &scb_data->hscbs[scb_data->numscbs];
5032		next_scb->hscb->tag = ahc->scb_data->numscbs;
5033		SLIST_INSERT_HEAD(&ahc->scb_data->free_scbs,
5034				  next_scb, links.sle);
5035		segs += AHC_NSEG;
5036		physaddr += (AHC_NSEG * sizeof(struct ahc_dma_seg));
5037		next_scb++;
5038		ahc->scb_data->numscbs++;
5039	}
5040}
5041
5042void
5043ahc_controller_info(struct ahc_softc *ahc, char *buf)
5044{
5045	int len;
5046
5047	len = sprintf(buf, "%s: ", ahc_chip_names[ahc->chip & AHC_CHIPID_MASK]);
5048	buf += len;
5049	if ((ahc->features & AHC_TWIN) != 0)
5050 		len = sprintf(buf, "Twin Channel, A SCSI Id=%d, "
5051			      "B SCSI Id=%d, primary %c, ",
5052			      ahc->our_id, ahc->our_id_b,
5053			      (ahc->flags & AHC_PRIMARY_CHANNEL) + 'A');
5054	else {
5055		const char *speed;
5056		const char *type;
5057
5058		speed = "";
5059		if ((ahc->features & AHC_ULTRA) != 0) {
5060			speed = "Ultra ";
5061		} else if ((ahc->features & AHC_DT) != 0) {
5062			speed = "Ultra160 ";
5063		} else if ((ahc->features & AHC_ULTRA2) != 0) {
5064			speed = "Ultra2 ";
5065		}
5066		if ((ahc->features & AHC_WIDE) != 0) {
5067			type = "Wide";
5068		} else {
5069			type = "Single";
5070		}
5071		len = sprintf(buf, "%s%s Channel %c, SCSI Id=%d, ",
5072			      speed, type, ahc->channel, ahc->our_id);
5073	}
5074	buf += len;
5075
5076	if ((ahc->flags & AHC_PAGESCBS) != 0)
5077		sprintf(buf, "%d/%d SCBs",
5078			ahc->scb_data->maxhscbs, AHC_MAX_QUEUE);
5079	else
5080		sprintf(buf, "%d SCBs", ahc->scb_data->maxhscbs);
5081}
5082
5083int
5084ahc_chip_init(struct ahc_softc *ahc)
5085{
5086	int	 term;
5087	int	 error;
5088	u_int	 i;
5089	u_int	 scsi_conf;
5090	u_int	 scsiseq_template;
5091	uint32_t physaddr;
5092
5093	ahc_outb(ahc, SEQ_FLAGS, 0);
5094	ahc_outb(ahc, SEQ_FLAGS2, 0);
5095
5096	/* Set the SCSI Id, SXFRCTL0, SXFRCTL1, and SIMODE1, for both channels*/
5097	if (ahc->features & AHC_TWIN) {
5098
5099		/*
5100		 * Setup Channel B first.
5101		 */
5102		ahc_outb(ahc, SBLKCTL, ahc_inb(ahc, SBLKCTL) | SELBUSB);
5103		term = (ahc->flags & AHC_TERM_ENB_B) != 0 ? STPWEN : 0;
5104		ahc_outb(ahc, SCSIID, ahc->our_id_b);
5105		scsi_conf = ahc_inb(ahc, SCSICONF + 1);
5106		ahc_outb(ahc, SXFRCTL1, (scsi_conf & (ENSPCHK|STIMESEL))
5107					|term|ahc->seltime_b|ENSTIMER|ACTNEGEN);
5108		if ((ahc->features & AHC_ULTRA2) != 0)
5109			ahc_outb(ahc, SIMODE0, ahc_inb(ahc, SIMODE0)|ENIOERR);
5110		ahc_outb(ahc, SIMODE1, ENSELTIMO|ENSCSIRST|ENSCSIPERR);
5111		ahc_outb(ahc, SXFRCTL0, DFON|SPIOEN);
5112
5113		/* Select Channel A */
5114		ahc_outb(ahc, SBLKCTL, ahc_inb(ahc, SBLKCTL) & ~SELBUSB);
5115	}
5116	term = (ahc->flags & AHC_TERM_ENB_A) != 0 ? STPWEN : 0;
5117	if ((ahc->features & AHC_ULTRA2) != 0)
5118		ahc_outb(ahc, SCSIID_ULTRA2, ahc->our_id);
5119	else
5120		ahc_outb(ahc, SCSIID, ahc->our_id);
5121	scsi_conf = ahc_inb(ahc, SCSICONF);
5122	ahc_outb(ahc, SXFRCTL1, (scsi_conf & (ENSPCHK|STIMESEL))
5123				|term|ahc->seltime
5124				|ENSTIMER|ACTNEGEN);
5125	if ((ahc->features & AHC_ULTRA2) != 0)
5126		ahc_outb(ahc, SIMODE0, ahc_inb(ahc, SIMODE0)|ENIOERR);
5127	ahc_outb(ahc, SIMODE1, ENSELTIMO|ENSCSIRST|ENSCSIPERR);
5128	ahc_outb(ahc, SXFRCTL0, DFON|SPIOEN);
5129
5130	/* There are no untagged SCBs active yet. */
5131	for (i = 0; i < 16; i++) {
5132		ahc_unbusy_tcl(ahc, BUILD_TCL(i << 4, 0));
5133		if ((ahc->flags & AHC_SCB_BTT) != 0) {
5134			int lun;
5135
5136			/*
5137			 * The SCB based BTT allows an entry per
5138			 * target and lun pair.
5139			 */
5140			for (lun = 1; lun < AHC_NUM_LUNS; lun++)
5141				ahc_unbusy_tcl(ahc, BUILD_TCL(i << 4, lun));
5142		}
5143	}
5144
5145	/* All of our queues are empty */
5146	for (i = 0; i < 256; i++)
5147		ahc->qoutfifo[i] = SCB_LIST_NULL;
5148	ahc_sync_qoutfifo(ahc, BUS_DMASYNC_PREREAD);
5149
5150	for (i = 0; i < 256; i++)
5151		ahc->qinfifo[i] = SCB_LIST_NULL;
5152
5153	if ((ahc->features & AHC_MULTI_TID) != 0) {
5154		ahc_outb(ahc, TARGID, 0);
5155		ahc_outb(ahc, TARGID + 1, 0);
5156	}
5157
5158	/*
5159	 * Tell the sequencer where it can find our arrays in memory.
5160	 */
5161	physaddr = ahc->scb_data->hscb_busaddr;
5162	ahc_outb(ahc, HSCB_ADDR, physaddr & 0xFF);
5163	ahc_outb(ahc, HSCB_ADDR + 1, (physaddr >> 8) & 0xFF);
5164	ahc_outb(ahc, HSCB_ADDR + 2, (physaddr >> 16) & 0xFF);
5165	ahc_outb(ahc, HSCB_ADDR + 3, (physaddr >> 24) & 0xFF);
5166
5167	physaddr = ahc->shared_data_busaddr;
5168	ahc_outb(ahc, SHARED_DATA_ADDR, physaddr & 0xFF);
5169	ahc_outb(ahc, SHARED_DATA_ADDR + 1, (physaddr >> 8) & 0xFF);
5170	ahc_outb(ahc, SHARED_DATA_ADDR + 2, (physaddr >> 16) & 0xFF);
5171	ahc_outb(ahc, SHARED_DATA_ADDR + 3, (physaddr >> 24) & 0xFF);
5172
5173	/*
5174	 * Initialize the group code to command length table.
5175	 * This overrides the values in TARG_SCSIRATE, so only
5176	 * setup the table after we have processed that information.
5177	 */
5178	ahc_outb(ahc, CMDSIZE_TABLE, 5);
5179	ahc_outb(ahc, CMDSIZE_TABLE + 1, 9);
5180	ahc_outb(ahc, CMDSIZE_TABLE + 2, 9);
5181	ahc_outb(ahc, CMDSIZE_TABLE + 3, 0);
5182	ahc_outb(ahc, CMDSIZE_TABLE + 4, 15);
5183	ahc_outb(ahc, CMDSIZE_TABLE + 5, 11);
5184	ahc_outb(ahc, CMDSIZE_TABLE + 6, 0);
5185	ahc_outb(ahc, CMDSIZE_TABLE + 7, 0);
5186
5187	if ((ahc->features & AHC_HS_MAILBOX) != 0)
5188		ahc_outb(ahc, HS_MAILBOX, 0);
5189
5190	/* Tell the sequencer of our initial queue positions */
5191	if ((ahc->features & AHC_TARGETMODE) != 0) {
5192		ahc->tqinfifonext = 1;
5193		ahc_outb(ahc, KERNEL_TQINPOS, ahc->tqinfifonext - 1);
5194		ahc_outb(ahc, TQINPOS, ahc->tqinfifonext);
5195	}
5196	ahc->qinfifonext = 0;
5197	ahc->qoutfifonext = 0;
5198	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
5199		ahc_outb(ahc, QOFF_CTLSTA, SCB_QSIZE_256);
5200		ahc_outb(ahc, HNSCB_QOFF, ahc->qinfifonext);
5201		ahc_outb(ahc, SNSCB_QOFF, ahc->qinfifonext);
5202		ahc_outb(ahc, SDSCB_QOFF, 0);
5203	} else {
5204		ahc_outb(ahc, KERNEL_QINPOS, ahc->qinfifonext);
5205		ahc_outb(ahc, QINPOS, ahc->qinfifonext);
5206		ahc_outb(ahc, QOUTPOS, ahc->qoutfifonext);
5207	}
5208
5209	/* We don't have any waiting selections */
5210	ahc_outb(ahc, WAITING_SCBH, SCB_LIST_NULL);
5211
5212	/* Our disconnection list is empty too */
5213	ahc_outb(ahc, DISCONNECTED_SCBH, SCB_LIST_NULL);
5214
5215	/* Message out buffer starts empty */
5216	ahc_outb(ahc, MSG_OUT, MSG_NOOP);
5217
5218	/*
5219	 * Setup the allowed SCSI Sequences based on operational mode.
5220	 * If we are a target, we'll enable select in operations once
5221	 * we've had a lun enabled.
5222	 */
5223	scsiseq_template = ENSELO|ENAUTOATNO|ENAUTOATNP;
5224	if ((ahc->flags & AHC_INITIATORROLE) != 0)
5225		scsiseq_template |= ENRSELI;
5226	ahc_outb(ahc, SCSISEQ_TEMPLATE, scsiseq_template);
5227
5228	/* Initialize our list of free SCBs. */
5229	ahc_build_free_scb_list(ahc);
5230
5231	/*
5232	 * Tell the sequencer which SCB will be the next one it receives.
5233	 */
5234	ahc_outb(ahc, NEXT_QUEUED_SCB, ahc->next_queued_scb->hscb->tag);
5235
5236	/*
5237	 * Load the Sequencer program and Enable the adapter
5238	 * in "fast" mode.
5239	 */
5240	if (bootverbose)
5241		printk("%s: Downloading Sequencer Program...",
5242		       ahc_name(ahc));
5243
5244	error = ahc_loadseq(ahc);
5245	if (error != 0)
5246		return (error);
5247
5248	if ((ahc->features & AHC_ULTRA2) != 0) {
5249		int wait;
5250
5251		/*
5252		 * Wait for up to 500ms for our transceivers
5253		 * to settle.  If the adapter does not have
5254		 * a cable attached, the transceivers may
5255		 * never settle, so don't complain if we
5256		 * fail here.
5257		 */
5258		for (wait = 5000;
5259		     (ahc_inb(ahc, SBLKCTL) & (ENAB40|ENAB20)) == 0 && wait;
5260		     wait--)
5261			ahc_delay(100);
5262	}
5263	ahc_restart(ahc);
5264	return (0);
5265}
5266
5267/*
5268 * Start the board, ready for normal operation
5269 */
5270int
5271ahc_init(struct ahc_softc *ahc)
5272{
5273	int	 max_targ;
5274	u_int	 i;
5275	u_int	 scsi_conf;
5276	u_int	 ultraenb;
5277	u_int	 discenable;
5278	u_int	 tagenable;
5279	size_t	 driver_data_size;
5280
5281#ifdef AHC_DEBUG
5282	if ((ahc_debug & AHC_DEBUG_SEQUENCER) != 0)
5283		ahc->flags |= AHC_SEQUENCER_DEBUG;
5284#endif
5285
5286#ifdef AHC_PRINT_SRAM
5287	printk("Scratch Ram:");
5288	for (i = 0x20; i < 0x5f; i++) {
5289		if (((i % 8) == 0) && (i != 0)) {
5290			printk ("\n              ");
5291		}
5292		printk (" 0x%x", ahc_inb(ahc, i));
5293	}
5294	if ((ahc->features & AHC_MORE_SRAM) != 0) {
5295		for (i = 0x70; i < 0x7f; i++) {
5296			if (((i % 8) == 0) && (i != 0)) {
5297				printk ("\n              ");
5298			}
5299			printk (" 0x%x", ahc_inb(ahc, i));
5300		}
5301	}
5302	printk ("\n");
5303	/*
5304	 * Reading uninitialized scratch ram may
5305	 * generate parity errors.
5306	 */
5307	ahc_outb(ahc, CLRINT, CLRPARERR);
5308	ahc_outb(ahc, CLRINT, CLRBRKADRINT);
5309#endif
5310	max_targ = 15;
5311
5312	/*
5313	 * Assume we have a board at this stage and it has been reset.
5314	 */
5315	if ((ahc->flags & AHC_USEDEFAULTS) != 0)
5316		ahc->our_id = ahc->our_id_b = 7;
5317
5318	/*
5319	 * Default to allowing initiator operations.
5320	 */
5321	ahc->flags |= AHC_INITIATORROLE;
5322
5323	/*
5324	 * Only allow target mode features if this unit has them enabled.
5325	 */
5326	if ((AHC_TMODE_ENABLE & (0x1 << ahc->unit)) == 0)
5327		ahc->features &= ~AHC_TARGETMODE;
5328
5329#ifndef __linux__
5330	/* DMA tag for mapping buffers into device visible space. */
5331	if (ahc_dma_tag_create(ahc, ahc->parent_dmat, /*alignment*/1,
5332			       /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1,
5333			       /*lowaddr*/ahc->flags & AHC_39BIT_ADDRESSING
5334					? (dma_addr_t)0x7FFFFFFFFFULL
5335					: BUS_SPACE_MAXADDR_32BIT,
5336			       /*highaddr*/BUS_SPACE_MAXADDR,
5337			       /*filter*/NULL, /*filterarg*/NULL,
5338			       /*maxsize*/(AHC_NSEG - 1) * PAGE_SIZE,
5339			       /*nsegments*/AHC_NSEG,
5340			       /*maxsegsz*/AHC_MAXTRANSFER_SIZE,
5341			       /*flags*/BUS_DMA_ALLOCNOW,
5342			       &ahc->buffer_dmat) != 0) {
5343		return (ENOMEM);
5344	}
5345#endif
5346
5347	ahc->init_level++;
5348
5349	/*
5350	 * DMA tag for our command fifos and other data in system memory
5351	 * the card's sequencer must be able to access.  For initiator
5352	 * roles, we need to allocate space for the qinfifo and qoutfifo.
5353	 * The qinfifo and qoutfifo are composed of 256 1 byte elements.
5354	 * When providing for the target mode role, we must additionally
5355	 * provide space for the incoming target command fifo and an extra
5356	 * byte to deal with a dma bug in some chip versions.
5357	 */
5358	driver_data_size = 2 * 256 * sizeof(uint8_t);
5359	if ((ahc->features & AHC_TARGETMODE) != 0)
5360		driver_data_size += AHC_TMODE_CMDS * sizeof(struct target_cmd)
5361				 + /*DMA WideOdd Bug Buffer*/1;
5362	if (ahc_dma_tag_create(ahc, ahc->parent_dmat, /*alignment*/1,
5363			       /*boundary*/BUS_SPACE_MAXADDR_32BIT + 1,
5364			       /*lowaddr*/BUS_SPACE_MAXADDR_32BIT,
5365			       /*highaddr*/BUS_SPACE_MAXADDR,
5366			       /*filter*/NULL, /*filterarg*/NULL,
5367			       driver_data_size,
5368			       /*nsegments*/1,
5369			       /*maxsegsz*/BUS_SPACE_MAXSIZE_32BIT,
5370			       /*flags*/0, &ahc->shared_data_dmat) != 0) {
5371		return (ENOMEM);
5372	}
5373
5374	ahc->init_level++;
5375
5376	/* Allocation of driver data */
5377	if (ahc_dmamem_alloc(ahc, ahc->shared_data_dmat,
5378			     (void **)&ahc->qoutfifo,
5379			     BUS_DMA_NOWAIT, &ahc->shared_data_dmamap) != 0) {
5380		return (ENOMEM);
5381	}
5382
5383	ahc->init_level++;
5384
5385	/* And permanently map it in */
5386	ahc_dmamap_load(ahc, ahc->shared_data_dmat, ahc->shared_data_dmamap,
5387			ahc->qoutfifo, driver_data_size, ahc_dmamap_cb,
5388			&ahc->shared_data_busaddr, /*flags*/0);
5389
5390	if ((ahc->features & AHC_TARGETMODE) != 0) {
5391		ahc->targetcmds = (struct target_cmd *)ahc->qoutfifo;
5392		ahc->qoutfifo = (uint8_t *)&ahc->targetcmds[AHC_TMODE_CMDS];
5393		ahc->dma_bug_buf = ahc->shared_data_busaddr
5394				 + driver_data_size - 1;
5395		/* All target command blocks start out invalid. */
5396		for (i = 0; i < AHC_TMODE_CMDS; i++)
5397			ahc->targetcmds[i].cmd_valid = 0;
5398		ahc_sync_tqinfifo(ahc, BUS_DMASYNC_PREREAD);
5399		ahc->qoutfifo = (uint8_t *)&ahc->targetcmds[256];
5400	}
5401	ahc->qinfifo = &ahc->qoutfifo[256];
5402
5403	ahc->init_level++;
5404
5405	/* Allocate SCB data now that buffer_dmat is initialized */
5406	if (ahc->scb_data->maxhscbs == 0)
5407		if (ahc_init_scbdata(ahc) != 0)
5408			return (ENOMEM);
5409
5410	/*
5411	 * Allocate a tstate to house information for our
5412	 * initiator presence on the bus as well as the user
5413	 * data for any target mode initiator.
5414	 */
5415	if (ahc_alloc_tstate(ahc, ahc->our_id, 'A') == NULL) {
5416		printk("%s: unable to allocate ahc_tmode_tstate.  "
5417		       "Failing attach\n", ahc_name(ahc));
5418		return (ENOMEM);
5419	}
5420
5421	if ((ahc->features & AHC_TWIN) != 0) {
5422		if (ahc_alloc_tstate(ahc, ahc->our_id_b, 'B') == NULL) {
5423			printk("%s: unable to allocate ahc_tmode_tstate.  "
5424			       "Failing attach\n", ahc_name(ahc));
5425			return (ENOMEM);
5426		}
5427	}
5428
5429	if (ahc->scb_data->maxhscbs < AHC_SCB_MAX_ALLOC) {
5430		ahc->flags |= AHC_PAGESCBS;
5431	} else {
5432		ahc->flags &= ~AHC_PAGESCBS;
5433	}
5434
5435#ifdef AHC_DEBUG
5436	if (ahc_debug & AHC_SHOW_MISC) {
5437		printk("%s: hardware scb %u bytes; kernel scb %u bytes; "
5438		       "ahc_dma %u bytes\n",
5439			ahc_name(ahc),
5440			(u_int)sizeof(struct hardware_scb),
5441			(u_int)sizeof(struct scb),
5442			(u_int)sizeof(struct ahc_dma_seg));
5443	}
5444#endif /* AHC_DEBUG */
5445
5446	/*
5447	 * Look at the information that board initialization or
5448	 * the board bios has left us.
5449	 */
5450	if (ahc->features & AHC_TWIN) {
5451		scsi_conf = ahc_inb(ahc, SCSICONF + 1);
5452		if ((scsi_conf & RESET_SCSI) != 0
5453		 && (ahc->flags & AHC_INITIATORROLE) != 0)
5454			ahc->flags |= AHC_RESET_BUS_B;
5455	}
5456
5457	scsi_conf = ahc_inb(ahc, SCSICONF);
5458	if ((scsi_conf & RESET_SCSI) != 0
5459	 && (ahc->flags & AHC_INITIATORROLE) != 0)
5460		ahc->flags |= AHC_RESET_BUS_A;
5461
5462	ultraenb = 0;
5463	tagenable = ALL_TARGETS_MASK;
5464
5465	/* Grab the disconnection disable table and invert it for our needs */
5466	if ((ahc->flags & AHC_USEDEFAULTS) != 0) {
5467		printk("%s: Host Adapter Bios disabled.  Using default SCSI "
5468			"device parameters\n", ahc_name(ahc));
5469		ahc->flags |= AHC_EXTENDED_TRANS_A|AHC_EXTENDED_TRANS_B|
5470			      AHC_TERM_ENB_A|AHC_TERM_ENB_B;
5471		discenable = ALL_TARGETS_MASK;
5472		if ((ahc->features & AHC_ULTRA) != 0)
5473			ultraenb = ALL_TARGETS_MASK;
5474	} else {
5475		discenable = ~((ahc_inb(ahc, DISC_DSB + 1) << 8)
5476			   | ahc_inb(ahc, DISC_DSB));
5477		if ((ahc->features & (AHC_ULTRA|AHC_ULTRA2)) != 0)
5478			ultraenb = (ahc_inb(ahc, ULTRA_ENB + 1) << 8)
5479				      | ahc_inb(ahc, ULTRA_ENB);
5480	}
5481
5482	if ((ahc->features & (AHC_WIDE|AHC_TWIN)) == 0)
5483		max_targ = 7;
5484
5485	for (i = 0; i <= max_targ; i++) {
5486		struct ahc_initiator_tinfo *tinfo;
5487		struct ahc_tmode_tstate *tstate;
5488		u_int our_id;
5489		u_int target_id;
5490		char channel;
5491
5492		channel = 'A';
5493		our_id = ahc->our_id;
5494		target_id = i;
5495		if (i > 7 && (ahc->features & AHC_TWIN) != 0) {
5496			channel = 'B';
5497			our_id = ahc->our_id_b;
5498			target_id = i % 8;
5499		}
5500		tinfo = ahc_fetch_transinfo(ahc, channel, our_id,
5501					    target_id, &tstate);
5502		/* Default to async narrow across the board */
5503		memset(tinfo, 0, sizeof(*tinfo));
5504		if (ahc->flags & AHC_USEDEFAULTS) {
5505			if ((ahc->features & AHC_WIDE) != 0)
5506				tinfo->user.width = MSG_EXT_WDTR_BUS_16_BIT;
5507
5508			/*
5509			 * These will be truncated when we determine the
5510			 * connection type we have with the target.
5511			 */
5512			tinfo->user.period = ahc_syncrates->period;
5513			tinfo->user.offset = MAX_OFFSET;
5514		} else {
5515			u_int scsirate;
5516			uint16_t mask;
5517
5518			/* Take the settings leftover in scratch RAM. */
5519			scsirate = ahc_inb(ahc, TARG_SCSIRATE + i);
5520			mask = (0x01 << i);
5521			if ((ahc->features & AHC_ULTRA2) != 0) {
5522				u_int offset;
5523				u_int maxsync;
5524
5525				if ((scsirate & SOFS) == 0x0F) {
5526					/*
5527					 * Haven't negotiated yet,
5528					 * so the format is different.
5529					 */
5530					scsirate = (scsirate & SXFR) >> 4
5531						 | (ultraenb & mask)
5532						  ? 0x08 : 0x0
5533						 | (scsirate & WIDEXFER);
5534					offset = MAX_OFFSET_ULTRA2;
5535				} else
5536					offset = ahc_inb(ahc, TARG_OFFSET + i);
5537				if ((scsirate & ~WIDEXFER) == 0 && offset != 0)
5538					/* Set to the lowest sync rate, 5MHz */
5539					scsirate |= 0x1c;
5540				maxsync = AHC_SYNCRATE_ULTRA2;
5541				if ((ahc->features & AHC_DT) != 0)
5542					maxsync = AHC_SYNCRATE_DT;
5543				tinfo->user.period =
5544				    ahc_find_period(ahc, scsirate, maxsync);
5545				if (offset == 0)
5546					tinfo->user.period = 0;
5547				else
5548					tinfo->user.offset = MAX_OFFSET;
5549				if ((scsirate & SXFR_ULTRA2) <= 8/*10MHz*/
5550				 && (ahc->features & AHC_DT) != 0)
5551					tinfo->user.ppr_options =
5552					    MSG_EXT_PPR_DT_REQ;
5553			} else if ((scsirate & SOFS) != 0) {
5554				if ((scsirate & SXFR) == 0x40
5555				 && (ultraenb & mask) != 0) {
5556					/* Treat 10MHz as a non-ultra speed */
5557					scsirate &= ~SXFR;
5558				 	ultraenb &= ~mask;
5559				}
5560				tinfo->user.period =
5561				    ahc_find_period(ahc, scsirate,
5562						    (ultraenb & mask)
5563						   ? AHC_SYNCRATE_ULTRA
5564						   : AHC_SYNCRATE_FAST);
5565				if (tinfo->user.period != 0)
5566					tinfo->user.offset = MAX_OFFSET;
5567			}
5568			if (tinfo->user.period == 0)
5569				tinfo->user.offset = 0;
5570			if ((scsirate & WIDEXFER) != 0
5571			 && (ahc->features & AHC_WIDE) != 0)
5572				tinfo->user.width = MSG_EXT_WDTR_BUS_16_BIT;
5573			tinfo->user.protocol_version = 4;
5574			if ((ahc->features & AHC_DT) != 0)
5575				tinfo->user.transport_version = 3;
5576			else
5577				tinfo->user.transport_version = 2;
5578			tinfo->goal.protocol_version = 2;
5579			tinfo->goal.transport_version = 2;
5580			tinfo->curr.protocol_version = 2;
5581			tinfo->curr.transport_version = 2;
5582		}
5583		tstate->ultraenb = 0;
5584	}
5585	ahc->user_discenable = discenable;
5586	ahc->user_tagenable = tagenable;
5587
5588	return (ahc->bus_chip_init(ahc));
5589}
5590
5591void
5592ahc_intr_enable(struct ahc_softc *ahc, int enable)
5593{
5594	u_int hcntrl;
5595
5596	hcntrl = ahc_inb(ahc, HCNTRL);
5597	hcntrl &= ~INTEN;
5598	ahc->pause &= ~INTEN;
5599	ahc->unpause &= ~INTEN;
5600	if (enable) {
5601		hcntrl |= INTEN;
5602		ahc->pause |= INTEN;
5603		ahc->unpause |= INTEN;
5604	}
5605	ahc_outb(ahc, HCNTRL, hcntrl);
5606}
5607
5608/*
5609 * Ensure that the card is paused in a location
5610 * outside of all critical sections and that all
5611 * pending work is completed prior to returning.
5612 * This routine should only be called from outside
5613 * an interrupt context.
5614 */
5615void
5616ahc_pause_and_flushwork(struct ahc_softc *ahc)
5617{
5618	int intstat;
5619	int maxloops;
5620	int paused;
5621
5622	maxloops = 1000;
5623	ahc->flags |= AHC_ALL_INTERRUPTS;
5624	paused = FALSE;
5625	do {
5626		if (paused) {
5627			ahc_unpause(ahc);
5628			/*
5629			 * Give the sequencer some time to service
5630			 * any active selections.
5631			 */
5632			ahc_delay(500);
5633		}
5634		ahc_intr(ahc);
5635		ahc_pause(ahc);
5636		paused = TRUE;
5637		ahc_outb(ahc, SCSISEQ, ahc_inb(ahc, SCSISEQ) & ~ENSELO);
5638		intstat = ahc_inb(ahc, INTSTAT);
5639		if ((intstat & INT_PEND) == 0) {
5640			ahc_clear_critical_section(ahc);
5641			intstat = ahc_inb(ahc, INTSTAT);
5642		}
5643	} while (--maxloops
5644	      && (intstat != 0xFF || (ahc->features & AHC_REMOVABLE) == 0)
5645	      && ((intstat & INT_PEND) != 0
5646	       || (ahc_inb(ahc, SSTAT0) & (SELDO|SELINGO)) != 0));
5647	if (maxloops == 0) {
5648		printk("Infinite interrupt loop, INTSTAT = %x",
5649		       ahc_inb(ahc, INTSTAT));
5650	}
5651	ahc_platform_flushwork(ahc);
5652	ahc->flags &= ~AHC_ALL_INTERRUPTS;
5653}
5654
5655#ifdef CONFIG_PM
5656int
5657ahc_suspend(struct ahc_softc *ahc)
5658{
5659
5660	ahc_pause_and_flushwork(ahc);
5661
5662	if (LIST_FIRST(&ahc->pending_scbs) != NULL) {
5663		ahc_unpause(ahc);
5664		return (EBUSY);
5665	}
5666
5667#ifdef AHC_TARGET_MODE
5668	/*
5669	 * XXX What about ATIOs that have not yet been serviced?
5670	 * Perhaps we should just refuse to be suspended if we
5671	 * are acting in a target role.
5672	 */
5673	if (ahc->pending_device != NULL) {
5674		ahc_unpause(ahc);
5675		return (EBUSY);
5676	}
5677#endif
5678	ahc_shutdown(ahc);
5679	return (0);
5680}
5681
5682int
5683ahc_resume(struct ahc_softc *ahc)
5684{
5685
5686	ahc_reset(ahc, /*reinit*/TRUE);
5687	ahc_intr_enable(ahc, TRUE);
5688	ahc_restart(ahc);
5689	return (0);
5690}
5691#endif
5692/************************** Busy Target Table *********************************/
5693/*
5694 * Return the untagged transaction id for a given target/channel lun.
5695 * Optionally, clear the entry.
5696 */
5697static u_int
5698ahc_index_busy_tcl(struct ahc_softc *ahc, u_int tcl)
5699{
5700	u_int scbid;
5701	u_int target_offset;
5702
5703	if ((ahc->flags & AHC_SCB_BTT) != 0) {
5704		u_int saved_scbptr;
5705
5706		saved_scbptr = ahc_inb(ahc, SCBPTR);
5707		ahc_outb(ahc, SCBPTR, TCL_LUN(tcl));
5708		scbid = ahc_inb(ahc, SCB_64_BTT + TCL_TARGET_OFFSET(tcl));
5709		ahc_outb(ahc, SCBPTR, saved_scbptr);
5710	} else {
5711		target_offset = TCL_TARGET_OFFSET(tcl);
5712		scbid = ahc_inb(ahc, BUSY_TARGETS + target_offset);
5713	}
5714
5715	return (scbid);
5716}
5717
5718static void
5719ahc_unbusy_tcl(struct ahc_softc *ahc, u_int tcl)
5720{
5721	u_int target_offset;
5722
5723	if ((ahc->flags & AHC_SCB_BTT) != 0) {
5724		u_int saved_scbptr;
5725
5726		saved_scbptr = ahc_inb(ahc, SCBPTR);
5727		ahc_outb(ahc, SCBPTR, TCL_LUN(tcl));
5728		ahc_outb(ahc, SCB_64_BTT+TCL_TARGET_OFFSET(tcl), SCB_LIST_NULL);
5729		ahc_outb(ahc, SCBPTR, saved_scbptr);
5730	} else {
5731		target_offset = TCL_TARGET_OFFSET(tcl);
5732		ahc_outb(ahc, BUSY_TARGETS + target_offset, SCB_LIST_NULL);
5733	}
5734}
5735
5736static void
5737ahc_busy_tcl(struct ahc_softc *ahc, u_int tcl, u_int scbid)
5738{
5739	u_int target_offset;
5740
5741	if ((ahc->flags & AHC_SCB_BTT) != 0) {
5742		u_int saved_scbptr;
5743
5744		saved_scbptr = ahc_inb(ahc, SCBPTR);
5745		ahc_outb(ahc, SCBPTR, TCL_LUN(tcl));
5746		ahc_outb(ahc, SCB_64_BTT + TCL_TARGET_OFFSET(tcl), scbid);
5747		ahc_outb(ahc, SCBPTR, saved_scbptr);
5748	} else {
5749		target_offset = TCL_TARGET_OFFSET(tcl);
5750		ahc_outb(ahc, BUSY_TARGETS + target_offset, scbid);
5751	}
5752}
5753
5754/************************** SCB and SCB queue management **********************/
5755int
5756ahc_match_scb(struct ahc_softc *ahc, struct scb *scb, int target,
5757	      char channel, int lun, u_int tag, role_t role)
5758{
5759	int targ = SCB_GET_TARGET(ahc, scb);
5760	char chan = SCB_GET_CHANNEL(ahc, scb);
5761	int slun = SCB_GET_LUN(scb);
5762	int match;
5763
5764	match = ((chan == channel) || (channel == ALL_CHANNELS));
5765	if (match != 0)
5766		match = ((targ == target) || (target == CAM_TARGET_WILDCARD));
5767	if (match != 0)
5768		match = ((lun == slun) || (lun == CAM_LUN_WILDCARD));
5769	if (match != 0) {
5770#ifdef AHC_TARGET_MODE
5771		int group;
5772
5773		group = XPT_FC_GROUP(scb->io_ctx->ccb_h.func_code);
5774		if (role == ROLE_INITIATOR) {
5775			match = (group != XPT_FC_GROUP_TMODE)
5776			      && ((tag == scb->hscb->tag)
5777			       || (tag == SCB_LIST_NULL));
5778		} else if (role == ROLE_TARGET) {
5779			match = (group == XPT_FC_GROUP_TMODE)
5780			      && ((tag == scb->io_ctx->csio.tag_id)
5781			       || (tag == SCB_LIST_NULL));
5782		}
5783#else /* !AHC_TARGET_MODE */
5784		match = ((tag == scb->hscb->tag) || (tag == SCB_LIST_NULL));
5785#endif /* AHC_TARGET_MODE */
5786	}
5787
5788	return match;
5789}
5790
5791static void
5792ahc_freeze_devq(struct ahc_softc *ahc, struct scb *scb)
5793{
5794	int	target;
5795	char	channel;
5796	int	lun;
5797
5798	target = SCB_GET_TARGET(ahc, scb);
5799	lun = SCB_GET_LUN(scb);
5800	channel = SCB_GET_CHANNEL(ahc, scb);
5801
5802	ahc_search_qinfifo(ahc, target, channel, lun,
5803			   /*tag*/SCB_LIST_NULL, ROLE_UNKNOWN,
5804			   CAM_REQUEUE_REQ, SEARCH_COMPLETE);
5805
5806	ahc_platform_freeze_devq(ahc, scb);
5807}
5808
5809void
5810ahc_qinfifo_requeue_tail(struct ahc_softc *ahc, struct scb *scb)
5811{
5812	struct scb *prev_scb;
5813
5814	prev_scb = NULL;
5815	if (ahc_qinfifo_count(ahc) != 0) {
5816		u_int prev_tag;
5817		uint8_t prev_pos;
5818
5819		prev_pos = ahc->qinfifonext - 1;
5820		prev_tag = ahc->qinfifo[prev_pos];
5821		prev_scb = ahc_lookup_scb(ahc, prev_tag);
5822	}
5823	ahc_qinfifo_requeue(ahc, prev_scb, scb);
5824	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
5825		ahc_outb(ahc, HNSCB_QOFF, ahc->qinfifonext);
5826	} else {
5827		ahc_outb(ahc, KERNEL_QINPOS, ahc->qinfifonext);
5828	}
5829}
5830
5831static void
5832ahc_qinfifo_requeue(struct ahc_softc *ahc, struct scb *prev_scb,
5833		    struct scb *scb)
5834{
5835	if (prev_scb == NULL) {
5836		ahc_outb(ahc, NEXT_QUEUED_SCB, scb->hscb->tag);
5837	} else {
5838		prev_scb->hscb->next = scb->hscb->tag;
5839		ahc_sync_scb(ahc, prev_scb,
5840			     BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
5841	}
5842	ahc->qinfifo[ahc->qinfifonext++] = scb->hscb->tag;
5843	scb->hscb->next = ahc->next_queued_scb->hscb->tag;
5844	ahc_sync_scb(ahc, scb, BUS_DMASYNC_PREREAD|BUS_DMASYNC_PREWRITE);
5845}
5846
5847static int
5848ahc_qinfifo_count(struct ahc_softc *ahc)
5849{
5850	uint8_t qinpos;
5851	uint8_t diff;
5852
5853	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
5854		qinpos = ahc_inb(ahc, SNSCB_QOFF);
5855		ahc_outb(ahc, SNSCB_QOFF, qinpos);
5856	} else
5857		qinpos = ahc_inb(ahc, QINPOS);
5858	diff = ahc->qinfifonext - qinpos;
5859	return (diff);
5860}
5861
5862int
5863ahc_search_qinfifo(struct ahc_softc *ahc, int target, char channel,
5864		   int lun, u_int tag, role_t role, uint32_t status,
5865		   ahc_search_action action)
5866{
5867	struct	scb *scb;
5868	struct	scb *prev_scb;
5869	uint8_t qinstart;
5870	uint8_t qinpos;
5871	uint8_t qintail;
5872	uint8_t next;
5873	uint8_t prev;
5874	uint8_t curscbptr;
5875	int	found;
5876	int	have_qregs;
5877
5878	qintail = ahc->qinfifonext;
5879	have_qregs = (ahc->features & AHC_QUEUE_REGS) != 0;
5880	if (have_qregs) {
5881		qinstart = ahc_inb(ahc, SNSCB_QOFF);
5882		ahc_outb(ahc, SNSCB_QOFF, qinstart);
5883	} else
5884		qinstart = ahc_inb(ahc, QINPOS);
5885	qinpos = qinstart;
5886	found = 0;
5887	prev_scb = NULL;
5888
5889	if (action == SEARCH_COMPLETE) {
5890		/*
5891		 * Don't attempt to run any queued untagged transactions
5892		 * until we are done with the abort process.
5893		 */
5894		ahc_freeze_untagged_queues(ahc);
5895	}
5896
5897	/*
5898	 * Start with an empty queue.  Entries that are not chosen
5899	 * for removal will be re-added to the queue as we go.
5900	 */
5901	ahc->qinfifonext = qinpos;
5902	ahc_outb(ahc, NEXT_QUEUED_SCB, ahc->next_queued_scb->hscb->tag);
5903
5904	while (qinpos != qintail) {
5905		scb = ahc_lookup_scb(ahc, ahc->qinfifo[qinpos]);
5906		if (scb == NULL) {
5907			printk("qinpos = %d, SCB index = %d\n",
5908				qinpos, ahc->qinfifo[qinpos]);
5909			panic("Loop 1\n");
5910		}
5911
5912		if (ahc_match_scb(ahc, scb, target, channel, lun, tag, role)) {
5913			/*
5914			 * We found an scb that needs to be acted on.
5915			 */
5916			found++;
5917			switch (action) {
5918			case SEARCH_COMPLETE:
5919			{
5920				cam_status ostat;
5921				cam_status cstat;
5922
5923				ostat = ahc_get_transaction_status(scb);
5924				if (ostat == CAM_REQ_INPROG)
5925					ahc_set_transaction_status(scb, status);
5926				cstat = ahc_get_transaction_status(scb);
5927				if (cstat != CAM_REQ_CMP)
5928					ahc_freeze_scb(scb);
5929				if ((scb->flags & SCB_ACTIVE) == 0)
5930					printk("Inactive SCB in qinfifo\n");
5931				ahc_done(ahc, scb);
5932
5933				/* FALLTHROUGH */
5934			}
5935			case SEARCH_REMOVE:
5936				break;
5937			case SEARCH_COUNT:
5938				ahc_qinfifo_requeue(ahc, prev_scb, scb);
5939				prev_scb = scb;
5940				break;
5941			}
5942		} else {
5943			ahc_qinfifo_requeue(ahc, prev_scb, scb);
5944			prev_scb = scb;
5945		}
5946		qinpos++;
5947	}
5948
5949	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
5950		ahc_outb(ahc, HNSCB_QOFF, ahc->qinfifonext);
5951	} else {
5952		ahc_outb(ahc, KERNEL_QINPOS, ahc->qinfifonext);
5953	}
5954
5955	if (action != SEARCH_COUNT
5956	 && (found != 0)
5957	 && (qinstart != ahc->qinfifonext)) {
5958		/*
5959		 * The sequencer may be in the process of dmaing
5960		 * down the SCB at the beginning of the queue.
5961		 * This could be problematic if either the first,
5962		 * or the second SCB is removed from the queue
5963		 * (the first SCB includes a pointer to the "next"
5964		 * SCB to dma). If we have removed any entries, swap
5965		 * the first element in the queue with the next HSCB
5966		 * so the sequencer will notice that NEXT_QUEUED_SCB
5967		 * has changed during its dma attempt and will retry
5968		 * the DMA.
5969		 */
5970		scb = ahc_lookup_scb(ahc, ahc->qinfifo[qinstart]);
5971
5972		if (scb == NULL) {
5973			printk("found = %d, qinstart = %d, qinfifionext = %d\n",
5974				found, qinstart, ahc->qinfifonext);
5975			panic("First/Second Qinfifo fixup\n");
5976		}
5977		/*
5978		 * ahc_swap_with_next_hscb forces our next pointer to
5979		 * point to the reserved SCB for future commands.  Save
5980		 * and restore our original next pointer to maintain
5981		 * queue integrity.
5982		 */
5983		next = scb->hscb->next;
5984		ahc->scb_data->scbindex[scb->hscb->tag] = NULL;
5985		ahc_swap_with_next_hscb(ahc, scb);
5986		scb->hscb->next = next;
5987		ahc->qinfifo[qinstart] = scb->hscb->tag;
5988
5989		/* Tell the card about the new head of the qinfifo. */
5990		ahc_outb(ahc, NEXT_QUEUED_SCB, scb->hscb->tag);
5991
5992		/* Fixup the tail "next" pointer. */
5993		qintail = ahc->qinfifonext - 1;
5994		scb = ahc_lookup_scb(ahc, ahc->qinfifo[qintail]);
5995		scb->hscb->next = ahc->next_queued_scb->hscb->tag;
5996	}
5997
5998	/*
5999	 * Search waiting for selection list.
6000	 */
6001	curscbptr = ahc_inb(ahc, SCBPTR);
6002	next = ahc_inb(ahc, WAITING_SCBH);  /* Start at head of list. */
6003	prev = SCB_LIST_NULL;
6004
6005	while (next != SCB_LIST_NULL) {
6006		uint8_t scb_index;
6007
6008		ahc_outb(ahc, SCBPTR, next);
6009		scb_index = ahc_inb(ahc, SCB_TAG);
6010		if (scb_index >= ahc->scb_data->numscbs) {
6011			printk("Waiting List inconsistency. "
6012			       "SCB index == %d, yet numscbs == %d.",
6013			       scb_index, ahc->scb_data->numscbs);
6014			ahc_dump_card_state(ahc);
6015			panic("for safety");
6016		}
6017		scb = ahc_lookup_scb(ahc, scb_index);
6018		if (scb == NULL) {
6019			printk("scb_index = %d, next = %d\n",
6020				scb_index, next);
6021			panic("Waiting List traversal\n");
6022		}
6023		if (ahc_match_scb(ahc, scb, target, channel,
6024				  lun, SCB_LIST_NULL, role)) {
6025			/*
6026			 * We found an scb that needs to be acted on.
6027			 */
6028			found++;
6029			switch (action) {
6030			case SEARCH_COMPLETE:
6031			{
6032				cam_status ostat;
6033				cam_status cstat;
6034
6035				ostat = ahc_get_transaction_status(scb);
6036				if (ostat == CAM_REQ_INPROG)
6037					ahc_set_transaction_status(scb,
6038								   status);
6039				cstat = ahc_get_transaction_status(scb);
6040				if (cstat != CAM_REQ_CMP)
6041					ahc_freeze_scb(scb);
6042				if ((scb->flags & SCB_ACTIVE) == 0)
6043					printk("Inactive SCB in Waiting List\n");
6044				ahc_done(ahc, scb);
6045				/* FALLTHROUGH */
6046			}
6047			case SEARCH_REMOVE:
6048				next = ahc_rem_wscb(ahc, next, prev);
6049				break;
6050			case SEARCH_COUNT:
6051				prev = next;
6052				next = ahc_inb(ahc, SCB_NEXT);
6053				break;
6054			}
6055		} else {
6056
6057			prev = next;
6058			next = ahc_inb(ahc, SCB_NEXT);
6059		}
6060	}
6061	ahc_outb(ahc, SCBPTR, curscbptr);
6062
6063	found += ahc_search_untagged_queues(ahc, /*ahc_io_ctx_t*/NULL, target,
6064					    channel, lun, status, action);
6065
6066	if (action == SEARCH_COMPLETE)
6067		ahc_release_untagged_queues(ahc);
6068	return (found);
6069}
6070
6071int
6072ahc_search_untagged_queues(struct ahc_softc *ahc, ahc_io_ctx_t ctx,
6073			   int target, char channel, int lun, uint32_t status,
6074			   ahc_search_action action)
6075{
6076	struct	scb *scb;
6077	int	maxtarget;
6078	int	found;
6079	int	i;
6080
6081	if (action == SEARCH_COMPLETE) {
6082		/*
6083		 * Don't attempt to run any queued untagged transactions
6084		 * until we are done with the abort process.
6085		 */
6086		ahc_freeze_untagged_queues(ahc);
6087	}
6088
6089	found = 0;
6090	i = 0;
6091	if ((ahc->flags & AHC_SCB_BTT) == 0) {
6092
6093		maxtarget = 16;
6094		if (target != CAM_TARGET_WILDCARD) {
6095
6096			i = target;
6097			if (channel == 'B')
6098				i += 8;
6099			maxtarget = i + 1;
6100		}
6101	} else {
6102		maxtarget = 0;
6103	}
6104
6105	for (; i < maxtarget; i++) {
6106		struct scb_tailq *untagged_q;
6107		struct scb *next_scb;
6108
6109		untagged_q = &(ahc->untagged_queues[i]);
6110		next_scb = TAILQ_FIRST(untagged_q);
6111		while (next_scb != NULL) {
6112
6113			scb = next_scb;
6114			next_scb = TAILQ_NEXT(scb, links.tqe);
6115
6116			/*
6117			 * The head of the list may be the currently
6118			 * active untagged command for a device.
6119			 * We're only searching for commands that
6120			 * have not been started.  A transaction
6121			 * marked active but still in the qinfifo
6122			 * is removed by the qinfifo scanning code
6123			 * above.
6124			 */
6125			if ((scb->flags & SCB_ACTIVE) != 0)
6126				continue;
6127
6128			if (ahc_match_scb(ahc, scb, target, channel, lun,
6129					  SCB_LIST_NULL, ROLE_INITIATOR) == 0
6130			 || (ctx != NULL && ctx != scb->io_ctx))
6131				continue;
6132
6133			/*
6134			 * We found an scb that needs to be acted on.
6135			 */
6136			found++;
6137			switch (action) {
6138			case SEARCH_COMPLETE:
6139			{
6140				cam_status ostat;
6141				cam_status cstat;
6142
6143				ostat = ahc_get_transaction_status(scb);
6144				if (ostat == CAM_REQ_INPROG)
6145					ahc_set_transaction_status(scb, status);
6146				cstat = ahc_get_transaction_status(scb);
6147				if (cstat != CAM_REQ_CMP)
6148					ahc_freeze_scb(scb);
6149				if ((scb->flags & SCB_ACTIVE) == 0)
6150					printk("Inactive SCB in untaggedQ\n");
6151				ahc_done(ahc, scb);
6152				break;
6153			}
6154			case SEARCH_REMOVE:
6155				scb->flags &= ~SCB_UNTAGGEDQ;
6156				TAILQ_REMOVE(untagged_q, scb, links.tqe);
6157				break;
6158			case SEARCH_COUNT:
6159				break;
6160			}
6161		}
6162	}
6163
6164	if (action == SEARCH_COMPLETE)
6165		ahc_release_untagged_queues(ahc);
6166	return (found);
6167}
6168
6169int
6170ahc_search_disc_list(struct ahc_softc *ahc, int target, char channel,
6171		     int lun, u_int tag, int stop_on_first, int remove,
6172		     int save_state)
6173{
6174	struct	scb *scbp;
6175	u_int	next;
6176	u_int	prev;
6177	u_int	count;
6178	u_int	active_scb;
6179
6180	count = 0;
6181	next = ahc_inb(ahc, DISCONNECTED_SCBH);
6182	prev = SCB_LIST_NULL;
6183
6184	if (save_state) {
6185		/* restore this when we're done */
6186		active_scb = ahc_inb(ahc, SCBPTR);
6187	} else
6188		/* Silence compiler */
6189		active_scb = SCB_LIST_NULL;
6190
6191	while (next != SCB_LIST_NULL) {
6192		u_int scb_index;
6193
6194		ahc_outb(ahc, SCBPTR, next);
6195		scb_index = ahc_inb(ahc, SCB_TAG);
6196		if (scb_index >= ahc->scb_data->numscbs) {
6197			printk("Disconnected List inconsistency. "
6198			       "SCB index == %d, yet numscbs == %d.",
6199			       scb_index, ahc->scb_data->numscbs);
6200			ahc_dump_card_state(ahc);
6201			panic("for safety");
6202		}
6203
6204		if (next == prev) {
6205			panic("Disconnected List Loop. "
6206			      "cur SCBPTR == %x, prev SCBPTR == %x.",
6207			      next, prev);
6208		}
6209		scbp = ahc_lookup_scb(ahc, scb_index);
6210		if (ahc_match_scb(ahc, scbp, target, channel, lun,
6211				  tag, ROLE_INITIATOR)) {
6212			count++;
6213			if (remove) {
6214				next =
6215				    ahc_rem_scb_from_disc_list(ahc, prev, next);
6216			} else {
6217				prev = next;
6218				next = ahc_inb(ahc, SCB_NEXT);
6219			}
6220			if (stop_on_first)
6221				break;
6222		} else {
6223			prev = next;
6224			next = ahc_inb(ahc, SCB_NEXT);
6225		}
6226	}
6227	if (save_state)
6228		ahc_outb(ahc, SCBPTR, active_scb);
6229	return (count);
6230}
6231
6232/*
6233 * Remove an SCB from the on chip list of disconnected transactions.
6234 * This is empty/unused if we are not performing SCB paging.
6235 */
6236static u_int
6237ahc_rem_scb_from_disc_list(struct ahc_softc *ahc, u_int prev, u_int scbptr)
6238{
6239	u_int next;
6240
6241	ahc_outb(ahc, SCBPTR, scbptr);
6242	next = ahc_inb(ahc, SCB_NEXT);
6243
6244	ahc_outb(ahc, SCB_CONTROL, 0);
6245
6246	ahc_add_curscb_to_free_list(ahc);
6247
6248	if (prev != SCB_LIST_NULL) {
6249		ahc_outb(ahc, SCBPTR, prev);
6250		ahc_outb(ahc, SCB_NEXT, next);
6251	} else
6252		ahc_outb(ahc, DISCONNECTED_SCBH, next);
6253
6254	return (next);
6255}
6256
6257/*
6258 * Add the SCB as selected by SCBPTR onto the on chip list of
6259 * free hardware SCBs.  This list is empty/unused if we are not
6260 * performing SCB paging.
6261 */
6262static void
6263ahc_add_curscb_to_free_list(struct ahc_softc *ahc)
6264{
6265	/*
6266	 * Invalidate the tag so that our abort
6267	 * routines don't think it's active.
6268	 */
6269	ahc_outb(ahc, SCB_TAG, SCB_LIST_NULL);
6270
6271	if ((ahc->flags & AHC_PAGESCBS) != 0) {
6272		ahc_outb(ahc, SCB_NEXT, ahc_inb(ahc, FREE_SCBH));
6273		ahc_outb(ahc, FREE_SCBH, ahc_inb(ahc, SCBPTR));
6274	}
6275}
6276
6277/*
6278 * Manipulate the waiting for selection list and return the
6279 * scb that follows the one that we remove.
6280 */
6281static u_int
6282ahc_rem_wscb(struct ahc_softc *ahc, u_int scbpos, u_int prev)
6283{
6284	u_int curscb, next;
6285
6286	/*
6287	 * Select the SCB we want to abort and
6288	 * pull the next pointer out of it.
6289	 */
6290	curscb = ahc_inb(ahc, SCBPTR);
6291	ahc_outb(ahc, SCBPTR, scbpos);
6292	next = ahc_inb(ahc, SCB_NEXT);
6293
6294	/* Clear the necessary fields */
6295	ahc_outb(ahc, SCB_CONTROL, 0);
6296
6297	ahc_add_curscb_to_free_list(ahc);
6298
6299	/* update the waiting list */
6300	if (prev == SCB_LIST_NULL) {
6301		/* First in the list */
6302		ahc_outb(ahc, WAITING_SCBH, next);
6303
6304		/*
6305		 * Ensure we aren't attempting to perform
6306		 * selection for this entry.
6307		 */
6308		ahc_outb(ahc, SCSISEQ, (ahc_inb(ahc, SCSISEQ) & ~ENSELO));
6309	} else {
6310		/*
6311		 * Select the scb that pointed to us
6312		 * and update its next pointer.
6313		 */
6314		ahc_outb(ahc, SCBPTR, prev);
6315		ahc_outb(ahc, SCB_NEXT, next);
6316	}
6317
6318	/*
6319	 * Point us back at the original scb position.
6320	 */
6321	ahc_outb(ahc, SCBPTR, curscb);
6322	return next;
6323}
6324
6325/******************************** Error Handling ******************************/
6326/*
6327 * Abort all SCBs that match the given description (target/channel/lun/tag),
6328 * setting their status to the passed in status if the status has not already
6329 * been modified from CAM_REQ_INPROG.  This routine assumes that the sequencer
6330 * is paused before it is called.
6331 */
6332static int
6333ahc_abort_scbs(struct ahc_softc *ahc, int target, char channel,
6334	       int lun, u_int tag, role_t role, uint32_t status)
6335{
6336	struct	scb *scbp;
6337	struct	scb *scbp_next;
6338	u_int	active_scb;
6339	int	i, j;
6340	int	maxtarget;
6341	int	minlun;
6342	int	maxlun;
6343
6344	int	found;
6345
6346	/*
6347	 * Don't attempt to run any queued untagged transactions
6348	 * until we are done with the abort process.
6349	 */
6350	ahc_freeze_untagged_queues(ahc);
6351
6352	/* restore this when we're done */
6353	active_scb = ahc_inb(ahc, SCBPTR);
6354
6355	found = ahc_search_qinfifo(ahc, target, channel, lun, SCB_LIST_NULL,
6356				   role, CAM_REQUEUE_REQ, SEARCH_COMPLETE);
6357
6358	/*
6359	 * Clean out the busy target table for any untagged commands.
6360	 */
6361	i = 0;
6362	maxtarget = 16;
6363	if (target != CAM_TARGET_WILDCARD) {
6364		i = target;
6365		if (channel == 'B')
6366			i += 8;
6367		maxtarget = i + 1;
6368	}
6369
6370	if (lun == CAM_LUN_WILDCARD) {
6371
6372		/*
6373		 * Unless we are using an SCB based
6374		 * busy targets table, there is only
6375		 * one table entry for all luns of
6376		 * a target.
6377		 */
6378		minlun = 0;
6379		maxlun = 1;
6380		if ((ahc->flags & AHC_SCB_BTT) != 0)
6381			maxlun = AHC_NUM_LUNS;
6382	} else {
6383		minlun = lun;
6384		maxlun = lun + 1;
6385	}
6386
6387	if (role != ROLE_TARGET) {
6388		for (;i < maxtarget; i++) {
6389			for (j = minlun;j < maxlun; j++) {
6390				u_int scbid;
6391				u_int tcl;
6392
6393				tcl = BUILD_TCL(i << 4, j);
6394				scbid = ahc_index_busy_tcl(ahc, tcl);
6395				scbp = ahc_lookup_scb(ahc, scbid);
6396				if (scbp == NULL
6397				 || ahc_match_scb(ahc, scbp, target, channel,
6398						  lun, tag, role) == 0)
6399					continue;
6400				ahc_unbusy_tcl(ahc, BUILD_TCL(i << 4, j));
6401			}
6402		}
6403
6404		/*
6405		 * Go through the disconnected list and remove any entries we
6406		 * have queued for completion, 0'ing their control byte too.
6407		 * We save the active SCB and restore it ourselves, so there
6408		 * is no reason for this search to restore it too.
6409		 */
6410		ahc_search_disc_list(ahc, target, channel, lun, tag,
6411				     /*stop_on_first*/FALSE, /*remove*/TRUE,
6412				     /*save_state*/FALSE);
6413	}
6414
6415	/*
6416	 * Go through the hardware SCB array looking for commands that
6417	 * were active but not on any list.  In some cases, these remnants
6418	 * might not still have mappings in the scbindex array (e.g. unexpected
6419	 * bus free with the same scb queued for an abort).  Don't hold this
6420	 * against them.
6421	 */
6422	for (i = 0; i < ahc->scb_data->maxhscbs; i++) {
6423		u_int scbid;
6424
6425		ahc_outb(ahc, SCBPTR, i);
6426		scbid = ahc_inb(ahc, SCB_TAG);
6427		scbp = ahc_lookup_scb(ahc, scbid);
6428		if ((scbp == NULL && scbid != SCB_LIST_NULL)
6429		 || (scbp != NULL
6430		  && ahc_match_scb(ahc, scbp, target, channel, lun, tag, role)))
6431			ahc_add_curscb_to_free_list(ahc);
6432	}
6433
6434	/*
6435	 * Go through the pending CCB list and look for
6436	 * commands for this target that are still active.
6437	 * These are other tagged commands that were
6438	 * disconnected when the reset occurred.
6439	 */
6440	scbp_next = LIST_FIRST(&ahc->pending_scbs);
6441	while (scbp_next != NULL) {
6442		scbp = scbp_next;
6443		scbp_next = LIST_NEXT(scbp, pending_links);
6444		if (ahc_match_scb(ahc, scbp, target, channel, lun, tag, role)) {
6445			cam_status ostat;
6446
6447			ostat = ahc_get_transaction_status(scbp);
6448			if (ostat == CAM_REQ_INPROG)
6449				ahc_set_transaction_status(scbp, status);
6450			if (ahc_get_transaction_status(scbp) != CAM_REQ_CMP)
6451				ahc_freeze_scb(scbp);
6452			if ((scbp->flags & SCB_ACTIVE) == 0)
6453				printk("Inactive SCB on pending list\n");
6454			ahc_done(ahc, scbp);
6455			found++;
6456		}
6457	}
6458	ahc_outb(ahc, SCBPTR, active_scb);
6459	ahc_platform_abort_scbs(ahc, target, channel, lun, tag, role, status);
6460	ahc_release_untagged_queues(ahc);
6461	return found;
6462}
6463
6464static void
6465ahc_reset_current_bus(struct ahc_softc *ahc)
6466{
6467	uint8_t scsiseq;
6468
6469	ahc_outb(ahc, SIMODE1, ahc_inb(ahc, SIMODE1) & ~ENSCSIRST);
6470	scsiseq = ahc_inb(ahc, SCSISEQ);
6471	ahc_outb(ahc, SCSISEQ, scsiseq | SCSIRSTO);
6472	ahc_flush_device_writes(ahc);
6473	ahc_delay(AHC_BUSRESET_DELAY);
6474	/* Turn off the bus reset */
6475	ahc_outb(ahc, SCSISEQ, scsiseq & ~SCSIRSTO);
6476
6477	ahc_clear_intstat(ahc);
6478
6479	/* Re-enable reset interrupts */
6480	ahc_outb(ahc, SIMODE1, ahc_inb(ahc, SIMODE1) | ENSCSIRST);
6481}
6482
6483int
6484ahc_reset_channel(struct ahc_softc *ahc, char channel, int initiate_reset)
6485{
6486	struct	ahc_devinfo devinfo;
6487	u_int	initiator, target, max_scsiid;
6488	u_int	sblkctl;
6489	u_int	scsiseq;
6490	u_int	simode1;
6491	int	found;
6492	int	restart_needed;
6493	char	cur_channel;
6494
6495	ahc->pending_device = NULL;
6496
6497	ahc_compile_devinfo(&devinfo,
6498			    CAM_TARGET_WILDCARD,
6499			    CAM_TARGET_WILDCARD,
6500			    CAM_LUN_WILDCARD,
6501			    channel, ROLE_UNKNOWN);
6502	ahc_pause(ahc);
6503
6504	/* Make sure the sequencer is in a safe location. */
6505	ahc_clear_critical_section(ahc);
6506
6507	/*
6508	 * Run our command complete fifos to ensure that we perform
6509	 * completion processing on any commands that 'completed'
6510	 * before the reset occurred.
6511	 */
6512	ahc_run_qoutfifo(ahc);
6513#ifdef AHC_TARGET_MODE
6514	/*
6515	 * XXX - In Twin mode, the tqinfifo may have commands
6516	 *	 for an unaffected channel in it.  However, if
6517	 *	 we have run out of ATIO resources to drain that
6518	 *	 queue, we may not get them all out here.  Further,
6519	 *	 the blocked transactions for the reset channel
6520	 *	 should just be killed off, irrespecitve of whether
6521	 *	 we are blocked on ATIO resources.  Write a routine
6522	 *	 to compact the tqinfifo appropriately.
6523	 */
6524	if ((ahc->flags & AHC_TARGETROLE) != 0) {
6525		ahc_run_tqinfifo(ahc, /*paused*/TRUE);
6526	}
6527#endif
6528
6529	/*
6530	 * Reset the bus if we are initiating this reset
6531	 */
6532	sblkctl = ahc_inb(ahc, SBLKCTL);
6533	cur_channel = 'A';
6534	if ((ahc->features & AHC_TWIN) != 0
6535	 && ((sblkctl & SELBUSB) != 0))
6536	    cur_channel = 'B';
6537	scsiseq = ahc_inb(ahc, SCSISEQ_TEMPLATE);
6538	if (cur_channel != channel) {
6539		/* Case 1: Command for another bus is active
6540		 * Stealthily reset the other bus without
6541		 * upsetting the current bus.
6542		 */
6543		ahc_outb(ahc, SBLKCTL, sblkctl ^ SELBUSB);
6544		simode1 = ahc_inb(ahc, SIMODE1) & ~(ENBUSFREE|ENSCSIRST);
6545#ifdef AHC_TARGET_MODE
6546		/*
6547		 * Bus resets clear ENSELI, so we cannot
6548		 * defer re-enabling bus reset interrupts
6549		 * if we are in target mode.
6550		 */
6551		if ((ahc->flags & AHC_TARGETROLE) != 0)
6552			simode1 |= ENSCSIRST;
6553#endif
6554		ahc_outb(ahc, SIMODE1, simode1);
6555		if (initiate_reset)
6556			ahc_reset_current_bus(ahc);
6557		ahc_clear_intstat(ahc);
6558		ahc_outb(ahc, SCSISEQ, scsiseq & (ENSELI|ENRSELI|ENAUTOATNP));
6559		ahc_outb(ahc, SBLKCTL, sblkctl);
6560		restart_needed = FALSE;
6561	} else {
6562		/* Case 2: A command from this bus is active or we're idle */
6563		simode1 = ahc_inb(ahc, SIMODE1) & ~(ENBUSFREE|ENSCSIRST);
6564#ifdef AHC_TARGET_MODE
6565		/*
6566		 * Bus resets clear ENSELI, so we cannot
6567		 * defer re-enabling bus reset interrupts
6568		 * if we are in target mode.
6569		 */
6570		if ((ahc->flags & AHC_TARGETROLE) != 0)
6571			simode1 |= ENSCSIRST;
6572#endif
6573		ahc_outb(ahc, SIMODE1, simode1);
6574		if (initiate_reset)
6575			ahc_reset_current_bus(ahc);
6576		ahc_clear_intstat(ahc);
6577		ahc_outb(ahc, SCSISEQ, scsiseq & (ENSELI|ENRSELI|ENAUTOATNP));
6578		restart_needed = TRUE;
6579	}
6580
6581	/*
6582	 * Clean up all the state information for the
6583	 * pending transactions on this bus.
6584	 */
6585	found = ahc_abort_scbs(ahc, CAM_TARGET_WILDCARD, channel,
6586			       CAM_LUN_WILDCARD, SCB_LIST_NULL,
6587			       ROLE_UNKNOWN, CAM_SCSI_BUS_RESET);
6588
6589	max_scsiid = (ahc->features & AHC_WIDE) ? 15 : 7;
6590
6591#ifdef AHC_TARGET_MODE
6592	/*
6593	 * Send an immediate notify ccb to all target more peripheral
6594	 * drivers affected by this action.
6595	 */
6596	for (target = 0; target <= max_scsiid; target++) {
6597		struct ahc_tmode_tstate* tstate;
6598		u_int lun;
6599
6600		tstate = ahc->enabled_targets[target];
6601		if (tstate == NULL)
6602			continue;
6603		for (lun = 0; lun < AHC_NUM_LUNS; lun++) {
6604			struct ahc_tmode_lstate* lstate;
6605
6606			lstate = tstate->enabled_luns[lun];
6607			if (lstate == NULL)
6608				continue;
6609
6610			ahc_queue_lstate_event(ahc, lstate, CAM_TARGET_WILDCARD,
6611					       EVENT_TYPE_BUS_RESET, /*arg*/0);
6612			ahc_send_lstate_events(ahc, lstate);
6613		}
6614	}
6615#endif
6616	/* Notify the XPT that a bus reset occurred */
6617	ahc_send_async(ahc, devinfo.channel, CAM_TARGET_WILDCARD,
6618		       CAM_LUN_WILDCARD, AC_BUS_RESET);
6619
6620	/*
6621	 * Revert to async/narrow transfers until we renegotiate.
6622	 */
6623	for (target = 0; target <= max_scsiid; target++) {
6624
6625		if (ahc->enabled_targets[target] == NULL)
6626			continue;
6627		for (initiator = 0; initiator <= max_scsiid; initiator++) {
6628			struct ahc_devinfo devinfo;
6629
6630			ahc_compile_devinfo(&devinfo, target, initiator,
6631					    CAM_LUN_WILDCARD,
6632					    channel, ROLE_UNKNOWN);
6633			ahc_set_width(ahc, &devinfo, MSG_EXT_WDTR_BUS_8_BIT,
6634				      AHC_TRANS_CUR, /*paused*/TRUE);
6635			ahc_set_syncrate(ahc, &devinfo, /*syncrate*/NULL,
6636					 /*period*/0, /*offset*/0,
6637					 /*ppr_options*/0, AHC_TRANS_CUR,
6638					 /*paused*/TRUE);
6639		}
6640	}
6641
6642	if (restart_needed)
6643		ahc_restart(ahc);
6644	else
6645		ahc_unpause(ahc);
6646	return found;
6647}
6648
6649
6650/***************************** Residual Processing ****************************/
6651/*
6652 * Calculate the residual for a just completed SCB.
6653 */
6654static void
6655ahc_calc_residual(struct ahc_softc *ahc, struct scb *scb)
6656{
6657	struct hardware_scb *hscb;
6658	struct status_pkt *spkt;
6659	uint32_t sgptr;
6660	uint32_t resid_sgptr;
6661	uint32_t resid;
6662
6663	/*
6664	 * 5 cases.
6665	 * 1) No residual.
6666	 *    SG_RESID_VALID clear in sgptr.
6667	 * 2) Transferless command
6668	 * 3) Never performed any transfers.
6669	 *    sgptr has SG_FULL_RESID set.
6670	 * 4) No residual but target did not
6671	 *    save data pointers after the
6672	 *    last transfer, so sgptr was
6673	 *    never updated.
6674	 * 5) We have a partial residual.
6675	 *    Use residual_sgptr to determine
6676	 *    where we are.
6677	 */
6678
6679	hscb = scb->hscb;
6680	sgptr = ahc_le32toh(hscb->sgptr);
6681	if ((sgptr & SG_RESID_VALID) == 0)
6682		/* Case 1 */
6683		return;
6684	sgptr &= ~SG_RESID_VALID;
6685
6686	if ((sgptr & SG_LIST_NULL) != 0)
6687		/* Case 2 */
6688		return;
6689
6690	spkt = &hscb->shared_data.status;
6691	resid_sgptr = ahc_le32toh(spkt->residual_sg_ptr);
6692	if ((sgptr & SG_FULL_RESID) != 0) {
6693		/* Case 3 */
6694		resid = ahc_get_transfer_length(scb);
6695	} else if ((resid_sgptr & SG_LIST_NULL) != 0) {
6696		/* Case 4 */
6697		return;
6698	} else if ((resid_sgptr & ~SG_PTR_MASK) != 0) {
6699		panic("Bogus resid sgptr value 0x%x\n", resid_sgptr);
6700	} else {
6701		struct ahc_dma_seg *sg;
6702
6703		/*
6704		 * Remainder of the SG where the transfer
6705		 * stopped.
6706		 */
6707		resid = ahc_le32toh(spkt->residual_datacnt) & AHC_SG_LEN_MASK;
6708		sg = ahc_sg_bus_to_virt(scb, resid_sgptr & SG_PTR_MASK);
6709
6710		/* The residual sg_ptr always points to the next sg */
6711		sg--;
6712
6713		/*
6714		 * Add up the contents of all residual
6715		 * SG segments that are after the SG where
6716		 * the transfer stopped.
6717		 */
6718		while ((ahc_le32toh(sg->len) & AHC_DMA_LAST_SEG) == 0) {
6719			sg++;
6720			resid += ahc_le32toh(sg->len) & AHC_SG_LEN_MASK;
6721		}
6722	}
6723	if ((scb->flags & SCB_SENSE) == 0)
6724		ahc_set_residual(scb, resid);
6725	else
6726		ahc_set_sense_residual(scb, resid);
6727
6728#ifdef AHC_DEBUG
6729	if ((ahc_debug & AHC_SHOW_MISC) != 0) {
6730		ahc_print_path(ahc, scb);
6731		printk("Handled %sResidual of %d bytes\n",
6732		       (scb->flags & SCB_SENSE) ? "Sense " : "", resid);
6733	}
6734#endif
6735}
6736
6737/******************************* Target Mode **********************************/
6738#ifdef AHC_TARGET_MODE
6739/*
6740 * Add a target mode event to this lun's queue
6741 */
6742static void
6743ahc_queue_lstate_event(struct ahc_softc *ahc, struct ahc_tmode_lstate *lstate,
6744		       u_int initiator_id, u_int event_type, u_int event_arg)
6745{
6746	struct ahc_tmode_event *event;
6747	int pending;
6748
6749	xpt_freeze_devq(lstate->path, /*count*/1);
6750	if (lstate->event_w_idx >= lstate->event_r_idx)
6751		pending = lstate->event_w_idx - lstate->event_r_idx;
6752	else
6753		pending = AHC_TMODE_EVENT_BUFFER_SIZE + 1
6754			- (lstate->event_r_idx - lstate->event_w_idx);
6755
6756	if (event_type == EVENT_TYPE_BUS_RESET
6757	 || event_type == MSG_BUS_DEV_RESET) {
6758		/*
6759		 * Any earlier events are irrelevant, so reset our buffer.
6760		 * This has the effect of allowing us to deal with reset
6761		 * floods (an external device holding down the reset line)
6762		 * without losing the event that is really interesting.
6763		 */
6764		lstate->event_r_idx = 0;
6765		lstate->event_w_idx = 0;
6766		xpt_release_devq(lstate->path, pending, /*runqueue*/FALSE);
6767	}
6768
6769	if (pending == AHC_TMODE_EVENT_BUFFER_SIZE) {
6770		xpt_print_path(lstate->path);
6771		printk("immediate event %x:%x lost\n",
6772		       lstate->event_buffer[lstate->event_r_idx].event_type,
6773		       lstate->event_buffer[lstate->event_r_idx].event_arg);
6774		lstate->event_r_idx++;
6775		if (lstate->event_r_idx == AHC_TMODE_EVENT_BUFFER_SIZE)
6776			lstate->event_r_idx = 0;
6777		xpt_release_devq(lstate->path, /*count*/1, /*runqueue*/FALSE);
6778	}
6779
6780	event = &lstate->event_buffer[lstate->event_w_idx];
6781	event->initiator_id = initiator_id;
6782	event->event_type = event_type;
6783	event->event_arg = event_arg;
6784	lstate->event_w_idx++;
6785	if (lstate->event_w_idx == AHC_TMODE_EVENT_BUFFER_SIZE)
6786		lstate->event_w_idx = 0;
6787}
6788
6789/*
6790 * Send any target mode events queued up waiting
6791 * for immediate notify resources.
6792 */
6793void
6794ahc_send_lstate_events(struct ahc_softc *ahc, struct ahc_tmode_lstate *lstate)
6795{
6796	struct ccb_hdr *ccbh;
6797	struct ccb_immed_notify *inot;
6798
6799	while (lstate->event_r_idx != lstate->event_w_idx
6800	    && (ccbh = SLIST_FIRST(&lstate->immed_notifies)) != NULL) {
6801		struct ahc_tmode_event *event;
6802
6803		event = &lstate->event_buffer[lstate->event_r_idx];
6804		SLIST_REMOVE_HEAD(&lstate->immed_notifies, sim_links.sle);
6805		inot = (struct ccb_immed_notify *)ccbh;
6806		switch (event->event_type) {
6807		case EVENT_TYPE_BUS_RESET:
6808			ccbh->status = CAM_SCSI_BUS_RESET|CAM_DEV_QFRZN;
6809			break;
6810		default:
6811			ccbh->status = CAM_MESSAGE_RECV|CAM_DEV_QFRZN;
6812			inot->message_args[0] = event->event_type;
6813			inot->message_args[1] = event->event_arg;
6814			break;
6815		}
6816		inot->initiator_id = event->initiator_id;
6817		inot->sense_len = 0;
6818		xpt_done((union ccb *)inot);
6819		lstate->event_r_idx++;
6820		if (lstate->event_r_idx == AHC_TMODE_EVENT_BUFFER_SIZE)
6821			lstate->event_r_idx = 0;
6822	}
6823}
6824#endif
6825
6826/******************** Sequencer Program Patching/Download *********************/
6827
6828#ifdef AHC_DUMP_SEQ
6829void
6830ahc_dumpseq(struct ahc_softc* ahc)
6831{
6832	int i;
6833
6834	ahc_outb(ahc, SEQCTL, PERRORDIS|FAILDIS|FASTMODE|LOADRAM);
6835	ahc_outb(ahc, SEQADDR0, 0);
6836	ahc_outb(ahc, SEQADDR1, 0);
6837	for (i = 0; i < ahc->instruction_ram_size; i++) {
6838		uint8_t ins_bytes[4];
6839
6840		ahc_insb(ahc, SEQRAM, ins_bytes, 4);
6841		printk("0x%08x\n", ins_bytes[0] << 24
6842				 | ins_bytes[1] << 16
6843				 | ins_bytes[2] << 8
6844				 | ins_bytes[3]);
6845	}
6846}
6847#endif
6848
6849static int
6850ahc_loadseq(struct ahc_softc *ahc)
6851{
6852	struct	cs cs_table[num_critical_sections];
6853	u_int	begin_set[num_critical_sections];
6854	u_int	end_set[num_critical_sections];
6855	const struct patch *cur_patch;
6856	u_int	cs_count;
6857	u_int	cur_cs;
6858	u_int	i;
6859	u_int	skip_addr;
6860	u_int	sg_prefetch_cnt;
6861	int	downloaded;
6862	uint8_t	download_consts[7];
6863
6864	/*
6865	 * Start out with 0 critical sections
6866	 * that apply to this firmware load.
6867	 */
6868	cs_count = 0;
6869	cur_cs = 0;
6870	memset(begin_set, 0, sizeof(begin_set));
6871	memset(end_set, 0, sizeof(end_set));
6872
6873	/* Setup downloadable constant table */
6874	download_consts[QOUTFIFO_OFFSET] = 0;
6875	if (ahc->targetcmds != NULL)
6876		download_consts[QOUTFIFO_OFFSET] += 32;
6877	download_consts[QINFIFO_OFFSET] = download_consts[QOUTFIFO_OFFSET] + 1;
6878	download_consts[CACHESIZE_MASK] = ahc->pci_cachesize - 1;
6879	download_consts[INVERTED_CACHESIZE_MASK] = ~(ahc->pci_cachesize - 1);
6880	sg_prefetch_cnt = ahc->pci_cachesize;
6881	if (sg_prefetch_cnt < (2 * sizeof(struct ahc_dma_seg)))
6882		sg_prefetch_cnt = 2 * sizeof(struct ahc_dma_seg);
6883	download_consts[SG_PREFETCH_CNT] = sg_prefetch_cnt;
6884	download_consts[SG_PREFETCH_ALIGN_MASK] = ~(sg_prefetch_cnt - 1);
6885	download_consts[SG_PREFETCH_ADDR_MASK] = (sg_prefetch_cnt - 1);
6886
6887	cur_patch = patches;
6888	downloaded = 0;
6889	skip_addr = 0;
6890	ahc_outb(ahc, SEQCTL, PERRORDIS|FAILDIS|FASTMODE|LOADRAM);
6891	ahc_outb(ahc, SEQADDR0, 0);
6892	ahc_outb(ahc, SEQADDR1, 0);
6893
6894	for (i = 0; i < sizeof(seqprog)/4; i++) {
6895		if (ahc_check_patch(ahc, &cur_patch, i, &skip_addr) == 0) {
6896			/*
6897			 * Don't download this instruction as it
6898			 * is in a patch that was removed.
6899			 */
6900			continue;
6901		}
6902
6903		if (downloaded == ahc->instruction_ram_size) {
6904			/*
6905			 * We're about to exceed the instruction
6906			 * storage capacity for this chip.  Fail
6907			 * the load.
6908			 */
6909			printk("\n%s: Program too large for instruction memory "
6910			       "size of %d!\n", ahc_name(ahc),
6911			       ahc->instruction_ram_size);
6912			return (ENOMEM);
6913		}
6914
6915		/*
6916		 * Move through the CS table until we find a CS
6917		 * that might apply to this instruction.
6918		 */
6919		for (; cur_cs < num_critical_sections; cur_cs++) {
6920			if (critical_sections[cur_cs].end <= i) {
6921				if (begin_set[cs_count] == TRUE
6922				 && end_set[cs_count] == FALSE) {
6923					cs_table[cs_count].end = downloaded;
6924				 	end_set[cs_count] = TRUE;
6925					cs_count++;
6926				}
6927				continue;
6928			}
6929			if (critical_sections[cur_cs].begin <= i
6930			 && begin_set[cs_count] == FALSE) {
6931				cs_table[cs_count].begin = downloaded;
6932				begin_set[cs_count] = TRUE;
6933			}
6934			break;
6935		}
6936		ahc_download_instr(ahc, i, download_consts);
6937		downloaded++;
6938	}
6939
6940	ahc->num_critical_sections = cs_count;
6941	if (cs_count != 0) {
6942
6943		cs_count *= sizeof(struct cs);
6944		ahc->critical_sections = kmalloc(cs_count, GFP_ATOMIC);
6945		if (ahc->critical_sections == NULL)
6946			panic("ahc_loadseq: Could not malloc");
6947		memcpy(ahc->critical_sections, cs_table, cs_count);
6948	}
6949	ahc_outb(ahc, SEQCTL, PERRORDIS|FAILDIS|FASTMODE);
6950
6951	if (bootverbose) {
6952		printk(" %d instructions downloaded\n", downloaded);
6953		printk("%s: Features 0x%x, Bugs 0x%x, Flags 0x%x\n",
6954		       ahc_name(ahc), ahc->features, ahc->bugs, ahc->flags);
6955	}
6956	return (0);
6957}
6958
6959static int
6960ahc_check_patch(struct ahc_softc *ahc, const struct patch **start_patch,
6961		u_int start_instr, u_int *skip_addr)
6962{
6963	const struct patch *cur_patch;
6964	const struct patch *last_patch;
6965	u_int	num_patches;
6966
6967	num_patches = ARRAY_SIZE(patches);
6968	last_patch = &patches[num_patches];
6969	cur_patch = *start_patch;
6970
6971	while (cur_patch < last_patch && start_instr == cur_patch->begin) {
6972
6973		if (cur_patch->patch_func(ahc) == 0) {
6974
6975			/* Start rejecting code */
6976			*skip_addr = start_instr + cur_patch->skip_instr;
6977			cur_patch += cur_patch->skip_patch;
6978		} else {
6979			/* Accepted this patch.  Advance to the next
6980			 * one and wait for our intruction pointer to
6981			 * hit this point.
6982			 */
6983			cur_patch++;
6984		}
6985	}
6986
6987	*start_patch = cur_patch;
6988	if (start_instr < *skip_addr)
6989		/* Still skipping */
6990		return (0);
6991
6992	return (1);
6993}
6994
6995static void
6996ahc_download_instr(struct ahc_softc *ahc, u_int instrptr, uint8_t *dconsts)
6997{
6998	union	ins_formats instr;
6999	struct	ins_format1 *fmt1_ins;
7000	struct	ins_format3 *fmt3_ins;
7001	u_int	opcode;
7002
7003	/*
7004	 * The firmware is always compiled into a little endian format.
7005	 */
7006	instr.integer = ahc_le32toh(*(uint32_t*)&seqprog[instrptr * 4]);
7007
7008	fmt1_ins = &instr.format1;
7009	fmt3_ins = NULL;
7010
7011	/* Pull the opcode */
7012	opcode = instr.format1.opcode;
7013	switch (opcode) {
7014	case AIC_OP_JMP:
7015	case AIC_OP_JC:
7016	case AIC_OP_JNC:
7017	case AIC_OP_CALL:
7018	case AIC_OP_JNE:
7019	case AIC_OP_JNZ:
7020	case AIC_OP_JE:
7021	case AIC_OP_JZ:
7022	{
7023		const struct patch *cur_patch;
7024		int address_offset;
7025		u_int address;
7026		u_int skip_addr;
7027		u_int i;
7028
7029		fmt3_ins = &instr.format3;
7030		address_offset = 0;
7031		address = fmt3_ins->address;
7032		cur_patch = patches;
7033		skip_addr = 0;
7034
7035		for (i = 0; i < address;) {
7036
7037			ahc_check_patch(ahc, &cur_patch, i, &skip_addr);
7038
7039			if (skip_addr > i) {
7040				int end_addr;
7041
7042				end_addr = min(address, skip_addr);
7043				address_offset += end_addr - i;
7044				i = skip_addr;
7045			} else {
7046				i++;
7047			}
7048		}
7049		address -= address_offset;
7050		fmt3_ins->address = address;
7051		/* FALLTHROUGH */
7052	}
7053	case AIC_OP_OR:
7054	case AIC_OP_AND:
7055	case AIC_OP_XOR:
7056	case AIC_OP_ADD:
7057	case AIC_OP_ADC:
7058	case AIC_OP_BMOV:
7059		if (fmt1_ins->parity != 0) {
7060			fmt1_ins->immediate = dconsts[fmt1_ins->immediate];
7061		}
7062		fmt1_ins->parity = 0;
7063		if ((ahc->features & AHC_CMD_CHAN) == 0
7064		 && opcode == AIC_OP_BMOV) {
7065			/*
7066			 * Block move was added at the same time
7067			 * as the command channel.  Verify that
7068			 * this is only a move of a single element
7069			 * and convert the BMOV to a MOV
7070			 * (AND with an immediate of FF).
7071			 */
7072			if (fmt1_ins->immediate != 1)
7073				panic("%s: BMOV not supported\n",
7074				      ahc_name(ahc));
7075			fmt1_ins->opcode = AIC_OP_AND;
7076			fmt1_ins->immediate = 0xff;
7077		}
7078		/* FALLTHROUGH */
7079	case AIC_OP_ROL:
7080		if ((ahc->features & AHC_ULTRA2) != 0) {
7081			int i, count;
7082
7083			/* Calculate odd parity for the instruction */
7084			for (i = 0, count = 0; i < 31; i++) {
7085				uint32_t mask;
7086
7087				mask = 0x01 << i;
7088				if ((instr.integer & mask) != 0)
7089					count++;
7090			}
7091			if ((count & 0x01) == 0)
7092				instr.format1.parity = 1;
7093		} else {
7094			/* Compress the instruction for older sequencers */
7095			if (fmt3_ins != NULL) {
7096				instr.integer =
7097					fmt3_ins->immediate
7098				      | (fmt3_ins->source << 8)
7099				      | (fmt3_ins->address << 16)
7100				      |	(fmt3_ins->opcode << 25);
7101			} else {
7102				instr.integer =
7103					fmt1_ins->immediate
7104				      | (fmt1_ins->source << 8)
7105				      | (fmt1_ins->destination << 16)
7106				      |	(fmt1_ins->ret << 24)
7107				      |	(fmt1_ins->opcode << 25);
7108			}
7109		}
7110		/* The sequencer is a little endian cpu */
7111		instr.integer = ahc_htole32(instr.integer);
7112		ahc_outsb(ahc, SEQRAM, instr.bytes, 4);
7113		break;
7114	default:
7115		panic("Unknown opcode encountered in seq program");
7116		break;
7117	}
7118}
7119
7120int
7121ahc_print_register(const ahc_reg_parse_entry_t *table, u_int num_entries,
7122		   const char *name, u_int address, u_int value,
7123		   u_int *cur_column, u_int wrap_point)
7124{
7125	int	printed;
7126	u_int	printed_mask;
7127
7128	if (cur_column != NULL && *cur_column >= wrap_point) {
7129		printk("\n");
7130		*cur_column = 0;
7131	}
7132	printed  = printk("%s[0x%x]", name, value);
7133	if (table == NULL) {
7134		printed += printk(" ");
7135		*cur_column += printed;
7136		return (printed);
7137	}
7138	printed_mask = 0;
7139	while (printed_mask != 0xFF) {
7140		int entry;
7141
7142		for (entry = 0; entry < num_entries; entry++) {
7143			if (((value & table[entry].mask)
7144			  != table[entry].value)
7145			 || ((printed_mask & table[entry].mask)
7146			  == table[entry].mask))
7147				continue;
7148
7149			printed += printk("%s%s",
7150					  printed_mask == 0 ? ":(" : "|",
7151					  table[entry].name);
7152			printed_mask |= table[entry].mask;
7153
7154			break;
7155		}
7156		if (entry >= num_entries)
7157			break;
7158	}
7159	if (printed_mask != 0)
7160		printed += printk(") ");
7161	else
7162		printed += printk(" ");
7163	if (cur_column != NULL)
7164		*cur_column += printed;
7165	return (printed);
7166}
7167
7168void
7169ahc_dump_card_state(struct ahc_softc *ahc)
7170{
7171	struct	scb *scb;
7172	struct	scb_tailq *untagged_q;
7173	u_int	cur_col;
7174	int	paused;
7175	int	target;
7176	int	maxtarget;
7177	int	i;
7178	uint8_t last_phase;
7179	uint8_t qinpos;
7180	uint8_t qintail;
7181	uint8_t qoutpos;
7182	uint8_t scb_index;
7183	uint8_t saved_scbptr;
7184
7185	if (ahc_is_paused(ahc)) {
7186		paused = 1;
7187	} else {
7188		paused = 0;
7189		ahc_pause(ahc);
7190	}
7191
7192	saved_scbptr = ahc_inb(ahc, SCBPTR);
7193	last_phase = ahc_inb(ahc, LASTPHASE);
7194	printk(">>>>>>>>>>>>>>>>>> Dump Card State Begins <<<<<<<<<<<<<<<<<\n"
7195	       "%s: Dumping Card State %s, at SEQADDR 0x%x\n",
7196	       ahc_name(ahc), ahc_lookup_phase_entry(last_phase)->phasemsg,
7197	       ahc_inb(ahc, SEQADDR0) | (ahc_inb(ahc, SEQADDR1) << 8));
7198	if (paused)
7199		printk("Card was paused\n");
7200	printk("ACCUM = 0x%x, SINDEX = 0x%x, DINDEX = 0x%x, ARG_2 = 0x%x\n",
7201	       ahc_inb(ahc, ACCUM), ahc_inb(ahc, SINDEX), ahc_inb(ahc, DINDEX),
7202	       ahc_inb(ahc, ARG_2));
7203	printk("HCNT = 0x%x SCBPTR = 0x%x\n", ahc_inb(ahc, HCNT),
7204	       ahc_inb(ahc, SCBPTR));
7205	cur_col = 0;
7206	if ((ahc->features & AHC_DT) != 0)
7207		ahc_scsiphase_print(ahc_inb(ahc, SCSIPHASE), &cur_col, 50);
7208	ahc_scsisigi_print(ahc_inb(ahc, SCSISIGI), &cur_col, 50);
7209	ahc_error_print(ahc_inb(ahc, ERROR), &cur_col, 50);
7210	ahc_scsibusl_print(ahc_inb(ahc, SCSIBUSL), &cur_col, 50);
7211	ahc_lastphase_print(ahc_inb(ahc, LASTPHASE), &cur_col, 50);
7212	ahc_scsiseq_print(ahc_inb(ahc, SCSISEQ), &cur_col, 50);
7213	ahc_sblkctl_print(ahc_inb(ahc, SBLKCTL), &cur_col, 50);
7214	ahc_scsirate_print(ahc_inb(ahc, SCSIRATE), &cur_col, 50);
7215	ahc_seqctl_print(ahc_inb(ahc, SEQCTL), &cur_col, 50);
7216	ahc_seq_flags_print(ahc_inb(ahc, SEQ_FLAGS), &cur_col, 50);
7217	ahc_sstat0_print(ahc_inb(ahc, SSTAT0), &cur_col, 50);
7218	ahc_sstat1_print(ahc_inb(ahc, SSTAT1), &cur_col, 50);
7219	ahc_sstat2_print(ahc_inb(ahc, SSTAT2), &cur_col, 50);
7220	ahc_sstat3_print(ahc_inb(ahc, SSTAT3), &cur_col, 50);
7221	ahc_simode0_print(ahc_inb(ahc, SIMODE0), &cur_col, 50);
7222	ahc_simode1_print(ahc_inb(ahc, SIMODE1), &cur_col, 50);
7223	ahc_sxfrctl0_print(ahc_inb(ahc, SXFRCTL0), &cur_col, 50);
7224	ahc_dfcntrl_print(ahc_inb(ahc, DFCNTRL), &cur_col, 50);
7225	ahc_dfstatus_print(ahc_inb(ahc, DFSTATUS), &cur_col, 50);
7226	if (cur_col != 0)
7227		printk("\n");
7228	printk("STACK:");
7229	for (i = 0; i < STACK_SIZE; i++)
7230		printk(" 0x%x", ahc_inb(ahc, STACK)|(ahc_inb(ahc, STACK) << 8));
7231	printk("\nSCB count = %d\n", ahc->scb_data->numscbs);
7232	printk("Kernel NEXTQSCB = %d\n", ahc->next_queued_scb->hscb->tag);
7233	printk("Card NEXTQSCB = %d\n", ahc_inb(ahc, NEXT_QUEUED_SCB));
7234	/* QINFIFO */
7235	printk("QINFIFO entries: ");
7236	if ((ahc->features & AHC_QUEUE_REGS) != 0) {
7237		qinpos = ahc_inb(ahc, SNSCB_QOFF);
7238		ahc_outb(ahc, SNSCB_QOFF, qinpos);
7239	} else
7240		qinpos = ahc_inb(ahc, QINPOS);
7241	qintail = ahc->qinfifonext;
7242	while (qinpos != qintail) {
7243		printk("%d ", ahc->qinfifo[qinpos]);
7244		qinpos++;
7245	}
7246	printk("\n");
7247
7248	printk("Waiting Queue entries: ");
7249	scb_index = ahc_inb(ahc, WAITING_SCBH);
7250	i = 0;
7251	while (scb_index != SCB_LIST_NULL && i++ < 256) {
7252		ahc_outb(ahc, SCBPTR, scb_index);
7253		printk("%d:%d ", scb_index, ahc_inb(ahc, SCB_TAG));
7254		scb_index = ahc_inb(ahc, SCB_NEXT);
7255	}
7256	printk("\n");
7257
7258	printk("Disconnected Queue entries: ");
7259	scb_index = ahc_inb(ahc, DISCONNECTED_SCBH);
7260	i = 0;
7261	while (scb_index != SCB_LIST_NULL && i++ < 256) {
7262		ahc_outb(ahc, SCBPTR, scb_index);
7263		printk("%d:%d ", scb_index, ahc_inb(ahc, SCB_TAG));
7264		scb_index = ahc_inb(ahc, SCB_NEXT);
7265	}
7266	printk("\n");
7267
7268	ahc_sync_qoutfifo(ahc, BUS_DMASYNC_POSTREAD);
7269	printk("QOUTFIFO entries: ");
7270	qoutpos = ahc->qoutfifonext;
7271	i = 0;
7272	while (ahc->qoutfifo[qoutpos] != SCB_LIST_NULL && i++ < 256) {
7273		printk("%d ", ahc->qoutfifo[qoutpos]);
7274		qoutpos++;
7275	}
7276	printk("\n");
7277
7278	printk("Sequencer Free SCB List: ");
7279	scb_index = ahc_inb(ahc, FREE_SCBH);
7280	i = 0;
7281	while (scb_index != SCB_LIST_NULL && i++ < 256) {
7282		ahc_outb(ahc, SCBPTR, scb_index);
7283		printk("%d ", scb_index);
7284		scb_index = ahc_inb(ahc, SCB_NEXT);
7285	}
7286	printk("\n");
7287
7288	printk("Sequencer SCB Info: ");
7289	for (i = 0; i < ahc->scb_data->maxhscbs; i++) {
7290		ahc_outb(ahc, SCBPTR, i);
7291		cur_col  = printk("\n%3d ", i);
7292
7293		ahc_scb_control_print(ahc_inb(ahc, SCB_CONTROL), &cur_col, 60);
7294		ahc_scb_scsiid_print(ahc_inb(ahc, SCB_SCSIID), &cur_col, 60);
7295		ahc_scb_lun_print(ahc_inb(ahc, SCB_LUN), &cur_col, 60);
7296		ahc_scb_tag_print(ahc_inb(ahc, SCB_TAG), &cur_col, 60);
7297	}
7298	printk("\n");
7299
7300	printk("Pending list: ");
7301	i = 0;
7302	LIST_FOREACH(scb, &ahc->pending_scbs, pending_links) {
7303		if (i++ > 256)
7304			break;
7305		cur_col  = printk("\n%3d ", scb->hscb->tag);
7306		ahc_scb_control_print(scb->hscb->control, &cur_col, 60);
7307		ahc_scb_scsiid_print(scb->hscb->scsiid, &cur_col, 60);
7308		ahc_scb_lun_print(scb->hscb->lun, &cur_col, 60);
7309		if ((ahc->flags & AHC_PAGESCBS) == 0) {
7310			ahc_outb(ahc, SCBPTR, scb->hscb->tag);
7311			printk("(");
7312			ahc_scb_control_print(ahc_inb(ahc, SCB_CONTROL),
7313					      &cur_col, 60);
7314			ahc_scb_tag_print(ahc_inb(ahc, SCB_TAG), &cur_col, 60);
7315			printk(")");
7316		}
7317	}
7318	printk("\n");
7319
7320	printk("Kernel Free SCB list: ");
7321	i = 0;
7322	SLIST_FOREACH(scb, &ahc->scb_data->free_scbs, links.sle) {
7323		if (i++ > 256)
7324			break;
7325		printk("%d ", scb->hscb->tag);
7326	}
7327	printk("\n");
7328
7329	maxtarget = (ahc->features & (AHC_WIDE|AHC_TWIN)) ? 15 : 7;
7330	for (target = 0; target <= maxtarget; target++) {
7331		untagged_q = &ahc->untagged_queues[target];
7332		if (TAILQ_FIRST(untagged_q) == NULL)
7333			continue;
7334		printk("Untagged Q(%d): ", target);
7335		i = 0;
7336		TAILQ_FOREACH(scb, untagged_q, links.tqe) {
7337			if (i++ > 256)
7338				break;
7339			printk("%d ", scb->hscb->tag);
7340		}
7341		printk("\n");
7342	}
7343
7344	ahc_platform_dump_card_state(ahc);
7345	printk("\n<<<<<<<<<<<<<<<<< Dump Card State Ends >>>>>>>>>>>>>>>>>>\n");
7346	ahc_outb(ahc, SCBPTR, saved_scbptr);
7347	if (paused == 0)
7348		ahc_unpause(ahc);
7349}
7350
7351/************************* Target Mode ****************************************/
7352#ifdef AHC_TARGET_MODE
7353cam_status
7354ahc_find_tmode_devs(struct ahc_softc *ahc, struct cam_sim *sim, union ccb *ccb,
7355		    struct ahc_tmode_tstate **tstate,
7356		    struct ahc_tmode_lstate **lstate,
7357		    int notfound_failure)
7358{
7359
7360	if ((ahc->features & AHC_TARGETMODE) == 0)
7361		return (CAM_REQ_INVALID);
7362
7363	/*
7364	 * Handle the 'black hole' device that sucks up
7365	 * requests to unattached luns on enabled targets.
7366	 */
7367	if (ccb->ccb_h.target_id == CAM_TARGET_WILDCARD
7368	 && ccb->ccb_h.target_lun == CAM_LUN_WILDCARD) {
7369		*tstate = NULL;
7370		*lstate = ahc->black_hole;
7371	} else {
7372		u_int max_id;
7373
7374		max_id = (ahc->features & AHC_WIDE) ? 16 : 8;
7375		if (ccb->ccb_h.target_id >= max_id)
7376			return (CAM_TID_INVALID);
7377
7378		if (ccb->ccb_h.target_lun >= AHC_NUM_LUNS)
7379			return (CAM_LUN_INVALID);
7380
7381		*tstate = ahc->enabled_targets[ccb->ccb_h.target_id];
7382		*lstate = NULL;
7383		if (*tstate != NULL)
7384			*lstate =
7385			    (*tstate)->enabled_luns[ccb->ccb_h.target_lun];
7386	}
7387
7388	if (notfound_failure != 0 && *lstate == NULL)
7389		return (CAM_PATH_INVALID);
7390
7391	return (CAM_REQ_CMP);
7392}
7393
7394void
7395ahc_handle_en_lun(struct ahc_softc *ahc, struct cam_sim *sim, union ccb *ccb)
7396{
7397	struct	   ahc_tmode_tstate *tstate;
7398	struct	   ahc_tmode_lstate *lstate;
7399	struct	   ccb_en_lun *cel;
7400	cam_status status;
7401	u_long	   s;
7402	u_int	   target;
7403	u_int	   lun;
7404	u_int	   target_mask;
7405	u_int	   our_id;
7406	int	   error;
7407	char	   channel;
7408
7409	status = ahc_find_tmode_devs(ahc, sim, ccb, &tstate, &lstate,
7410				     /*notfound_failure*/FALSE);
7411
7412	if (status != CAM_REQ_CMP) {
7413		ccb->ccb_h.status = status;
7414		return;
7415	}
7416
7417	if (cam_sim_bus(sim) == 0)
7418		our_id = ahc->our_id;
7419	else
7420		our_id = ahc->our_id_b;
7421
7422	if (ccb->ccb_h.target_id != our_id) {
7423		/*
7424		 * our_id represents our initiator ID, or
7425		 * the ID of the first target to have an
7426		 * enabled lun in target mode.  There are
7427		 * two cases that may preclude enabling a
7428		 * target id other than our_id.
7429		 *
7430		 *   o our_id is for an active initiator role.
7431		 *     Since the hardware does not support
7432		 *     reselections to the initiator role at
7433		 *     anything other than our_id, and our_id
7434		 *     is used by the hardware to indicate the
7435		 *     ID to use for both select-out and
7436		 *     reselect-out operations, the only target
7437		 *     ID we can support in this mode is our_id.
7438		 *
7439		 *   o The MULTARGID feature is not available and
7440		 *     a previous target mode ID has been enabled.
7441		 */
7442		if ((ahc->features & AHC_MULTIROLE) != 0) {
7443
7444			if ((ahc->features & AHC_MULTI_TID) != 0
7445		   	 && (ahc->flags & AHC_INITIATORROLE) != 0) {
7446				/*
7447				 * Only allow additional targets if
7448				 * the initiator role is disabled.
7449				 * The hardware cannot handle a re-select-in
7450				 * on the initiator id during a re-select-out
7451				 * on a different target id.
7452				 */
7453				status = CAM_TID_INVALID;
7454			} else if ((ahc->flags & AHC_INITIATORROLE) != 0
7455				|| ahc->enabled_luns > 0) {
7456				/*
7457				 * Only allow our target id to change
7458				 * if the initiator role is not configured
7459				 * and there are no enabled luns which
7460				 * are attached to the currently registered
7461				 * scsi id.
7462				 */
7463				status = CAM_TID_INVALID;
7464			}
7465		} else if ((ahc->features & AHC_MULTI_TID) == 0
7466			&& ahc->enabled_luns > 0) {
7467
7468			status = CAM_TID_INVALID;
7469		}
7470	}
7471
7472	if (status != CAM_REQ_CMP) {
7473		ccb->ccb_h.status = status;
7474		return;
7475	}
7476
7477	/*
7478	 * We now have an id that is valid.
7479	 * If we aren't in target mode, switch modes.
7480	 */
7481	if ((ahc->flags & AHC_TARGETROLE) == 0
7482	 && ccb->ccb_h.target_id != CAM_TARGET_WILDCARD) {
7483		u_long	 s;
7484		ahc_flag saved_flags;
7485
7486		printk("Configuring Target Mode\n");
7487		ahc_lock(ahc, &s);
7488		if (LIST_FIRST(&ahc->pending_scbs) != NULL) {
7489			ccb->ccb_h.status = CAM_BUSY;
7490			ahc_unlock(ahc, &s);
7491			return;
7492		}
7493		saved_flags = ahc->flags;
7494		ahc->flags |= AHC_TARGETROLE;
7495		if ((ahc->features & AHC_MULTIROLE) == 0)
7496			ahc->flags &= ~AHC_INITIATORROLE;
7497		ahc_pause(ahc);
7498		error = ahc_loadseq(ahc);
7499		if (error != 0) {
7500			/*
7501			 * Restore original configuration and notify
7502			 * the caller that we cannot support target mode.
7503			 * Since the adapter started out in this
7504			 * configuration, the firmware load will succeed,
7505			 * so there is no point in checking ahc_loadseq's
7506			 * return value.
7507			 */
7508			ahc->flags = saved_flags;
7509			(void)ahc_loadseq(ahc);
7510			ahc_restart(ahc);
7511			ahc_unlock(ahc, &s);
7512			ccb->ccb_h.status = CAM_FUNC_NOTAVAIL;
7513			return;
7514		}
7515		ahc_restart(ahc);
7516		ahc_unlock(ahc, &s);
7517	}
7518	cel = &ccb->cel;
7519	target = ccb->ccb_h.target_id;
7520	lun = ccb->ccb_h.target_lun;
7521	channel = SIM_CHANNEL(ahc, sim);
7522	target_mask = 0x01 << target;
7523	if (channel == 'B')
7524		target_mask <<= 8;
7525
7526	if (cel->enable != 0) {
7527		u_int scsiseq;
7528
7529		/* Are we already enabled?? */
7530		if (lstate != NULL) {
7531			xpt_print_path(ccb->ccb_h.path);
7532			printk("Lun already enabled\n");
7533			ccb->ccb_h.status = CAM_LUN_ALRDY_ENA;
7534			return;
7535		}
7536
7537		if (cel->grp6_len != 0
7538		 || cel->grp7_len != 0) {
7539			/*
7540			 * Don't (yet?) support vendor
7541			 * specific commands.
7542			 */
7543			ccb->ccb_h.status = CAM_REQ_INVALID;
7544			printk("Non-zero Group Codes\n");
7545			return;
7546		}
7547
7548		/*
7549		 * Seems to be okay.
7550		 * Setup our data structures.
7551		 */
7552		if (target != CAM_TARGET_WILDCARD && tstate == NULL) {
7553			tstate = ahc_alloc_tstate(ahc, target, channel);
7554			if (tstate == NULL) {
7555				xpt_print_path(ccb->ccb_h.path);
7556				printk("Couldn't allocate tstate\n");
7557				ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
7558				return;
7559			}
7560		}
7561		lstate = kmalloc(sizeof(*lstate), GFP_ATOMIC);
7562		if (lstate == NULL) {
7563			xpt_print_path(ccb->ccb_h.path);
7564			printk("Couldn't allocate lstate\n");
7565			ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
7566			return;
7567		}
7568		memset(lstate, 0, sizeof(*lstate));
7569		status = xpt_create_path(&lstate->path, /*periph*/NULL,
7570					 xpt_path_path_id(ccb->ccb_h.path),
7571					 xpt_path_target_id(ccb->ccb_h.path),
7572					 xpt_path_lun_id(ccb->ccb_h.path));
7573		if (status != CAM_REQ_CMP) {
7574			kfree(lstate);
7575			xpt_print_path(ccb->ccb_h.path);
7576			printk("Couldn't allocate path\n");
7577			ccb->ccb_h.status = CAM_RESRC_UNAVAIL;
7578			return;
7579		}
7580		SLIST_INIT(&lstate->accept_tios);
7581		SLIST_INIT(&lstate->immed_notifies);
7582		ahc_lock(ahc, &s);
7583		ahc_pause(ahc);
7584		if (target != CAM_TARGET_WILDCARD) {
7585			tstate->enabled_luns[lun] = lstate;
7586			ahc->enabled_luns++;
7587
7588			if ((ahc->features & AHC_MULTI_TID) != 0) {
7589				u_int targid_mask;
7590
7591				targid_mask = ahc_inb(ahc, TARGID)
7592					    | (ahc_inb(ahc, TARGID + 1) << 8);
7593
7594				targid_mask |= target_mask;
7595				ahc_outb(ahc, TARGID, targid_mask);
7596				ahc_outb(ahc, TARGID+1, (targid_mask >> 8));
7597
7598				ahc_update_scsiid(ahc, targid_mask);
7599			} else {
7600				u_int our_id;
7601				char  channel;
7602
7603				channel = SIM_CHANNEL(ahc, sim);
7604				our_id = SIM_SCSI_ID(ahc, sim);
7605
7606				/*
7607				 * This can only happen if selections
7608				 * are not enabled
7609				 */
7610				if (target != our_id) {
7611					u_int sblkctl;
7612					char  cur_channel;
7613					int   swap;
7614
7615					sblkctl = ahc_inb(ahc, SBLKCTL);
7616					cur_channel = (sblkctl & SELBUSB)
7617						    ? 'B' : 'A';
7618					if ((ahc->features & AHC_TWIN) == 0)
7619						cur_channel = 'A';
7620					swap = cur_channel != channel;
7621					if (channel == 'A')
7622						ahc->our_id = target;
7623					else
7624						ahc->our_id_b = target;
7625
7626					if (swap)
7627						ahc_outb(ahc, SBLKCTL,
7628							 sblkctl ^ SELBUSB);
7629
7630					ahc_outb(ahc, SCSIID, target);
7631
7632					if (swap)
7633						ahc_outb(ahc, SBLKCTL, sblkctl);
7634				}
7635			}
7636		} else
7637			ahc->black_hole = lstate;
7638		/* Allow select-in operations */
7639		if (ahc->black_hole != NULL && ahc->enabled_luns > 0) {
7640			scsiseq = ahc_inb(ahc, SCSISEQ_TEMPLATE);
7641			scsiseq |= ENSELI;
7642			ahc_outb(ahc, SCSISEQ_TEMPLATE, scsiseq);
7643			scsiseq = ahc_inb(ahc, SCSISEQ);
7644			scsiseq |= ENSELI;
7645			ahc_outb(ahc, SCSISEQ, scsiseq);
7646		}
7647		ahc_unpause(ahc);
7648		ahc_unlock(ahc, &s);
7649		ccb->ccb_h.status = CAM_REQ_CMP;
7650		xpt_print_path(ccb->ccb_h.path);
7651		printk("Lun now enabled for target mode\n");
7652	} else {
7653		struct scb *scb;
7654		int i, empty;
7655
7656		if (lstate == NULL) {
7657			ccb->ccb_h.status = CAM_LUN_INVALID;
7658			return;
7659		}
7660
7661		ahc_lock(ahc, &s);
7662
7663		ccb->ccb_h.status = CAM_REQ_CMP;
7664		LIST_FOREACH(scb, &ahc->pending_scbs, pending_links) {
7665			struct ccb_hdr *ccbh;
7666
7667			ccbh = &scb->io_ctx->ccb_h;
7668			if (ccbh->func_code == XPT_CONT_TARGET_IO
7669			 && !xpt_path_comp(ccbh->path, ccb->ccb_h.path)){
7670				printk("CTIO pending\n");
7671				ccb->ccb_h.status = CAM_REQ_INVALID;
7672				ahc_unlock(ahc, &s);
7673				return;
7674			}
7675		}
7676
7677		if (SLIST_FIRST(&lstate->accept_tios) != NULL) {
7678			printk("ATIOs pending\n");
7679			ccb->ccb_h.status = CAM_REQ_INVALID;
7680		}
7681
7682		if (SLIST_FIRST(&lstate->immed_notifies) != NULL) {
7683			printk("INOTs pending\n");
7684			ccb->ccb_h.status = CAM_REQ_INVALID;
7685		}
7686
7687		if (ccb->ccb_h.status != CAM_REQ_CMP) {
7688			ahc_unlock(ahc, &s);
7689			return;
7690		}
7691
7692		xpt_print_path(ccb->ccb_h.path);
7693		printk("Target mode disabled\n");
7694		xpt_free_path(lstate->path);
7695		kfree(lstate);
7696
7697		ahc_pause(ahc);
7698		/* Can we clean up the target too? */
7699		if (target != CAM_TARGET_WILDCARD) {
7700			tstate->enabled_luns[lun] = NULL;
7701			ahc->enabled_luns--;
7702			for (empty = 1, i = 0; i < 8; i++)
7703				if (tstate->enabled_luns[i] != NULL) {
7704					empty = 0;
7705					break;
7706				}
7707
7708			if (empty) {
7709				ahc_free_tstate(ahc, target, channel,
7710						/*force*/FALSE);
7711				if (ahc->features & AHC_MULTI_TID) {
7712					u_int targid_mask;
7713
7714					targid_mask = ahc_inb(ahc, TARGID)
7715						    | (ahc_inb(ahc, TARGID + 1)
7716						       << 8);
7717
7718					targid_mask &= ~target_mask;
7719					ahc_outb(ahc, TARGID, targid_mask);
7720					ahc_outb(ahc, TARGID+1,
7721					 	 (targid_mask >> 8));
7722					ahc_update_scsiid(ahc, targid_mask);
7723				}
7724			}
7725		} else {
7726
7727			ahc->black_hole = NULL;
7728
7729			/*
7730			 * We can't allow selections without
7731			 * our black hole device.
7732			 */
7733			empty = TRUE;
7734		}
7735		if (ahc->enabled_luns == 0) {
7736			/* Disallow select-in */
7737			u_int scsiseq;
7738
7739			scsiseq = ahc_inb(ahc, SCSISEQ_TEMPLATE);
7740			scsiseq &= ~ENSELI;
7741			ahc_outb(ahc, SCSISEQ_TEMPLATE, scsiseq);
7742			scsiseq = ahc_inb(ahc, SCSISEQ);
7743			scsiseq &= ~ENSELI;
7744			ahc_outb(ahc, SCSISEQ, scsiseq);
7745
7746			if ((ahc->features & AHC_MULTIROLE) == 0) {
7747				printk("Configuring Initiator Mode\n");
7748				ahc->flags &= ~AHC_TARGETROLE;
7749				ahc->flags |= AHC_INITIATORROLE;
7750				/*
7751				 * Returning to a configuration that
7752				 * fit previously will always succeed.
7753				 */
7754				(void)ahc_loadseq(ahc);
7755				ahc_restart(ahc);
7756				/*
7757				 * Unpaused.  The extra unpause
7758				 * that follows is harmless.
7759				 */
7760			}
7761		}
7762		ahc_unpause(ahc);
7763		ahc_unlock(ahc, &s);
7764	}
7765}
7766
7767static void
7768ahc_update_scsiid(struct ahc_softc *ahc, u_int targid_mask)
7769{
7770	u_int scsiid_mask;
7771	u_int scsiid;
7772
7773	if ((ahc->features & AHC_MULTI_TID) == 0)
7774		panic("ahc_update_scsiid called on non-multitid unit\n");
7775
7776	/*
7777	 * Since we will rely on the TARGID mask
7778	 * for selection enables, ensure that OID
7779	 * in SCSIID is not set to some other ID
7780	 * that we don't want to allow selections on.
7781	 */
7782	if ((ahc->features & AHC_ULTRA2) != 0)
7783		scsiid = ahc_inb(ahc, SCSIID_ULTRA2);
7784	else
7785		scsiid = ahc_inb(ahc, SCSIID);
7786	scsiid_mask = 0x1 << (scsiid & OID);
7787	if ((targid_mask & scsiid_mask) == 0) {
7788		u_int our_id;
7789
7790		/* ffs counts from 1 */
7791		our_id = ffs(targid_mask);
7792		if (our_id == 0)
7793			our_id = ahc->our_id;
7794		else
7795			our_id--;
7796		scsiid &= TID;
7797		scsiid |= our_id;
7798	}
7799	if ((ahc->features & AHC_ULTRA2) != 0)
7800		ahc_outb(ahc, SCSIID_ULTRA2, scsiid);
7801	else
7802		ahc_outb(ahc, SCSIID, scsiid);
7803}
7804
7805static void
7806ahc_run_tqinfifo(struct ahc_softc *ahc, int paused)
7807{
7808	struct target_cmd *cmd;
7809
7810	/*
7811	 * If the card supports auto-access pause,
7812	 * we can access the card directly regardless
7813	 * of whether it is paused or not.
7814	 */
7815	if ((ahc->features & AHC_AUTOPAUSE) != 0)
7816		paused = TRUE;
7817
7818	ahc_sync_tqinfifo(ahc, BUS_DMASYNC_POSTREAD);
7819	while ((cmd = &ahc->targetcmds[ahc->tqinfifonext])->cmd_valid != 0) {
7820
7821		/*
7822		 * Only advance through the queue if we
7823		 * have the resources to process the command.
7824		 */
7825		if (ahc_handle_target_cmd(ahc, cmd) != 0)
7826			break;
7827
7828		cmd->cmd_valid = 0;
7829		ahc_dmamap_sync(ahc, ahc->shared_data_dmat,
7830				ahc->shared_data_dmamap,
7831				ahc_targetcmd_offset(ahc, ahc->tqinfifonext),
7832				sizeof(struct target_cmd),
7833				BUS_DMASYNC_PREREAD);
7834		ahc->tqinfifonext++;
7835
7836		/*
7837		 * Lazily update our position in the target mode incoming
7838		 * command queue as seen by the sequencer.
7839		 */
7840		if ((ahc->tqinfifonext & (HOST_TQINPOS - 1)) == 1) {
7841			if ((ahc->features & AHC_HS_MAILBOX) != 0) {
7842				u_int hs_mailbox;
7843
7844				hs_mailbox = ahc_inb(ahc, HS_MAILBOX);
7845				hs_mailbox &= ~HOST_TQINPOS;
7846				hs_mailbox |= ahc->tqinfifonext & HOST_TQINPOS;
7847				ahc_outb(ahc, HS_MAILBOX, hs_mailbox);
7848			} else {
7849				if (!paused)
7850					ahc_pause(ahc);
7851				ahc_outb(ahc, KERNEL_TQINPOS,
7852					 ahc->tqinfifonext & HOST_TQINPOS);
7853				if (!paused)
7854					ahc_unpause(ahc);
7855			}
7856		}
7857	}
7858}
7859
7860static int
7861ahc_handle_target_cmd(struct ahc_softc *ahc, struct target_cmd *cmd)
7862{
7863	struct	  ahc_tmode_tstate *tstate;
7864	struct	  ahc_tmode_lstate *lstate;
7865	struct	  ccb_accept_tio *atio;
7866	uint8_t *byte;
7867	int	  initiator;
7868	int	  target;
7869	int	  lun;
7870
7871	initiator = SCSIID_TARGET(ahc, cmd->scsiid);
7872	target = SCSIID_OUR_ID(cmd->scsiid);
7873	lun    = (cmd->identify & MSG_IDENTIFY_LUNMASK);
7874
7875	byte = cmd->bytes;
7876	tstate = ahc->enabled_targets[target];
7877	lstate = NULL;
7878	if (tstate != NULL)
7879		lstate = tstate->enabled_luns[lun];
7880
7881	/*
7882	 * Commands for disabled luns go to the black hole driver.
7883	 */
7884	if (lstate == NULL)
7885		lstate = ahc->black_hole;
7886
7887	atio = (struct ccb_accept_tio*)SLIST_FIRST(&lstate->accept_tios);
7888	if (atio == NULL) {
7889		ahc->flags |= AHC_TQINFIFO_BLOCKED;
7890		/*
7891		 * Wait for more ATIOs from the peripheral driver for this lun.
7892		 */
7893		if (bootverbose)
7894			printk("%s: ATIOs exhausted\n", ahc_name(ahc));
7895		return (1);
7896	} else
7897		ahc->flags &= ~AHC_TQINFIFO_BLOCKED;
7898#if 0
7899	printk("Incoming command from %d for %d:%d%s\n",
7900	       initiator, target, lun,
7901	       lstate == ahc->black_hole ? "(Black Holed)" : "");
7902#endif
7903	SLIST_REMOVE_HEAD(&lstate->accept_tios, sim_links.sle);
7904
7905	if (lstate == ahc->black_hole) {
7906		/* Fill in the wildcards */
7907		atio->ccb_h.target_id = target;
7908		atio->ccb_h.target_lun = lun;
7909	}
7910
7911	/*
7912	 * Package it up and send it off to
7913	 * whomever has this lun enabled.
7914	 */
7915	atio->sense_len = 0;
7916	atio->init_id = initiator;
7917	if (byte[0] != 0xFF) {
7918		/* Tag was included */
7919		atio->tag_action = *byte++;
7920		atio->tag_id = *byte++;
7921		atio->ccb_h.flags = CAM_TAG_ACTION_VALID;
7922	} else {
7923		atio->ccb_h.flags = 0;
7924	}
7925	byte++;
7926
7927	/* Okay.  Now determine the cdb size based on the command code */
7928	switch (*byte >> CMD_GROUP_CODE_SHIFT) {
7929	case 0:
7930		atio->cdb_len = 6;
7931		break;
7932	case 1:
7933	case 2:
7934		atio->cdb_len = 10;
7935		break;
7936	case 4:
7937		atio->cdb_len = 16;
7938		break;
7939	case 5:
7940		atio->cdb_len = 12;
7941		break;
7942	case 3:
7943	default:
7944		/* Only copy the opcode. */
7945		atio->cdb_len = 1;
7946		printk("Reserved or VU command code type encountered\n");
7947		break;
7948	}
7949
7950	memcpy(atio->cdb_io.cdb_bytes, byte, atio->cdb_len);
7951
7952	atio->ccb_h.status |= CAM_CDB_RECVD;
7953
7954	if ((cmd->identify & MSG_IDENTIFY_DISCFLAG) == 0) {
7955		/*
7956		 * We weren't allowed to disconnect.
7957		 * We're hanging on the bus until a
7958		 * continue target I/O comes in response
7959		 * to this accept tio.
7960		 */
7961#if 0
7962		printk("Received Immediate Command %d:%d:%d - %p\n",
7963		       initiator, target, lun, ahc->pending_device);
7964#endif
7965		ahc->pending_device = lstate;
7966		ahc_freeze_ccb((union ccb *)atio);
7967		atio->ccb_h.flags |= CAM_DIS_DISCONNECT;
7968	}
7969	xpt_done((union ccb*)atio);
7970	return (0);
7971}
7972
7973#endif
7974