1 | /* $NetBSD: ld_iop.c,v 1.36 2016/09/16 15:20:50 jdolecek Exp $ */ |
2 | |
3 | /*- |
4 | * Copyright (c) 2000, 2001 The NetBSD Foundation, Inc. |
5 | * All rights reserved. |
6 | * |
7 | * This code is derived from software contributed to The NetBSD Foundation |
8 | * by Andrew Doran. |
9 | * |
10 | * Redistribution and use in source and binary forms, with or without |
11 | * modification, are permitted provided that the following conditions |
12 | * are met: |
13 | * 1. Redistributions of source code must retain the above copyright |
14 | * notice, this list of conditions and the following disclaimer. |
15 | * 2. Redistributions in binary form must reproduce the above copyright |
16 | * notice, this list of conditions and the following disclaimer in the |
17 | * documentation and/or other materials provided with the distribution. |
18 | * |
19 | * THIS SOFTWARE IS PROVIDED BY THE NETBSD FOUNDATION, INC. AND CONTRIBUTORS |
20 | * ``AS IS'' AND ANY EXPRESS OR IMPLIED WARRANTIES, INCLUDING, BUT NOT LIMITED |
21 | * TO, THE IMPLIED WARRANTIES OF MERCHANTABILITY AND FITNESS FOR A PARTICULAR |
22 | * PURPOSE ARE DISCLAIMED. IN NO EVENT SHALL THE FOUNDATION OR CONTRIBUTORS |
23 | * BE LIABLE FOR ANY DIRECT, INDIRECT, INCIDENTAL, SPECIAL, EXEMPLARY, OR |
24 | * CONSEQUENTIAL DAMAGES (INCLUDING, BUT NOT LIMITED TO, PROCUREMENT OF |
25 | * SUBSTITUTE GOODS OR SERVICES; LOSS OF USE, DATA, OR PROFITS; OR BUSINESS |
26 | * INTERRUPTION) HOWEVER CAUSED AND ON ANY THEORY OF LIABILITY, WHETHER IN |
27 | * CONTRACT, STRICT LIABILITY, OR TORT (INCLUDING NEGLIGENCE OR OTHERWISE) |
28 | * ARISING IN ANY WAY OUT OF THE USE OF THIS SOFTWARE, EVEN IF ADVISED OF THE |
29 | * POSSIBILITY OF SUCH DAMAGE. |
30 | */ |
31 | |
32 | /* |
33 | * I2O front-end for ld(4) driver, supporting random block storage class |
34 | * devices. Currently, this doesn't handle anything more complex than |
35 | * fixed direct-access devices. |
36 | */ |
37 | |
38 | #include <sys/cdefs.h> |
39 | __KERNEL_RCSID(0, "$NetBSD: ld_iop.c,v 1.36 2016/09/16 15:20:50 jdolecek Exp $" ); |
40 | |
41 | #include <sys/param.h> |
42 | #include <sys/systm.h> |
43 | #include <sys/kernel.h> |
44 | #include <sys/device.h> |
45 | #include <sys/buf.h> |
46 | #include <sys/bufq.h> |
47 | #include <sys/endian.h> |
48 | #include <sys/dkio.h> |
49 | #include <sys/disk.h> |
50 | #include <sys/proc.h> |
51 | |
52 | #include <sys/bus.h> |
53 | |
54 | #include <dev/ldvar.h> |
55 | |
56 | #include <dev/i2o/i2o.h> |
57 | #include <dev/i2o/iopio.h> |
58 | #include <dev/i2o/iopvar.h> |
59 | |
60 | #define LD_IOP_TIMEOUT 30*1000 |
61 | |
62 | #define LD_IOP_CLAIMED 0x01 |
63 | #define LD_IOP_NEW_EVTMASK 0x02 |
64 | |
65 | struct ld_iop_softc { |
66 | struct ld_softc sc_ld; |
67 | struct iop_initiator sc_ii; |
68 | struct iop_initiator sc_eventii; |
69 | int sc_flags; |
70 | }; |
71 | |
72 | static void ld_iop_adjqparam(device_t, int); |
73 | static void ld_iop_attach(device_t, device_t, void *); |
74 | static int ld_iop_detach(device_t, int); |
75 | static int ld_iop_dump(struct ld_softc *, void *, int, int); |
76 | static int ld_iop_flush(struct ld_softc *, int); |
77 | static void ld_iop_intr(device_t, struct iop_msg *, void *); |
78 | static void ld_iop_intr_event(device_t, struct iop_msg *, void *); |
79 | static int ld_iop_match(device_t, cfdata_t, void *); |
80 | static int ld_iop_start(struct ld_softc *, struct buf *); |
81 | static void ld_iop_unconfig(struct ld_iop_softc *, int); |
82 | |
83 | CFATTACH_DECL_NEW(ld_iop, sizeof(struct ld_iop_softc), |
84 | ld_iop_match, ld_iop_attach, ld_iop_detach, NULL); |
85 | |
86 | static const char * const ld_iop_errors[] = { |
87 | "success" , |
88 | "media error" , |
89 | "access error" , |
90 | "device failure" , |
91 | "device not ready" , |
92 | "media not present" , |
93 | "media locked" , |
94 | "media failure" , |
95 | "protocol failure" , |
96 | "bus failure" , |
97 | "access violation" , |
98 | "media write protected" , |
99 | "device reset" , |
100 | "volume changed, waiting for acknowledgement" , |
101 | "timeout" , |
102 | }; |
103 | |
104 | static int |
105 | ld_iop_match(device_t parent, cfdata_t match, void *aux) |
106 | { |
107 | struct iop_attach_args *ia; |
108 | |
109 | ia = aux; |
110 | |
111 | return (ia->ia_class == I2O_CLASS_RANDOM_BLOCK_STORAGE); |
112 | } |
113 | |
114 | static void |
115 | ld_iop_attach(device_t parent, device_t self, void *aux) |
116 | { |
117 | struct iop_attach_args *ia = aux; |
118 | struct ld_iop_softc *sc = device_private(self); |
119 | struct iop_softc *iop = device_private(parent); |
120 | struct ld_softc *ld = &sc->sc_ld; |
121 | int rv, evreg, enable; |
122 | const char *typestr, *fixedstr; |
123 | u_int cachesz; |
124 | u_int32_t timeoutbase, rwvtimeoutbase, rwvtimeout; |
125 | struct { |
126 | struct i2o_param_op_results pr; |
127 | struct i2o_param_read_results prr; |
128 | union { |
129 | struct i2o_param_rbs_cache_control cc; |
130 | struct i2o_param_rbs_device_info bdi; |
131 | } p; |
132 | } __packed param; |
133 | |
134 | ld->sc_dv = self; |
135 | evreg = 0; |
136 | |
137 | /* Register us as an initiator. */ |
138 | sc->sc_ii.ii_dv = self; |
139 | sc->sc_ii.ii_intr = ld_iop_intr; |
140 | sc->sc_ii.ii_adjqparam = ld_iop_adjqparam; |
141 | sc->sc_ii.ii_flags = 0; |
142 | sc->sc_ii.ii_tid = ia->ia_tid; |
143 | iop_initiator_register(iop, &sc->sc_ii); |
144 | |
145 | /* Register another initiator to handle events from the device. */ |
146 | sc->sc_eventii.ii_dv = self; |
147 | sc->sc_eventii.ii_intr = ld_iop_intr_event; |
148 | sc->sc_eventii.ii_flags = II_NOTCTX | II_UTILITY; |
149 | sc->sc_eventii.ii_tid = ia->ia_tid; |
150 | iop_initiator_register(iop, &sc->sc_eventii); |
151 | |
152 | rv = iop_util_eventreg(iop, &sc->sc_eventii, |
153 | I2O_EVENT_GEN_EVENT_MASK_MODIFIED | |
154 | I2O_EVENT_GEN_DEVICE_RESET | |
155 | I2O_EVENT_GEN_STATE_CHANGE | |
156 | I2O_EVENT_GEN_GENERAL_WARNING); |
157 | if (rv != 0) { |
158 | aprint_error_dev(self, "unable to register for events" ); |
159 | goto bad; |
160 | } |
161 | evreg = 1; |
162 | |
163 | /* |
164 | * Start out with one queued command. The `iop' driver will adjust |
165 | * the queue parameters once we're up and running. |
166 | */ |
167 | ld->sc_maxqueuecnt = 1; |
168 | |
169 | ld->sc_maxxfer = IOP_MAX_XFER; |
170 | ld->sc_dump = ld_iop_dump; |
171 | ld->sc_flush = ld_iop_flush; |
172 | ld->sc_start = ld_iop_start; |
173 | |
174 | /* Say what the device is. */ |
175 | printf(":" ); |
176 | iop_print_ident(iop, ia->ia_tid); |
177 | |
178 | /* |
179 | * Claim the device so that we don't get any nasty surprises. Allow |
180 | * failure. |
181 | */ |
182 | rv = iop_util_claim(iop, &sc->sc_ii, 0, |
183 | I2O_UTIL_CLAIM_CAPACITY_SENSITIVE | |
184 | I2O_UTIL_CLAIM_NO_PEER_SERVICE | |
185 | I2O_UTIL_CLAIM_NO_MANAGEMENT_SERVICE | |
186 | I2O_UTIL_CLAIM_PRIMARY_USER); |
187 | sc->sc_flags = rv ? 0 : LD_IOP_CLAIMED; |
188 | |
189 | rv = iop_field_get_all(iop, ia->ia_tid, I2O_PARAM_RBS_DEVICE_INFO, |
190 | ¶m, sizeof(param), NULL); |
191 | if (rv != 0) |
192 | goto bad; |
193 | |
194 | ld->sc_secsize = le32toh(param.p.bdi.blocksize); |
195 | ld->sc_secperunit = (int) |
196 | (le64toh(param.p.bdi.capacity) / ld->sc_secsize); |
197 | |
198 | switch (param.p.bdi.type) { |
199 | case I2O_RBS_TYPE_DIRECT: |
200 | typestr = "direct access" ; |
201 | enable = 1; |
202 | break; |
203 | case I2O_RBS_TYPE_WORM: |
204 | typestr = "WORM" ; |
205 | enable = 0; |
206 | break; |
207 | case I2O_RBS_TYPE_CDROM: |
208 | typestr = "CD-ROM" ; |
209 | enable = 0; |
210 | break; |
211 | case I2O_RBS_TYPE_OPTICAL: |
212 | typestr = "optical" ; |
213 | enable = 0; |
214 | break; |
215 | default: |
216 | typestr = "unknown" ; |
217 | enable = 0; |
218 | break; |
219 | } |
220 | |
221 | if ((le32toh(param.p.bdi.capabilities) & I2O_RBS_CAP_REMOVABLE_MEDIA) |
222 | != 0) { |
223 | /* ld->sc_flags = LDF_REMOVABLE; */ |
224 | fixedstr = "removable" ; |
225 | enable = 0; |
226 | } else |
227 | fixedstr = "fixed" ; |
228 | |
229 | printf(" %s, %s" , typestr, fixedstr); |
230 | |
231 | /* |
232 | * Determine if the device has an private cache. If so, print the |
233 | * cache size. Even if the device doesn't appear to have a cache, |
234 | * we perform a flush at shutdown. |
235 | */ |
236 | rv = iop_field_get_all(iop, ia->ia_tid, I2O_PARAM_RBS_CACHE_CONTROL, |
237 | ¶m, sizeof(param), NULL); |
238 | if (rv != 0) |
239 | goto bad; |
240 | |
241 | if ((cachesz = le32toh(param.p.cc.totalcachesize)) != 0) |
242 | printf(", %dkB cache" , cachesz >> 10); |
243 | |
244 | printf("\n" ); |
245 | |
246 | /* |
247 | * Configure the DDM's timeout functions to time out all commands |
248 | * after 30 seconds. |
249 | */ |
250 | timeoutbase = htole32(LD_IOP_TIMEOUT * 1000); |
251 | rwvtimeoutbase = htole32(LD_IOP_TIMEOUT * 1000); |
252 | rwvtimeout = 0; |
253 | |
254 | iop_field_set(iop, ia->ia_tid, I2O_PARAM_RBS_OPERATION, |
255 | &timeoutbase, sizeof(timeoutbase), |
256 | I2O_PARAM_RBS_OPERATION_timeoutbase); |
257 | iop_field_set(iop, ia->ia_tid, I2O_PARAM_RBS_OPERATION, |
258 | &rwvtimeoutbase, sizeof(rwvtimeoutbase), |
259 | I2O_PARAM_RBS_OPERATION_rwvtimeoutbase); |
260 | iop_field_set(iop, ia->ia_tid, I2O_PARAM_RBS_OPERATION, |
261 | &rwvtimeout, sizeof(rwvtimeout), |
262 | I2O_PARAM_RBS_OPERATION_rwvtimeoutbase); |
263 | |
264 | if (enable) |
265 | ld->sc_flags |= LDF_ENABLED; |
266 | else |
267 | aprint_error_dev(self, "device not yet supported\n" ); |
268 | |
269 | ldattach(ld, BUFQ_DISK_DEFAULT_STRAT); |
270 | return; |
271 | |
272 | bad: |
273 | ld_iop_unconfig(sc, evreg); |
274 | } |
275 | |
276 | static void |
277 | ld_iop_unconfig(struct ld_iop_softc *sc, int evreg) |
278 | { |
279 | struct iop_softc *iop; |
280 | |
281 | iop = device_private(device_parent(sc->sc_ld.sc_dv)); |
282 | |
283 | if ((sc->sc_flags & LD_IOP_CLAIMED) != 0) |
284 | iop_util_claim(iop, &sc->sc_ii, 1, |
285 | I2O_UTIL_CLAIM_PRIMARY_USER); |
286 | |
287 | if (evreg) { |
288 | /* |
289 | * Mask off events, and wait up to 5 seconds for a reply. |
290 | * Note that some adapters won't reply to this (XXX We |
291 | * should check the event capabilities). |
292 | */ |
293 | mutex_spin_enter(&iop->sc_intrlock); |
294 | sc->sc_flags &= ~LD_IOP_NEW_EVTMASK; |
295 | mutex_spin_exit(&iop->sc_intrlock); |
296 | |
297 | iop_util_eventreg(iop, &sc->sc_eventii, |
298 | I2O_EVENT_GEN_EVENT_MASK_MODIFIED); |
299 | |
300 | mutex_spin_enter(&iop->sc_intrlock); |
301 | if ((sc->sc_flags & LD_IOP_NEW_EVTMASK) == 0) |
302 | cv_timedwait(&sc->sc_eventii.ii_cv, |
303 | &iop->sc_intrlock, hz * 5); |
304 | mutex_spin_exit(&iop->sc_intrlock); |
305 | } |
306 | |
307 | iop_initiator_unregister(iop, &sc->sc_eventii); |
308 | iop_initiator_unregister(iop, &sc->sc_ii); |
309 | } |
310 | |
311 | static int |
312 | ld_iop_detach(device_t self, int flags) |
313 | { |
314 | struct ld_iop_softc *sc; |
315 | struct iop_softc *iop; |
316 | int rv; |
317 | |
318 | sc = device_private(self); |
319 | iop = device_private(device_parent(self)); |
320 | |
321 | if ((rv = ldbegindetach(&sc->sc_ld, flags)) != 0) |
322 | return (rv); |
323 | |
324 | /* |
325 | * Abort any requests queued with the IOP, but allow requests that |
326 | * are already in progress to complete. |
327 | */ |
328 | if ((sc->sc_ld.sc_flags & LDF_ENABLED) != 0) |
329 | iop_util_abort(iop, &sc->sc_ii, 0, 0, |
330 | I2O_UTIL_ABORT_WILD | I2O_UTIL_ABORT_CLEAN); |
331 | |
332 | ldenddetach(&sc->sc_ld); |
333 | |
334 | /* Un-claim the target, and un-register our initiators. */ |
335 | if ((sc->sc_ld.sc_flags & LDF_ENABLED) != 0) |
336 | ld_iop_unconfig(sc, 1); |
337 | |
338 | return (0); |
339 | } |
340 | |
341 | static int |
342 | ld_iop_start(struct ld_softc *ld, struct buf *bp) |
343 | { |
344 | struct iop_msg *im; |
345 | struct iop_softc *iop; |
346 | struct ld_iop_softc *sc; |
347 | struct i2o_rbs_block_read *mf; |
348 | u_int rv, flags, write; |
349 | u_int64_t ba; |
350 | u_int32_t mb[IOP_MAX_MSG_SIZE / sizeof(u_int32_t)]; |
351 | |
352 | sc = device_private(ld->sc_dv); |
353 | iop = device_private(device_parent(ld->sc_dv)); |
354 | |
355 | im = iop_msg_alloc(iop, 0); |
356 | im->im_dvcontext = bp; |
357 | |
358 | write = ((bp->b_flags & B_READ) == 0); |
359 | ba = (u_int64_t)bp->b_rawblkno * ld->sc_secsize; |
360 | |
361 | /* |
362 | * Write through the cache when performing synchronous writes. When |
363 | * performing a read, we don't request that the DDM cache the data, |
364 | * as there's little advantage to it. |
365 | */ |
366 | if (write) { |
367 | if ((bp->b_flags & B_ASYNC) == 0) |
368 | flags = I2O_RBS_BLOCK_WRITE_CACHE_WT; |
369 | else |
370 | flags = I2O_RBS_BLOCK_WRITE_CACHE_WB; |
371 | } else |
372 | flags = 0; |
373 | |
374 | /* |
375 | * Fill the message frame. We can use the block_read structure for |
376 | * both reads and writes, as it's almost identical to the |
377 | * block_write structure. |
378 | */ |
379 | mf = (struct i2o_rbs_block_read *)mb; |
380 | mf->msgflags = I2O_MSGFLAGS(i2o_rbs_block_read); |
381 | mf->msgfunc = I2O_MSGFUNC(sc->sc_ii.ii_tid, |
382 | write ? I2O_RBS_BLOCK_WRITE : I2O_RBS_BLOCK_READ); |
383 | mf->msgictx = sc->sc_ii.ii_ictx; |
384 | mf->msgtctx = im->im_tctx; |
385 | mf->flags = flags | (1 << 16); /* flags & time multiplier */ |
386 | mf->datasize = bp->b_bcount; |
387 | mf->lowoffset = (u_int32_t)ba; |
388 | mf->highoffset = (u_int32_t)(ba >> 32); |
389 | |
390 | /* Map the data transfer and enqueue the command. */ |
391 | rv = iop_msg_map_bio(iop, im, mb, bp->b_data, bp->b_bcount, write); |
392 | if (rv == 0) { |
393 | if ((rv = iop_post(iop, mb)) != 0) { |
394 | iop_msg_unmap(iop, im); |
395 | iop_msg_free(iop, im); |
396 | } |
397 | } |
398 | return (rv); |
399 | } |
400 | |
401 | static int |
402 | ld_iop_dump(struct ld_softc *ld, void *data, int blkno, int blkcnt) |
403 | { |
404 | struct iop_msg *im; |
405 | struct iop_softc *iop; |
406 | struct ld_iop_softc *sc; |
407 | struct i2o_rbs_block_write *mf; |
408 | int rv, bcount; |
409 | u_int64_t ba; |
410 | u_int32_t mb[IOP_MAX_MSG_SIZE / sizeof(u_int32_t)]; |
411 | |
412 | sc = device_private(ld->sc_dv); |
413 | iop = device_private(device_parent(ld->sc_dv)); |
414 | bcount = blkcnt * ld->sc_secsize; |
415 | ba = (u_int64_t)blkno * ld->sc_secsize; |
416 | im = iop_msg_alloc(iop, IM_POLL); |
417 | |
418 | mf = (struct i2o_rbs_block_write *)mb; |
419 | mf->msgflags = I2O_MSGFLAGS(i2o_rbs_block_write); |
420 | mf->msgfunc = I2O_MSGFUNC(sc->sc_ii.ii_tid, I2O_RBS_BLOCK_WRITE); |
421 | mf->msgictx = sc->sc_ii.ii_ictx; |
422 | mf->msgtctx = im->im_tctx; |
423 | mf->flags = I2O_RBS_BLOCK_WRITE_CACHE_WT | (1 << 16); |
424 | mf->datasize = bcount; |
425 | mf->lowoffset = (u_int32_t)ba; |
426 | mf->highoffset = (u_int32_t)(ba >> 32); |
427 | |
428 | if ((rv = iop_msg_map(iop, im, mb, data, bcount, 1, NULL)) != 0) { |
429 | iop_msg_free(iop, im); |
430 | return (rv); |
431 | } |
432 | |
433 | rv = iop_msg_post(iop, im, mb, LD_IOP_TIMEOUT * 2); |
434 | iop_msg_unmap(iop, im); |
435 | iop_msg_free(iop, im); |
436 | return (rv); |
437 | } |
438 | |
439 | static int |
440 | ld_iop_flush(struct ld_softc *ld, int flags) |
441 | { |
442 | struct iop_msg *im; |
443 | struct iop_softc *iop; |
444 | struct ld_iop_softc *sc; |
445 | struct i2o_rbs_cache_flush mf; |
446 | int rv; |
447 | |
448 | sc = device_private(ld->sc_dv); |
449 | iop = device_private(device_parent(ld->sc_dv)); |
450 | im = iop_msg_alloc(iop, IM_WAIT); |
451 | |
452 | mf.msgflags = I2O_MSGFLAGS(i2o_rbs_cache_flush); |
453 | mf.msgfunc = I2O_MSGFUNC(sc->sc_ii.ii_tid, I2O_RBS_CACHE_FLUSH); |
454 | mf.msgictx = sc->sc_ii.ii_ictx; |
455 | mf.msgtctx = im->im_tctx; |
456 | mf.flags = 1 << 16; /* time multiplier */ |
457 | |
458 | /* Ancient disks will return an error here. */ |
459 | rv = iop_msg_post(iop, im, &mf, LD_IOP_TIMEOUT * 2); |
460 | iop_msg_free(iop, im); |
461 | return (rv); |
462 | } |
463 | |
464 | void |
465 | ld_iop_intr(device_t dv, struct iop_msg *im, void *reply) |
466 | { |
467 | struct i2o_rbs_reply *rb; |
468 | struct buf *bp; |
469 | struct ld_iop_softc *sc; |
470 | struct iop_softc *iop; |
471 | int err, detail; |
472 | const char *errstr; |
473 | |
474 | rb = reply; |
475 | bp = im->im_dvcontext; |
476 | sc = device_private(dv); |
477 | iop = device_private(device_parent(dv)); |
478 | |
479 | err = ((rb->msgflags & I2O_MSGFLAGS_FAIL) != 0); |
480 | |
481 | if (!err && rb->reqstatus != I2O_STATUS_SUCCESS) { |
482 | detail = le16toh(rb->detail); |
483 | if (detail >= __arraycount(ld_iop_errors)) |
484 | errstr = "<unknown>" ; |
485 | else |
486 | errstr = ld_iop_errors[detail]; |
487 | aprint_error_dev(dv, "error 0x%04x: %s\n" , detail, errstr); |
488 | err = 1; |
489 | } |
490 | |
491 | if (err) { |
492 | bp->b_error = EIO; |
493 | bp->b_resid = bp->b_bcount; |
494 | } else |
495 | bp->b_resid = bp->b_bcount - le32toh(rb->transfercount); |
496 | |
497 | iop_msg_unmap(iop, im); |
498 | iop_msg_free(iop, im); |
499 | lddone(&sc->sc_ld, bp); |
500 | } |
501 | |
502 | static void |
503 | ld_iop_intr_event(device_t dv, struct iop_msg *im, void *reply) |
504 | { |
505 | struct i2o_util_event_register_reply *rb; |
506 | struct ld_iop_softc *sc; |
507 | struct iop_softc *iop; |
508 | u_int event; |
509 | |
510 | rb = reply; |
511 | |
512 | if ((rb->msgflags & I2O_MSGFLAGS_FAIL) != 0) |
513 | return; |
514 | |
515 | event = le32toh(rb->event); |
516 | sc = device_private(dv); |
517 | |
518 | if (event == I2O_EVENT_GEN_EVENT_MASK_MODIFIED) { |
519 | iop = device_private(device_parent(dv)); |
520 | mutex_spin_enter(&iop->sc_intrlock); |
521 | sc->sc_flags |= LD_IOP_NEW_EVTMASK; |
522 | cv_broadcast(&sc->sc_eventii.ii_cv); |
523 | mutex_spin_exit(&iop->sc_intrlock); |
524 | return; |
525 | } |
526 | |
527 | printf("%s: event 0x%08x received\n" , device_xname(dv), event); |
528 | } |
529 | |
530 | static void |
531 | ld_iop_adjqparam(device_t dv, int mpi) |
532 | { |
533 | struct ld_iop_softc *sc = device_private(dv); |
534 | struct iop_softc *iop = device_private(device_parent(dv)); |
535 | struct ld_softc *ld = &sc->sc_ld; |
536 | |
537 | /* |
538 | * AMI controllers seem to loose the plot if you hand off lots of |
539 | * queued commands. |
540 | */ |
541 | if (le16toh(I2O_ORG_AMI) == iop->sc_status.orgid && mpi > 64) |
542 | mpi = 64; |
543 | |
544 | ldadjqparam(ld, mpi); |
545 | } |
546 | |