ieee1394: sbp2: recheck node generation in sbp2_update
[deliverable/linux.git] / drivers / ieee1394 / sbp2.c
1 /*
2 * sbp2.c - SBP-2 protocol driver for IEEE-1394
3 *
4 * Copyright (C) 2000 James Goodwin, Filanet Corporation (www.filanet.com)
5 * jamesg@filanet.com (JSG)
6 *
7 * Copyright (C) 2003 Ben Collins <bcollins@debian.org>
8 *
9 * This program is free software; you can redistribute it and/or modify
10 * it under the terms of the GNU General Public License as published by
11 * the Free Software Foundation; either version 2 of the License, or
12 * (at your option) any later version.
13 *
14 * This program is distributed in the hope that it will be useful,
15 * but WITHOUT ANY WARRANTY; without even the implied warranty of
16 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
17 * GNU General Public License for more details.
18 *
19 * You should have received a copy of the GNU General Public License
20 * along with this program; if not, write to the Free Software Foundation,
21 * Inc., 59 Temple Place - Suite 330, Boston, MA 02111-1307, USA.
22 */
23
24 /*
25 * Brief Description:
26 *
27 * This driver implements the Serial Bus Protocol 2 (SBP-2) over IEEE-1394
28 * under Linux. The SBP-2 driver is implemented as an IEEE-1394 high-level
29 * driver. It also registers as a SCSI lower-level driver in order to accept
30 * SCSI commands for transport using SBP-2.
31 *
32 * You may access any attached SBP-2 storage devices as if they were SCSI
33 * devices (e.g. mount /dev/sda1, fdisk, mkfs, etc.).
34 *
35 * Current Issues:
36 *
37 * - Error Handling: SCSI aborts and bus reset requests are handled somewhat
38 * but the code needs additional debugging.
39 */
40
41 #include <linux/kernel.h>
42 #include <linux/list.h>
43 #include <linux/string.h>
44 #include <linux/stringify.h>
45 #include <linux/slab.h>
46 #include <linux/interrupt.h>
47 #include <linux/fs.h>
48 #include <linux/poll.h>
49 #include <linux/module.h>
50 #include <linux/moduleparam.h>
51 #include <linux/types.h>
52 #include <linux/delay.h>
53 #include <linux/sched.h>
54 #include <linux/blkdev.h>
55 #include <linux/smp_lock.h>
56 #include <linux/init.h>
57 #include <linux/pci.h>
58 #include <linux/wait.h>
59
60 #include <asm/current.h>
61 #include <asm/uaccess.h>
62 #include <asm/io.h>
63 #include <asm/byteorder.h>
64 #include <asm/system.h>
65 #include <asm/scatterlist.h>
66
67 #include <scsi/scsi.h>
68 #include <scsi/scsi_cmnd.h>
69 #include <scsi/scsi_dbg.h>
70 #include <scsi/scsi_device.h>
71 #include <scsi/scsi_host.h>
72
73 #include "csr1212.h"
74 #include "ieee1394.h"
75 #include "ieee1394_types.h"
76 #include "ieee1394_core.h"
77 #include "nodemgr.h"
78 #include "hosts.h"
79 #include "highlevel.h"
80 #include "ieee1394_transactions.h"
81 #include "sbp2.h"
82
83 /*
84 * Module load parameter definitions
85 */
86
87 /*
88 * Change max_speed on module load if you have a bad IEEE-1394
89 * controller that has trouble running 2KB packets at 400mb.
90 *
91 * NOTE: On certain OHCI parts I have seen short packets on async transmit
92 * (probably due to PCI latency/throughput issues with the part). You can
93 * bump down the speed if you are running into problems.
94 */
95 static int max_speed = IEEE1394_SPEED_MAX;
96 module_param(max_speed, int, 0644);
97 MODULE_PARM_DESC(max_speed, "Force max speed (3 = 800mb, 2 = 400mb, 1 = 200mb, 0 = 100mb)");
98
99 /*
100 * Set serialize_io to 1 if you'd like only one scsi command sent
101 * down to us at a time (debugging). This might be necessary for very
102 * badly behaved sbp2 devices.
103 *
104 * TODO: Make this configurable per device.
105 */
106 static int serialize_io = 1;
107 module_param(serialize_io, int, 0444);
108 MODULE_PARM_DESC(serialize_io, "Serialize I/O coming from scsi drivers (default = 1, faster = 0)");
109
110 /*
111 * Bump up max_sectors if you'd like to support very large sized
112 * transfers. Please note that some older sbp2 bridge chips are broken for
113 * transfers greater or equal to 128KB. Default is a value of 255
114 * sectors, or just under 128KB (at 512 byte sector size). I can note that
115 * the Oxsemi sbp2 chipsets have no problems supporting very large
116 * transfer sizes.
117 */
118 static int max_sectors = SBP2_MAX_SECTORS;
119 module_param(max_sectors, int, 0444);
120 MODULE_PARM_DESC(max_sectors, "Change max sectors per I/O supported (default = "
121 __stringify(SBP2_MAX_SECTORS) ")");
122
123 /*
124 * Exclusive login to sbp2 device? In most cases, the sbp2 driver should
125 * do an exclusive login, as it's generally unsafe to have two hosts
126 * talking to a single sbp2 device at the same time (filesystem coherency,
127 * etc.). If you're running an sbp2 device that supports multiple logins,
128 * and you're either running read-only filesystems or some sort of special
129 * filesystem supporting multiple hosts, e.g. OpenGFS, Oracle Cluster
130 * File System, or Lustre, then set exclusive_login to zero.
131 *
132 * So far only bridges from Oxford Semiconductor are known to support
133 * concurrent logins. Depending on firmware, four or two concurrent logins
134 * are possible on OXFW911 and newer Oxsemi bridges.
135 */
136 static int exclusive_login = 1;
137 module_param(exclusive_login, int, 0644);
138 MODULE_PARM_DESC(exclusive_login, "Exclusive login to sbp2 device (default = 1)");
139
140 /*
141 * If any of the following workarounds is required for your device to work,
142 * please submit the kernel messages logged by sbp2 to the linux1394-devel
143 * mailing list.
144 *
145 * - 128kB max transfer
146 * Limit transfer size. Necessary for some old bridges.
147 *
148 * - 36 byte inquiry
149 * When scsi_mod probes the device, let the inquiry command look like that
150 * from MS Windows.
151 *
152 * - skip mode page 8
153 * Suppress sending of mode_sense for mode page 8 if the device pretends to
154 * support the SCSI Primary Block commands instead of Reduced Block Commands.
155 *
156 * - fix capacity
157 * Tell sd_mod to correct the last sector number reported by read_capacity.
158 * Avoids access beyond actual disk limits on devices with an off-by-one bug.
159 * Don't use this with devices which don't have this bug.
160 *
161 * - override internal blacklist
162 * Instead of adding to the built-in blacklist, use only the workarounds
163 * specified in the module load parameter.
164 * Useful if a blacklist entry interfered with a non-broken device.
165 */
166 static int sbp2_default_workarounds;
167 module_param_named(workarounds, sbp2_default_workarounds, int, 0644);
168 MODULE_PARM_DESC(workarounds, "Work around device bugs (default = 0"
169 ", 128kB max transfer = " __stringify(SBP2_WORKAROUND_128K_MAX_TRANS)
170 ", 36 byte inquiry = " __stringify(SBP2_WORKAROUND_INQUIRY_36)
171 ", skip mode page 8 = " __stringify(SBP2_WORKAROUND_MODE_SENSE_8)
172 ", fix capacity = " __stringify(SBP2_WORKAROUND_FIX_CAPACITY)
173 ", override internal blacklist = " __stringify(SBP2_WORKAROUND_OVERRIDE)
174 ", or a combination)");
175
176 /*
177 * Export information about protocols/devices supported by this driver.
178 */
179 static struct ieee1394_device_id sbp2_id_table[] = {
180 {
181 .match_flags = IEEE1394_MATCH_SPECIFIER_ID | IEEE1394_MATCH_VERSION,
182 .specifier_id = SBP2_UNIT_SPEC_ID_ENTRY & 0xffffff,
183 .version = SBP2_SW_VERSION_ENTRY & 0xffffff},
184 {}
185 };
186
187 MODULE_DEVICE_TABLE(ieee1394, sbp2_id_table);
188
189 /*
190 * Debug levels, configured via kernel config, or enable here.
191 */
192
193 #define CONFIG_IEEE1394_SBP2_DEBUG 0
194 /* #define CONFIG_IEEE1394_SBP2_DEBUG_ORBS */
195 /* #define CONFIG_IEEE1394_SBP2_DEBUG_DMA */
196 /* #define CONFIG_IEEE1394_SBP2_DEBUG 1 */
197 /* #define CONFIG_IEEE1394_SBP2_DEBUG 2 */
198 /* #define CONFIG_IEEE1394_SBP2_PACKET_DUMP */
199
200 #ifdef CONFIG_IEEE1394_SBP2_DEBUG_ORBS
201 #define SBP2_ORB_DEBUG(fmt, args...) HPSB_ERR("sbp2(%s): "fmt, __FUNCTION__, ## args)
202 static u32 global_outstanding_command_orbs = 0;
203 #define outstanding_orb_incr global_outstanding_command_orbs++
204 #define outstanding_orb_decr global_outstanding_command_orbs--
205 #else
206 #define SBP2_ORB_DEBUG(fmt, args...) do {} while (0)
207 #define outstanding_orb_incr do {} while (0)
208 #define outstanding_orb_decr do {} while (0)
209 #endif
210
211 #ifdef CONFIG_IEEE1394_SBP2_DEBUG_DMA
212 #define SBP2_DMA_ALLOC(fmt, args...) \
213 HPSB_ERR("sbp2(%s)alloc(%d): "fmt, __FUNCTION__, \
214 ++global_outstanding_dmas, ## args)
215 #define SBP2_DMA_FREE(fmt, args...) \
216 HPSB_ERR("sbp2(%s)free(%d): "fmt, __FUNCTION__, \
217 --global_outstanding_dmas, ## args)
218 static u32 global_outstanding_dmas = 0;
219 #else
220 #define SBP2_DMA_ALLOC(fmt, args...) do {} while (0)
221 #define SBP2_DMA_FREE(fmt, args...) do {} while (0)
222 #endif
223
224 #if CONFIG_IEEE1394_SBP2_DEBUG >= 2
225 #define SBP2_DEBUG(fmt, args...) HPSB_ERR("sbp2: "fmt, ## args)
226 #define SBP2_INFO(fmt, args...) HPSB_ERR("sbp2: "fmt, ## args)
227 #define SBP2_NOTICE(fmt, args...) HPSB_ERR("sbp2: "fmt, ## args)
228 #define SBP2_WARN(fmt, args...) HPSB_ERR("sbp2: "fmt, ## args)
229 #elif CONFIG_IEEE1394_SBP2_DEBUG == 1
230 #define SBP2_DEBUG(fmt, args...) HPSB_DEBUG("sbp2: "fmt, ## args)
231 #define SBP2_INFO(fmt, args...) HPSB_INFO("sbp2: "fmt, ## args)
232 #define SBP2_NOTICE(fmt, args...) HPSB_NOTICE("sbp2: "fmt, ## args)
233 #define SBP2_WARN(fmt, args...) HPSB_WARN("sbp2: "fmt, ## args)
234 #else
235 #define SBP2_DEBUG(fmt, args...) do {} while (0)
236 #define SBP2_INFO(fmt, args...) HPSB_INFO("sbp2: "fmt, ## args)
237 #define SBP2_NOTICE(fmt, args...) HPSB_NOTICE("sbp2: "fmt, ## args)
238 #define SBP2_WARN(fmt, args...) HPSB_WARN("sbp2: "fmt, ## args)
239 #endif
240
241 #define SBP2_ERR(fmt, args...) HPSB_ERR("sbp2: "fmt, ## args)
242 #define SBP2_DEBUG_ENTER() SBP2_DEBUG("%s", __FUNCTION__)
243
244 /*
245 * Globals
246 */
247
248 static void sbp2scsi_complete_all_commands(struct scsi_id_instance_data *scsi_id,
249 u32 status);
250
251 static void sbp2scsi_complete_command(struct scsi_id_instance_data *scsi_id,
252 u32 scsi_status, struct scsi_cmnd *SCpnt,
253 void (*done)(struct scsi_cmnd *));
254
255 static struct scsi_host_template scsi_driver_template;
256
257 static const u8 sbp2_speedto_max_payload[] = { 0x7, 0x8, 0x9, 0xA, 0xB, 0xC };
258
259 static void sbp2_host_reset(struct hpsb_host *host);
260
261 static int sbp2_probe(struct device *dev);
262 static int sbp2_remove(struct device *dev);
263 static int sbp2_update(struct unit_directory *ud);
264
265 static struct hpsb_highlevel sbp2_highlevel = {
266 .name = SBP2_DEVICE_NAME,
267 .host_reset = sbp2_host_reset,
268 };
269
270 static struct hpsb_address_ops sbp2_ops = {
271 .write = sbp2_handle_status_write
272 };
273
274 #ifdef CONFIG_IEEE1394_SBP2_PHYS_DMA
275 static struct hpsb_address_ops sbp2_physdma_ops = {
276 .read = sbp2_handle_physdma_read,
277 .write = sbp2_handle_physdma_write,
278 };
279 #endif
280
281 static struct hpsb_protocol_driver sbp2_driver = {
282 .name = "SBP2 Driver",
283 .id_table = sbp2_id_table,
284 .update = sbp2_update,
285 .driver = {
286 .name = SBP2_DEVICE_NAME,
287 .bus = &ieee1394_bus_type,
288 .probe = sbp2_probe,
289 .remove = sbp2_remove,
290 },
291 };
292
293 /*
294 * List of devices with known bugs.
295 *
296 * The firmware_revision field, masked with 0xffff00, is the best indicator
297 * for the type of bridge chip of a device. It yields a few false positives
298 * but this did not break correctly behaving devices so far.
299 */
300 static const struct {
301 u32 firmware_revision;
302 u32 model_id;
303 unsigned workarounds;
304 } sbp2_workarounds_table[] = {
305 /* DViCO Momobay CX-1 with TSB42AA9 bridge */ {
306 .firmware_revision = 0x002800,
307 .model_id = 0x001010,
308 .workarounds = SBP2_WORKAROUND_INQUIRY_36 |
309 SBP2_WORKAROUND_MODE_SENSE_8,
310 },
311 /* Initio bridges, actually only needed for some older ones */ {
312 .firmware_revision = 0x000200,
313 .workarounds = SBP2_WORKAROUND_INQUIRY_36,
314 },
315 /* Symbios bridge */ {
316 .firmware_revision = 0xa0b800,
317 .workarounds = SBP2_WORKAROUND_128K_MAX_TRANS,
318 },
319 /*
320 * Note about the following Apple iPod blacklist entries:
321 *
322 * There are iPods (2nd gen, 3rd gen) with model_id==0. Since our
323 * matching logic treats 0 as a wildcard, we cannot match this ID
324 * without rewriting the matching routine. Fortunately these iPods
325 * do not feature the read_capacity bug according to one report.
326 * Read_capacity behaviour as well as model_id could change due to
327 * Apple-supplied firmware updates though.
328 */
329 /* iPod 4th generation */ {
330 .firmware_revision = 0x0a2700,
331 .model_id = 0x000021,
332 .workarounds = SBP2_WORKAROUND_FIX_CAPACITY,
333 },
334 /* iPod mini */ {
335 .firmware_revision = 0x0a2700,
336 .model_id = 0x000023,
337 .workarounds = SBP2_WORKAROUND_FIX_CAPACITY,
338 },
339 /* iPod Photo */ {
340 .firmware_revision = 0x0a2700,
341 .model_id = 0x00007e,
342 .workarounds = SBP2_WORKAROUND_FIX_CAPACITY,
343 }
344 };
345
346 /**************************************
347 * General utility functions
348 **************************************/
349
350 #ifndef __BIG_ENDIAN
351 /*
352 * Converts a buffer from be32 to cpu byte ordering. Length is in bytes.
353 */
354 static inline void sbp2util_be32_to_cpu_buffer(void *buffer, int length)
355 {
356 u32 *temp = buffer;
357
358 for (length = (length >> 2); length--; )
359 temp[length] = be32_to_cpu(temp[length]);
360
361 return;
362 }
363
364 /*
365 * Converts a buffer from cpu to be32 byte ordering. Length is in bytes.
366 */
367 static inline void sbp2util_cpu_to_be32_buffer(void *buffer, int length)
368 {
369 u32 *temp = buffer;
370
371 for (length = (length >> 2); length--; )
372 temp[length] = cpu_to_be32(temp[length]);
373
374 return;
375 }
376 #else /* BIG_ENDIAN */
377 /* Why waste the cpu cycles? */
378 #define sbp2util_be32_to_cpu_buffer(x,y) do {} while (0)
379 #define sbp2util_cpu_to_be32_buffer(x,y) do {} while (0)
380 #endif
381
382 #ifdef CONFIG_IEEE1394_SBP2_PACKET_DUMP
383 /*
384 * Debug packet dump routine. Length is in bytes.
385 */
386 static void sbp2util_packet_dump(void *buffer, int length, char *dump_name,
387 u32 dump_phys_addr)
388 {
389 int i;
390 unsigned char *dump = buffer;
391
392 if (!dump || !length || !dump_name)
393 return;
394
395 if (dump_phys_addr)
396 printk("[%s, 0x%x]", dump_name, dump_phys_addr);
397 else
398 printk("[%s]", dump_name);
399 for (i = 0; i < length; i++) {
400 if (i > 0x3f) {
401 printk("\n ...");
402 break;
403 }
404 if ((i & 0x3) == 0)
405 printk(" ");
406 if ((i & 0xf) == 0)
407 printk("\n ");
408 printk("%02x ", (int)dump[i]);
409 }
410 printk("\n");
411
412 return;
413 }
414 #else
415 #define sbp2util_packet_dump(w,x,y,z) do {} while (0)
416 #endif
417
418 static DECLARE_WAIT_QUEUE_HEAD(access_wq);
419
420 /*
421 * Waits for completion of an SBP-2 access request.
422 * Returns nonzero if timed out or prematurely interrupted.
423 */
424 static int sbp2util_access_timeout(struct scsi_id_instance_data *scsi_id,
425 int timeout)
426 {
427 long leftover = wait_event_interruptible_timeout(
428 access_wq, scsi_id->access_complete, timeout);
429
430 scsi_id->access_complete = 0;
431 return leftover <= 0;
432 }
433
434 /* Frees an allocated packet */
435 static void sbp2_free_packet(struct hpsb_packet *packet)
436 {
437 hpsb_free_tlabel(packet);
438 hpsb_free_packet(packet);
439 }
440
441 /* This is much like hpsb_node_write(), except it ignores the response
442 * subaction and returns immediately. Can be used from interrupts.
443 */
444 static int sbp2util_node_write_no_wait(struct node_entry *ne, u64 addr,
445 quadlet_t *buffer, size_t length)
446 {
447 struct hpsb_packet *packet;
448
449 packet = hpsb_make_writepacket(ne->host, ne->nodeid,
450 addr, buffer, length);
451 if (!packet)
452 return -ENOMEM;
453
454 hpsb_set_packet_complete_task(packet,
455 (void (*)(void *))sbp2_free_packet,
456 packet);
457
458 hpsb_node_fill_packet(ne, packet);
459
460 if (hpsb_send_packet(packet) < 0) {
461 sbp2_free_packet(packet);
462 return -EIO;
463 }
464
465 return 0;
466 }
467
468 static void sbp2util_notify_fetch_agent(struct scsi_id_instance_data *scsi_id,
469 u64 offset, quadlet_t *data, size_t len)
470 {
471 /*
472 * There is a small window after a bus reset within which the node
473 * entry's generation is current but the reconnect wasn't completed.
474 */
475 if (atomic_read(&scsi_id->unfinished_reset))
476 return;
477
478 if (hpsb_node_write(scsi_id->ne,
479 scsi_id->sbp2_command_block_agent_addr + offset,
480 data, len))
481 SBP2_ERR("sbp2util_notify_fetch_agent failed.");
482 /*
483 * Now accept new SCSI commands, unless a bus reset happended during
484 * hpsb_node_write.
485 */
486 if (!atomic_read(&scsi_id->unfinished_reset))
487 scsi_unblock_requests(scsi_id->scsi_host);
488 }
489
490 static void sbp2util_write_orb_pointer(void *p)
491 {
492 quadlet_t data[2];
493
494 data[0] = ORB_SET_NODE_ID(
495 ((struct scsi_id_instance_data *)p)->hi->host->node_id);
496 data[1] = ((struct scsi_id_instance_data *)p)->last_orb_dma;
497 sbp2util_cpu_to_be32_buffer(data, 8);
498 sbp2util_notify_fetch_agent(p, SBP2_ORB_POINTER_OFFSET, data, 8);
499 }
500
501 static void sbp2util_write_doorbell(void *p)
502 {
503 sbp2util_notify_fetch_agent(p, SBP2_DOORBELL_OFFSET, NULL, 4);
504 }
505
506 /*
507 * This function is called to create a pool of command orbs used for
508 * command processing. It is called when a new sbp2 device is detected.
509 */
510 static int sbp2util_create_command_orb_pool(struct scsi_id_instance_data *scsi_id)
511 {
512 struct sbp2scsi_host_info *hi = scsi_id->hi;
513 int i;
514 unsigned long flags, orbs;
515 struct sbp2_command_info *command;
516
517 orbs = serialize_io ? 2 : SBP2_MAX_CMDS;
518
519 spin_lock_irqsave(&scsi_id->sbp2_command_orb_lock, flags);
520 for (i = 0; i < orbs; i++) {
521 command = kzalloc(sizeof(*command), GFP_ATOMIC);
522 if (!command) {
523 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock,
524 flags);
525 return -ENOMEM;
526 }
527 command->command_orb_dma =
528 pci_map_single(hi->host->pdev, &command->command_orb,
529 sizeof(struct sbp2_command_orb),
530 PCI_DMA_TODEVICE);
531 SBP2_DMA_ALLOC("single command orb DMA");
532 command->sge_dma =
533 pci_map_single(hi->host->pdev,
534 &command->scatter_gather_element,
535 sizeof(command->scatter_gather_element),
536 PCI_DMA_BIDIRECTIONAL);
537 SBP2_DMA_ALLOC("scatter_gather_element");
538 INIT_LIST_HEAD(&command->list);
539 list_add_tail(&command->list, &scsi_id->sbp2_command_orb_completed);
540 }
541 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
542 return 0;
543 }
544
545 /*
546 * This function is called to delete a pool of command orbs.
547 */
548 static void sbp2util_remove_command_orb_pool(struct scsi_id_instance_data *scsi_id)
549 {
550 struct hpsb_host *host = scsi_id->hi->host;
551 struct list_head *lh, *next;
552 struct sbp2_command_info *command;
553 unsigned long flags;
554
555 spin_lock_irqsave(&scsi_id->sbp2_command_orb_lock, flags);
556 if (!list_empty(&scsi_id->sbp2_command_orb_completed)) {
557 list_for_each_safe(lh, next, &scsi_id->sbp2_command_orb_completed) {
558 command = list_entry(lh, struct sbp2_command_info, list);
559
560 /* Release our generic DMA's */
561 pci_unmap_single(host->pdev, command->command_orb_dma,
562 sizeof(struct sbp2_command_orb),
563 PCI_DMA_TODEVICE);
564 SBP2_DMA_FREE("single command orb DMA");
565 pci_unmap_single(host->pdev, command->sge_dma,
566 sizeof(command->scatter_gather_element),
567 PCI_DMA_BIDIRECTIONAL);
568 SBP2_DMA_FREE("scatter_gather_element");
569
570 kfree(command);
571 }
572 }
573 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
574 return;
575 }
576
577 /*
578 * This function finds the sbp2_command for a given outstanding command
579 * orb.Only looks at the inuse list.
580 */
581 static struct sbp2_command_info *sbp2util_find_command_for_orb(
582 struct scsi_id_instance_data *scsi_id, dma_addr_t orb)
583 {
584 struct sbp2_command_info *command;
585 unsigned long flags;
586
587 spin_lock_irqsave(&scsi_id->sbp2_command_orb_lock, flags);
588 if (!list_empty(&scsi_id->sbp2_command_orb_inuse)) {
589 list_for_each_entry(command, &scsi_id->sbp2_command_orb_inuse, list) {
590 if (command->command_orb_dma == orb) {
591 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
592 return command;
593 }
594 }
595 }
596 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
597
598 SBP2_ORB_DEBUG("could not match command orb %x", (unsigned int)orb);
599
600 return NULL;
601 }
602
603 /*
604 * This function finds the sbp2_command for a given outstanding SCpnt.
605 * Only looks at the inuse list.
606 * Must be called with scsi_id->sbp2_command_orb_lock held.
607 */
608 static struct sbp2_command_info *sbp2util_find_command_for_SCpnt(
609 struct scsi_id_instance_data *scsi_id, void *SCpnt)
610 {
611 struct sbp2_command_info *command;
612
613 if (!list_empty(&scsi_id->sbp2_command_orb_inuse))
614 list_for_each_entry(command, &scsi_id->sbp2_command_orb_inuse, list)
615 if (command->Current_SCpnt == SCpnt)
616 return command;
617 return NULL;
618 }
619
620 /*
621 * This function allocates a command orb used to send a scsi command.
622 */
623 static struct sbp2_command_info *sbp2util_allocate_command_orb(
624 struct scsi_id_instance_data *scsi_id,
625 struct scsi_cmnd *Current_SCpnt,
626 void (*Current_done)(struct scsi_cmnd *))
627 {
628 struct list_head *lh;
629 struct sbp2_command_info *command = NULL;
630 unsigned long flags;
631
632 spin_lock_irqsave(&scsi_id->sbp2_command_orb_lock, flags);
633 if (!list_empty(&scsi_id->sbp2_command_orb_completed)) {
634 lh = scsi_id->sbp2_command_orb_completed.next;
635 list_del(lh);
636 command = list_entry(lh, struct sbp2_command_info, list);
637 command->Current_done = Current_done;
638 command->Current_SCpnt = Current_SCpnt;
639 list_add_tail(&command->list, &scsi_id->sbp2_command_orb_inuse);
640 } else {
641 SBP2_ERR("%s: no orbs available", __FUNCTION__);
642 }
643 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
644 return command;
645 }
646
647 /* Free our DMA's */
648 static void sbp2util_free_command_dma(struct sbp2_command_info *command)
649 {
650 struct scsi_id_instance_data *scsi_id =
651 (struct scsi_id_instance_data *)command->Current_SCpnt->device->host->hostdata[0];
652 struct hpsb_host *host;
653
654 if (!scsi_id) {
655 SBP2_ERR("%s: scsi_id == NULL", __FUNCTION__);
656 return;
657 }
658
659 host = scsi_id->ud->ne->host;
660
661 if (command->cmd_dma) {
662 if (command->dma_type == CMD_DMA_SINGLE) {
663 pci_unmap_single(host->pdev, command->cmd_dma,
664 command->dma_size, command->dma_dir);
665 SBP2_DMA_FREE("single bulk");
666 } else if (command->dma_type == CMD_DMA_PAGE) {
667 pci_unmap_page(host->pdev, command->cmd_dma,
668 command->dma_size, command->dma_dir);
669 SBP2_DMA_FREE("single page");
670 } /* XXX: Check for CMD_DMA_NONE bug */
671 command->dma_type = CMD_DMA_NONE;
672 command->cmd_dma = 0;
673 }
674
675 if (command->sge_buffer) {
676 pci_unmap_sg(host->pdev, command->sge_buffer,
677 command->dma_size, command->dma_dir);
678 SBP2_DMA_FREE("scatter list");
679 command->sge_buffer = NULL;
680 }
681 }
682
683 /*
684 * This function moves a command to the completed orb list.
685 * Must be called with scsi_id->sbp2_command_orb_lock held.
686 */
687 static void sbp2util_mark_command_completed(
688 struct scsi_id_instance_data *scsi_id,
689 struct sbp2_command_info *command)
690 {
691 list_del(&command->list);
692 sbp2util_free_command_dma(command);
693 list_add_tail(&command->list, &scsi_id->sbp2_command_orb_completed);
694 }
695
696 /*
697 * Is scsi_id valid? Is the 1394 node still present?
698 */
699 static inline int sbp2util_node_is_available(struct scsi_id_instance_data *scsi_id)
700 {
701 return scsi_id && scsi_id->ne && !scsi_id->ne->in_limbo;
702 }
703
704 /*********************************************
705 * IEEE-1394 core driver stack related section
706 *********************************************/
707 static struct scsi_id_instance_data *sbp2_alloc_device(struct unit_directory *ud);
708
709 static int sbp2_probe(struct device *dev)
710 {
711 struct unit_directory *ud;
712 struct scsi_id_instance_data *scsi_id;
713
714 SBP2_DEBUG_ENTER();
715
716 ud = container_of(dev, struct unit_directory, device);
717
718 /* Don't probe UD's that have the LUN flag. We'll probe the LUN(s)
719 * instead. */
720 if (ud->flags & UNIT_DIRECTORY_HAS_LUN_DIRECTORY)
721 return -ENODEV;
722
723 scsi_id = sbp2_alloc_device(ud);
724
725 if (!scsi_id)
726 return -ENOMEM;
727
728 sbp2_parse_unit_directory(scsi_id, ud);
729
730 return sbp2_start_device(scsi_id);
731 }
732
733 static int sbp2_remove(struct device *dev)
734 {
735 struct unit_directory *ud;
736 struct scsi_id_instance_data *scsi_id;
737 struct scsi_device *sdev;
738
739 SBP2_DEBUG_ENTER();
740
741 ud = container_of(dev, struct unit_directory, device);
742 scsi_id = ud->device.driver_data;
743 if (!scsi_id)
744 return 0;
745
746 if (scsi_id->scsi_host) {
747 /* Get rid of enqueued commands if there is no chance to
748 * send them. */
749 if (!sbp2util_node_is_available(scsi_id))
750 sbp2scsi_complete_all_commands(scsi_id, DID_NO_CONNECT);
751 /* scsi_remove_device() will trigger shutdown functions of SCSI
752 * highlevel drivers which would deadlock if blocked. */
753 atomic_set(&scsi_id->unfinished_reset, 0);
754 scsi_unblock_requests(scsi_id->scsi_host);
755 }
756 sdev = scsi_id->sdev;
757 if (sdev) {
758 scsi_id->sdev = NULL;
759 scsi_remove_device(sdev);
760 }
761
762 sbp2_logout_device(scsi_id);
763 sbp2_remove_device(scsi_id);
764
765 return 0;
766 }
767
768 static int sbp2_update(struct unit_directory *ud)
769 {
770 struct scsi_id_instance_data *scsi_id = ud->device.driver_data;
771
772 SBP2_DEBUG_ENTER();
773
774 if (sbp2_reconnect_device(scsi_id)) {
775
776 /*
777 * Ok, reconnect has failed. Perhaps we didn't
778 * reconnect fast enough. Try doing a regular login, but
779 * first do a logout just in case of any weirdness.
780 */
781 sbp2_logout_device(scsi_id);
782
783 if (sbp2_login_device(scsi_id)) {
784 /* Login failed too, just fail, and the backend
785 * will call our sbp2_remove for us */
786 SBP2_ERR("Failed to reconnect to sbp2 device!");
787 return -EBUSY;
788 }
789 }
790
791 /* Set max retries to something large on the device. */
792 sbp2_set_busy_timeout(scsi_id);
793
794 /* Do a SBP-2 fetch agent reset. */
795 sbp2_agent_reset(scsi_id, 1);
796
797 /* Get the max speed and packet size that we can use. */
798 sbp2_max_speed_and_size(scsi_id);
799
800 /* Complete any pending commands with busy (so they get
801 * retried) and remove them from our queue
802 */
803 sbp2scsi_complete_all_commands(scsi_id, DID_BUS_BUSY);
804
805 /* Accept new commands unless there was another bus reset in the
806 * meantime. */
807 if (hpsb_node_entry_valid(scsi_id->ne)) {
808 atomic_set(&scsi_id->unfinished_reset, 0);
809 scsi_unblock_requests(scsi_id->scsi_host);
810 }
811 return 0;
812 }
813
814 /* This functions is called by the sbp2_probe, for each new device. We now
815 * allocate one scsi host for each scsi_id (unit directory). */
816 static struct scsi_id_instance_data *sbp2_alloc_device(struct unit_directory *ud)
817 {
818 struct sbp2scsi_host_info *hi;
819 struct Scsi_Host *scsi_host = NULL;
820 struct scsi_id_instance_data *scsi_id = NULL;
821
822 SBP2_DEBUG_ENTER();
823
824 scsi_id = kzalloc(sizeof(*scsi_id), GFP_KERNEL);
825 if (!scsi_id) {
826 SBP2_ERR("failed to create scsi_id");
827 goto failed_alloc;
828 }
829
830 scsi_id->ne = ud->ne;
831 scsi_id->ud = ud;
832 scsi_id->speed_code = IEEE1394_SPEED_100;
833 scsi_id->max_payload_size = sbp2_speedto_max_payload[IEEE1394_SPEED_100];
834 scsi_id->status_fifo_addr = CSR1212_INVALID_ADDR_SPACE;
835 INIT_LIST_HEAD(&scsi_id->sbp2_command_orb_inuse);
836 INIT_LIST_HEAD(&scsi_id->sbp2_command_orb_completed);
837 INIT_LIST_HEAD(&scsi_id->scsi_list);
838 spin_lock_init(&scsi_id->sbp2_command_orb_lock);
839 atomic_set(&scsi_id->unfinished_reset, 0);
840 INIT_WORK(&scsi_id->protocol_work, NULL, NULL);
841
842 ud->device.driver_data = scsi_id;
843
844 hi = hpsb_get_hostinfo(&sbp2_highlevel, ud->ne->host);
845 if (!hi) {
846 hi = hpsb_create_hostinfo(&sbp2_highlevel, ud->ne->host, sizeof(*hi));
847 if (!hi) {
848 SBP2_ERR("failed to allocate hostinfo");
849 goto failed_alloc;
850 }
851 SBP2_DEBUG("sbp2_alloc_device: allocated hostinfo");
852 hi->host = ud->ne->host;
853 INIT_LIST_HEAD(&hi->scsi_ids);
854
855 #ifdef CONFIG_IEEE1394_SBP2_PHYS_DMA
856 /* Handle data movement if physical dma is not
857 * enabled or not supported on host controller */
858 if (!hpsb_register_addrspace(&sbp2_highlevel, ud->ne->host,
859 &sbp2_physdma_ops,
860 0x0ULL, 0xfffffffcULL)) {
861 SBP2_ERR("failed to register lower 4GB address range");
862 goto failed_alloc;
863 }
864 #endif
865 }
866
867 /* Prevent unloading of the 1394 host */
868 if (!try_module_get(hi->host->driver->owner)) {
869 SBP2_ERR("failed to get a reference on 1394 host driver");
870 goto failed_alloc;
871 }
872
873 scsi_id->hi = hi;
874
875 list_add_tail(&scsi_id->scsi_list, &hi->scsi_ids);
876
877 /* Register the status FIFO address range. We could use the same FIFO
878 * for targets at different nodes. However we need different FIFOs per
879 * target in order to support multi-unit devices.
880 * The FIFO is located out of the local host controller's physical range
881 * but, if possible, within the posted write area. Status writes will
882 * then be performed as unified transactions. This slightly reduces
883 * bandwidth usage, and some Prolific based devices seem to require it.
884 */
885 scsi_id->status_fifo_addr = hpsb_allocate_and_register_addrspace(
886 &sbp2_highlevel, ud->ne->host, &sbp2_ops,
887 sizeof(struct sbp2_status_block), sizeof(quadlet_t),
888 ud->ne->host->low_addr_space, CSR1212_ALL_SPACE_END);
889 if (scsi_id->status_fifo_addr == CSR1212_INVALID_ADDR_SPACE) {
890 SBP2_ERR("failed to allocate status FIFO address range");
891 goto failed_alloc;
892 }
893
894 /* Register our host with the SCSI stack. */
895 scsi_host = scsi_host_alloc(&scsi_driver_template,
896 sizeof(unsigned long));
897 if (!scsi_host) {
898 SBP2_ERR("failed to register scsi host");
899 goto failed_alloc;
900 }
901
902 scsi_host->hostdata[0] = (unsigned long)scsi_id;
903
904 if (!scsi_add_host(scsi_host, &ud->device)) {
905 scsi_id->scsi_host = scsi_host;
906 return scsi_id;
907 }
908
909 SBP2_ERR("failed to add scsi host");
910 scsi_host_put(scsi_host);
911
912 failed_alloc:
913 sbp2_remove_device(scsi_id);
914 return NULL;
915 }
916
917 static void sbp2_host_reset(struct hpsb_host *host)
918 {
919 struct sbp2scsi_host_info *hi;
920 struct scsi_id_instance_data *scsi_id;
921
922 hi = hpsb_get_hostinfo(&sbp2_highlevel, host);
923
924 if (hi) {
925 list_for_each_entry(scsi_id, &hi->scsi_ids, scsi_list) {
926 atomic_set(&scsi_id->unfinished_reset, 1);
927 scsi_block_requests(scsi_id->scsi_host);
928 }
929 }
930 }
931
932 /*
933 * This function is where we first pull the node unique ids, and then
934 * allocate memory and register a SBP-2 device.
935 */
936 static int sbp2_start_device(struct scsi_id_instance_data *scsi_id)
937 {
938 struct sbp2scsi_host_info *hi = scsi_id->hi;
939 int error;
940
941 SBP2_DEBUG_ENTER();
942
943 /* Login FIFO DMA */
944 scsi_id->login_response =
945 pci_alloc_consistent(hi->host->pdev,
946 sizeof(struct sbp2_login_response),
947 &scsi_id->login_response_dma);
948 if (!scsi_id->login_response)
949 goto alloc_fail;
950 SBP2_DMA_ALLOC("consistent DMA region for login FIFO");
951
952 /* Query logins ORB DMA */
953 scsi_id->query_logins_orb =
954 pci_alloc_consistent(hi->host->pdev,
955 sizeof(struct sbp2_query_logins_orb),
956 &scsi_id->query_logins_orb_dma);
957 if (!scsi_id->query_logins_orb)
958 goto alloc_fail;
959 SBP2_DMA_ALLOC("consistent DMA region for query logins ORB");
960
961 /* Query logins response DMA */
962 scsi_id->query_logins_response =
963 pci_alloc_consistent(hi->host->pdev,
964 sizeof(struct sbp2_query_logins_response),
965 &scsi_id->query_logins_response_dma);
966 if (!scsi_id->query_logins_response)
967 goto alloc_fail;
968 SBP2_DMA_ALLOC("consistent DMA region for query logins response");
969
970 /* Reconnect ORB DMA */
971 scsi_id->reconnect_orb =
972 pci_alloc_consistent(hi->host->pdev,
973 sizeof(struct sbp2_reconnect_orb),
974 &scsi_id->reconnect_orb_dma);
975 if (!scsi_id->reconnect_orb)
976 goto alloc_fail;
977 SBP2_DMA_ALLOC("consistent DMA region for reconnect ORB");
978
979 /* Logout ORB DMA */
980 scsi_id->logout_orb =
981 pci_alloc_consistent(hi->host->pdev,
982 sizeof(struct sbp2_logout_orb),
983 &scsi_id->logout_orb_dma);
984 if (!scsi_id->logout_orb)
985 goto alloc_fail;
986 SBP2_DMA_ALLOC("consistent DMA region for logout ORB");
987
988 /* Login ORB DMA */
989 scsi_id->login_orb =
990 pci_alloc_consistent(hi->host->pdev,
991 sizeof(struct sbp2_login_orb),
992 &scsi_id->login_orb_dma);
993 if (!scsi_id->login_orb)
994 goto alloc_fail;
995 SBP2_DMA_ALLOC("consistent DMA region for login ORB");
996
997 SBP2_DEBUG("New SBP-2 device inserted, SCSI ID = %x", scsi_id->ud->id);
998
999 /*
1000 * Create our command orb pool
1001 */
1002 if (sbp2util_create_command_orb_pool(scsi_id)) {
1003 SBP2_ERR("sbp2util_create_command_orb_pool failed!");
1004 sbp2_remove_device(scsi_id);
1005 return -ENOMEM;
1006 }
1007
1008 /* Schedule a timeout here. The reason is that we may be so close
1009 * to a bus reset, that the device is not available for logins.
1010 * This can happen when the bus reset is caused by the host
1011 * connected to the sbp2 device being removed. That host would
1012 * have a certain amount of time to relogin before the sbp2 device
1013 * allows someone else to login instead. One second makes sense. */
1014 msleep_interruptible(1000);
1015 if (signal_pending(current)) {
1016 sbp2_remove_device(scsi_id);
1017 return -EINTR;
1018 }
1019
1020 /*
1021 * Login to the sbp-2 device
1022 */
1023 if (sbp2_login_device(scsi_id)) {
1024 /* Login failed, just remove the device. */
1025 sbp2_remove_device(scsi_id);
1026 return -EBUSY;
1027 }
1028
1029 /*
1030 * Set max retries to something large on the device
1031 */
1032 sbp2_set_busy_timeout(scsi_id);
1033
1034 /*
1035 * Do a SBP-2 fetch agent reset
1036 */
1037 sbp2_agent_reset(scsi_id, 1);
1038
1039 /*
1040 * Get the max speed and packet size that we can use
1041 */
1042 sbp2_max_speed_and_size(scsi_id);
1043
1044 /* Add this device to the scsi layer now */
1045 error = scsi_add_device(scsi_id->scsi_host, 0, scsi_id->ud->id, 0);
1046 if (error) {
1047 SBP2_ERR("scsi_add_device failed");
1048 sbp2_logout_device(scsi_id);
1049 sbp2_remove_device(scsi_id);
1050 return error;
1051 }
1052
1053 return 0;
1054
1055 alloc_fail:
1056 SBP2_ERR("Could not allocate memory for scsi_id");
1057 sbp2_remove_device(scsi_id);
1058 return -ENOMEM;
1059 }
1060
1061 /*
1062 * This function removes an sbp2 device from the sbp2scsi_host_info struct.
1063 */
1064 static void sbp2_remove_device(struct scsi_id_instance_data *scsi_id)
1065 {
1066 struct sbp2scsi_host_info *hi;
1067
1068 SBP2_DEBUG_ENTER();
1069
1070 if (!scsi_id)
1071 return;
1072
1073 hi = scsi_id->hi;
1074
1075 /* This will remove our scsi device aswell */
1076 if (scsi_id->scsi_host) {
1077 scsi_remove_host(scsi_id->scsi_host);
1078 scsi_host_put(scsi_id->scsi_host);
1079 }
1080 flush_scheduled_work();
1081 sbp2util_remove_command_orb_pool(scsi_id);
1082
1083 list_del(&scsi_id->scsi_list);
1084
1085 if (scsi_id->login_response) {
1086 pci_free_consistent(hi->host->pdev,
1087 sizeof(struct sbp2_login_response),
1088 scsi_id->login_response,
1089 scsi_id->login_response_dma);
1090 SBP2_DMA_FREE("single login FIFO");
1091 }
1092
1093 if (scsi_id->login_orb) {
1094 pci_free_consistent(hi->host->pdev,
1095 sizeof(struct sbp2_login_orb),
1096 scsi_id->login_orb,
1097 scsi_id->login_orb_dma);
1098 SBP2_DMA_FREE("single login ORB");
1099 }
1100
1101 if (scsi_id->reconnect_orb) {
1102 pci_free_consistent(hi->host->pdev,
1103 sizeof(struct sbp2_reconnect_orb),
1104 scsi_id->reconnect_orb,
1105 scsi_id->reconnect_orb_dma);
1106 SBP2_DMA_FREE("single reconnect orb");
1107 }
1108
1109 if (scsi_id->logout_orb) {
1110 pci_free_consistent(hi->host->pdev,
1111 sizeof(struct sbp2_logout_orb),
1112 scsi_id->logout_orb,
1113 scsi_id->logout_orb_dma);
1114 SBP2_DMA_FREE("single logout orb");
1115 }
1116
1117 if (scsi_id->query_logins_orb) {
1118 pci_free_consistent(hi->host->pdev,
1119 sizeof(struct sbp2_query_logins_orb),
1120 scsi_id->query_logins_orb,
1121 scsi_id->query_logins_orb_dma);
1122 SBP2_DMA_FREE("single query logins orb");
1123 }
1124
1125 if (scsi_id->query_logins_response) {
1126 pci_free_consistent(hi->host->pdev,
1127 sizeof(struct sbp2_query_logins_response),
1128 scsi_id->query_logins_response,
1129 scsi_id->query_logins_response_dma);
1130 SBP2_DMA_FREE("single query logins data");
1131 }
1132
1133 if (scsi_id->status_fifo_addr != CSR1212_INVALID_ADDR_SPACE)
1134 hpsb_unregister_addrspace(&sbp2_highlevel, hi->host,
1135 scsi_id->status_fifo_addr);
1136
1137 scsi_id->ud->device.driver_data = NULL;
1138
1139 if (hi)
1140 module_put(hi->host->driver->owner);
1141
1142 SBP2_DEBUG("SBP-2 device removed, SCSI ID = %d", scsi_id->ud->id);
1143
1144 kfree(scsi_id);
1145 }
1146
1147 #ifdef CONFIG_IEEE1394_SBP2_PHYS_DMA
1148 /*
1149 * This function deals with physical dma write requests (for adapters that do not support
1150 * physical dma in hardware). Mostly just here for debugging...
1151 */
1152 static int sbp2_handle_physdma_write(struct hpsb_host *host, int nodeid,
1153 int destid, quadlet_t *data, u64 addr,
1154 size_t length, u16 flags)
1155 {
1156
1157 /*
1158 * Manually put the data in the right place.
1159 */
1160 memcpy(bus_to_virt((u32) addr), data, length);
1161 sbp2util_packet_dump(data, length, "sbp2 phys dma write by device",
1162 (u32) addr);
1163 return RCODE_COMPLETE;
1164 }
1165
1166 /*
1167 * This function deals with physical dma read requests (for adapters that do not support
1168 * physical dma in hardware). Mostly just here for debugging...
1169 */
1170 static int sbp2_handle_physdma_read(struct hpsb_host *host, int nodeid,
1171 quadlet_t *data, u64 addr, size_t length,
1172 u16 flags)
1173 {
1174
1175 /*
1176 * Grab data from memory and send a read response.
1177 */
1178 memcpy(data, bus_to_virt((u32) addr), length);
1179 sbp2util_packet_dump(data, length, "sbp2 phys dma read by device",
1180 (u32) addr);
1181 return RCODE_COMPLETE;
1182 }
1183 #endif
1184
1185 /**************************************
1186 * SBP-2 protocol related section
1187 **************************************/
1188
1189 /*
1190 * This function queries the device for the maximum concurrent logins it
1191 * supports.
1192 */
1193 static int sbp2_query_logins(struct scsi_id_instance_data *scsi_id)
1194 {
1195 struct sbp2scsi_host_info *hi = scsi_id->hi;
1196 quadlet_t data[2];
1197 int max_logins;
1198 int active_logins;
1199
1200 SBP2_DEBUG_ENTER();
1201
1202 scsi_id->query_logins_orb->reserved1 = 0x0;
1203 scsi_id->query_logins_orb->reserved2 = 0x0;
1204
1205 scsi_id->query_logins_orb->query_response_lo = scsi_id->query_logins_response_dma;
1206 scsi_id->query_logins_orb->query_response_hi = ORB_SET_NODE_ID(hi->host->node_id);
1207
1208 scsi_id->query_logins_orb->lun_misc = ORB_SET_FUNCTION(SBP2_QUERY_LOGINS_REQUEST);
1209 scsi_id->query_logins_orb->lun_misc |= ORB_SET_NOTIFY(1);
1210 scsi_id->query_logins_orb->lun_misc |= ORB_SET_LUN(scsi_id->sbp2_lun);
1211
1212 scsi_id->query_logins_orb->reserved_resp_length =
1213 ORB_SET_QUERY_LOGINS_RESP_LENGTH(sizeof(struct sbp2_query_logins_response));
1214
1215 scsi_id->query_logins_orb->status_fifo_hi =
1216 ORB_SET_STATUS_FIFO_HI(scsi_id->status_fifo_addr, hi->host->node_id);
1217 scsi_id->query_logins_orb->status_fifo_lo =
1218 ORB_SET_STATUS_FIFO_LO(scsi_id->status_fifo_addr);
1219
1220 sbp2util_cpu_to_be32_buffer(scsi_id->query_logins_orb, sizeof(struct sbp2_query_logins_orb));
1221
1222 sbp2util_packet_dump(scsi_id->query_logins_orb, sizeof(struct sbp2_query_logins_orb),
1223 "sbp2 query logins orb", scsi_id->query_logins_orb_dma);
1224
1225 memset(scsi_id->query_logins_response, 0, sizeof(struct sbp2_query_logins_response));
1226
1227 data[0] = ORB_SET_NODE_ID(hi->host->node_id);
1228 data[1] = scsi_id->query_logins_orb_dma;
1229 sbp2util_cpu_to_be32_buffer(data, 8);
1230
1231 hpsb_node_write(scsi_id->ne, scsi_id->sbp2_management_agent_addr, data, 8);
1232
1233 if (sbp2util_access_timeout(scsi_id, 2*HZ)) {
1234 SBP2_INFO("Error querying logins to SBP-2 device - timed out");
1235 return -EIO;
1236 }
1237
1238 if (scsi_id->status_block.ORB_offset_lo != scsi_id->query_logins_orb_dma) {
1239 SBP2_INFO("Error querying logins to SBP-2 device - timed out");
1240 return -EIO;
1241 }
1242
1243 if (STATUS_TEST_RDS(scsi_id->status_block.ORB_offset_hi_misc)) {
1244 SBP2_INFO("Error querying logins to SBP-2 device - failed");
1245 return -EIO;
1246 }
1247
1248 sbp2util_cpu_to_be32_buffer(scsi_id->query_logins_response, sizeof(struct sbp2_query_logins_response));
1249
1250 SBP2_DEBUG("length_max_logins = %x",
1251 (unsigned int)scsi_id->query_logins_response->length_max_logins);
1252
1253 max_logins = RESPONSE_GET_MAX_LOGINS(scsi_id->query_logins_response->length_max_logins);
1254 SBP2_INFO("Maximum concurrent logins supported: %d", max_logins);
1255
1256 active_logins = RESPONSE_GET_ACTIVE_LOGINS(scsi_id->query_logins_response->length_max_logins);
1257 SBP2_INFO("Number of active logins: %d", active_logins);
1258
1259 if (active_logins >= max_logins) {
1260 return -EIO;
1261 }
1262
1263 return 0;
1264 }
1265
1266 /*
1267 * This function is called in order to login to a particular SBP-2 device,
1268 * after a bus reset.
1269 */
1270 static int sbp2_login_device(struct scsi_id_instance_data *scsi_id)
1271 {
1272 struct sbp2scsi_host_info *hi = scsi_id->hi;
1273 quadlet_t data[2];
1274
1275 SBP2_DEBUG_ENTER();
1276
1277 if (!scsi_id->login_orb) {
1278 SBP2_DEBUG("%s: login_orb not alloc'd!", __FUNCTION__);
1279 return -EIO;
1280 }
1281
1282 if (!exclusive_login) {
1283 if (sbp2_query_logins(scsi_id)) {
1284 SBP2_INFO("Device does not support any more concurrent logins");
1285 return -EIO;
1286 }
1287 }
1288
1289 /* Set-up login ORB, assume no password */
1290 scsi_id->login_orb->password_hi = 0;
1291 scsi_id->login_orb->password_lo = 0;
1292
1293 scsi_id->login_orb->login_response_lo = scsi_id->login_response_dma;
1294 scsi_id->login_orb->login_response_hi = ORB_SET_NODE_ID(hi->host->node_id);
1295
1296 scsi_id->login_orb->lun_misc = ORB_SET_FUNCTION(SBP2_LOGIN_REQUEST);
1297 scsi_id->login_orb->lun_misc |= ORB_SET_RECONNECT(0); /* One second reconnect time */
1298 scsi_id->login_orb->lun_misc |= ORB_SET_EXCLUSIVE(exclusive_login); /* Exclusive access to device */
1299 scsi_id->login_orb->lun_misc |= ORB_SET_NOTIFY(1); /* Notify us of login complete */
1300 scsi_id->login_orb->lun_misc |= ORB_SET_LUN(scsi_id->sbp2_lun);
1301
1302 scsi_id->login_orb->passwd_resp_lengths =
1303 ORB_SET_LOGIN_RESP_LENGTH(sizeof(struct sbp2_login_response));
1304
1305 scsi_id->login_orb->status_fifo_hi =
1306 ORB_SET_STATUS_FIFO_HI(scsi_id->status_fifo_addr, hi->host->node_id);
1307 scsi_id->login_orb->status_fifo_lo =
1308 ORB_SET_STATUS_FIFO_LO(scsi_id->status_fifo_addr);
1309
1310 sbp2util_cpu_to_be32_buffer(scsi_id->login_orb, sizeof(struct sbp2_login_orb));
1311
1312 sbp2util_packet_dump(scsi_id->login_orb, sizeof(struct sbp2_login_orb),
1313 "sbp2 login orb", scsi_id->login_orb_dma);
1314
1315 memset(scsi_id->login_response, 0, sizeof(struct sbp2_login_response));
1316
1317 data[0] = ORB_SET_NODE_ID(hi->host->node_id);
1318 data[1] = scsi_id->login_orb_dma;
1319 sbp2util_cpu_to_be32_buffer(data, 8);
1320
1321 hpsb_node_write(scsi_id->ne, scsi_id->sbp2_management_agent_addr, data, 8);
1322
1323 /*
1324 * Wait for login status (up to 20 seconds)...
1325 */
1326 if (sbp2util_access_timeout(scsi_id, 20*HZ)) {
1327 SBP2_ERR("Error logging into SBP-2 device - timed out");
1328 return -EIO;
1329 }
1330
1331 /*
1332 * Sanity. Make sure status returned matches login orb.
1333 */
1334 if (scsi_id->status_block.ORB_offset_lo != scsi_id->login_orb_dma) {
1335 SBP2_ERR("Error logging into SBP-2 device - timed out");
1336 return -EIO;
1337 }
1338
1339 if (STATUS_TEST_RDS(scsi_id->status_block.ORB_offset_hi_misc)) {
1340 SBP2_ERR("Error logging into SBP-2 device - failed");
1341 return -EIO;
1342 }
1343
1344 /*
1345 * Byte swap the login response, for use when reconnecting or
1346 * logging out.
1347 */
1348 sbp2util_cpu_to_be32_buffer(scsi_id->login_response, sizeof(struct sbp2_login_response));
1349
1350 /*
1351 * Grab our command block agent address from the login response.
1352 */
1353 SBP2_DEBUG("command_block_agent_hi = %x",
1354 (unsigned int)scsi_id->login_response->command_block_agent_hi);
1355 SBP2_DEBUG("command_block_agent_lo = %x",
1356 (unsigned int)scsi_id->login_response->command_block_agent_lo);
1357
1358 scsi_id->sbp2_command_block_agent_addr =
1359 ((u64)scsi_id->login_response->command_block_agent_hi) << 32;
1360 scsi_id->sbp2_command_block_agent_addr |= ((u64)scsi_id->login_response->command_block_agent_lo);
1361 scsi_id->sbp2_command_block_agent_addr &= 0x0000ffffffffffffULL;
1362
1363 SBP2_INFO("Logged into SBP-2 device");
1364 return 0;
1365 }
1366
1367 /*
1368 * This function is called in order to logout from a particular SBP-2
1369 * device, usually called during driver unload.
1370 */
1371 static int sbp2_logout_device(struct scsi_id_instance_data *scsi_id)
1372 {
1373 struct sbp2scsi_host_info *hi = scsi_id->hi;
1374 quadlet_t data[2];
1375 int error;
1376
1377 SBP2_DEBUG_ENTER();
1378
1379 /*
1380 * Set-up logout ORB
1381 */
1382 scsi_id->logout_orb->reserved1 = 0x0;
1383 scsi_id->logout_orb->reserved2 = 0x0;
1384 scsi_id->logout_orb->reserved3 = 0x0;
1385 scsi_id->logout_orb->reserved4 = 0x0;
1386
1387 scsi_id->logout_orb->login_ID_misc = ORB_SET_FUNCTION(SBP2_LOGOUT_REQUEST);
1388 scsi_id->logout_orb->login_ID_misc |= ORB_SET_LOGIN_ID(scsi_id->login_response->length_login_ID);
1389
1390 /* Notify us when complete */
1391 scsi_id->logout_orb->login_ID_misc |= ORB_SET_NOTIFY(1);
1392
1393 scsi_id->logout_orb->reserved5 = 0x0;
1394 scsi_id->logout_orb->status_fifo_hi =
1395 ORB_SET_STATUS_FIFO_HI(scsi_id->status_fifo_addr, hi->host->node_id);
1396 scsi_id->logout_orb->status_fifo_lo =
1397 ORB_SET_STATUS_FIFO_LO(scsi_id->status_fifo_addr);
1398
1399 /*
1400 * Byte swap ORB if necessary
1401 */
1402 sbp2util_cpu_to_be32_buffer(scsi_id->logout_orb, sizeof(struct sbp2_logout_orb));
1403
1404 sbp2util_packet_dump(scsi_id->logout_orb, sizeof(struct sbp2_logout_orb),
1405 "sbp2 logout orb", scsi_id->logout_orb_dma);
1406
1407 /*
1408 * Ok, let's write to the target's management agent register
1409 */
1410 data[0] = ORB_SET_NODE_ID(hi->host->node_id);
1411 data[1] = scsi_id->logout_orb_dma;
1412 sbp2util_cpu_to_be32_buffer(data, 8);
1413
1414 error = hpsb_node_write(scsi_id->ne,
1415 scsi_id->sbp2_management_agent_addr, data, 8);
1416 if (error)
1417 return error;
1418
1419 /* Wait for device to logout...1 second. */
1420 if (sbp2util_access_timeout(scsi_id, HZ))
1421 return -EIO;
1422
1423 SBP2_INFO("Logged out of SBP-2 device");
1424 return 0;
1425 }
1426
1427 /*
1428 * This function is called in order to reconnect to a particular SBP-2
1429 * device, after a bus reset.
1430 */
1431 static int sbp2_reconnect_device(struct scsi_id_instance_data *scsi_id)
1432 {
1433 struct sbp2scsi_host_info *hi = scsi_id->hi;
1434 quadlet_t data[2];
1435 int error;
1436
1437 SBP2_DEBUG_ENTER();
1438
1439 /*
1440 * Set-up reconnect ORB
1441 */
1442 scsi_id->reconnect_orb->reserved1 = 0x0;
1443 scsi_id->reconnect_orb->reserved2 = 0x0;
1444 scsi_id->reconnect_orb->reserved3 = 0x0;
1445 scsi_id->reconnect_orb->reserved4 = 0x0;
1446
1447 scsi_id->reconnect_orb->login_ID_misc = ORB_SET_FUNCTION(SBP2_RECONNECT_REQUEST);
1448 scsi_id->reconnect_orb->login_ID_misc |=
1449 ORB_SET_LOGIN_ID(scsi_id->login_response->length_login_ID);
1450
1451 /* Notify us when complete */
1452 scsi_id->reconnect_orb->login_ID_misc |= ORB_SET_NOTIFY(1);
1453
1454 scsi_id->reconnect_orb->reserved5 = 0x0;
1455 scsi_id->reconnect_orb->status_fifo_hi =
1456 ORB_SET_STATUS_FIFO_HI(scsi_id->status_fifo_addr, hi->host->node_id);
1457 scsi_id->reconnect_orb->status_fifo_lo =
1458 ORB_SET_STATUS_FIFO_LO(scsi_id->status_fifo_addr);
1459
1460 /*
1461 * Byte swap ORB if necessary
1462 */
1463 sbp2util_cpu_to_be32_buffer(scsi_id->reconnect_orb, sizeof(struct sbp2_reconnect_orb));
1464
1465 sbp2util_packet_dump(scsi_id->reconnect_orb, sizeof(struct sbp2_reconnect_orb),
1466 "sbp2 reconnect orb", scsi_id->reconnect_orb_dma);
1467
1468 data[0] = ORB_SET_NODE_ID(hi->host->node_id);
1469 data[1] = scsi_id->reconnect_orb_dma;
1470 sbp2util_cpu_to_be32_buffer(data, 8);
1471
1472 error = hpsb_node_write(scsi_id->ne,
1473 scsi_id->sbp2_management_agent_addr, data, 8);
1474 if (error)
1475 return error;
1476
1477 /*
1478 * Wait for reconnect status (up to 1 second)...
1479 */
1480 if (sbp2util_access_timeout(scsi_id, HZ)) {
1481 SBP2_ERR("Error reconnecting to SBP-2 device - timed out");
1482 return -EIO;
1483 }
1484
1485 /*
1486 * Sanity. Make sure status returned matches reconnect orb.
1487 */
1488 if (scsi_id->status_block.ORB_offset_lo != scsi_id->reconnect_orb_dma) {
1489 SBP2_ERR("Error reconnecting to SBP-2 device - timed out");
1490 return -EIO;
1491 }
1492
1493 if (STATUS_TEST_RDS(scsi_id->status_block.ORB_offset_hi_misc)) {
1494 SBP2_ERR("Error reconnecting to SBP-2 device - failed");
1495 return -EIO;
1496 }
1497
1498 HPSB_DEBUG("Reconnected to SBP-2 device");
1499 return 0;
1500 }
1501
1502 /*
1503 * This function is called in order to set the busy timeout (number of
1504 * retries to attempt) on the sbp2 device.
1505 */
1506 static int sbp2_set_busy_timeout(struct scsi_id_instance_data *scsi_id)
1507 {
1508 quadlet_t data;
1509
1510 SBP2_DEBUG_ENTER();
1511
1512 data = cpu_to_be32(SBP2_BUSY_TIMEOUT_VALUE);
1513 if (hpsb_node_write(scsi_id->ne, SBP2_BUSY_TIMEOUT_ADDRESS, &data, 4))
1514 SBP2_ERR("%s error", __FUNCTION__);
1515 return 0;
1516 }
1517
1518 /*
1519 * This function is called to parse sbp2 device's config rom unit
1520 * directory. Used to determine things like sbp2 management agent offset,
1521 * and command set used (SCSI or RBC).
1522 */
1523 static void sbp2_parse_unit_directory(struct scsi_id_instance_data *scsi_id,
1524 struct unit_directory *ud)
1525 {
1526 struct csr1212_keyval *kv;
1527 struct csr1212_dentry *dentry;
1528 u64 management_agent_addr;
1529 u32 command_set_spec_id, command_set, unit_characteristics,
1530 firmware_revision;
1531 unsigned workarounds;
1532 int i;
1533
1534 SBP2_DEBUG_ENTER();
1535
1536 management_agent_addr = 0x0;
1537 command_set_spec_id = 0x0;
1538 command_set = 0x0;
1539 unit_characteristics = 0x0;
1540 firmware_revision = 0x0;
1541
1542 /* Handle different fields in the unit directory, based on keys */
1543 csr1212_for_each_dir_entry(ud->ne->csr, kv, ud->ud_kv, dentry) {
1544 switch (kv->key.id) {
1545 case CSR1212_KV_ID_DEPENDENT_INFO:
1546 if (kv->key.type == CSR1212_KV_TYPE_CSR_OFFSET) {
1547 /* Save off the management agent address */
1548 management_agent_addr =
1549 CSR1212_REGISTER_SPACE_BASE +
1550 (kv->value.csr_offset << 2);
1551
1552 SBP2_DEBUG("sbp2_management_agent_addr = %x",
1553 (unsigned int)management_agent_addr);
1554 } else if (kv->key.type == CSR1212_KV_TYPE_IMMEDIATE) {
1555 scsi_id->sbp2_lun =
1556 ORB_SET_LUN(kv->value.immediate);
1557 }
1558 break;
1559
1560 case SBP2_COMMAND_SET_SPEC_ID_KEY:
1561 /* Command spec organization */
1562 command_set_spec_id = kv->value.immediate;
1563 SBP2_DEBUG("sbp2_command_set_spec_id = %x",
1564 (unsigned int)command_set_spec_id);
1565 break;
1566
1567 case SBP2_COMMAND_SET_KEY:
1568 /* Command set used by sbp2 device */
1569 command_set = kv->value.immediate;
1570 SBP2_DEBUG("sbp2_command_set = %x",
1571 (unsigned int)command_set);
1572 break;
1573
1574 case SBP2_UNIT_CHARACTERISTICS_KEY:
1575 /*
1576 * Unit characterisitcs (orb related stuff
1577 * that I'm not yet paying attention to)
1578 */
1579 unit_characteristics = kv->value.immediate;
1580 SBP2_DEBUG("sbp2_unit_characteristics = %x",
1581 (unsigned int)unit_characteristics);
1582 break;
1583
1584 case SBP2_FIRMWARE_REVISION_KEY:
1585 /* Firmware revision */
1586 firmware_revision = kv->value.immediate;
1587 SBP2_DEBUG("sbp2_firmware_revision = %x",
1588 (unsigned int)firmware_revision);
1589 break;
1590
1591 default:
1592 break;
1593 }
1594 }
1595
1596 workarounds = sbp2_default_workarounds;
1597
1598 if (!(workarounds & SBP2_WORKAROUND_OVERRIDE))
1599 for (i = 0; i < ARRAY_SIZE(sbp2_workarounds_table); i++) {
1600 if (sbp2_workarounds_table[i].firmware_revision &&
1601 sbp2_workarounds_table[i].firmware_revision !=
1602 (firmware_revision & 0xffff00))
1603 continue;
1604 if (sbp2_workarounds_table[i].model_id &&
1605 sbp2_workarounds_table[i].model_id != ud->model_id)
1606 continue;
1607 workarounds |= sbp2_workarounds_table[i].workarounds;
1608 break;
1609 }
1610
1611 if (workarounds)
1612 SBP2_INFO("Workarounds for node " NODE_BUS_FMT ": 0x%x "
1613 "(firmware_revision 0x%06x, vendor_id 0x%06x,"
1614 " model_id 0x%06x)",
1615 NODE_BUS_ARGS(ud->ne->host, ud->ne->nodeid),
1616 workarounds, firmware_revision,
1617 ud->vendor_id ? ud->vendor_id : ud->ne->vendor_id,
1618 ud->model_id);
1619
1620 /* We would need one SCSI host template for each target to adjust
1621 * max_sectors on the fly, therefore warn only. */
1622 if (workarounds & SBP2_WORKAROUND_128K_MAX_TRANS &&
1623 (max_sectors * 512) > (128 * 1024))
1624 SBP2_WARN("Node " NODE_BUS_FMT ": Bridge only supports 128KB "
1625 "max transfer size. WARNING: Current max_sectors "
1626 "setting is larger than 128KB (%d sectors)",
1627 NODE_BUS_ARGS(ud->ne->host, ud->ne->nodeid),
1628 max_sectors);
1629
1630 /* If this is a logical unit directory entry, process the parent
1631 * to get the values. */
1632 if (ud->flags & UNIT_DIRECTORY_LUN_DIRECTORY) {
1633 struct unit_directory *parent_ud =
1634 container_of(ud->device.parent, struct unit_directory, device);
1635 sbp2_parse_unit_directory(scsi_id, parent_ud);
1636 } else {
1637 scsi_id->sbp2_management_agent_addr = management_agent_addr;
1638 scsi_id->sbp2_command_set_spec_id = command_set_spec_id;
1639 scsi_id->sbp2_command_set = command_set;
1640 scsi_id->sbp2_unit_characteristics = unit_characteristics;
1641 scsi_id->sbp2_firmware_revision = firmware_revision;
1642 scsi_id->workarounds = workarounds;
1643 if (ud->flags & UNIT_DIRECTORY_HAS_LUN)
1644 scsi_id->sbp2_lun = ORB_SET_LUN(ud->lun);
1645 }
1646 }
1647
1648 #define SBP2_PAYLOAD_TO_BYTES(p) (1 << ((p) + 2))
1649
1650 /*
1651 * This function is called in order to determine the max speed and packet
1652 * size we can use in our ORBs. Note, that we (the driver and host) only
1653 * initiate the transaction. The SBP-2 device actually transfers the data
1654 * (by reading from the DMA area we tell it). This means that the SBP-2
1655 * device decides the actual maximum data it can transfer. We just tell it
1656 * the speed that it needs to use, and the max_rec the host supports, and
1657 * it takes care of the rest.
1658 */
1659 static int sbp2_max_speed_and_size(struct scsi_id_instance_data *scsi_id)
1660 {
1661 struct sbp2scsi_host_info *hi = scsi_id->hi;
1662 u8 payload;
1663
1664 SBP2_DEBUG_ENTER();
1665
1666 scsi_id->speed_code =
1667 hi->host->speed[NODEID_TO_NODE(scsi_id->ne->nodeid)];
1668
1669 /* Bump down our speed if the user requested it */
1670 if (scsi_id->speed_code > max_speed) {
1671 scsi_id->speed_code = max_speed;
1672 SBP2_ERR("Forcing SBP-2 max speed down to %s",
1673 hpsb_speedto_str[scsi_id->speed_code]);
1674 }
1675
1676 /* Payload size is the lesser of what our speed supports and what
1677 * our host supports. */
1678 payload = min(sbp2_speedto_max_payload[scsi_id->speed_code],
1679 (u8) (hi->host->csr.max_rec - 1));
1680
1681 /* If physical DMA is off, work around limitation in ohci1394:
1682 * packet size must not exceed PAGE_SIZE */
1683 if (scsi_id->ne->host->low_addr_space < (1ULL << 32))
1684 while (SBP2_PAYLOAD_TO_BYTES(payload) + 24 > PAGE_SIZE &&
1685 payload)
1686 payload--;
1687
1688 HPSB_DEBUG("Node " NODE_BUS_FMT ": Max speed [%s] - Max payload [%u]",
1689 NODE_BUS_ARGS(hi->host, scsi_id->ne->nodeid),
1690 hpsb_speedto_str[scsi_id->speed_code],
1691 SBP2_PAYLOAD_TO_BYTES(payload));
1692
1693 scsi_id->max_payload_size = payload;
1694 return 0;
1695 }
1696
1697 /*
1698 * This function is called in order to perform a SBP-2 agent reset.
1699 */
1700 static int sbp2_agent_reset(struct scsi_id_instance_data *scsi_id, int wait)
1701 {
1702 quadlet_t data;
1703 u64 addr;
1704 int retval;
1705 unsigned long flags;
1706
1707 SBP2_DEBUG_ENTER();
1708
1709 cancel_delayed_work(&scsi_id->protocol_work);
1710 if (wait)
1711 flush_scheduled_work();
1712
1713 data = ntohl(SBP2_AGENT_RESET_DATA);
1714 addr = scsi_id->sbp2_command_block_agent_addr + SBP2_AGENT_RESET_OFFSET;
1715
1716 if (wait)
1717 retval = hpsb_node_write(scsi_id->ne, addr, &data, 4);
1718 else
1719 retval = sbp2util_node_write_no_wait(scsi_id->ne, addr, &data, 4);
1720
1721 if (retval < 0) {
1722 SBP2_ERR("hpsb_node_write failed.\n");
1723 return -EIO;
1724 }
1725
1726 /*
1727 * Need to make sure orb pointer is written on next command
1728 */
1729 spin_lock_irqsave(&scsi_id->sbp2_command_orb_lock, flags);
1730 scsi_id->last_orb = NULL;
1731 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
1732
1733 return 0;
1734 }
1735
1736 static void sbp2_prep_command_orb_sg(struct sbp2_command_orb *orb,
1737 struct sbp2scsi_host_info *hi,
1738 struct sbp2_command_info *command,
1739 unsigned int scsi_use_sg,
1740 struct scatterlist *sgpnt,
1741 u32 orb_direction,
1742 enum dma_data_direction dma_dir)
1743 {
1744 command->dma_dir = dma_dir;
1745 orb->data_descriptor_hi = ORB_SET_NODE_ID(hi->host->node_id);
1746 orb->misc |= ORB_SET_DIRECTION(orb_direction);
1747
1748 /* Special case if only one element (and less than 64KB in size) */
1749 if ((scsi_use_sg == 1) &&
1750 (sgpnt[0].length <= SBP2_MAX_SG_ELEMENT_LENGTH)) {
1751
1752 SBP2_DEBUG("Only one s/g element");
1753 command->dma_size = sgpnt[0].length;
1754 command->dma_type = CMD_DMA_PAGE;
1755 command->cmd_dma = pci_map_page(hi->host->pdev,
1756 sgpnt[0].page,
1757 sgpnt[0].offset,
1758 command->dma_size,
1759 command->dma_dir);
1760 SBP2_DMA_ALLOC("single page scatter element");
1761
1762 orb->data_descriptor_lo = command->cmd_dma;
1763 orb->misc |= ORB_SET_DATA_SIZE(command->dma_size);
1764
1765 } else {
1766 struct sbp2_unrestricted_page_table *sg_element =
1767 &command->scatter_gather_element[0];
1768 u32 sg_count, sg_len;
1769 dma_addr_t sg_addr;
1770 int i, count = pci_map_sg(hi->host->pdev, sgpnt, scsi_use_sg,
1771 dma_dir);
1772
1773 SBP2_DMA_ALLOC("scatter list");
1774
1775 command->dma_size = scsi_use_sg;
1776 command->sge_buffer = sgpnt;
1777
1778 /* use page tables (s/g) */
1779 orb->misc |= ORB_SET_PAGE_TABLE_PRESENT(0x1);
1780 orb->data_descriptor_lo = command->sge_dma;
1781
1782 /*
1783 * Loop through and fill out our sbp-2 page tables
1784 * (and split up anything too large)
1785 */
1786 for (i = 0, sg_count = 0 ; i < count; i++, sgpnt++) {
1787 sg_len = sg_dma_len(sgpnt);
1788 sg_addr = sg_dma_address(sgpnt);
1789 while (sg_len) {
1790 sg_element[sg_count].segment_base_lo = sg_addr;
1791 if (sg_len > SBP2_MAX_SG_ELEMENT_LENGTH) {
1792 sg_element[sg_count].length_segment_base_hi =
1793 PAGE_TABLE_SET_SEGMENT_LENGTH(SBP2_MAX_SG_ELEMENT_LENGTH);
1794 sg_addr += SBP2_MAX_SG_ELEMENT_LENGTH;
1795 sg_len -= SBP2_MAX_SG_ELEMENT_LENGTH;
1796 } else {
1797 sg_element[sg_count].length_segment_base_hi =
1798 PAGE_TABLE_SET_SEGMENT_LENGTH(sg_len);
1799 sg_len = 0;
1800 }
1801 sg_count++;
1802 }
1803 }
1804
1805 /* Number of page table (s/g) elements */
1806 orb->misc |= ORB_SET_DATA_SIZE(sg_count);
1807
1808 sbp2util_packet_dump(sg_element,
1809 (sizeof(struct sbp2_unrestricted_page_table)) * sg_count,
1810 "sbp2 s/g list", command->sge_dma);
1811
1812 /* Byte swap page tables if necessary */
1813 sbp2util_cpu_to_be32_buffer(sg_element,
1814 (sizeof(struct sbp2_unrestricted_page_table)) *
1815 sg_count);
1816 }
1817 }
1818
1819 static void sbp2_prep_command_orb_no_sg(struct sbp2_command_orb *orb,
1820 struct sbp2scsi_host_info *hi,
1821 struct sbp2_command_info *command,
1822 struct scatterlist *sgpnt,
1823 u32 orb_direction,
1824 unsigned int scsi_request_bufflen,
1825 void *scsi_request_buffer,
1826 enum dma_data_direction dma_dir)
1827 {
1828 command->dma_dir = dma_dir;
1829 command->dma_size = scsi_request_bufflen;
1830 command->dma_type = CMD_DMA_SINGLE;
1831 command->cmd_dma = pci_map_single(hi->host->pdev, scsi_request_buffer,
1832 command->dma_size, command->dma_dir);
1833 orb->data_descriptor_hi = ORB_SET_NODE_ID(hi->host->node_id);
1834 orb->misc |= ORB_SET_DIRECTION(orb_direction);
1835
1836 SBP2_DMA_ALLOC("single bulk");
1837
1838 /*
1839 * Handle case where we get a command w/o s/g enabled (but
1840 * check for transfers larger than 64K)
1841 */
1842 if (scsi_request_bufflen <= SBP2_MAX_SG_ELEMENT_LENGTH) {
1843
1844 orb->data_descriptor_lo = command->cmd_dma;
1845 orb->misc |= ORB_SET_DATA_SIZE(scsi_request_bufflen);
1846
1847 } else {
1848 struct sbp2_unrestricted_page_table *sg_element =
1849 &command->scatter_gather_element[0];
1850 u32 sg_count, sg_len;
1851 dma_addr_t sg_addr;
1852
1853 /*
1854 * Need to turn this into page tables, since the
1855 * buffer is too large.
1856 */
1857 orb->data_descriptor_lo = command->sge_dma;
1858
1859 /* Use page tables (s/g) */
1860 orb->misc |= ORB_SET_PAGE_TABLE_PRESENT(0x1);
1861
1862 /*
1863 * fill out our sbp-2 page tables (and split up
1864 * the large buffer)
1865 */
1866 sg_count = 0;
1867 sg_len = scsi_request_bufflen;
1868 sg_addr = command->cmd_dma;
1869 while (sg_len) {
1870 sg_element[sg_count].segment_base_lo = sg_addr;
1871 if (sg_len > SBP2_MAX_SG_ELEMENT_LENGTH) {
1872 sg_element[sg_count].length_segment_base_hi =
1873 PAGE_TABLE_SET_SEGMENT_LENGTH(SBP2_MAX_SG_ELEMENT_LENGTH);
1874 sg_addr += SBP2_MAX_SG_ELEMENT_LENGTH;
1875 sg_len -= SBP2_MAX_SG_ELEMENT_LENGTH;
1876 } else {
1877 sg_element[sg_count].length_segment_base_hi =
1878 PAGE_TABLE_SET_SEGMENT_LENGTH(sg_len);
1879 sg_len = 0;
1880 }
1881 sg_count++;
1882 }
1883
1884 /* Number of page table (s/g) elements */
1885 orb->misc |= ORB_SET_DATA_SIZE(sg_count);
1886
1887 sbp2util_packet_dump(sg_element,
1888 (sizeof(struct sbp2_unrestricted_page_table)) * sg_count,
1889 "sbp2 s/g list", command->sge_dma);
1890
1891 /* Byte swap page tables if necessary */
1892 sbp2util_cpu_to_be32_buffer(sg_element,
1893 (sizeof(struct sbp2_unrestricted_page_table)) *
1894 sg_count);
1895 }
1896 }
1897
1898 /*
1899 * This function is called to create the actual command orb and s/g list
1900 * out of the scsi command itself.
1901 */
1902 static void sbp2_create_command_orb(struct scsi_id_instance_data *scsi_id,
1903 struct sbp2_command_info *command,
1904 unchar *scsi_cmd,
1905 unsigned int scsi_use_sg,
1906 unsigned int scsi_request_bufflen,
1907 void *scsi_request_buffer,
1908 enum dma_data_direction dma_dir)
1909 {
1910 struct sbp2scsi_host_info *hi = scsi_id->hi;
1911 struct scatterlist *sgpnt = (struct scatterlist *)scsi_request_buffer;
1912 struct sbp2_command_orb *command_orb = &command->command_orb;
1913 u32 orb_direction;
1914
1915 /*
1916 * Set-up our command ORB..
1917 *
1918 * NOTE: We're doing unrestricted page tables (s/g), as this is
1919 * best performance (at least with the devices I have). This means
1920 * that data_size becomes the number of s/g elements, and
1921 * page_size should be zero (for unrestricted).
1922 */
1923 command_orb->next_ORB_hi = ORB_SET_NULL_PTR(1);
1924 command_orb->next_ORB_lo = 0x0;
1925 command_orb->misc = ORB_SET_MAX_PAYLOAD(scsi_id->max_payload_size);
1926 command_orb->misc |= ORB_SET_SPEED(scsi_id->speed_code);
1927 command_orb->misc |= ORB_SET_NOTIFY(1); /* Notify us when complete */
1928
1929 if (dma_dir == DMA_NONE)
1930 orb_direction = ORB_DIRECTION_NO_DATA_TRANSFER;
1931 else if (dma_dir == DMA_TO_DEVICE && scsi_request_bufflen)
1932 orb_direction = ORB_DIRECTION_WRITE_TO_MEDIA;
1933 else if (dma_dir == DMA_FROM_DEVICE && scsi_request_bufflen)
1934 orb_direction = ORB_DIRECTION_READ_FROM_MEDIA;
1935 else {
1936 SBP2_WARN("Falling back to DMA_NONE");
1937 orb_direction = ORB_DIRECTION_NO_DATA_TRANSFER;
1938 }
1939
1940 /* Set-up our pagetable stuff */
1941 if (orb_direction == ORB_DIRECTION_NO_DATA_TRANSFER) {
1942 SBP2_DEBUG("No data transfer");
1943 command_orb->data_descriptor_hi = 0x0;
1944 command_orb->data_descriptor_lo = 0x0;
1945 command_orb->misc |= ORB_SET_DIRECTION(1);
1946 } else if (scsi_use_sg) {
1947 SBP2_DEBUG("Use scatter/gather");
1948 sbp2_prep_command_orb_sg(command_orb, hi, command, scsi_use_sg,
1949 sgpnt, orb_direction, dma_dir);
1950 } else {
1951 SBP2_DEBUG("No scatter/gather");
1952 sbp2_prep_command_orb_no_sg(command_orb, hi, command, sgpnt,
1953 orb_direction, scsi_request_bufflen,
1954 scsi_request_buffer, dma_dir);
1955 }
1956
1957 /* Byte swap command ORB if necessary */
1958 sbp2util_cpu_to_be32_buffer(command_orb, sizeof(struct sbp2_command_orb));
1959
1960 /* Put our scsi command in the command ORB */
1961 memset(command_orb->cdb, 0, 12);
1962 memcpy(command_orb->cdb, scsi_cmd, COMMAND_SIZE(*scsi_cmd));
1963 }
1964
1965 /*
1966 * This function is called in order to begin a regular SBP-2 command.
1967 */
1968 static void sbp2_link_orb_command(struct scsi_id_instance_data *scsi_id,
1969 struct sbp2_command_info *command)
1970 {
1971 struct sbp2scsi_host_info *hi = scsi_id->hi;
1972 struct sbp2_command_orb *command_orb = &command->command_orb;
1973 struct sbp2_command_orb *last_orb;
1974 dma_addr_t last_orb_dma;
1975 u64 addr = scsi_id->sbp2_command_block_agent_addr;
1976 quadlet_t data[2];
1977 size_t length;
1978 unsigned long flags;
1979
1980 outstanding_orb_incr;
1981 SBP2_ORB_DEBUG("sending command orb %p, total orbs = %x",
1982 command_orb, global_outstanding_command_orbs);
1983
1984 pci_dma_sync_single_for_device(hi->host->pdev, command->command_orb_dma,
1985 sizeof(struct sbp2_command_orb),
1986 PCI_DMA_TODEVICE);
1987 pci_dma_sync_single_for_device(hi->host->pdev, command->sge_dma,
1988 sizeof(command->scatter_gather_element),
1989 PCI_DMA_BIDIRECTIONAL);
1990 /*
1991 * Check to see if there are any previous orbs to use
1992 */
1993 spin_lock_irqsave(&scsi_id->sbp2_command_orb_lock, flags);
1994 last_orb = scsi_id->last_orb;
1995 last_orb_dma = scsi_id->last_orb_dma;
1996 if (!last_orb) {
1997 /*
1998 * last_orb == NULL means: We know that the target's fetch agent
1999 * is not active right now.
2000 */
2001 addr += SBP2_ORB_POINTER_OFFSET;
2002 data[0] = ORB_SET_NODE_ID(hi->host->node_id);
2003 data[1] = command->command_orb_dma;
2004 sbp2util_cpu_to_be32_buffer(data, 8);
2005 length = 8;
2006 } else {
2007 /*
2008 * last_orb != NULL means: We know that the target's fetch agent
2009 * is (very probably) not dead or in reset state right now.
2010 * We have an ORB already sent that we can append a new one to.
2011 * The target's fetch agent may or may not have read this
2012 * previous ORB yet.
2013 */
2014 pci_dma_sync_single_for_cpu(hi->host->pdev, last_orb_dma,
2015 sizeof(struct sbp2_command_orb),
2016 PCI_DMA_TODEVICE);
2017 last_orb->next_ORB_lo = cpu_to_be32(command->command_orb_dma);
2018 wmb();
2019 /* Tells hardware that this pointer is valid */
2020 last_orb->next_ORB_hi = 0;
2021 pci_dma_sync_single_for_device(hi->host->pdev, last_orb_dma,
2022 sizeof(struct sbp2_command_orb),
2023 PCI_DMA_TODEVICE);
2024 addr += SBP2_DOORBELL_OFFSET;
2025 data[0] = 0;
2026 length = 4;
2027 }
2028 scsi_id->last_orb = command_orb;
2029 scsi_id->last_orb_dma = command->command_orb_dma;
2030 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
2031
2032 SBP2_ORB_DEBUG("write to %s register, command orb %p",
2033 last_orb ? "DOORBELL" : "ORB_POINTER", command_orb);
2034 if (sbp2util_node_write_no_wait(scsi_id->ne, addr, data, length)) {
2035 /*
2036 * sbp2util_node_write_no_wait failed. We certainly ran out
2037 * of transaction labels, perhaps just because there were no
2038 * context switches which gave khpsbpkt a chance to collect
2039 * free tlabels. Try again in non-atomic context. If necessary,
2040 * the workqueue job will sleep to guaranteedly get a tlabel.
2041 * We do not accept new commands until the job is over.
2042 */
2043 scsi_block_requests(scsi_id->scsi_host);
2044 PREPARE_WORK(&scsi_id->protocol_work,
2045 last_orb ? sbp2util_write_doorbell:
2046 sbp2util_write_orb_pointer,
2047 scsi_id);
2048 schedule_work(&scsi_id->protocol_work);
2049 }
2050 }
2051
2052 /*
2053 * This function is called in order to begin a regular SBP-2 command.
2054 */
2055 static int sbp2_send_command(struct scsi_id_instance_data *scsi_id,
2056 struct scsi_cmnd *SCpnt,
2057 void (*done)(struct scsi_cmnd *))
2058 {
2059 unchar *cmd = (unchar *) SCpnt->cmnd;
2060 unsigned int request_bufflen = SCpnt->request_bufflen;
2061 struct sbp2_command_info *command;
2062
2063 SBP2_DEBUG_ENTER();
2064 SBP2_DEBUG("SCSI transfer size = %x", request_bufflen);
2065 SBP2_DEBUG("SCSI s/g elements = %x", (unsigned int)SCpnt->use_sg);
2066
2067 /*
2068 * Allocate a command orb and s/g structure
2069 */
2070 command = sbp2util_allocate_command_orb(scsi_id, SCpnt, done);
2071 if (!command) {
2072 return -EIO;
2073 }
2074
2075 /*
2076 * Now actually fill in the comamnd orb and sbp2 s/g list
2077 */
2078 sbp2_create_command_orb(scsi_id, command, cmd, SCpnt->use_sg,
2079 request_bufflen, SCpnt->request_buffer,
2080 SCpnt->sc_data_direction);
2081
2082 sbp2util_packet_dump(&command->command_orb, sizeof(struct sbp2_command_orb),
2083 "sbp2 command orb", command->command_orb_dma);
2084
2085 /*
2086 * Link up the orb, and ring the doorbell if needed
2087 */
2088 sbp2_link_orb_command(scsi_id, command);
2089
2090 return 0;
2091 }
2092
2093 /*
2094 * Translates SBP-2 status into SCSI sense data for check conditions
2095 */
2096 static unsigned int sbp2_status_to_sense_data(unchar *sbp2_status, unchar *sense_data)
2097 {
2098 SBP2_DEBUG_ENTER();
2099
2100 /*
2101 * Ok, it's pretty ugly... ;-)
2102 */
2103 sense_data[0] = 0x70;
2104 sense_data[1] = 0x0;
2105 sense_data[2] = sbp2_status[9];
2106 sense_data[3] = sbp2_status[12];
2107 sense_data[4] = sbp2_status[13];
2108 sense_data[5] = sbp2_status[14];
2109 sense_data[6] = sbp2_status[15];
2110 sense_data[7] = 10;
2111 sense_data[8] = sbp2_status[16];
2112 sense_data[9] = sbp2_status[17];
2113 sense_data[10] = sbp2_status[18];
2114 sense_data[11] = sbp2_status[19];
2115 sense_data[12] = sbp2_status[10];
2116 sense_data[13] = sbp2_status[11];
2117 sense_data[14] = sbp2_status[20];
2118 sense_data[15] = sbp2_status[21];
2119
2120 return sbp2_status[8] & 0x3f; /* return scsi status */
2121 }
2122
2123 /*
2124 * This function deals with status writes from the SBP-2 device
2125 */
2126 static int sbp2_handle_status_write(struct hpsb_host *host, int nodeid,
2127 int destid, quadlet_t *data, u64 addr,
2128 size_t length, u16 fl)
2129 {
2130 struct sbp2scsi_host_info *hi;
2131 struct scsi_id_instance_data *scsi_id = NULL, *scsi_id_tmp;
2132 struct scsi_cmnd *SCpnt = NULL;
2133 struct sbp2_status_block *sb;
2134 u32 scsi_status = SBP2_SCSI_STATUS_GOOD;
2135 struct sbp2_command_info *command;
2136 unsigned long flags;
2137
2138 SBP2_DEBUG_ENTER();
2139
2140 sbp2util_packet_dump(data, length, "sbp2 status write by device", (u32)addr);
2141
2142 if (unlikely(length < 8 || length > sizeof(struct sbp2_status_block))) {
2143 SBP2_ERR("Wrong size of status block");
2144 return RCODE_ADDRESS_ERROR;
2145 }
2146 if (unlikely(!host)) {
2147 SBP2_ERR("host is NULL - this is bad!");
2148 return RCODE_ADDRESS_ERROR;
2149 }
2150 hi = hpsb_get_hostinfo(&sbp2_highlevel, host);
2151 if (unlikely(!hi)) {
2152 SBP2_ERR("host info is NULL - this is bad!");
2153 return RCODE_ADDRESS_ERROR;
2154 }
2155 /*
2156 * Find our scsi_id structure by looking at the status fifo address
2157 * written to by the sbp2 device.
2158 */
2159 list_for_each_entry(scsi_id_tmp, &hi->scsi_ids, scsi_list) {
2160 if (scsi_id_tmp->ne->nodeid == nodeid &&
2161 scsi_id_tmp->status_fifo_addr == addr) {
2162 scsi_id = scsi_id_tmp;
2163 break;
2164 }
2165 }
2166 if (unlikely(!scsi_id)) {
2167 SBP2_ERR("scsi_id is NULL - device is gone?");
2168 return RCODE_ADDRESS_ERROR;
2169 }
2170
2171 /*
2172 * Put response into scsi_id status fifo buffer. The first two bytes
2173 * come in big endian bit order. Often the target writes only a
2174 * truncated status block, minimally the first two quadlets. The rest
2175 * is implied to be zeros.
2176 */
2177 sb = &scsi_id->status_block;
2178 memset(sb->command_set_dependent, 0, sizeof(sb->command_set_dependent));
2179 memcpy(sb, data, length);
2180 sbp2util_be32_to_cpu_buffer(sb, 8);
2181
2182 /*
2183 * Ignore unsolicited status. Handle command ORB status.
2184 */
2185 if (unlikely(STATUS_GET_SRC(sb->ORB_offset_hi_misc) == 2))
2186 command = NULL;
2187 else
2188 command = sbp2util_find_command_for_orb(scsi_id,
2189 sb->ORB_offset_lo);
2190 if (command) {
2191 SBP2_DEBUG("Found status for command ORB");
2192 pci_dma_sync_single_for_cpu(hi->host->pdev, command->command_orb_dma,
2193 sizeof(struct sbp2_command_orb),
2194 PCI_DMA_TODEVICE);
2195 pci_dma_sync_single_for_cpu(hi->host->pdev, command->sge_dma,
2196 sizeof(command->scatter_gather_element),
2197 PCI_DMA_BIDIRECTIONAL);
2198
2199 SBP2_ORB_DEBUG("matched command orb %p", &command->command_orb);
2200 outstanding_orb_decr;
2201
2202 /*
2203 * Matched status with command, now grab scsi command pointers
2204 * and check status.
2205 */
2206 /*
2207 * FIXME: If the src field in the status is 1, the ORB DMA must
2208 * not be reused until status for a subsequent ORB is received.
2209 */
2210 SCpnt = command->Current_SCpnt;
2211 spin_lock_irqsave(&scsi_id->sbp2_command_orb_lock, flags);
2212 sbp2util_mark_command_completed(scsi_id, command);
2213 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
2214
2215 if (SCpnt) {
2216 if (STATUS_TEST_RS(sb->ORB_offset_hi_misc))
2217 scsi_status =
2218 SBP2_SCSI_STATUS_COMMAND_TERMINATED;
2219 /*
2220 * See if the target stored any scsi status information.
2221 */
2222 if (STATUS_GET_LEN(sb->ORB_offset_hi_misc) > 1) {
2223 SBP2_DEBUG("CHECK CONDITION");
2224 scsi_status = sbp2_status_to_sense_data(
2225 (unchar *)sb, SCpnt->sense_buffer);
2226 }
2227
2228 /*
2229 * Check to see if the dead bit is set. If so, we'll
2230 * have to initiate a fetch agent reset.
2231 */
2232 if (STATUS_TEST_D(sb->ORB_offset_hi_misc)) {
2233 SBP2_DEBUG("Dead bit set - "
2234 "initiating fetch agent reset");
2235 sbp2_agent_reset(scsi_id, 0);
2236 }
2237
2238 SBP2_ORB_DEBUG("completing command orb %p", &command->command_orb);
2239 }
2240
2241 /*
2242 * Check here to see if there are no commands in-use. If there
2243 * are none, we know that the fetch agent left the active state
2244 * _and_ that we did not reactivate it yet. Therefore clear
2245 * last_orb so that next time we write directly to the
2246 * ORB_POINTER register. That way the fetch agent does not need
2247 * to refetch the next_ORB.
2248 */
2249 spin_lock_irqsave(&scsi_id->sbp2_command_orb_lock, flags);
2250 if (list_empty(&scsi_id->sbp2_command_orb_inuse))
2251 scsi_id->last_orb = NULL;
2252 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
2253
2254 } else {
2255 /*
2256 * It's probably a login/logout/reconnect status.
2257 */
2258 if ((sb->ORB_offset_lo == scsi_id->reconnect_orb_dma) ||
2259 (sb->ORB_offset_lo == scsi_id->login_orb_dma) ||
2260 (sb->ORB_offset_lo == scsi_id->query_logins_orb_dma) ||
2261 (sb->ORB_offset_lo == scsi_id->logout_orb_dma)) {
2262 scsi_id->access_complete = 1;
2263 wake_up_interruptible(&access_wq);
2264 }
2265 }
2266
2267 if (SCpnt) {
2268 SBP2_DEBUG("Completing SCSI command");
2269 sbp2scsi_complete_command(scsi_id, scsi_status, SCpnt,
2270 command->Current_done);
2271 SBP2_ORB_DEBUG("command orb completed");
2272 }
2273
2274 return RCODE_COMPLETE;
2275 }
2276
2277 /**************************************
2278 * SCSI interface related section
2279 **************************************/
2280
2281 /*
2282 * This routine is the main request entry routine for doing I/O. It is
2283 * called from the scsi stack directly.
2284 */
2285 static int sbp2scsi_queuecommand(struct scsi_cmnd *SCpnt,
2286 void (*done)(struct scsi_cmnd *))
2287 {
2288 struct scsi_id_instance_data *scsi_id =
2289 (struct scsi_id_instance_data *)SCpnt->device->host->hostdata[0];
2290 struct sbp2scsi_host_info *hi;
2291 int result = DID_NO_CONNECT << 16;
2292
2293 SBP2_DEBUG_ENTER();
2294 #if (CONFIG_IEEE1394_SBP2_DEBUG >= 2) || defined(CONFIG_IEEE1394_SBP2_PACKET_DUMP)
2295 scsi_print_command(SCpnt);
2296 #endif
2297
2298 if (!sbp2util_node_is_available(scsi_id))
2299 goto done;
2300
2301 hi = scsi_id->hi;
2302
2303 if (!hi) {
2304 SBP2_ERR("sbp2scsi_host_info is NULL - this is bad!");
2305 goto done;
2306 }
2307
2308 /*
2309 * Until we handle multiple luns, just return selection time-out
2310 * to any IO directed at non-zero LUNs
2311 */
2312 if (SCpnt->device->lun)
2313 goto done;
2314
2315 /*
2316 * Check for request sense command, and handle it here
2317 * (autorequest sense)
2318 */
2319 if (SCpnt->cmnd[0] == REQUEST_SENSE) {
2320 SBP2_DEBUG("REQUEST_SENSE");
2321 memcpy(SCpnt->request_buffer, SCpnt->sense_buffer, SCpnt->request_bufflen);
2322 memset(SCpnt->sense_buffer, 0, sizeof(SCpnt->sense_buffer));
2323 sbp2scsi_complete_command(scsi_id, SBP2_SCSI_STATUS_GOOD, SCpnt, done);
2324 return 0;
2325 }
2326
2327 /*
2328 * Check to see if we are in the middle of a bus reset.
2329 */
2330 if (!hpsb_node_entry_valid(scsi_id->ne)) {
2331 SBP2_ERR("Bus reset in progress - rejecting command");
2332 result = DID_BUS_BUSY << 16;
2333 goto done;
2334 }
2335
2336 /*
2337 * Bidirectional commands are not yet implemented,
2338 * and unknown transfer direction not handled.
2339 */
2340 if (SCpnt->sc_data_direction == DMA_BIDIRECTIONAL) {
2341 SBP2_ERR("Cannot handle DMA_BIDIRECTIONAL - rejecting command");
2342 result = DID_ERROR << 16;
2343 goto done;
2344 }
2345
2346 /*
2347 * Try and send our SCSI command
2348 */
2349 if (sbp2_send_command(scsi_id, SCpnt, done)) {
2350 SBP2_ERR("Error sending SCSI command");
2351 sbp2scsi_complete_command(scsi_id, SBP2_SCSI_STATUS_SELECTION_TIMEOUT,
2352 SCpnt, done);
2353 }
2354 return 0;
2355
2356 done:
2357 SCpnt->result = result;
2358 done(SCpnt);
2359 return 0;
2360 }
2361
2362 /*
2363 * This function is called in order to complete all outstanding SBP-2
2364 * commands (in case of resets, etc.).
2365 */
2366 static void sbp2scsi_complete_all_commands(struct scsi_id_instance_data *scsi_id,
2367 u32 status)
2368 {
2369 struct sbp2scsi_host_info *hi = scsi_id->hi;
2370 struct list_head *lh;
2371 struct sbp2_command_info *command;
2372 unsigned long flags;
2373
2374 SBP2_DEBUG_ENTER();
2375
2376 spin_lock_irqsave(&scsi_id->sbp2_command_orb_lock, flags);
2377 while (!list_empty(&scsi_id->sbp2_command_orb_inuse)) {
2378 SBP2_DEBUG("Found pending command to complete");
2379 lh = scsi_id->sbp2_command_orb_inuse.next;
2380 command = list_entry(lh, struct sbp2_command_info, list);
2381 pci_dma_sync_single_for_cpu(hi->host->pdev, command->command_orb_dma,
2382 sizeof(struct sbp2_command_orb),
2383 PCI_DMA_TODEVICE);
2384 pci_dma_sync_single_for_cpu(hi->host->pdev, command->sge_dma,
2385 sizeof(command->scatter_gather_element),
2386 PCI_DMA_BIDIRECTIONAL);
2387 sbp2util_mark_command_completed(scsi_id, command);
2388 if (command->Current_SCpnt) {
2389 command->Current_SCpnt->result = status << 16;
2390 command->Current_done(command->Current_SCpnt);
2391 }
2392 }
2393 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
2394
2395 return;
2396 }
2397
2398 /*
2399 * This function is called in order to complete a regular SBP-2 command.
2400 *
2401 * This can be called in interrupt context.
2402 */
2403 static void sbp2scsi_complete_command(struct scsi_id_instance_data *scsi_id,
2404 u32 scsi_status, struct scsi_cmnd *SCpnt,
2405 void (*done)(struct scsi_cmnd *))
2406 {
2407 SBP2_DEBUG_ENTER();
2408
2409 /*
2410 * Sanity
2411 */
2412 if (!SCpnt) {
2413 SBP2_ERR("SCpnt is NULL");
2414 return;
2415 }
2416
2417 /*
2418 * If a bus reset is in progress and there was an error, don't
2419 * complete the command, just let it get retried at the end of the
2420 * bus reset.
2421 */
2422 if (!hpsb_node_entry_valid(scsi_id->ne)
2423 && (scsi_status != SBP2_SCSI_STATUS_GOOD)) {
2424 SBP2_ERR("Bus reset in progress - retry command later");
2425 return;
2426 }
2427
2428 /*
2429 * Switch on scsi status
2430 */
2431 switch (scsi_status) {
2432 case SBP2_SCSI_STATUS_GOOD:
2433 SCpnt->result = DID_OK << 16;
2434 break;
2435
2436 case SBP2_SCSI_STATUS_BUSY:
2437 SBP2_ERR("SBP2_SCSI_STATUS_BUSY");
2438 SCpnt->result = DID_BUS_BUSY << 16;
2439 break;
2440
2441 case SBP2_SCSI_STATUS_CHECK_CONDITION:
2442 SBP2_DEBUG("SBP2_SCSI_STATUS_CHECK_CONDITION");
2443 SCpnt->result = CHECK_CONDITION << 1 | DID_OK << 16;
2444 #if CONFIG_IEEE1394_SBP2_DEBUG >= 1
2445 scsi_print_command(SCpnt);
2446 scsi_print_sense(SBP2_DEVICE_NAME, SCpnt);
2447 #endif
2448 break;
2449
2450 case SBP2_SCSI_STATUS_SELECTION_TIMEOUT:
2451 SBP2_ERR("SBP2_SCSI_STATUS_SELECTION_TIMEOUT");
2452 SCpnt->result = DID_NO_CONNECT << 16;
2453 scsi_print_command(SCpnt);
2454 break;
2455
2456 case SBP2_SCSI_STATUS_CONDITION_MET:
2457 case SBP2_SCSI_STATUS_RESERVATION_CONFLICT:
2458 case SBP2_SCSI_STATUS_COMMAND_TERMINATED:
2459 SBP2_ERR("Bad SCSI status = %x", scsi_status);
2460 SCpnt->result = DID_ERROR << 16;
2461 scsi_print_command(SCpnt);
2462 break;
2463
2464 default:
2465 SBP2_ERR("Unsupported SCSI status = %x", scsi_status);
2466 SCpnt->result = DID_ERROR << 16;
2467 }
2468
2469 /*
2470 * If a bus reset is in progress and there was an error, complete
2471 * the command as busy so that it will get retried.
2472 */
2473 if (!hpsb_node_entry_valid(scsi_id->ne)
2474 && (scsi_status != SBP2_SCSI_STATUS_GOOD)) {
2475 SBP2_ERR("Completing command with busy (bus reset)");
2476 SCpnt->result = DID_BUS_BUSY << 16;
2477 }
2478
2479 /*
2480 * If a unit attention occurs, return busy status so it gets
2481 * retried... it could have happened because of a 1394 bus reset
2482 * or hot-plug...
2483 * XXX DID_BUS_BUSY is actually a bad idea because it will defy
2484 * the scsi layer's retry logic.
2485 */
2486 #if 0
2487 if ((scsi_status == SBP2_SCSI_STATUS_CHECK_CONDITION) &&
2488 (SCpnt->sense_buffer[2] == UNIT_ATTENTION)) {
2489 SBP2_DEBUG("UNIT ATTENTION - return busy");
2490 SCpnt->result = DID_BUS_BUSY << 16;
2491 }
2492 #endif
2493
2494 /*
2495 * Tell scsi stack that we're done with this command
2496 */
2497 done(SCpnt);
2498 }
2499
2500 static int sbp2scsi_slave_alloc(struct scsi_device *sdev)
2501 {
2502 struct scsi_id_instance_data *scsi_id =
2503 (struct scsi_id_instance_data *)sdev->host->hostdata[0];
2504
2505 scsi_id->sdev = sdev;
2506
2507 if (scsi_id->workarounds & SBP2_WORKAROUND_INQUIRY_36)
2508 sdev->inquiry_len = 36;
2509 return 0;
2510 }
2511
2512 static int sbp2scsi_slave_configure(struct scsi_device *sdev)
2513 {
2514 struct scsi_id_instance_data *scsi_id =
2515 (struct scsi_id_instance_data *)sdev->host->hostdata[0];
2516
2517 blk_queue_dma_alignment(sdev->request_queue, (512 - 1));
2518 sdev->use_10_for_rw = 1;
2519 sdev->use_10_for_ms = 1;
2520
2521 if (sdev->type == TYPE_DISK &&
2522 scsi_id->workarounds & SBP2_WORKAROUND_MODE_SENSE_8)
2523 sdev->skip_ms_page_8 = 1;
2524 if (scsi_id->workarounds & SBP2_WORKAROUND_FIX_CAPACITY)
2525 sdev->fix_capacity = 1;
2526 if (scsi_id->ne->guid_vendor_id == 0x0010b9 && /* Maxtor's OUI */
2527 (sdev->type == TYPE_DISK || sdev->type == TYPE_RBC))
2528 sdev->allow_restart = 1;
2529 return 0;
2530 }
2531
2532 static void sbp2scsi_slave_destroy(struct scsi_device *sdev)
2533 {
2534 ((struct scsi_id_instance_data *)sdev->host->hostdata[0])->sdev = NULL;
2535 return;
2536 }
2537
2538 /*
2539 * Called by scsi stack when something has really gone wrong. Usually
2540 * called when a command has timed-out for some reason.
2541 */
2542 static int sbp2scsi_abort(struct scsi_cmnd *SCpnt)
2543 {
2544 struct scsi_id_instance_data *scsi_id =
2545 (struct scsi_id_instance_data *)SCpnt->device->host->hostdata[0];
2546 struct sbp2scsi_host_info *hi = scsi_id->hi;
2547 struct sbp2_command_info *command;
2548 unsigned long flags;
2549
2550 SBP2_ERR("aborting sbp2 command");
2551 scsi_print_command(SCpnt);
2552
2553 if (sbp2util_node_is_available(scsi_id)) {
2554
2555 /*
2556 * Right now, just return any matching command structures
2557 * to the free pool.
2558 */
2559 spin_lock_irqsave(&scsi_id->sbp2_command_orb_lock, flags);
2560 command = sbp2util_find_command_for_SCpnt(scsi_id, SCpnt);
2561 if (command) {
2562 SBP2_DEBUG("Found command to abort");
2563 pci_dma_sync_single_for_cpu(hi->host->pdev,
2564 command->command_orb_dma,
2565 sizeof(struct sbp2_command_orb),
2566 PCI_DMA_TODEVICE);
2567 pci_dma_sync_single_for_cpu(hi->host->pdev,
2568 command->sge_dma,
2569 sizeof(command->scatter_gather_element),
2570 PCI_DMA_BIDIRECTIONAL);
2571 sbp2util_mark_command_completed(scsi_id, command);
2572 if (command->Current_SCpnt) {
2573 command->Current_SCpnt->result = DID_ABORT << 16;
2574 command->Current_done(command->Current_SCpnt);
2575 }
2576 }
2577 spin_unlock_irqrestore(&scsi_id->sbp2_command_orb_lock, flags);
2578
2579 /*
2580 * Initiate a fetch agent reset.
2581 */
2582 sbp2_agent_reset(scsi_id, 1);
2583 sbp2scsi_complete_all_commands(scsi_id, DID_BUS_BUSY);
2584 }
2585
2586 return SUCCESS;
2587 }
2588
2589 /*
2590 * Called by scsi stack when something has really gone wrong.
2591 */
2592 static int sbp2scsi_reset(struct scsi_cmnd *SCpnt)
2593 {
2594 struct scsi_id_instance_data *scsi_id =
2595 (struct scsi_id_instance_data *)SCpnt->device->host->hostdata[0];
2596
2597 SBP2_ERR("reset requested");
2598
2599 if (sbp2util_node_is_available(scsi_id)) {
2600 SBP2_ERR("Generating sbp2 fetch agent reset");
2601 sbp2_agent_reset(scsi_id, 1);
2602 }
2603
2604 return SUCCESS;
2605 }
2606
2607 static ssize_t sbp2_sysfs_ieee1394_id_show(struct device *dev,
2608 struct device_attribute *attr,
2609 char *buf)
2610 {
2611 struct scsi_device *sdev;
2612 struct scsi_id_instance_data *scsi_id;
2613 int lun;
2614
2615 if (!(sdev = to_scsi_device(dev)))
2616 return 0;
2617
2618 if (!(scsi_id = (struct scsi_id_instance_data *)sdev->host->hostdata[0]))
2619 return 0;
2620
2621 lun = ORB_SET_LUN(scsi_id->sbp2_lun);
2622
2623 return sprintf(buf, "%016Lx:%d:%d\n", (unsigned long long)scsi_id->ne->guid,
2624 scsi_id->ud->id, lun);
2625 }
2626 static DEVICE_ATTR(ieee1394_id, S_IRUGO, sbp2_sysfs_ieee1394_id_show, NULL);
2627
2628 static struct device_attribute *sbp2_sysfs_sdev_attrs[] = {
2629 &dev_attr_ieee1394_id,
2630 NULL
2631 };
2632
2633 MODULE_AUTHOR("Ben Collins <bcollins@debian.org>");
2634 MODULE_DESCRIPTION("IEEE-1394 SBP-2 protocol driver");
2635 MODULE_SUPPORTED_DEVICE(SBP2_DEVICE_NAME);
2636 MODULE_LICENSE("GPL");
2637
2638 /* SCSI host template */
2639 static struct scsi_host_template scsi_driver_template = {
2640 .module = THIS_MODULE,
2641 .name = "SBP-2 IEEE-1394",
2642 .proc_name = SBP2_DEVICE_NAME,
2643 .queuecommand = sbp2scsi_queuecommand,
2644 .eh_abort_handler = sbp2scsi_abort,
2645 .eh_device_reset_handler = sbp2scsi_reset,
2646 .slave_alloc = sbp2scsi_slave_alloc,
2647 .slave_configure = sbp2scsi_slave_configure,
2648 .slave_destroy = sbp2scsi_slave_destroy,
2649 .this_id = -1,
2650 .sg_tablesize = SG_ALL,
2651 .use_clustering = ENABLE_CLUSTERING,
2652 .cmd_per_lun = SBP2_MAX_CMDS,
2653 .can_queue = SBP2_MAX_CMDS,
2654 .emulated = 1,
2655 .sdev_attrs = sbp2_sysfs_sdev_attrs,
2656 };
2657
2658 static int sbp2_module_init(void)
2659 {
2660 int ret;
2661
2662 SBP2_DEBUG_ENTER();
2663
2664 /* Module load debug option to force one command at a time (serializing I/O) */
2665 if (serialize_io) {
2666 SBP2_INFO("Driver forced to serialize I/O (serialize_io=1)");
2667 SBP2_INFO("Try serialize_io=0 for better performance");
2668 scsi_driver_template.can_queue = 1;
2669 scsi_driver_template.cmd_per_lun = 1;
2670 }
2671
2672 if (sbp2_default_workarounds & SBP2_WORKAROUND_128K_MAX_TRANS &&
2673 (max_sectors * 512) > (128 * 1024))
2674 max_sectors = 128 * 1024 / 512;
2675 scsi_driver_template.max_sectors = max_sectors;
2676
2677 /* Register our high level driver with 1394 stack */
2678 hpsb_register_highlevel(&sbp2_highlevel);
2679
2680 ret = hpsb_register_protocol(&sbp2_driver);
2681 if (ret) {
2682 SBP2_ERR("Failed to register protocol");
2683 hpsb_unregister_highlevel(&sbp2_highlevel);
2684 return ret;
2685 }
2686
2687 return 0;
2688 }
2689
2690 static void __exit sbp2_module_exit(void)
2691 {
2692 SBP2_DEBUG_ENTER();
2693
2694 hpsb_unregister_protocol(&sbp2_driver);
2695
2696 hpsb_unregister_highlevel(&sbp2_highlevel);
2697 }
2698
2699 module_init(sbp2_module_init);
2700 module_exit(sbp2_module_exit);
This page took 0.084927 seconds and 6 git commands to generate.