Commit | Line | Data |
---|---|---|
072b91f9 BK |
1 | /* |
2 | * ibmvfc.c -- driver for IBM Power Virtual Fibre Channel Adapter | |
3 | * | |
4 | * Written By: Brian King <brking@linux.vnet.ibm.com>, IBM Corporation | |
5 | * | |
6 | * Copyright (C) IBM Corporation, 2008 | |
7 | * | |
8 | * This program is free software; you can redistribute it and/or modify | |
9 | * it under the terms of the GNU General Public License as published by | |
10 | * the Free Software Foundation; either version 2 of the License, or | |
11 | * (at your option) any later version. | |
12 | * | |
13 | * This program is distributed in the hope that it will be useful, | |
14 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
15 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
16 | * GNU General Public License for more details. | |
17 | * | |
18 | * You should have received a copy of the GNU General Public License | |
19 | * along with this program; if not, write to the Free Software | |
20 | * Foundation, Inc., 59 Temple Place, Suite 330, Boston, MA 02111-1307 USA | |
21 | * | |
22 | */ | |
23 | ||
24 | #include <linux/module.h> | |
25 | #include <linux/moduleparam.h> | |
26 | #include <linux/dma-mapping.h> | |
27 | #include <linux/dmapool.h> | |
28 | #include <linux/delay.h> | |
29 | #include <linux/interrupt.h> | |
30 | #include <linux/kthread.h> | |
5a0e3ad6 | 31 | #include <linux/slab.h> |
072b91f9 | 32 | #include <linux/of.h> |
b0f4d4cf | 33 | #include <linux/pm.h> |
072b91f9 BK |
34 | #include <linux/stringify.h> |
35 | #include <asm/firmware.h> | |
36 | #include <asm/irq.h> | |
37 | #include <asm/vio.h> | |
38 | #include <scsi/scsi.h> | |
39 | #include <scsi/scsi_cmnd.h> | |
40 | #include <scsi/scsi_host.h> | |
41 | #include <scsi/scsi_device.h> | |
42 | #include <scsi/scsi_tcq.h> | |
43 | #include <scsi/scsi_transport_fc.h> | |
d31429e1 | 44 | #include <scsi/scsi_bsg_fc.h> |
072b91f9 BK |
45 | #include "ibmvfc.h" |
46 | ||
47 | static unsigned int init_timeout = IBMVFC_INIT_TIMEOUT; | |
48 | static unsigned int default_timeout = IBMVFC_DEFAULT_TIMEOUT; | |
1abf635d | 49 | static u64 max_lun = IBMVFC_MAX_LUN; |
072b91f9 BK |
50 | static unsigned int max_targets = IBMVFC_MAX_TARGETS; |
51 | static unsigned int max_requests = IBMVFC_MAX_REQUESTS_DEFAULT; | |
52 | static unsigned int disc_threads = IBMVFC_MAX_DISC_THREADS; | |
072b91f9 BK |
53 | static unsigned int ibmvfc_debug = IBMVFC_DEBUG; |
54 | static unsigned int log_level = IBMVFC_DEFAULT_LOG_LEVEL; | |
55 | static LIST_HEAD(ibmvfc_head); | |
56 | static DEFINE_SPINLOCK(ibmvfc_driver_lock); | |
57 | static struct scsi_transport_template *ibmvfc_transport_template; | |
58 | ||
59 | MODULE_DESCRIPTION("IBM Virtual Fibre Channel Driver"); | |
60 | MODULE_AUTHOR("Brian King <brking@linux.vnet.ibm.com>"); | |
61 | MODULE_LICENSE("GPL"); | |
62 | MODULE_VERSION(IBMVFC_DRIVER_VERSION); | |
63 | ||
64 | module_param_named(init_timeout, init_timeout, uint, S_IRUGO | S_IWUSR); | |
65 | MODULE_PARM_DESC(init_timeout, "Initialization timeout in seconds. " | |
66 | "[Default=" __stringify(IBMVFC_INIT_TIMEOUT) "]"); | |
67 | module_param_named(default_timeout, default_timeout, uint, S_IRUGO | S_IWUSR); | |
68 | MODULE_PARM_DESC(default_timeout, | |
69 | "Default timeout in seconds for initialization and EH commands. " | |
70 | "[Default=" __stringify(IBMVFC_DEFAULT_TIMEOUT) "]"); | |
71 | module_param_named(max_requests, max_requests, uint, S_IRUGO); | |
72 | MODULE_PARM_DESC(max_requests, "Maximum requests for this adapter. " | |
73 | "[Default=" __stringify(IBMVFC_MAX_REQUESTS_DEFAULT) "]"); | |
1abf635d | 74 | module_param_named(max_lun, max_lun, ullong, S_IRUGO); |
072b91f9 BK |
75 | MODULE_PARM_DESC(max_lun, "Maximum allowed LUN. " |
76 | "[Default=" __stringify(IBMVFC_MAX_LUN) "]"); | |
77 | module_param_named(max_targets, max_targets, uint, S_IRUGO); | |
78 | MODULE_PARM_DESC(max_targets, "Maximum allowed targets. " | |
79 | "[Default=" __stringify(IBMVFC_MAX_TARGETS) "]"); | |
545ef9a2 | 80 | module_param_named(disc_threads, disc_threads, uint, S_IRUGO); |
072b91f9 BK |
81 | MODULE_PARM_DESC(disc_threads, "Number of device discovery threads to use. " |
82 | "[Default=" __stringify(IBMVFC_MAX_DISC_THREADS) "]"); | |
83 | module_param_named(debug, ibmvfc_debug, uint, S_IRUGO | S_IWUSR); | |
84 | MODULE_PARM_DESC(debug, "Enable driver debug information. " | |
85 | "[Default=" __stringify(IBMVFC_DEBUG) "]"); | |
072b91f9 BK |
86 | module_param_named(log_level, log_level, uint, 0); |
87 | MODULE_PARM_DESC(log_level, "Set to 0 - 4 for increasing verbosity of device driver. " | |
88 | "[Default=" __stringify(IBMVFC_DEFAULT_LOG_LEVEL) "]"); | |
89 | ||
90 | static const struct { | |
91 | u16 status; | |
92 | u16 error; | |
93 | u8 result; | |
94 | u8 retry; | |
95 | int log; | |
96 | char *name; | |
97 | } cmd_status [] = { | |
98 | { IBMVFC_FABRIC_MAPPED, IBMVFC_UNABLE_TO_ESTABLISH, DID_ERROR, 1, 1, "unable to establish" }, | |
99 | { IBMVFC_FABRIC_MAPPED, IBMVFC_XPORT_FAULT, DID_OK, 1, 0, "transport fault" }, | |
100 | { IBMVFC_FABRIC_MAPPED, IBMVFC_CMD_TIMEOUT, DID_TIME_OUT, 1, 1, "command timeout" }, | |
752b3232 | 101 | { IBMVFC_FABRIC_MAPPED, IBMVFC_ENETDOWN, DID_TRANSPORT_DISRUPTED, 1, 1, "network down" }, |
072b91f9 BK |
102 | { IBMVFC_FABRIC_MAPPED, IBMVFC_HW_FAILURE, DID_ERROR, 1, 1, "hardware failure" }, |
103 | { IBMVFC_FABRIC_MAPPED, IBMVFC_LINK_DOWN_ERR, DID_REQUEUE, 0, 0, "link down" }, | |
104 | { IBMVFC_FABRIC_MAPPED, IBMVFC_LINK_DEAD_ERR, DID_ERROR, 0, 0, "link dead" }, | |
105 | { IBMVFC_FABRIC_MAPPED, IBMVFC_UNABLE_TO_REGISTER, DID_ERROR, 1, 1, "unable to register" }, | |
106 | { IBMVFC_FABRIC_MAPPED, IBMVFC_XPORT_BUSY, DID_BUS_BUSY, 1, 0, "transport busy" }, | |
107 | { IBMVFC_FABRIC_MAPPED, IBMVFC_XPORT_DEAD, DID_ERROR, 0, 1, "transport dead" }, | |
108 | { IBMVFC_FABRIC_MAPPED, IBMVFC_CONFIG_ERROR, DID_ERROR, 1, 1, "configuration error" }, | |
109 | { IBMVFC_FABRIC_MAPPED, IBMVFC_NAME_SERVER_FAIL, DID_ERROR, 1, 1, "name server failure" }, | |
497f9c50 | 110 | { IBMVFC_FABRIC_MAPPED, IBMVFC_LINK_HALTED, DID_REQUEUE, 1, 0, "link halted" }, |
072b91f9 BK |
111 | { IBMVFC_FABRIC_MAPPED, IBMVFC_XPORT_GENERAL, DID_OK, 1, 0, "general transport error" }, |
112 | ||
113 | { IBMVFC_VIOS_FAILURE, IBMVFC_CRQ_FAILURE, DID_REQUEUE, 1, 1, "CRQ failure" }, | |
114 | { IBMVFC_VIOS_FAILURE, IBMVFC_SW_FAILURE, DID_ERROR, 0, 1, "software failure" }, | |
752b3232 BK |
115 | { IBMVFC_VIOS_FAILURE, IBMVFC_INVALID_PARAMETER, DID_ERROR, 0, 1, "invalid parameter" }, |
116 | { IBMVFC_VIOS_FAILURE, IBMVFC_MISSING_PARAMETER, DID_ERROR, 0, 1, "missing parameter" }, | |
072b91f9 | 117 | { IBMVFC_VIOS_FAILURE, IBMVFC_HOST_IO_BUS, DID_ERROR, 1, 1, "host I/O bus failure" }, |
752b3232 BK |
118 | { IBMVFC_VIOS_FAILURE, IBMVFC_TRANS_CANCELLED, DID_ERROR, 0, 1, "transaction cancelled" }, |
119 | { IBMVFC_VIOS_FAILURE, IBMVFC_TRANS_CANCELLED_IMPLICIT, DID_ERROR, 0, 1, "transaction cancelled implicit" }, | |
072b91f9 | 120 | { IBMVFC_VIOS_FAILURE, IBMVFC_INSUFFICIENT_RESOURCE, DID_REQUEUE, 1, 1, "insufficient resources" }, |
646d3857 | 121 | { IBMVFC_VIOS_FAILURE, IBMVFC_PLOGI_REQUIRED, DID_ERROR, 0, 1, "port login required" }, |
072b91f9 BK |
122 | { IBMVFC_VIOS_FAILURE, IBMVFC_COMMAND_FAILED, DID_ERROR, 1, 1, "command failed" }, |
123 | ||
124 | { IBMVFC_FC_FAILURE, IBMVFC_INVALID_ELS_CMD_CODE, DID_ERROR, 0, 1, "invalid ELS command code" }, | |
125 | { IBMVFC_FC_FAILURE, IBMVFC_INVALID_VERSION, DID_ERROR, 0, 1, "invalid version level" }, | |
126 | { IBMVFC_FC_FAILURE, IBMVFC_LOGICAL_ERROR, DID_ERROR, 1, 1, "logical error" }, | |
127 | { IBMVFC_FC_FAILURE, IBMVFC_INVALID_CT_IU_SIZE, DID_ERROR, 0, 1, "invalid CT_IU size" }, | |
128 | { IBMVFC_FC_FAILURE, IBMVFC_LOGICAL_BUSY, DID_REQUEUE, 1, 0, "logical busy" }, | |
129 | { IBMVFC_FC_FAILURE, IBMVFC_PROTOCOL_ERROR, DID_ERROR, 1, 1, "protocol error" }, | |
130 | { IBMVFC_FC_FAILURE, IBMVFC_UNABLE_TO_PERFORM_REQ, DID_ERROR, 1, 1, "unable to perform request" }, | |
131 | { IBMVFC_FC_FAILURE, IBMVFC_CMD_NOT_SUPPORTED, DID_ERROR, 0, 0, "command not supported" }, | |
132 | { IBMVFC_FC_FAILURE, IBMVFC_SERVER_NOT_AVAIL, DID_ERROR, 0, 1, "server not available" }, | |
133 | { IBMVFC_FC_FAILURE, IBMVFC_CMD_IN_PROGRESS, DID_ERROR, 0, 1, "command already in progress" }, | |
134 | { IBMVFC_FC_FAILURE, IBMVFC_VENDOR_SPECIFIC, DID_ERROR, 1, 1, "vendor specific" }, | |
135 | ||
136 | { IBMVFC_FC_SCSI_ERROR, 0, DID_OK, 1, 0, "SCSI error" }, | |
137 | }; | |
138 | ||
139 | static void ibmvfc_npiv_login(struct ibmvfc_host *); | |
140 | static void ibmvfc_tgt_send_prli(struct ibmvfc_target *); | |
141 | static void ibmvfc_tgt_send_plogi(struct ibmvfc_target *); | |
142 | static void ibmvfc_tgt_query_target(struct ibmvfc_target *); | |
79111d08 | 143 | static void ibmvfc_npiv_logout(struct ibmvfc_host *); |
072b91f9 BK |
144 | |
145 | static const char *unknown_error = "unknown error"; | |
146 | ||
147 | #ifdef CONFIG_SCSI_IBMVFC_TRACE | |
148 | /** | |
149 | * ibmvfc_trc_start - Log a start trace entry | |
150 | * @evt: ibmvfc event struct | |
151 | * | |
152 | **/ | |
153 | static void ibmvfc_trc_start(struct ibmvfc_event *evt) | |
154 | { | |
155 | struct ibmvfc_host *vhost = evt->vhost; | |
156 | struct ibmvfc_cmd *vfc_cmd = &evt->iu.cmd; | |
157 | struct ibmvfc_mad_common *mad = &evt->iu.mad_common; | |
158 | struct ibmvfc_trace_entry *entry; | |
159 | ||
160 | entry = &vhost->trace[vhost->trace_index++]; | |
161 | entry->evt = evt; | |
162 | entry->time = jiffies; | |
163 | entry->fmt = evt->crq.format; | |
164 | entry->type = IBMVFC_TRC_START; | |
165 | ||
166 | switch (entry->fmt) { | |
167 | case IBMVFC_CMD_FORMAT: | |
168 | entry->op_code = vfc_cmd->iu.cdb[0]; | |
0aab6c3f | 169 | entry->scsi_id = be64_to_cpu(vfc_cmd->tgt_scsi_id); |
072b91f9 BK |
170 | entry->lun = scsilun_to_int(&vfc_cmd->iu.lun); |
171 | entry->tmf_flags = vfc_cmd->iu.tmf_flags; | |
0aab6c3f | 172 | entry->u.start.xfer_len = be32_to_cpu(vfc_cmd->iu.xfer_len); |
072b91f9 BK |
173 | break; |
174 | case IBMVFC_MAD_FORMAT: | |
0aab6c3f | 175 | entry->op_code = be32_to_cpu(mad->opcode); |
072b91f9 BK |
176 | break; |
177 | default: | |
178 | break; | |
179 | }; | |
180 | } | |
181 | ||
182 | /** | |
183 | * ibmvfc_trc_end - Log an end trace entry | |
184 | * @evt: ibmvfc event struct | |
185 | * | |
186 | **/ | |
187 | static void ibmvfc_trc_end(struct ibmvfc_event *evt) | |
188 | { | |
189 | struct ibmvfc_host *vhost = evt->vhost; | |
190 | struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; | |
191 | struct ibmvfc_mad_common *mad = &evt->xfer_iu->mad_common; | |
192 | struct ibmvfc_trace_entry *entry = &vhost->trace[vhost->trace_index++]; | |
193 | ||
194 | entry->evt = evt; | |
195 | entry->time = jiffies; | |
196 | entry->fmt = evt->crq.format; | |
197 | entry->type = IBMVFC_TRC_END; | |
198 | ||
199 | switch (entry->fmt) { | |
200 | case IBMVFC_CMD_FORMAT: | |
201 | entry->op_code = vfc_cmd->iu.cdb[0]; | |
0aab6c3f | 202 | entry->scsi_id = be64_to_cpu(vfc_cmd->tgt_scsi_id); |
072b91f9 BK |
203 | entry->lun = scsilun_to_int(&vfc_cmd->iu.lun); |
204 | entry->tmf_flags = vfc_cmd->iu.tmf_flags; | |
0aab6c3f TD |
205 | entry->u.end.status = be16_to_cpu(vfc_cmd->status); |
206 | entry->u.end.error = be16_to_cpu(vfc_cmd->error); | |
072b91f9 BK |
207 | entry->u.end.fcp_rsp_flags = vfc_cmd->rsp.flags; |
208 | entry->u.end.rsp_code = vfc_cmd->rsp.data.info.rsp_code; | |
209 | entry->u.end.scsi_status = vfc_cmd->rsp.scsi_status; | |
210 | break; | |
211 | case IBMVFC_MAD_FORMAT: | |
0aab6c3f TD |
212 | entry->op_code = be32_to_cpu(mad->opcode); |
213 | entry->u.end.status = be16_to_cpu(mad->status); | |
072b91f9 BK |
214 | break; |
215 | default: | |
216 | break; | |
217 | ||
218 | }; | |
219 | } | |
220 | ||
221 | #else | |
222 | #define ibmvfc_trc_start(evt) do { } while (0) | |
223 | #define ibmvfc_trc_end(evt) do { } while (0) | |
224 | #endif | |
225 | ||
226 | /** | |
227 | * ibmvfc_get_err_index - Find the index into cmd_status for the fcp response | |
228 | * @status: status / error class | |
229 | * @error: error | |
230 | * | |
231 | * Return value: | |
232 | * index into cmd_status / -EINVAL on failure | |
233 | **/ | |
234 | static int ibmvfc_get_err_index(u16 status, u16 error) | |
235 | { | |
236 | int i; | |
237 | ||
238 | for (i = 0; i < ARRAY_SIZE(cmd_status); i++) | |
239 | if ((cmd_status[i].status & status) == cmd_status[i].status && | |
240 | cmd_status[i].error == error) | |
241 | return i; | |
242 | ||
243 | return -EINVAL; | |
244 | } | |
245 | ||
246 | /** | |
247 | * ibmvfc_get_cmd_error - Find the error description for the fcp response | |
248 | * @status: status / error class | |
249 | * @error: error | |
250 | * | |
251 | * Return value: | |
252 | * error description string | |
253 | **/ | |
254 | static const char *ibmvfc_get_cmd_error(u16 status, u16 error) | |
255 | { | |
256 | int rc = ibmvfc_get_err_index(status, error); | |
257 | if (rc >= 0) | |
258 | return cmd_status[rc].name; | |
259 | return unknown_error; | |
260 | } | |
261 | ||
262 | /** | |
263 | * ibmvfc_get_err_result - Find the scsi status to return for the fcp response | |
264 | * @vfc_cmd: ibmvfc command struct | |
265 | * | |
266 | * Return value: | |
267 | * SCSI result value to return for completed command | |
268 | **/ | |
269 | static int ibmvfc_get_err_result(struct ibmvfc_cmd *vfc_cmd) | |
270 | { | |
271 | int err; | |
272 | struct ibmvfc_fcp_rsp *rsp = &vfc_cmd->rsp; | |
0aab6c3f | 273 | int fc_rsp_len = be32_to_cpu(rsp->fcp_rsp_len); |
072b91f9 BK |
274 | |
275 | if ((rsp->flags & FCP_RSP_LEN_VALID) && | |
4a2837d4 | 276 | ((fc_rsp_len && fc_rsp_len != 4 && fc_rsp_len != 8) || |
072b91f9 BK |
277 | rsp->data.info.rsp_code)) |
278 | return DID_ERROR << 16; | |
279 | ||
0aab6c3f | 280 | err = ibmvfc_get_err_index(be16_to_cpu(vfc_cmd->status), be16_to_cpu(vfc_cmd->error)); |
072b91f9 BK |
281 | if (err >= 0) |
282 | return rsp->scsi_status | (cmd_status[err].result << 16); | |
283 | return rsp->scsi_status | (DID_ERROR << 16); | |
284 | } | |
285 | ||
286 | /** | |
287 | * ibmvfc_retry_cmd - Determine if error status is retryable | |
288 | * @status: status / error class | |
289 | * @error: error | |
290 | * | |
291 | * Return value: | |
292 | * 1 if error should be retried / 0 if it should not | |
293 | **/ | |
294 | static int ibmvfc_retry_cmd(u16 status, u16 error) | |
295 | { | |
296 | int rc = ibmvfc_get_err_index(status, error); | |
297 | ||
298 | if (rc >= 0) | |
299 | return cmd_status[rc].retry; | |
300 | return 1; | |
301 | } | |
302 | ||
303 | static const char *unknown_fc_explain = "unknown fc explain"; | |
304 | ||
305 | static const struct { | |
306 | u16 fc_explain; | |
307 | char *name; | |
308 | } ls_explain [] = { | |
309 | { 0x00, "no additional explanation" }, | |
310 | { 0x01, "service parameter error - options" }, | |
311 | { 0x03, "service parameter error - initiator control" }, | |
312 | { 0x05, "service parameter error - recipient control" }, | |
313 | { 0x07, "service parameter error - received data field size" }, | |
314 | { 0x09, "service parameter error - concurrent seq" }, | |
315 | { 0x0B, "service parameter error - credit" }, | |
316 | { 0x0D, "invalid N_Port/F_Port_Name" }, | |
317 | { 0x0E, "invalid node/Fabric Name" }, | |
318 | { 0x0F, "invalid common service parameters" }, | |
319 | { 0x11, "invalid association header" }, | |
320 | { 0x13, "association header required" }, | |
321 | { 0x15, "invalid originator S_ID" }, | |
322 | { 0x17, "invalid OX_ID-RX-ID combination" }, | |
323 | { 0x19, "command (request) already in progress" }, | |
324 | { 0x1E, "N_Port Login requested" }, | |
325 | { 0x1F, "Invalid N_Port_ID" }, | |
326 | }; | |
327 | ||
328 | static const struct { | |
329 | u16 fc_explain; | |
330 | char *name; | |
331 | } gs_explain [] = { | |
332 | { 0x00, "no additional explanation" }, | |
333 | { 0x01, "port identifier not registered" }, | |
334 | { 0x02, "port name not registered" }, | |
335 | { 0x03, "node name not registered" }, | |
336 | { 0x04, "class of service not registered" }, | |
337 | { 0x06, "initial process associator not registered" }, | |
338 | { 0x07, "FC-4 TYPEs not registered" }, | |
339 | { 0x08, "symbolic port name not registered" }, | |
340 | { 0x09, "symbolic node name not registered" }, | |
341 | { 0x0A, "port type not registered" }, | |
342 | { 0xF0, "authorization exception" }, | |
343 | { 0xF1, "authentication exception" }, | |
344 | { 0xF2, "data base full" }, | |
345 | { 0xF3, "data base empty" }, | |
346 | { 0xF4, "processing request" }, | |
347 | { 0xF5, "unable to verify connection" }, | |
348 | { 0xF6, "devices not in a common zone" }, | |
349 | }; | |
350 | ||
351 | /** | |
352 | * ibmvfc_get_ls_explain - Return the FC Explain description text | |
353 | * @status: FC Explain status | |
354 | * | |
355 | * Returns: | |
356 | * error string | |
357 | **/ | |
358 | static const char *ibmvfc_get_ls_explain(u16 status) | |
359 | { | |
360 | int i; | |
361 | ||
362 | for (i = 0; i < ARRAY_SIZE(ls_explain); i++) | |
363 | if (ls_explain[i].fc_explain == status) | |
364 | return ls_explain[i].name; | |
365 | ||
366 | return unknown_fc_explain; | |
367 | } | |
368 | ||
369 | /** | |
370 | * ibmvfc_get_gs_explain - Return the FC Explain description text | |
371 | * @status: FC Explain status | |
372 | * | |
373 | * Returns: | |
374 | * error string | |
375 | **/ | |
376 | static const char *ibmvfc_get_gs_explain(u16 status) | |
377 | { | |
378 | int i; | |
379 | ||
380 | for (i = 0; i < ARRAY_SIZE(gs_explain); i++) | |
381 | if (gs_explain[i].fc_explain == status) | |
382 | return gs_explain[i].name; | |
383 | ||
384 | return unknown_fc_explain; | |
385 | } | |
386 | ||
387 | static const struct { | |
388 | enum ibmvfc_fc_type fc_type; | |
389 | char *name; | |
390 | } fc_type [] = { | |
391 | { IBMVFC_FABRIC_REJECT, "fabric reject" }, | |
392 | { IBMVFC_PORT_REJECT, "port reject" }, | |
393 | { IBMVFC_LS_REJECT, "ELS reject" }, | |
394 | { IBMVFC_FABRIC_BUSY, "fabric busy" }, | |
395 | { IBMVFC_PORT_BUSY, "port busy" }, | |
396 | { IBMVFC_BASIC_REJECT, "basic reject" }, | |
397 | }; | |
398 | ||
399 | static const char *unknown_fc_type = "unknown fc type"; | |
400 | ||
401 | /** | |
402 | * ibmvfc_get_fc_type - Return the FC Type description text | |
403 | * @status: FC Type error status | |
404 | * | |
405 | * Returns: | |
406 | * error string | |
407 | **/ | |
408 | static const char *ibmvfc_get_fc_type(u16 status) | |
409 | { | |
410 | int i; | |
411 | ||
412 | for (i = 0; i < ARRAY_SIZE(fc_type); i++) | |
413 | if (fc_type[i].fc_type == status) | |
414 | return fc_type[i].name; | |
415 | ||
416 | return unknown_fc_type; | |
417 | } | |
418 | ||
419 | /** | |
420 | * ibmvfc_set_tgt_action - Set the next init action for the target | |
421 | * @tgt: ibmvfc target struct | |
422 | * @action: action to perform | |
423 | * | |
424 | **/ | |
425 | static void ibmvfc_set_tgt_action(struct ibmvfc_target *tgt, | |
426 | enum ibmvfc_target_action action) | |
427 | { | |
428 | switch (tgt->action) { | |
429 | case IBMVFC_TGT_ACTION_DEL_RPORT: | |
d5da3040 BK |
430 | if (action == IBMVFC_TGT_ACTION_DELETED_RPORT) |
431 | tgt->action = action; | |
432 | case IBMVFC_TGT_ACTION_DELETED_RPORT: | |
072b91f9 BK |
433 | break; |
434 | default: | |
43c8da90 BK |
435 | if (action == IBMVFC_TGT_ACTION_DEL_RPORT) |
436 | tgt->add_rport = 0; | |
072b91f9 BK |
437 | tgt->action = action; |
438 | break; | |
439 | } | |
440 | } | |
441 | ||
442 | /** | |
443 | * ibmvfc_set_host_state - Set the state for the host | |
444 | * @vhost: ibmvfc host struct | |
445 | * @state: state to set host to | |
446 | * | |
447 | * Returns: | |
448 | * 0 if state changed / non-zero if not changed | |
449 | **/ | |
450 | static int ibmvfc_set_host_state(struct ibmvfc_host *vhost, | |
451 | enum ibmvfc_host_state state) | |
452 | { | |
453 | int rc = 0; | |
454 | ||
455 | switch (vhost->state) { | |
456 | case IBMVFC_HOST_OFFLINE: | |
457 | rc = -EINVAL; | |
458 | break; | |
459 | default: | |
460 | vhost->state = state; | |
461 | break; | |
462 | }; | |
463 | ||
464 | return rc; | |
465 | } | |
466 | ||
467 | /** | |
468 | * ibmvfc_set_host_action - Set the next init action for the host | |
469 | * @vhost: ibmvfc host struct | |
470 | * @action: action to perform | |
471 | * | |
472 | **/ | |
473 | static void ibmvfc_set_host_action(struct ibmvfc_host *vhost, | |
474 | enum ibmvfc_host_action action) | |
475 | { | |
476 | switch (action) { | |
477 | case IBMVFC_HOST_ACTION_ALLOC_TGTS: | |
478 | if (vhost->action == IBMVFC_HOST_ACTION_INIT_WAIT) | |
479 | vhost->action = action; | |
480 | break; | |
79111d08 BK |
481 | case IBMVFC_HOST_ACTION_LOGO_WAIT: |
482 | if (vhost->action == IBMVFC_HOST_ACTION_LOGO) | |
483 | vhost->action = action; | |
484 | break; | |
072b91f9 BK |
485 | case IBMVFC_HOST_ACTION_INIT_WAIT: |
486 | if (vhost->action == IBMVFC_HOST_ACTION_INIT) | |
487 | vhost->action = action; | |
488 | break; | |
489 | case IBMVFC_HOST_ACTION_QUERY: | |
490 | switch (vhost->action) { | |
491 | case IBMVFC_HOST_ACTION_INIT_WAIT: | |
492 | case IBMVFC_HOST_ACTION_NONE: | |
43c8da90 | 493 | case IBMVFC_HOST_ACTION_TGT_DEL_FAILED: |
072b91f9 BK |
494 | vhost->action = action; |
495 | break; | |
496 | default: | |
497 | break; | |
498 | }; | |
499 | break; | |
500 | case IBMVFC_HOST_ACTION_TGT_INIT: | |
501 | if (vhost->action == IBMVFC_HOST_ACTION_ALLOC_TGTS) | |
502 | vhost->action = action; | |
503 | break; | |
504 | case IBMVFC_HOST_ACTION_INIT: | |
505 | case IBMVFC_HOST_ACTION_TGT_DEL: | |
73ee5d86 BK |
506 | switch (vhost->action) { |
507 | case IBMVFC_HOST_ACTION_RESET: | |
508 | case IBMVFC_HOST_ACTION_REENABLE: | |
509 | break; | |
510 | default: | |
511 | vhost->action = action; | |
512 | break; | |
513 | }; | |
514 | break; | |
515 | case IBMVFC_HOST_ACTION_LOGO: | |
072b91f9 | 516 | case IBMVFC_HOST_ACTION_QUERY_TGTS: |
10e79499 | 517 | case IBMVFC_HOST_ACTION_TGT_DEL_FAILED: |
072b91f9 | 518 | case IBMVFC_HOST_ACTION_NONE: |
73ee5d86 BK |
519 | case IBMVFC_HOST_ACTION_RESET: |
520 | case IBMVFC_HOST_ACTION_REENABLE: | |
072b91f9 BK |
521 | default: |
522 | vhost->action = action; | |
523 | break; | |
524 | }; | |
525 | } | |
526 | ||
527 | /** | |
528 | * ibmvfc_reinit_host - Re-start host initialization (no NPIV Login) | |
529 | * @vhost: ibmvfc host struct | |
530 | * | |
531 | * Return value: | |
532 | * nothing | |
533 | **/ | |
534 | static void ibmvfc_reinit_host(struct ibmvfc_host *vhost) | |
535 | { | |
536 | if (vhost->action == IBMVFC_HOST_ACTION_NONE) { | |
2d0da2a4 BK |
537 | if (!ibmvfc_set_host_state(vhost, IBMVFC_INITIALIZING)) { |
538 | scsi_block_requests(vhost->host); | |
539 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_QUERY); | |
540 | } | |
072b91f9 BK |
541 | } else |
542 | vhost->reinit = 1; | |
543 | ||
544 | wake_up(&vhost->work_wait_q); | |
545 | } | |
546 | ||
547 | /** | |
548 | * ibmvfc_link_down - Handle a link down event from the adapter | |
549 | * @vhost: ibmvfc host struct | |
550 | * @state: ibmvfc host state to enter | |
551 | * | |
552 | **/ | |
553 | static void ibmvfc_link_down(struct ibmvfc_host *vhost, | |
554 | enum ibmvfc_host_state state) | |
555 | { | |
556 | struct ibmvfc_target *tgt; | |
557 | ||
558 | ENTER; | |
559 | scsi_block_requests(vhost->host); | |
560 | list_for_each_entry(tgt, &vhost->targets, queue) | |
561 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); | |
562 | ibmvfc_set_host_state(vhost, state); | |
563 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_TGT_DEL); | |
564 | vhost->events_to_log |= IBMVFC_AE_LINKDOWN; | |
565 | wake_up(&vhost->work_wait_q); | |
566 | LEAVE; | |
567 | } | |
568 | ||
569 | /** | |
570 | * ibmvfc_init_host - Start host initialization | |
571 | * @vhost: ibmvfc host struct | |
572 | * | |
573 | * Return value: | |
574 | * nothing | |
575 | **/ | |
861890c6 | 576 | static void ibmvfc_init_host(struct ibmvfc_host *vhost) |
072b91f9 BK |
577 | { |
578 | struct ibmvfc_target *tgt; | |
579 | ||
580 | if (vhost->action == IBMVFC_HOST_ACTION_INIT_WAIT) { | |
1c41fa82 | 581 | if (++vhost->init_retries > IBMVFC_MAX_HOST_INIT_RETRIES) { |
072b91f9 BK |
582 | dev_err(vhost->dev, |
583 | "Host initialization retries exceeded. Taking adapter offline\n"); | |
584 | ibmvfc_link_down(vhost, IBMVFC_HOST_OFFLINE); | |
585 | return; | |
586 | } | |
587 | } | |
588 | ||
589 | if (!ibmvfc_set_host_state(vhost, IBMVFC_INITIALIZING)) { | |
861890c6 BK |
590 | memset(vhost->async_crq.msgs, 0, PAGE_SIZE); |
591 | vhost->async_crq.cur = 0; | |
cf6f10d7 | 592 | |
072b91f9 | 593 | list_for_each_entry(tgt, &vhost->targets, queue) |
5e47167b | 594 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); |
072b91f9 BK |
595 | scsi_block_requests(vhost->host); |
596 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_INIT); | |
597 | vhost->job_step = ibmvfc_npiv_login; | |
598 | wake_up(&vhost->work_wait_q); | |
599 | } | |
600 | } | |
601 | ||
602 | /** | |
603 | * ibmvfc_send_crq - Send a CRQ | |
604 | * @vhost: ibmvfc host struct | |
605 | * @word1: the first 64 bits of the data | |
606 | * @word2: the second 64 bits of the data | |
607 | * | |
608 | * Return value: | |
609 | * 0 on success / other on failure | |
610 | **/ | |
611 | static int ibmvfc_send_crq(struct ibmvfc_host *vhost, u64 word1, u64 word2) | |
612 | { | |
613 | struct vio_dev *vdev = to_vio_dev(vhost->dev); | |
614 | return plpar_hcall_norets(H_SEND_CRQ, vdev->unit_address, word1, word2); | |
615 | } | |
616 | ||
617 | /** | |
618 | * ibmvfc_send_crq_init - Send a CRQ init message | |
619 | * @vhost: ibmvfc host struct | |
620 | * | |
621 | * Return value: | |
622 | * 0 on success / other on failure | |
623 | **/ | |
624 | static int ibmvfc_send_crq_init(struct ibmvfc_host *vhost) | |
625 | { | |
626 | ibmvfc_dbg(vhost, "Sending CRQ init\n"); | |
627 | return ibmvfc_send_crq(vhost, 0xC001000000000000LL, 0); | |
628 | } | |
629 | ||
630 | /** | |
631 | * ibmvfc_send_crq_init_complete - Send a CRQ init complete message | |
632 | * @vhost: ibmvfc host struct | |
633 | * | |
634 | * Return value: | |
635 | * 0 on success / other on failure | |
636 | **/ | |
637 | static int ibmvfc_send_crq_init_complete(struct ibmvfc_host *vhost) | |
638 | { | |
639 | ibmvfc_dbg(vhost, "Sending CRQ init complete\n"); | |
640 | return ibmvfc_send_crq(vhost, 0xC002000000000000LL, 0); | |
641 | } | |
642 | ||
643 | /** | |
644 | * ibmvfc_release_crq_queue - Deallocates data and unregisters CRQ | |
645 | * @vhost: ibmvfc host struct | |
646 | * | |
647 | * Frees irq, deallocates a page for messages, unmaps dma, and unregisters | |
648 | * the crq with the hypervisor. | |
649 | **/ | |
650 | static void ibmvfc_release_crq_queue(struct ibmvfc_host *vhost) | |
651 | { | |
73ee5d86 | 652 | long rc = 0; |
072b91f9 BK |
653 | struct vio_dev *vdev = to_vio_dev(vhost->dev); |
654 | struct ibmvfc_crq_queue *crq = &vhost->crq; | |
655 | ||
656 | ibmvfc_dbg(vhost, "Releasing CRQ\n"); | |
657 | free_irq(vdev->irq, vhost); | |
039a0898 | 658 | tasklet_kill(&vhost->tasklet); |
072b91f9 | 659 | do { |
73ee5d86 BK |
660 | if (rc) |
661 | msleep(100); | |
072b91f9 BK |
662 | rc = plpar_hcall_norets(H_FREE_CRQ, vdev->unit_address); |
663 | } while (rc == H_BUSY || H_IS_LONG_BUSY(rc)); | |
664 | ||
665 | vhost->state = IBMVFC_NO_CRQ; | |
79111d08 | 666 | vhost->logged_in = 0; |
072b91f9 BK |
667 | dma_unmap_single(vhost->dev, crq->msg_token, PAGE_SIZE, DMA_BIDIRECTIONAL); |
668 | free_page((unsigned long)crq->msgs); | |
669 | } | |
670 | ||
671 | /** | |
672 | * ibmvfc_reenable_crq_queue - reenables the CRQ | |
673 | * @vhost: ibmvfc host struct | |
674 | * | |
675 | * Return value: | |
676 | * 0 on success / other on failure | |
677 | **/ | |
678 | static int ibmvfc_reenable_crq_queue(struct ibmvfc_host *vhost) | |
679 | { | |
73ee5d86 | 680 | int rc = 0; |
072b91f9 BK |
681 | struct vio_dev *vdev = to_vio_dev(vhost->dev); |
682 | ||
683 | /* Re-enable the CRQ */ | |
684 | do { | |
73ee5d86 BK |
685 | if (rc) |
686 | msleep(100); | |
072b91f9 BK |
687 | rc = plpar_hcall_norets(H_ENABLE_CRQ, vdev->unit_address); |
688 | } while (rc == H_IN_PROGRESS || rc == H_BUSY || H_IS_LONG_BUSY(rc)); | |
689 | ||
690 | if (rc) | |
691 | dev_err(vhost->dev, "Error enabling adapter (rc=%d)\n", rc); | |
692 | ||
693 | return rc; | |
694 | } | |
695 | ||
696 | /** | |
697 | * ibmvfc_reset_crq - resets a crq after a failure | |
698 | * @vhost: ibmvfc host struct | |
699 | * | |
700 | * Return value: | |
701 | * 0 on success / other on failure | |
702 | **/ | |
703 | static int ibmvfc_reset_crq(struct ibmvfc_host *vhost) | |
704 | { | |
73ee5d86 BK |
705 | int rc = 0; |
706 | unsigned long flags; | |
072b91f9 BK |
707 | struct vio_dev *vdev = to_vio_dev(vhost->dev); |
708 | struct ibmvfc_crq_queue *crq = &vhost->crq; | |
709 | ||
710 | /* Close the CRQ */ | |
711 | do { | |
73ee5d86 BK |
712 | if (rc) |
713 | msleep(100); | |
072b91f9 BK |
714 | rc = plpar_hcall_norets(H_FREE_CRQ, vdev->unit_address); |
715 | } while (rc == H_BUSY || H_IS_LONG_BUSY(rc)); | |
716 | ||
73ee5d86 | 717 | spin_lock_irqsave(vhost->host->host_lock, flags); |
072b91f9 | 718 | vhost->state = IBMVFC_NO_CRQ; |
79111d08 | 719 | vhost->logged_in = 0; |
072b91f9 BK |
720 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_NONE); |
721 | ||
722 | /* Clean out the queue */ | |
723 | memset(crq->msgs, 0, PAGE_SIZE); | |
724 | crq->cur = 0; | |
725 | ||
726 | /* And re-open it again */ | |
727 | rc = plpar_hcall_norets(H_REG_CRQ, vdev->unit_address, | |
728 | crq->msg_token, PAGE_SIZE); | |
729 | ||
730 | if (rc == H_CLOSED) | |
731 | /* Adapter is good, but other end is not ready */ | |
732 | dev_warn(vhost->dev, "Partner adapter not ready\n"); | |
733 | else if (rc != 0) | |
734 | dev_warn(vhost->dev, "Couldn't register crq (rc=%d)\n", rc); | |
73ee5d86 | 735 | spin_unlock_irqrestore(vhost->host->host_lock, flags); |
072b91f9 BK |
736 | |
737 | return rc; | |
738 | } | |
739 | ||
740 | /** | |
741 | * ibmvfc_valid_event - Determines if event is valid. | |
742 | * @pool: event_pool that contains the event | |
743 | * @evt: ibmvfc event to be checked for validity | |
744 | * | |
745 | * Return value: | |
746 | * 1 if event is valid / 0 if event is not valid | |
747 | **/ | |
748 | static int ibmvfc_valid_event(struct ibmvfc_event_pool *pool, | |
749 | struct ibmvfc_event *evt) | |
750 | { | |
751 | int index = evt - pool->events; | |
752 | if (index < 0 || index >= pool->size) /* outside of bounds */ | |
753 | return 0; | |
754 | if (evt != pool->events + index) /* unaligned */ | |
755 | return 0; | |
756 | return 1; | |
757 | } | |
758 | ||
759 | /** | |
760 | * ibmvfc_free_event - Free the specified event | |
761 | * @evt: ibmvfc_event to be freed | |
762 | * | |
763 | **/ | |
764 | static void ibmvfc_free_event(struct ibmvfc_event *evt) | |
765 | { | |
766 | struct ibmvfc_host *vhost = evt->vhost; | |
767 | struct ibmvfc_event_pool *pool = &vhost->pool; | |
768 | ||
769 | BUG_ON(!ibmvfc_valid_event(pool, evt)); | |
770 | BUG_ON(atomic_inc_return(&evt->free) != 1); | |
771 | list_add_tail(&evt->queue, &vhost->free); | |
772 | } | |
773 | ||
774 | /** | |
775 | * ibmvfc_scsi_eh_done - EH done function for queuecommand commands | |
776 | * @evt: ibmvfc event struct | |
777 | * | |
778 | * This function does not setup any error status, that must be done | |
779 | * before this function gets called. | |
780 | **/ | |
781 | static void ibmvfc_scsi_eh_done(struct ibmvfc_event *evt) | |
782 | { | |
783 | struct scsi_cmnd *cmnd = evt->cmnd; | |
784 | ||
785 | if (cmnd) { | |
786 | scsi_dma_unmap(cmnd); | |
787 | cmnd->scsi_done(cmnd); | |
788 | } | |
789 | ||
ad8dcffa BK |
790 | if (evt->eh_comp) |
791 | complete(evt->eh_comp); | |
792 | ||
072b91f9 BK |
793 | ibmvfc_free_event(evt); |
794 | } | |
795 | ||
796 | /** | |
797 | * ibmvfc_fail_request - Fail request with specified error code | |
798 | * @evt: ibmvfc event struct | |
799 | * @error_code: error code to fail request with | |
800 | * | |
801 | * Return value: | |
802 | * none | |
803 | **/ | |
804 | static void ibmvfc_fail_request(struct ibmvfc_event *evt, int error_code) | |
805 | { | |
806 | if (evt->cmnd) { | |
807 | evt->cmnd->result = (error_code << 16); | |
808 | evt->done = ibmvfc_scsi_eh_done; | |
809 | } else | |
0aab6c3f | 810 | evt->xfer_iu->mad_common.status = cpu_to_be16(IBMVFC_MAD_DRIVER_FAILED); |
072b91f9 BK |
811 | |
812 | list_del(&evt->queue); | |
813 | del_timer(&evt->timer); | |
814 | ibmvfc_trc_end(evt); | |
815 | evt->done(evt); | |
816 | } | |
817 | ||
818 | /** | |
819 | * ibmvfc_purge_requests - Our virtual adapter just shut down. Purge any sent requests | |
820 | * @vhost: ibmvfc host struct | |
821 | * @error_code: error code to fail requests with | |
822 | * | |
823 | * Return value: | |
824 | * none | |
825 | **/ | |
826 | static void ibmvfc_purge_requests(struct ibmvfc_host *vhost, int error_code) | |
827 | { | |
828 | struct ibmvfc_event *evt, *pos; | |
829 | ||
830 | ibmvfc_dbg(vhost, "Purging all requests\n"); | |
831 | list_for_each_entry_safe(evt, pos, &vhost->sent, queue) | |
832 | ibmvfc_fail_request(evt, error_code); | |
833 | } | |
834 | ||
835 | /** | |
79111d08 | 836 | * ibmvfc_hard_reset_host - Reset the connection to the server by breaking the CRQ |
072b91f9 BK |
837 | * @vhost: struct ibmvfc host to reset |
838 | **/ | |
79111d08 | 839 | static void ibmvfc_hard_reset_host(struct ibmvfc_host *vhost) |
072b91f9 | 840 | { |
072b91f9 | 841 | ibmvfc_purge_requests(vhost, DID_ERROR); |
73ee5d86 BK |
842 | ibmvfc_link_down(vhost, IBMVFC_LINK_DOWN); |
843 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_RESET); | |
072b91f9 BK |
844 | } |
845 | ||
846 | /** | |
79111d08 | 847 | * __ibmvfc_reset_host - Reset the connection to the server (no locking) |
072b91f9 BK |
848 | * @vhost: struct ibmvfc host to reset |
849 | **/ | |
79111d08 BK |
850 | static void __ibmvfc_reset_host(struct ibmvfc_host *vhost) |
851 | { | |
852 | if (vhost->logged_in && vhost->action != IBMVFC_HOST_ACTION_LOGO_WAIT && | |
853 | !ibmvfc_set_host_state(vhost, IBMVFC_INITIALIZING)) { | |
854 | scsi_block_requests(vhost->host); | |
855 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_LOGO); | |
856 | vhost->job_step = ibmvfc_npiv_logout; | |
857 | wake_up(&vhost->work_wait_q); | |
858 | } else | |
859 | ibmvfc_hard_reset_host(vhost); | |
860 | } | |
861 | ||
862 | /** | |
863 | * ibmvfc_reset_host - Reset the connection to the server | |
864 | * @vhost: ibmvfc host struct | |
865 | **/ | |
072b91f9 BK |
866 | static void ibmvfc_reset_host(struct ibmvfc_host *vhost) |
867 | { | |
868 | unsigned long flags; | |
869 | ||
870 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
871 | __ibmvfc_reset_host(vhost); | |
872 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
873 | } | |
874 | ||
875 | /** | |
876 | * ibmvfc_retry_host_init - Retry host initialization if allowed | |
877 | * @vhost: ibmvfc host struct | |
878 | * | |
7d0e4622 BK |
879 | * Returns: 1 if init will be retried / 0 if not |
880 | * | |
072b91f9 | 881 | **/ |
7d0e4622 | 882 | static int ibmvfc_retry_host_init(struct ibmvfc_host *vhost) |
072b91f9 | 883 | { |
7d0e4622 BK |
884 | int retry = 0; |
885 | ||
072b91f9 | 886 | if (vhost->action == IBMVFC_HOST_ACTION_INIT_WAIT) { |
1c41fa82 BK |
887 | vhost->delay_init = 1; |
888 | if (++vhost->init_retries > IBMVFC_MAX_HOST_INIT_RETRIES) { | |
072b91f9 BK |
889 | dev_err(vhost->dev, |
890 | "Host initialization retries exceeded. Taking adapter offline\n"); | |
891 | ibmvfc_link_down(vhost, IBMVFC_HOST_OFFLINE); | |
1c41fa82 | 892 | } else if (vhost->init_retries == IBMVFC_MAX_HOST_INIT_RETRIES) |
072b91f9 | 893 | __ibmvfc_reset_host(vhost); |
7d0e4622 | 894 | else { |
072b91f9 | 895 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_INIT); |
7d0e4622 BK |
896 | retry = 1; |
897 | } | |
072b91f9 BK |
898 | } |
899 | ||
900 | wake_up(&vhost->work_wait_q); | |
7d0e4622 | 901 | return retry; |
072b91f9 BK |
902 | } |
903 | ||
904 | /** | |
b3c10489 | 905 | * __ibmvfc_get_target - Find the specified scsi_target (no locking) |
072b91f9 BK |
906 | * @starget: scsi target struct |
907 | * | |
908 | * Return value: | |
909 | * ibmvfc_target struct / NULL if not found | |
910 | **/ | |
b3c10489 | 911 | static struct ibmvfc_target *__ibmvfc_get_target(struct scsi_target *starget) |
072b91f9 BK |
912 | { |
913 | struct Scsi_Host *shost = dev_to_shost(starget->dev.parent); | |
914 | struct ibmvfc_host *vhost = shost_priv(shost); | |
915 | struct ibmvfc_target *tgt; | |
916 | ||
917 | list_for_each_entry(tgt, &vhost->targets, queue) | |
b3c10489 BK |
918 | if (tgt->target_id == starget->id) { |
919 | kref_get(&tgt->kref); | |
072b91f9 | 920 | return tgt; |
b3c10489 | 921 | } |
072b91f9 BK |
922 | return NULL; |
923 | } | |
924 | ||
925 | /** | |
b3c10489 | 926 | * ibmvfc_get_target - Find the specified scsi_target |
072b91f9 BK |
927 | * @starget: scsi target struct |
928 | * | |
929 | * Return value: | |
930 | * ibmvfc_target struct / NULL if not found | |
931 | **/ | |
b3c10489 | 932 | static struct ibmvfc_target *ibmvfc_get_target(struct scsi_target *starget) |
072b91f9 BK |
933 | { |
934 | struct Scsi_Host *shost = dev_to_shost(starget->dev.parent); | |
935 | struct ibmvfc_target *tgt; | |
936 | unsigned long flags; | |
937 | ||
938 | spin_lock_irqsave(shost->host_lock, flags); | |
b3c10489 | 939 | tgt = __ibmvfc_get_target(starget); |
072b91f9 BK |
940 | spin_unlock_irqrestore(shost->host_lock, flags); |
941 | return tgt; | |
942 | } | |
943 | ||
944 | /** | |
945 | * ibmvfc_get_host_speed - Get host port speed | |
946 | * @shost: scsi host struct | |
947 | * | |
948 | * Return value: | |
949 | * none | |
950 | **/ | |
951 | static void ibmvfc_get_host_speed(struct Scsi_Host *shost) | |
952 | { | |
953 | struct ibmvfc_host *vhost = shost_priv(shost); | |
954 | unsigned long flags; | |
955 | ||
956 | spin_lock_irqsave(shost->host_lock, flags); | |
957 | if (vhost->state == IBMVFC_ACTIVE) { | |
0aab6c3f | 958 | switch (be64_to_cpu(vhost->login_buf->resp.link_speed) / 100) { |
072b91f9 BK |
959 | case 1: |
960 | fc_host_speed(shost) = FC_PORTSPEED_1GBIT; | |
961 | break; | |
962 | case 2: | |
963 | fc_host_speed(shost) = FC_PORTSPEED_2GBIT; | |
964 | break; | |
965 | case 4: | |
966 | fc_host_speed(shost) = FC_PORTSPEED_4GBIT; | |
967 | break; | |
968 | case 8: | |
969 | fc_host_speed(shost) = FC_PORTSPEED_8GBIT; | |
970 | break; | |
971 | case 10: | |
972 | fc_host_speed(shost) = FC_PORTSPEED_10GBIT; | |
973 | break; | |
974 | case 16: | |
975 | fc_host_speed(shost) = FC_PORTSPEED_16GBIT; | |
976 | break; | |
977 | default: | |
775a42ec | 978 | ibmvfc_log(vhost, 3, "Unknown port speed: %lld Gbit\n", |
0aab6c3f | 979 | be64_to_cpu(vhost->login_buf->resp.link_speed) / 100); |
072b91f9 BK |
980 | fc_host_speed(shost) = FC_PORTSPEED_UNKNOWN; |
981 | break; | |
982 | } | |
983 | } else | |
984 | fc_host_speed(shost) = FC_PORTSPEED_UNKNOWN; | |
985 | spin_unlock_irqrestore(shost->host_lock, flags); | |
986 | } | |
987 | ||
988 | /** | |
989 | * ibmvfc_get_host_port_state - Get host port state | |
990 | * @shost: scsi host struct | |
991 | * | |
992 | * Return value: | |
993 | * none | |
994 | **/ | |
995 | static void ibmvfc_get_host_port_state(struct Scsi_Host *shost) | |
996 | { | |
997 | struct ibmvfc_host *vhost = shost_priv(shost); | |
998 | unsigned long flags; | |
999 | ||
1000 | spin_lock_irqsave(shost->host_lock, flags); | |
1001 | switch (vhost->state) { | |
1002 | case IBMVFC_INITIALIZING: | |
1003 | case IBMVFC_ACTIVE: | |
1004 | fc_host_port_state(shost) = FC_PORTSTATE_ONLINE; | |
1005 | break; | |
1006 | case IBMVFC_LINK_DOWN: | |
1007 | fc_host_port_state(shost) = FC_PORTSTATE_LINKDOWN; | |
1008 | break; | |
1009 | case IBMVFC_LINK_DEAD: | |
1010 | case IBMVFC_HOST_OFFLINE: | |
1011 | fc_host_port_state(shost) = FC_PORTSTATE_OFFLINE; | |
1012 | break; | |
1013 | case IBMVFC_HALTED: | |
1014 | fc_host_port_state(shost) = FC_PORTSTATE_BLOCKED; | |
1015 | break; | |
0ae808e0 BK |
1016 | case IBMVFC_NO_CRQ: |
1017 | fc_host_port_state(shost) = FC_PORTSTATE_UNKNOWN; | |
1018 | break; | |
072b91f9 BK |
1019 | default: |
1020 | ibmvfc_log(vhost, 3, "Unknown port state: %d\n", vhost->state); | |
1021 | fc_host_port_state(shost) = FC_PORTSTATE_UNKNOWN; | |
1022 | break; | |
1023 | } | |
1024 | spin_unlock_irqrestore(shost->host_lock, flags); | |
1025 | } | |
1026 | ||
1027 | /** | |
1028 | * ibmvfc_set_rport_dev_loss_tmo - Set rport's device loss timeout | |
1029 | * @rport: rport struct | |
1030 | * @timeout: timeout value | |
1031 | * | |
1032 | * Return value: | |
1033 | * none | |
1034 | **/ | |
1035 | static void ibmvfc_set_rport_dev_loss_tmo(struct fc_rport *rport, u32 timeout) | |
1036 | { | |
1037 | if (timeout) | |
1038 | rport->dev_loss_tmo = timeout; | |
1039 | else | |
1040 | rport->dev_loss_tmo = 1; | |
1041 | } | |
1042 | ||
b3c10489 BK |
1043 | /** |
1044 | * ibmvfc_release_tgt - Free memory allocated for a target | |
1045 | * @kref: kref struct | |
1046 | * | |
1047 | **/ | |
1048 | static void ibmvfc_release_tgt(struct kref *kref) | |
1049 | { | |
1050 | struct ibmvfc_target *tgt = container_of(kref, struct ibmvfc_target, kref); | |
1051 | kfree(tgt); | |
1052 | } | |
1053 | ||
072b91f9 BK |
1054 | /** |
1055 | * ibmvfc_get_starget_node_name - Get SCSI target's node name | |
1056 | * @starget: scsi target struct | |
1057 | * | |
1058 | * Return value: | |
1059 | * none | |
1060 | **/ | |
1061 | static void ibmvfc_get_starget_node_name(struct scsi_target *starget) | |
1062 | { | |
b3c10489 | 1063 | struct ibmvfc_target *tgt = ibmvfc_get_target(starget); |
072b91f9 | 1064 | fc_starget_port_name(starget) = tgt ? tgt->ids.node_name : 0; |
b3c10489 BK |
1065 | if (tgt) |
1066 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
072b91f9 BK |
1067 | } |
1068 | ||
1069 | /** | |
1070 | * ibmvfc_get_starget_port_name - Get SCSI target's port name | |
1071 | * @starget: scsi target struct | |
1072 | * | |
1073 | * Return value: | |
1074 | * none | |
1075 | **/ | |
1076 | static void ibmvfc_get_starget_port_name(struct scsi_target *starget) | |
1077 | { | |
b3c10489 | 1078 | struct ibmvfc_target *tgt = ibmvfc_get_target(starget); |
072b91f9 | 1079 | fc_starget_port_name(starget) = tgt ? tgt->ids.port_name : 0; |
b3c10489 BK |
1080 | if (tgt) |
1081 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
072b91f9 BK |
1082 | } |
1083 | ||
1084 | /** | |
1085 | * ibmvfc_get_starget_port_id - Get SCSI target's port ID | |
1086 | * @starget: scsi target struct | |
1087 | * | |
1088 | * Return value: | |
1089 | * none | |
1090 | **/ | |
1091 | static void ibmvfc_get_starget_port_id(struct scsi_target *starget) | |
1092 | { | |
b3c10489 | 1093 | struct ibmvfc_target *tgt = ibmvfc_get_target(starget); |
072b91f9 | 1094 | fc_starget_port_id(starget) = tgt ? tgt->scsi_id : -1; |
b3c10489 BK |
1095 | if (tgt) |
1096 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
072b91f9 BK |
1097 | } |
1098 | ||
1099 | /** | |
1100 | * ibmvfc_wait_while_resetting - Wait while the host resets | |
1101 | * @vhost: ibmvfc host struct | |
1102 | * | |
1103 | * Return value: | |
1104 | * 0 on success / other on failure | |
1105 | **/ | |
1106 | static int ibmvfc_wait_while_resetting(struct ibmvfc_host *vhost) | |
1107 | { | |
1108 | long timeout = wait_event_timeout(vhost->init_wait_q, | |
3eddc569 BK |
1109 | ((vhost->state == IBMVFC_ACTIVE || |
1110 | vhost->state == IBMVFC_HOST_OFFLINE || | |
1111 | vhost->state == IBMVFC_LINK_DEAD) && | |
1112 | vhost->action == IBMVFC_HOST_ACTION_NONE), | |
072b91f9 BK |
1113 | (init_timeout * HZ)); |
1114 | ||
1115 | return timeout ? 0 : -EIO; | |
1116 | } | |
1117 | ||
1118 | /** | |
1119 | * ibmvfc_issue_fc_host_lip - Re-initiate link initialization | |
1120 | * @shost: scsi host struct | |
1121 | * | |
1122 | * Return value: | |
1123 | * 0 on success / other on failure | |
1124 | **/ | |
1125 | static int ibmvfc_issue_fc_host_lip(struct Scsi_Host *shost) | |
1126 | { | |
1127 | struct ibmvfc_host *vhost = shost_priv(shost); | |
1128 | ||
1129 | dev_err(vhost->dev, "Initiating host LIP. Resetting connection\n"); | |
1130 | ibmvfc_reset_host(vhost); | |
1131 | return ibmvfc_wait_while_resetting(vhost); | |
1132 | } | |
1133 | ||
1134 | /** | |
1135 | * ibmvfc_gather_partition_info - Gather info about the LPAR | |
1136 | * | |
1137 | * Return value: | |
1138 | * none | |
1139 | **/ | |
1140 | static void ibmvfc_gather_partition_info(struct ibmvfc_host *vhost) | |
1141 | { | |
1142 | struct device_node *rootdn; | |
1143 | const char *name; | |
1144 | const unsigned int *num; | |
1145 | ||
1146 | rootdn = of_find_node_by_path("/"); | |
1147 | if (!rootdn) | |
1148 | return; | |
1149 | ||
1150 | name = of_get_property(rootdn, "ibm,partition-name", NULL); | |
1151 | if (name) | |
1152 | strncpy(vhost->partition_name, name, sizeof(vhost->partition_name)); | |
1153 | num = of_get_property(rootdn, "ibm,partition-no", NULL); | |
1154 | if (num) | |
1155 | vhost->partition_number = *num; | |
1156 | of_node_put(rootdn); | |
1157 | } | |
1158 | ||
1159 | /** | |
1160 | * ibmvfc_set_login_info - Setup info for NPIV login | |
1161 | * @vhost: ibmvfc host struct | |
1162 | * | |
1163 | * Return value: | |
1164 | * none | |
1165 | **/ | |
1166 | static void ibmvfc_set_login_info(struct ibmvfc_host *vhost) | |
1167 | { | |
1168 | struct ibmvfc_npiv_login *login_info = &vhost->login_info; | |
61c7a080 | 1169 | struct device_node *of_node = vhost->dev->of_node; |
072b91f9 BK |
1170 | const char *location; |
1171 | ||
1172 | memset(login_info, 0, sizeof(*login_info)); | |
1173 | ||
0aab6c3f TD |
1174 | login_info->ostype = cpu_to_be32(IBMVFC_OS_LINUX); |
1175 | login_info->max_dma_len = cpu_to_be64(IBMVFC_MAX_SECTORS << 9); | |
1176 | login_info->max_payload = cpu_to_be32(sizeof(struct ibmvfc_fcp_cmd_iu)); | |
1177 | login_info->max_response = cpu_to_be32(sizeof(struct ibmvfc_fcp_rsp)); | |
1178 | login_info->partition_num = cpu_to_be32(vhost->partition_number); | |
1179 | login_info->vfc_frame_version = cpu_to_be32(1); | |
1180 | login_info->fcp_version = cpu_to_be16(3); | |
1181 | login_info->flags = cpu_to_be16(IBMVFC_FLUSH_ON_HALT); | |
072b91f9 | 1182 | if (vhost->client_migrated) |
0aab6c3f | 1183 | login_info->flags |= cpu_to_be16(IBMVFC_CLIENT_MIGRATED); |
072b91f9 | 1184 | |
0aab6c3f TD |
1185 | login_info->max_cmds = cpu_to_be32(max_requests + IBMVFC_NUM_INTERNAL_REQ); |
1186 | login_info->capabilities = cpu_to_be64(IBMVFC_CAN_MIGRATE); | |
1187 | login_info->async.va = cpu_to_be64(vhost->async_crq.msg_token); | |
1188 | login_info->async.len = cpu_to_be32(vhost->async_crq.size * sizeof(*vhost->async_crq.msgs)); | |
072b91f9 BK |
1189 | strncpy(login_info->partition_name, vhost->partition_name, IBMVFC_MAX_NAME); |
1190 | strncpy(login_info->device_name, | |
71610f55 | 1191 | dev_name(&vhost->host->shost_gendev), IBMVFC_MAX_NAME); |
072b91f9 BK |
1192 | |
1193 | location = of_get_property(of_node, "ibm,loc-code", NULL); | |
71610f55 | 1194 | location = location ? location : dev_name(vhost->dev); |
072b91f9 BK |
1195 | strncpy(login_info->drc_name, location, IBMVFC_MAX_NAME); |
1196 | } | |
1197 | ||
1198 | /** | |
1199 | * ibmvfc_init_event_pool - Allocates and initializes the event pool for a host | |
1200 | * @vhost: ibmvfc host who owns the event pool | |
1201 | * | |
1202 | * Returns zero on success. | |
1203 | **/ | |
1204 | static int ibmvfc_init_event_pool(struct ibmvfc_host *vhost) | |
1205 | { | |
1206 | int i; | |
1207 | struct ibmvfc_event_pool *pool = &vhost->pool; | |
1208 | ||
1209 | ENTER; | |
1210 | pool->size = max_requests + IBMVFC_NUM_INTERNAL_REQ; | |
1211 | pool->events = kcalloc(pool->size, sizeof(*pool->events), GFP_KERNEL); | |
1212 | if (!pool->events) | |
1213 | return -ENOMEM; | |
1214 | ||
1215 | pool->iu_storage = dma_alloc_coherent(vhost->dev, | |
1216 | pool->size * sizeof(*pool->iu_storage), | |
1217 | &pool->iu_token, 0); | |
1218 | ||
1219 | if (!pool->iu_storage) { | |
1220 | kfree(pool->events); | |
1221 | return -ENOMEM; | |
1222 | } | |
1223 | ||
1224 | for (i = 0; i < pool->size; ++i) { | |
1225 | struct ibmvfc_event *evt = &pool->events[i]; | |
1226 | atomic_set(&evt->free, 1); | |
1227 | evt->crq.valid = 0x80; | |
0aab6c3f | 1228 | evt->crq.ioba = cpu_to_be64(pool->iu_token + (sizeof(*evt->xfer_iu) * i)); |
072b91f9 BK |
1229 | evt->xfer_iu = pool->iu_storage + i; |
1230 | evt->vhost = vhost; | |
1231 | evt->ext_list = NULL; | |
1232 | list_add_tail(&evt->queue, &vhost->free); | |
1233 | } | |
1234 | ||
1235 | LEAVE; | |
1236 | return 0; | |
1237 | } | |
1238 | ||
1239 | /** | |
1240 | * ibmvfc_free_event_pool - Frees memory of the event pool of a host | |
1241 | * @vhost: ibmvfc host who owns the event pool | |
1242 | * | |
1243 | **/ | |
1244 | static void ibmvfc_free_event_pool(struct ibmvfc_host *vhost) | |
1245 | { | |
1246 | int i; | |
1247 | struct ibmvfc_event_pool *pool = &vhost->pool; | |
1248 | ||
1249 | ENTER; | |
1250 | for (i = 0; i < pool->size; ++i) { | |
1251 | list_del(&pool->events[i].queue); | |
1252 | BUG_ON(atomic_read(&pool->events[i].free) != 1); | |
1253 | if (pool->events[i].ext_list) | |
1254 | dma_pool_free(vhost->sg_pool, | |
1255 | pool->events[i].ext_list, | |
1256 | pool->events[i].ext_list_token); | |
1257 | } | |
1258 | ||
1259 | kfree(pool->events); | |
1260 | dma_free_coherent(vhost->dev, | |
1261 | pool->size * sizeof(*pool->iu_storage), | |
1262 | pool->iu_storage, pool->iu_token); | |
1263 | LEAVE; | |
1264 | } | |
1265 | ||
1266 | /** | |
1267 | * ibmvfc_get_event - Gets the next free event in pool | |
1268 | * @vhost: ibmvfc host struct | |
1269 | * | |
1270 | * Returns a free event from the pool. | |
1271 | **/ | |
1272 | static struct ibmvfc_event *ibmvfc_get_event(struct ibmvfc_host *vhost) | |
1273 | { | |
1274 | struct ibmvfc_event *evt; | |
1275 | ||
1276 | BUG_ON(list_empty(&vhost->free)); | |
1277 | evt = list_entry(vhost->free.next, struct ibmvfc_event, queue); | |
1278 | atomic_set(&evt->free, 0); | |
1279 | list_del(&evt->queue); | |
1280 | return evt; | |
1281 | } | |
1282 | ||
1283 | /** | |
1284 | * ibmvfc_init_event - Initialize fields in an event struct that are always | |
1285 | * required. | |
1286 | * @evt: The event | |
1287 | * @done: Routine to call when the event is responded to | |
1288 | * @format: SRP or MAD format | |
1289 | **/ | |
1290 | static void ibmvfc_init_event(struct ibmvfc_event *evt, | |
1291 | void (*done) (struct ibmvfc_event *), u8 format) | |
1292 | { | |
1293 | evt->cmnd = NULL; | |
1294 | evt->sync_iu = NULL; | |
1295 | evt->crq.format = format; | |
1296 | evt->done = done; | |
ad8dcffa | 1297 | evt->eh_comp = NULL; |
072b91f9 BK |
1298 | } |
1299 | ||
1300 | /** | |
1301 | * ibmvfc_map_sg_list - Initialize scatterlist | |
1302 | * @scmd: scsi command struct | |
1303 | * @nseg: number of scatterlist segments | |
1304 | * @md: memory descriptor list to initialize | |
1305 | **/ | |
1306 | static void ibmvfc_map_sg_list(struct scsi_cmnd *scmd, int nseg, | |
1307 | struct srp_direct_buf *md) | |
1308 | { | |
1309 | int i; | |
1310 | struct scatterlist *sg; | |
1311 | ||
1312 | scsi_for_each_sg(scmd, sg, nseg, i) { | |
0aab6c3f TD |
1313 | md[i].va = cpu_to_be64(sg_dma_address(sg)); |
1314 | md[i].len = cpu_to_be32(sg_dma_len(sg)); | |
072b91f9 BK |
1315 | md[i].key = 0; |
1316 | } | |
1317 | } | |
1318 | ||
1319 | /** | |
1320 | * ibmvfc_map_sg_data - Maps dma for a scatterlist and initializes decriptor fields | |
1321 | * @scmd: Scsi_Cmnd with the scatterlist | |
1322 | * @evt: ibmvfc event struct | |
1323 | * @vfc_cmd: vfc_cmd that contains the memory descriptor | |
1324 | * @dev: device for which to map dma memory | |
1325 | * | |
1326 | * Returns: | |
1327 | * 0 on success / non-zero on failure | |
1328 | **/ | |
1329 | static int ibmvfc_map_sg_data(struct scsi_cmnd *scmd, | |
1330 | struct ibmvfc_event *evt, | |
1331 | struct ibmvfc_cmd *vfc_cmd, struct device *dev) | |
1332 | { | |
1333 | ||
1334 | int sg_mapped; | |
1335 | struct srp_direct_buf *data = &vfc_cmd->ioba; | |
1336 | struct ibmvfc_host *vhost = dev_get_drvdata(dev); | |
1337 | ||
1338 | sg_mapped = scsi_dma_map(scmd); | |
1339 | if (!sg_mapped) { | |
0aab6c3f | 1340 | vfc_cmd->flags |= cpu_to_be16(IBMVFC_NO_MEM_DESC); |
072b91f9 BK |
1341 | return 0; |
1342 | } else if (unlikely(sg_mapped < 0)) { | |
1343 | if (vhost->log_level > IBMVFC_DEFAULT_LOG_LEVEL) | |
1344 | scmd_printk(KERN_ERR, scmd, "Failed to map DMA buffer for command\n"); | |
1345 | return sg_mapped; | |
1346 | } | |
1347 | ||
1348 | if (scmd->sc_data_direction == DMA_TO_DEVICE) { | |
0aab6c3f | 1349 | vfc_cmd->flags |= cpu_to_be16(IBMVFC_WRITE); |
072b91f9 BK |
1350 | vfc_cmd->iu.add_cdb_len |= IBMVFC_WRDATA; |
1351 | } else { | |
0aab6c3f | 1352 | vfc_cmd->flags |= cpu_to_be16(IBMVFC_READ); |
072b91f9 BK |
1353 | vfc_cmd->iu.add_cdb_len |= IBMVFC_RDDATA; |
1354 | } | |
1355 | ||
1356 | if (sg_mapped == 1) { | |
1357 | ibmvfc_map_sg_list(scmd, sg_mapped, data); | |
1358 | return 0; | |
1359 | } | |
1360 | ||
0aab6c3f | 1361 | vfc_cmd->flags |= cpu_to_be16(IBMVFC_SCATTERLIST); |
072b91f9 BK |
1362 | |
1363 | if (!evt->ext_list) { | |
1364 | evt->ext_list = dma_pool_alloc(vhost->sg_pool, GFP_ATOMIC, | |
1365 | &evt->ext_list_token); | |
1366 | ||
1367 | if (!evt->ext_list) { | |
64b840dd BK |
1368 | scsi_dma_unmap(scmd); |
1369 | if (vhost->log_level > IBMVFC_DEFAULT_LOG_LEVEL) | |
1370 | scmd_printk(KERN_ERR, scmd, "Can't allocate memory for scatterlist\n"); | |
072b91f9 BK |
1371 | return -ENOMEM; |
1372 | } | |
1373 | } | |
1374 | ||
1375 | ibmvfc_map_sg_list(scmd, sg_mapped, evt->ext_list); | |
1376 | ||
0aab6c3f TD |
1377 | data->va = cpu_to_be64(evt->ext_list_token); |
1378 | data->len = cpu_to_be32(sg_mapped * sizeof(struct srp_direct_buf)); | |
072b91f9 BK |
1379 | data->key = 0; |
1380 | return 0; | |
1381 | } | |
1382 | ||
1383 | /** | |
1384 | * ibmvfc_timeout - Internal command timeout handler | |
1385 | * @evt: struct ibmvfc_event that timed out | |
1386 | * | |
1387 | * Called when an internally generated command times out | |
1388 | **/ | |
1389 | static void ibmvfc_timeout(struct ibmvfc_event *evt) | |
1390 | { | |
1391 | struct ibmvfc_host *vhost = evt->vhost; | |
1392 | dev_err(vhost->dev, "Command timed out (%p). Resetting connection\n", evt); | |
1393 | ibmvfc_reset_host(vhost); | |
1394 | } | |
1395 | ||
1396 | /** | |
1397 | * ibmvfc_send_event - Transforms event to u64 array and calls send_crq() | |
1398 | * @evt: event to be sent | |
1399 | * @vhost: ibmvfc host struct | |
1400 | * @timeout: timeout in seconds - 0 means do not time command | |
1401 | * | |
1402 | * Returns the value returned from ibmvfc_send_crq(). (Zero for success) | |
1403 | **/ | |
1404 | static int ibmvfc_send_event(struct ibmvfc_event *evt, | |
1405 | struct ibmvfc_host *vhost, unsigned long timeout) | |
1406 | { | |
0aab6c3f | 1407 | __be64 *crq_as_u64 = (__be64 *) &evt->crq; |
072b91f9 BK |
1408 | int rc; |
1409 | ||
1410 | /* Copy the IU into the transfer area */ | |
1411 | *evt->xfer_iu = evt->iu; | |
1412 | if (evt->crq.format == IBMVFC_CMD_FORMAT) | |
0aab6c3f | 1413 | evt->xfer_iu->cmd.tag = cpu_to_be64((u64)evt); |
072b91f9 | 1414 | else if (evt->crq.format == IBMVFC_MAD_FORMAT) |
0aab6c3f | 1415 | evt->xfer_iu->mad_common.tag = cpu_to_be64((u64)evt); |
072b91f9 BK |
1416 | else |
1417 | BUG(); | |
1418 | ||
1419 | list_add_tail(&evt->queue, &vhost->sent); | |
1420 | init_timer(&evt->timer); | |
1421 | ||
1422 | if (timeout) { | |
1423 | evt->timer.data = (unsigned long) evt; | |
1424 | evt->timer.expires = jiffies + (timeout * HZ); | |
1425 | evt->timer.function = (void (*)(unsigned long))ibmvfc_timeout; | |
1426 | add_timer(&evt->timer); | |
1427 | } | |
1428 | ||
a528ab7a BK |
1429 | mb(); |
1430 | ||
0aab6c3f TD |
1431 | if ((rc = ibmvfc_send_crq(vhost, be64_to_cpu(crq_as_u64[0]), |
1432 | be64_to_cpu(crq_as_u64[1])))) { | |
072b91f9 BK |
1433 | list_del(&evt->queue); |
1434 | del_timer(&evt->timer); | |
1435 | ||
1436 | /* If send_crq returns H_CLOSED, return SCSI_MLQUEUE_HOST_BUSY. | |
1437 | * Firmware will send a CRQ with a transport event (0xFF) to | |
1438 | * tell this client what has happened to the transport. This | |
1439 | * will be handled in ibmvfc_handle_crq() | |
1440 | */ | |
1441 | if (rc == H_CLOSED) { | |
1442 | if (printk_ratelimit()) | |
1443 | dev_warn(vhost->dev, "Send warning. Receive queue closed, will retry.\n"); | |
1444 | if (evt->cmnd) | |
1445 | scsi_dma_unmap(evt->cmnd); | |
1446 | ibmvfc_free_event(evt); | |
1447 | return SCSI_MLQUEUE_HOST_BUSY; | |
1448 | } | |
1449 | ||
1450 | dev_err(vhost->dev, "Send error (rc=%d)\n", rc); | |
1451 | if (evt->cmnd) { | |
1452 | evt->cmnd->result = DID_ERROR << 16; | |
1453 | evt->done = ibmvfc_scsi_eh_done; | |
1454 | } else | |
0aab6c3f | 1455 | evt->xfer_iu->mad_common.status = cpu_to_be16(IBMVFC_MAD_CRQ_ERROR); |
072b91f9 BK |
1456 | |
1457 | evt->done(evt); | |
1458 | } else | |
1459 | ibmvfc_trc_start(evt); | |
1460 | ||
1461 | return 0; | |
1462 | } | |
1463 | ||
1464 | /** | |
1465 | * ibmvfc_log_error - Log an error for the failed command if appropriate | |
1466 | * @evt: ibmvfc event to log | |
1467 | * | |
1468 | **/ | |
1469 | static void ibmvfc_log_error(struct ibmvfc_event *evt) | |
1470 | { | |
1471 | struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; | |
1472 | struct ibmvfc_host *vhost = evt->vhost; | |
1473 | struct ibmvfc_fcp_rsp *rsp = &vfc_cmd->rsp; | |
1474 | struct scsi_cmnd *cmnd = evt->cmnd; | |
1475 | const char *err = unknown_error; | |
0aab6c3f | 1476 | int index = ibmvfc_get_err_index(be16_to_cpu(vfc_cmd->status), be16_to_cpu(vfc_cmd->error)); |
072b91f9 BK |
1477 | int logerr = 0; |
1478 | int rsp_code = 0; | |
1479 | ||
1480 | if (index >= 0) { | |
1481 | logerr = cmd_status[index].log; | |
1482 | err = cmd_status[index].name; | |
1483 | } | |
1484 | ||
0ae808e0 | 1485 | if (!logerr && (vhost->log_level <= (IBMVFC_DEFAULT_LOG_LEVEL + 1))) |
072b91f9 BK |
1486 | return; |
1487 | ||
1488 | if (rsp->flags & FCP_RSP_LEN_VALID) | |
1489 | rsp_code = rsp->data.info.rsp_code; | |
1490 | ||
1491 | scmd_printk(KERN_ERR, cmnd, "Command (%02X) failed: %s (%x:%x) " | |
1492 | "flags: %x fcp_rsp: %x, resid=%d, scsi_status: %x\n", | |
1493 | cmnd->cmnd[0], err, vfc_cmd->status, vfc_cmd->error, | |
1494 | rsp->flags, rsp_code, scsi_get_resid(cmnd), rsp->scsi_status); | |
1495 | } | |
1496 | ||
6d29cc56 BK |
1497 | /** |
1498 | * ibmvfc_relogin - Log back into the specified device | |
1499 | * @sdev: scsi device struct | |
1500 | * | |
1501 | **/ | |
1502 | static void ibmvfc_relogin(struct scsi_device *sdev) | |
1503 | { | |
1504 | struct ibmvfc_host *vhost = shost_priv(sdev->host); | |
1505 | struct fc_rport *rport = starget_to_rport(scsi_target(sdev)); | |
1506 | struct ibmvfc_target *tgt; | |
1507 | ||
1508 | list_for_each_entry(tgt, &vhost->targets, queue) { | |
1509 | if (rport == tgt->rport) { | |
1510 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); | |
1511 | break; | |
1512 | } | |
1513 | } | |
1514 | ||
1515 | ibmvfc_reinit_host(vhost); | |
1516 | } | |
1517 | ||
072b91f9 BK |
1518 | /** |
1519 | * ibmvfc_scsi_done - Handle responses from commands | |
1520 | * @evt: ibmvfc event to be handled | |
1521 | * | |
1522 | * Used as a callback when sending scsi cmds. | |
1523 | **/ | |
1524 | static void ibmvfc_scsi_done(struct ibmvfc_event *evt) | |
1525 | { | |
1526 | struct ibmvfc_cmd *vfc_cmd = &evt->xfer_iu->cmd; | |
1527 | struct ibmvfc_fcp_rsp *rsp = &vfc_cmd->rsp; | |
1528 | struct scsi_cmnd *cmnd = evt->cmnd; | |
2bac406d | 1529 | u32 rsp_len = 0; |
0aab6c3f | 1530 | u32 sense_len = be32_to_cpu(rsp->fcp_sense_len); |
072b91f9 BK |
1531 | |
1532 | if (cmnd) { | |
0aab6c3f TD |
1533 | if (be16_to_cpu(vfc_cmd->response_flags) & IBMVFC_ADAPTER_RESID_VALID) |
1534 | scsi_set_resid(cmnd, be32_to_cpu(vfc_cmd->adapter_resid)); | |
072b91f9 | 1535 | else if (rsp->flags & FCP_RESID_UNDER) |
0aab6c3f | 1536 | scsi_set_resid(cmnd, be32_to_cpu(rsp->fcp_resid)); |
072b91f9 BK |
1537 | else |
1538 | scsi_set_resid(cmnd, 0); | |
1539 | ||
1540 | if (vfc_cmd->status) { | |
1541 | cmnd->result = ibmvfc_get_err_result(vfc_cmd); | |
1542 | ||
1543 | if (rsp->flags & FCP_RSP_LEN_VALID) | |
0aab6c3f | 1544 | rsp_len = be32_to_cpu(rsp->fcp_rsp_len); |
072b91f9 BK |
1545 | if ((sense_len + rsp_len) > SCSI_SENSE_BUFFERSIZE) |
1546 | sense_len = SCSI_SENSE_BUFFERSIZE - rsp_len; | |
2bac406d | 1547 | if ((rsp->flags & FCP_SNS_LEN_VALID) && rsp->fcp_sense_len && rsp_len <= 8) |
072b91f9 | 1548 | memcpy(cmnd->sense_buffer, rsp->data.sense + rsp_len, sense_len); |
0aab6c3f TD |
1549 | if ((be16_to_cpu(vfc_cmd->status) & IBMVFC_VIOS_FAILURE) && |
1550 | (be16_to_cpu(vfc_cmd->error) == IBMVFC_PLOGI_REQUIRED)) | |
6d29cc56 | 1551 | ibmvfc_relogin(cmnd->device); |
072b91f9 | 1552 | |
50119dad BK |
1553 | if (!cmnd->result && (!scsi_get_resid(cmnd) || (rsp->flags & FCP_RESID_OVER))) |
1554 | cmnd->result = (DID_ERROR << 16); | |
1555 | ||
072b91f9 BK |
1556 | ibmvfc_log_error(evt); |
1557 | } | |
1558 | ||
1559 | if (!cmnd->result && | |
1560 | (scsi_bufflen(cmnd) - scsi_get_resid(cmnd) < cmnd->underflow)) | |
1561 | cmnd->result = (DID_ERROR << 16); | |
1562 | ||
1563 | scsi_dma_unmap(cmnd); | |
1564 | cmnd->scsi_done(cmnd); | |
1565 | } | |
1566 | ||
ad8dcffa BK |
1567 | if (evt->eh_comp) |
1568 | complete(evt->eh_comp); | |
1569 | ||
072b91f9 BK |
1570 | ibmvfc_free_event(evt); |
1571 | } | |
1572 | ||
1573 | /** | |
1574 | * ibmvfc_host_chkready - Check if the host can accept commands | |
1575 | * @vhost: struct ibmvfc host | |
1576 | * | |
1577 | * Returns: | |
1578 | * 1 if host can accept command / 0 if not | |
1579 | **/ | |
1580 | static inline int ibmvfc_host_chkready(struct ibmvfc_host *vhost) | |
1581 | { | |
1582 | int result = 0; | |
1583 | ||
1584 | switch (vhost->state) { | |
1585 | case IBMVFC_LINK_DEAD: | |
1586 | case IBMVFC_HOST_OFFLINE: | |
1587 | result = DID_NO_CONNECT << 16; | |
1588 | break; | |
1589 | case IBMVFC_NO_CRQ: | |
1590 | case IBMVFC_INITIALIZING: | |
1591 | case IBMVFC_HALTED: | |
1592 | case IBMVFC_LINK_DOWN: | |
1593 | result = DID_REQUEUE << 16; | |
1594 | break; | |
1595 | case IBMVFC_ACTIVE: | |
1596 | result = 0; | |
1597 | break; | |
1598 | }; | |
1599 | ||
1600 | return result; | |
1601 | } | |
1602 | ||
1603 | /** | |
1604 | * ibmvfc_queuecommand - The queuecommand function of the scsi template | |
1605 | * @cmnd: struct scsi_cmnd to be executed | |
1606 | * @done: Callback function to be called when cmnd is completed | |
1607 | * | |
1608 | * Returns: | |
1609 | * 0 on success / other on failure | |
1610 | **/ | |
f281233d | 1611 | static int ibmvfc_queuecommand_lck(struct scsi_cmnd *cmnd, |
072b91f9 BK |
1612 | void (*done) (struct scsi_cmnd *)) |
1613 | { | |
1614 | struct ibmvfc_host *vhost = shost_priv(cmnd->device->host); | |
1615 | struct fc_rport *rport = starget_to_rport(scsi_target(cmnd->device)); | |
1616 | struct ibmvfc_cmd *vfc_cmd; | |
1617 | struct ibmvfc_event *evt; | |
1618 | u8 tag[2]; | |
1619 | int rc; | |
1620 | ||
1621 | if (unlikely((rc = fc_remote_port_chkready(rport))) || | |
1622 | unlikely((rc = ibmvfc_host_chkready(vhost)))) { | |
1623 | cmnd->result = rc; | |
1624 | done(cmnd); | |
1625 | return 0; | |
1626 | } | |
1627 | ||
1628 | cmnd->result = (DID_OK << 16); | |
1629 | evt = ibmvfc_get_event(vhost); | |
1630 | ibmvfc_init_event(evt, ibmvfc_scsi_done, IBMVFC_CMD_FORMAT); | |
1631 | evt->cmnd = cmnd; | |
1632 | cmnd->scsi_done = done; | |
1633 | vfc_cmd = &evt->iu.cmd; | |
1634 | memset(vfc_cmd, 0, sizeof(*vfc_cmd)); | |
0aab6c3f TD |
1635 | vfc_cmd->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); |
1636 | vfc_cmd->resp.len = cpu_to_be32(sizeof(vfc_cmd->rsp)); | |
1637 | vfc_cmd->frame_type = cpu_to_be32(IBMVFC_SCSI_FCP_TYPE); | |
1638 | vfc_cmd->payload_len = cpu_to_be32(sizeof(vfc_cmd->iu)); | |
1639 | vfc_cmd->resp_len = cpu_to_be32(sizeof(vfc_cmd->rsp)); | |
1640 | vfc_cmd->cancel_key = cpu_to_be32((unsigned long)cmnd->device->hostdata); | |
1641 | vfc_cmd->tgt_scsi_id = cpu_to_be64(rport->port_id); | |
1642 | vfc_cmd->iu.xfer_len = cpu_to_be32(scsi_bufflen(cmnd)); | |
072b91f9 BK |
1643 | int_to_scsilun(cmnd->device->lun, &vfc_cmd->iu.lun); |
1644 | memcpy(vfc_cmd->iu.cdb, cmnd->cmnd, cmnd->cmd_len); | |
1645 | ||
1646 | if (scsi_populate_tag_msg(cmnd, tag)) { | |
0aab6c3f | 1647 | vfc_cmd->task_tag = cpu_to_be64(tag[1]); |
072b91f9 BK |
1648 | switch (tag[0]) { |
1649 | case MSG_SIMPLE_TAG: | |
1650 | vfc_cmd->iu.pri_task_attr = IBMVFC_SIMPLE_TASK; | |
1651 | break; | |
1652 | case MSG_HEAD_TAG: | |
1653 | vfc_cmd->iu.pri_task_attr = IBMVFC_HEAD_OF_QUEUE; | |
1654 | break; | |
1655 | case MSG_ORDERED_TAG: | |
1656 | vfc_cmd->iu.pri_task_attr = IBMVFC_ORDERED_TASK; | |
1657 | break; | |
1658 | }; | |
1659 | } | |
1660 | ||
1661 | if (likely(!(rc = ibmvfc_map_sg_data(cmnd, evt, vfc_cmd, vhost->dev)))) | |
1662 | return ibmvfc_send_event(evt, vhost, 0); | |
1663 | ||
1664 | ibmvfc_free_event(evt); | |
1665 | if (rc == -ENOMEM) | |
1666 | return SCSI_MLQUEUE_HOST_BUSY; | |
1667 | ||
1668 | if (vhost->log_level > IBMVFC_DEFAULT_LOG_LEVEL) | |
1669 | scmd_printk(KERN_ERR, cmnd, | |
1670 | "Failed to map DMA buffer for command. rc=%d\n", rc); | |
1671 | ||
1672 | cmnd->result = DID_ERROR << 16; | |
1673 | done(cmnd); | |
1674 | return 0; | |
1675 | } | |
1676 | ||
f281233d JG |
1677 | static DEF_SCSI_QCMD(ibmvfc_queuecommand) |
1678 | ||
072b91f9 BK |
1679 | /** |
1680 | * ibmvfc_sync_completion - Signal that a synchronous command has completed | |
1681 | * @evt: ibmvfc event struct | |
1682 | * | |
1683 | **/ | |
1684 | static void ibmvfc_sync_completion(struct ibmvfc_event *evt) | |
1685 | { | |
1686 | /* copy the response back */ | |
1687 | if (evt->sync_iu) | |
1688 | *evt->sync_iu = *evt->xfer_iu; | |
1689 | ||
1690 | complete(&evt->comp); | |
1691 | } | |
1692 | ||
d31429e1 BK |
1693 | /** |
1694 | * ibmvfc_bsg_timeout_done - Completion handler for cancelling BSG commands | |
1695 | * @evt: struct ibmvfc_event | |
1696 | * | |
1697 | **/ | |
1698 | static void ibmvfc_bsg_timeout_done(struct ibmvfc_event *evt) | |
1699 | { | |
1700 | struct ibmvfc_host *vhost = evt->vhost; | |
1701 | ||
1702 | ibmvfc_free_event(evt); | |
1703 | vhost->aborting_passthru = 0; | |
1704 | dev_info(vhost->dev, "Passthru command cancelled\n"); | |
1705 | } | |
1706 | ||
1707 | /** | |
1708 | * ibmvfc_bsg_timeout - Handle a BSG timeout | |
1709 | * @job: struct fc_bsg_job that timed out | |
1710 | * | |
1711 | * Returns: | |
1712 | * 0 on success / other on failure | |
1713 | **/ | |
1714 | static int ibmvfc_bsg_timeout(struct fc_bsg_job *job) | |
1715 | { | |
1716 | struct ibmvfc_host *vhost = shost_priv(job->shost); | |
1717 | unsigned long port_id = (unsigned long)job->dd_data; | |
1718 | struct ibmvfc_event *evt; | |
1719 | struct ibmvfc_tmf *tmf; | |
1720 | unsigned long flags; | |
1721 | int rc; | |
1722 | ||
1723 | ENTER; | |
1724 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
1725 | if (vhost->aborting_passthru || vhost->state != IBMVFC_ACTIVE) { | |
1726 | __ibmvfc_reset_host(vhost); | |
1727 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
1728 | return 0; | |
1729 | } | |
1730 | ||
1731 | vhost->aborting_passthru = 1; | |
1732 | evt = ibmvfc_get_event(vhost); | |
1733 | ibmvfc_init_event(evt, ibmvfc_bsg_timeout_done, IBMVFC_MAD_FORMAT); | |
1734 | ||
1735 | tmf = &evt->iu.tmf; | |
1736 | memset(tmf, 0, sizeof(*tmf)); | |
0aab6c3f TD |
1737 | tmf->common.version = cpu_to_be32(1); |
1738 | tmf->common.opcode = cpu_to_be32(IBMVFC_TMF_MAD); | |
1739 | tmf->common.length = cpu_to_be16(sizeof(*tmf)); | |
1740 | tmf->scsi_id = cpu_to_be64(port_id); | |
1741 | tmf->cancel_key = cpu_to_be32(IBMVFC_PASSTHRU_CANCEL_KEY); | |
1742 | tmf->my_cancel_key = cpu_to_be32(IBMVFC_INTERNAL_CANCEL_KEY); | |
d31429e1 BK |
1743 | rc = ibmvfc_send_event(evt, vhost, default_timeout); |
1744 | ||
1745 | if (rc != 0) { | |
1746 | vhost->aborting_passthru = 0; | |
1747 | dev_err(vhost->dev, "Failed to send cancel event. rc=%d\n", rc); | |
1748 | rc = -EIO; | |
1749 | } else | |
1750 | dev_info(vhost->dev, "Cancelling passthru command to port id 0x%lx\n", | |
1751 | port_id); | |
1752 | ||
1753 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
1754 | ||
1755 | LEAVE; | |
1756 | return rc; | |
1757 | } | |
1758 | ||
1759 | /** | |
1760 | * ibmvfc_bsg_plogi - PLOGI into a target to handle a BSG command | |
1761 | * @vhost: struct ibmvfc_host to send command | |
1762 | * @port_id: port ID to send command | |
1763 | * | |
1764 | * Returns: | |
1765 | * 0 on success / other on failure | |
1766 | **/ | |
1767 | static int ibmvfc_bsg_plogi(struct ibmvfc_host *vhost, unsigned int port_id) | |
1768 | { | |
1769 | struct ibmvfc_port_login *plogi; | |
1770 | struct ibmvfc_target *tgt; | |
1771 | struct ibmvfc_event *evt; | |
1772 | union ibmvfc_iu rsp_iu; | |
1773 | unsigned long flags; | |
1774 | int rc = 0, issue_login = 1; | |
1775 | ||
1776 | ENTER; | |
1777 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
1778 | list_for_each_entry(tgt, &vhost->targets, queue) { | |
1779 | if (tgt->scsi_id == port_id) { | |
1780 | issue_login = 0; | |
1781 | break; | |
1782 | } | |
1783 | } | |
1784 | ||
1785 | if (!issue_login) | |
1786 | goto unlock_out; | |
1787 | if (unlikely((rc = ibmvfc_host_chkready(vhost)))) | |
1788 | goto unlock_out; | |
1789 | ||
1790 | evt = ibmvfc_get_event(vhost); | |
1791 | ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); | |
1792 | plogi = &evt->iu.plogi; | |
1793 | memset(plogi, 0, sizeof(*plogi)); | |
0aab6c3f TD |
1794 | plogi->common.version = cpu_to_be32(1); |
1795 | plogi->common.opcode = cpu_to_be32(IBMVFC_PORT_LOGIN); | |
1796 | plogi->common.length = cpu_to_be16(sizeof(*plogi)); | |
1797 | plogi->scsi_id = cpu_to_be64(port_id); | |
d31429e1 BK |
1798 | evt->sync_iu = &rsp_iu; |
1799 | init_completion(&evt->comp); | |
1800 | ||
1801 | rc = ibmvfc_send_event(evt, vhost, default_timeout); | |
1802 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
1803 | ||
1804 | if (rc) | |
1805 | return -EIO; | |
1806 | ||
1807 | wait_for_completion(&evt->comp); | |
1808 | ||
1809 | if (rsp_iu.plogi.common.status) | |
1810 | rc = -EIO; | |
1811 | ||
1812 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
1813 | ibmvfc_free_event(evt); | |
1814 | unlock_out: | |
1815 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
1816 | LEAVE; | |
1817 | return rc; | |
1818 | } | |
1819 | ||
1820 | /** | |
1821 | * ibmvfc_bsg_request - Handle a BSG request | |
1822 | * @job: struct fc_bsg_job to be executed | |
1823 | * | |
1824 | * Returns: | |
1825 | * 0 on success / other on failure | |
1826 | **/ | |
1827 | static int ibmvfc_bsg_request(struct fc_bsg_job *job) | |
1828 | { | |
1829 | struct ibmvfc_host *vhost = shost_priv(job->shost); | |
1830 | struct fc_rport *rport = job->rport; | |
1831 | struct ibmvfc_passthru_mad *mad; | |
1832 | struct ibmvfc_event *evt; | |
1833 | union ibmvfc_iu rsp_iu; | |
1834 | unsigned long flags, port_id = -1; | |
1835 | unsigned int code = job->request->msgcode; | |
1836 | int rc = 0, req_seg, rsp_seg, issue_login = 0; | |
1837 | u32 fc_flags, rsp_len; | |
1838 | ||
1839 | ENTER; | |
1840 | job->reply->reply_payload_rcv_len = 0; | |
1841 | if (rport) | |
1842 | port_id = rport->port_id; | |
1843 | ||
1844 | switch (code) { | |
1845 | case FC_BSG_HST_ELS_NOLOGIN: | |
1846 | port_id = (job->request->rqst_data.h_els.port_id[0] << 16) | | |
1847 | (job->request->rqst_data.h_els.port_id[1] << 8) | | |
1848 | job->request->rqst_data.h_els.port_id[2]; | |
1849 | case FC_BSG_RPT_ELS: | |
1850 | fc_flags = IBMVFC_FC_ELS; | |
1851 | break; | |
1852 | case FC_BSG_HST_CT: | |
1853 | issue_login = 1; | |
1854 | port_id = (job->request->rqst_data.h_ct.port_id[0] << 16) | | |
1855 | (job->request->rqst_data.h_ct.port_id[1] << 8) | | |
1856 | job->request->rqst_data.h_ct.port_id[2]; | |
1857 | case FC_BSG_RPT_CT: | |
1858 | fc_flags = IBMVFC_FC_CT_IU; | |
1859 | break; | |
1860 | default: | |
1861 | return -ENOTSUPP; | |
1862 | }; | |
1863 | ||
1864 | if (port_id == -1) | |
1865 | return -EINVAL; | |
1866 | if (!mutex_trylock(&vhost->passthru_mutex)) | |
1867 | return -EBUSY; | |
1868 | ||
1869 | job->dd_data = (void *)port_id; | |
1870 | req_seg = dma_map_sg(vhost->dev, job->request_payload.sg_list, | |
1871 | job->request_payload.sg_cnt, DMA_TO_DEVICE); | |
1872 | ||
1873 | if (!req_seg) { | |
1874 | mutex_unlock(&vhost->passthru_mutex); | |
1875 | return -ENOMEM; | |
1876 | } | |
1877 | ||
1878 | rsp_seg = dma_map_sg(vhost->dev, job->reply_payload.sg_list, | |
1879 | job->reply_payload.sg_cnt, DMA_FROM_DEVICE); | |
1880 | ||
1881 | if (!rsp_seg) { | |
1882 | dma_unmap_sg(vhost->dev, job->request_payload.sg_list, | |
1883 | job->request_payload.sg_cnt, DMA_TO_DEVICE); | |
1884 | mutex_unlock(&vhost->passthru_mutex); | |
1885 | return -ENOMEM; | |
1886 | } | |
1887 | ||
1888 | if (req_seg > 1 || rsp_seg > 1) { | |
1889 | rc = -EINVAL; | |
1890 | goto out; | |
1891 | } | |
1892 | ||
1893 | if (issue_login) | |
1894 | rc = ibmvfc_bsg_plogi(vhost, port_id); | |
1895 | ||
1896 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
1897 | ||
1898 | if (unlikely(rc || (rport && (rc = fc_remote_port_chkready(rport)))) || | |
1899 | unlikely((rc = ibmvfc_host_chkready(vhost)))) { | |
1900 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
1901 | goto out; | |
1902 | } | |
1903 | ||
1904 | evt = ibmvfc_get_event(vhost); | |
1905 | ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); | |
1906 | mad = &evt->iu.passthru; | |
1907 | ||
1908 | memset(mad, 0, sizeof(*mad)); | |
0aab6c3f TD |
1909 | mad->common.version = cpu_to_be32(1); |
1910 | mad->common.opcode = cpu_to_be32(IBMVFC_PASSTHRU); | |
1911 | mad->common.length = cpu_to_be16(sizeof(*mad) - sizeof(mad->fc_iu) - sizeof(mad->iu)); | |
1912 | ||
1913 | mad->cmd_ioba.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + | |
1914 | offsetof(struct ibmvfc_passthru_mad, iu)); | |
1915 | mad->cmd_ioba.len = cpu_to_be32(sizeof(mad->iu)); | |
1916 | ||
1917 | mad->iu.cmd_len = cpu_to_be32(job->request_payload.payload_len); | |
1918 | mad->iu.rsp_len = cpu_to_be32(job->reply_payload.payload_len); | |
1919 | mad->iu.flags = cpu_to_be32(fc_flags); | |
1920 | mad->iu.cancel_key = cpu_to_be32(IBMVFC_PASSTHRU_CANCEL_KEY); | |
1921 | ||
1922 | mad->iu.cmd.va = cpu_to_be64(sg_dma_address(job->request_payload.sg_list)); | |
1923 | mad->iu.cmd.len = cpu_to_be32(sg_dma_len(job->request_payload.sg_list)); | |
1924 | mad->iu.rsp.va = cpu_to_be64(sg_dma_address(job->reply_payload.sg_list)); | |
1925 | mad->iu.rsp.len = cpu_to_be32(sg_dma_len(job->reply_payload.sg_list)); | |
1926 | mad->iu.scsi_id = cpu_to_be64(port_id); | |
1927 | mad->iu.tag = cpu_to_be64((u64)evt); | |
1928 | rsp_len = be32_to_cpu(mad->iu.rsp.len); | |
d31429e1 BK |
1929 | |
1930 | evt->sync_iu = &rsp_iu; | |
1931 | init_completion(&evt->comp); | |
1932 | rc = ibmvfc_send_event(evt, vhost, 0); | |
1933 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
1934 | ||
1935 | if (rc) { | |
1936 | rc = -EIO; | |
1937 | goto out; | |
1938 | } | |
1939 | ||
1940 | wait_for_completion(&evt->comp); | |
1941 | ||
1942 | if (rsp_iu.passthru.common.status) | |
1943 | rc = -EIO; | |
1944 | else | |
1945 | job->reply->reply_payload_rcv_len = rsp_len; | |
1946 | ||
1947 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
1948 | ibmvfc_free_event(evt); | |
1949 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
1950 | job->reply->result = rc; | |
1951 | job->job_done(job); | |
1952 | rc = 0; | |
1953 | out: | |
1954 | dma_unmap_sg(vhost->dev, job->request_payload.sg_list, | |
1955 | job->request_payload.sg_cnt, DMA_TO_DEVICE); | |
1956 | dma_unmap_sg(vhost->dev, job->reply_payload.sg_list, | |
1957 | job->reply_payload.sg_cnt, DMA_FROM_DEVICE); | |
1958 | mutex_unlock(&vhost->passthru_mutex); | |
1959 | LEAVE; | |
1960 | return rc; | |
1961 | } | |
1962 | ||
072b91f9 BK |
1963 | /** |
1964 | * ibmvfc_reset_device - Reset the device with the specified reset type | |
1965 | * @sdev: scsi device to reset | |
1966 | * @type: reset type | |
1967 | * @desc: reset type description for log messages | |
1968 | * | |
1969 | * Returns: | |
1970 | * 0 on success / other on failure | |
1971 | **/ | |
1972 | static int ibmvfc_reset_device(struct scsi_device *sdev, int type, char *desc) | |
1973 | { | |
1974 | struct ibmvfc_host *vhost = shost_priv(sdev->host); | |
1975 | struct fc_rport *rport = starget_to_rport(scsi_target(sdev)); | |
1976 | struct ibmvfc_cmd *tmf; | |
50ed9a00 | 1977 | struct ibmvfc_event *evt = NULL; |
072b91f9 BK |
1978 | union ibmvfc_iu rsp_iu; |
1979 | struct ibmvfc_fcp_rsp *fc_rsp = &rsp_iu.cmd.rsp; | |
1980 | int rsp_rc = -EBUSY; | |
1981 | unsigned long flags; | |
1982 | int rsp_code = 0; | |
1983 | ||
1984 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
1985 | if (vhost->state == IBMVFC_ACTIVE) { | |
1986 | evt = ibmvfc_get_event(vhost); | |
1987 | ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_CMD_FORMAT); | |
1988 | ||
1989 | tmf = &evt->iu.cmd; | |
1990 | memset(tmf, 0, sizeof(*tmf)); | |
0aab6c3f TD |
1991 | tmf->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); |
1992 | tmf->resp.len = cpu_to_be32(sizeof(tmf->rsp)); | |
1993 | tmf->frame_type = cpu_to_be32(IBMVFC_SCSI_FCP_TYPE); | |
1994 | tmf->payload_len = cpu_to_be32(sizeof(tmf->iu)); | |
1995 | tmf->resp_len = cpu_to_be32(sizeof(tmf->rsp)); | |
1996 | tmf->cancel_key = cpu_to_be32((unsigned long)sdev->hostdata); | |
1997 | tmf->tgt_scsi_id = cpu_to_be64(rport->port_id); | |
072b91f9 | 1998 | int_to_scsilun(sdev->lun, &tmf->iu.lun); |
0aab6c3f | 1999 | tmf->flags = cpu_to_be16((IBMVFC_NO_MEM_DESC | IBMVFC_TMF)); |
072b91f9 BK |
2000 | tmf->iu.tmf_flags = type; |
2001 | evt->sync_iu = &rsp_iu; | |
2002 | ||
2003 | init_completion(&evt->comp); | |
2004 | rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); | |
2005 | } | |
2006 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
2007 | ||
2008 | if (rsp_rc != 0) { | |
2009 | sdev_printk(KERN_ERR, sdev, "Failed to send %s reset event. rc=%d\n", | |
2010 | desc, rsp_rc); | |
2011 | return -EIO; | |
2012 | } | |
2013 | ||
2014 | sdev_printk(KERN_INFO, sdev, "Resetting %s\n", desc); | |
2015 | wait_for_completion(&evt->comp); | |
2016 | ||
230934a6 BK |
2017 | if (rsp_iu.cmd.status) |
2018 | rsp_code = ibmvfc_get_err_result(&rsp_iu.cmd); | |
2019 | ||
2020 | if (rsp_code) { | |
072b91f9 BK |
2021 | if (fc_rsp->flags & FCP_RSP_LEN_VALID) |
2022 | rsp_code = fc_rsp->data.info.rsp_code; | |
2023 | ||
2024 | sdev_printk(KERN_ERR, sdev, "%s reset failed: %s (%x:%x) " | |
0aab6c3f TD |
2025 | "flags: %x fcp_rsp: %x, scsi_status: %x\n", desc, |
2026 | ibmvfc_get_cmd_error(be16_to_cpu(rsp_iu.cmd.status), be16_to_cpu(rsp_iu.cmd.error)), | |
072b91f9 BK |
2027 | rsp_iu.cmd.status, rsp_iu.cmd.error, fc_rsp->flags, rsp_code, |
2028 | fc_rsp->scsi_status); | |
2029 | rsp_rc = -EIO; | |
2030 | } else | |
2031 | sdev_printk(KERN_INFO, sdev, "%s reset successful\n", desc); | |
2032 | ||
2033 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
2034 | ibmvfc_free_event(evt); | |
2035 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
2036 | return rsp_rc; | |
2037 | } | |
2038 | ||
2039 | /** | |
d2fab5cf BK |
2040 | * ibmvfc_match_rport - Match function for specified remote port |
2041 | * @evt: ibmvfc event struct | |
2042 | * @device: device to match (rport) | |
072b91f9 BK |
2043 | * |
2044 | * Returns: | |
d2fab5cf | 2045 | * 1 if event matches rport / 0 if event does not match rport |
072b91f9 | 2046 | **/ |
d2fab5cf | 2047 | static int ibmvfc_match_rport(struct ibmvfc_event *evt, void *rport) |
072b91f9 | 2048 | { |
d2fab5cf | 2049 | struct fc_rport *cmd_rport; |
072b91f9 | 2050 | |
d2fab5cf BK |
2051 | if (evt->cmnd) { |
2052 | cmd_rport = starget_to_rport(scsi_target(evt->cmnd->device)); | |
2053 | if (cmd_rport == rport) | |
2054 | return 1; | |
072b91f9 | 2055 | } |
d2fab5cf BK |
2056 | return 0; |
2057 | } | |
072b91f9 | 2058 | |
d2fab5cf BK |
2059 | /** |
2060 | * ibmvfc_match_target - Match function for specified target | |
2061 | * @evt: ibmvfc event struct | |
2062 | * @device: device to match (starget) | |
2063 | * | |
2064 | * Returns: | |
2065 | * 1 if event matches starget / 0 if event does not match starget | |
2066 | **/ | |
2067 | static int ibmvfc_match_target(struct ibmvfc_event *evt, void *device) | |
2068 | { | |
2069 | if (evt->cmnd && scsi_target(evt->cmnd->device) == device) | |
2070 | return 1; | |
2071 | return 0; | |
2072 | } | |
072b91f9 | 2073 | |
d2fab5cf BK |
2074 | /** |
2075 | * ibmvfc_match_lun - Match function for specified LUN | |
2076 | * @evt: ibmvfc event struct | |
2077 | * @device: device to match (sdev) | |
2078 | * | |
2079 | * Returns: | |
2080 | * 1 if event matches sdev / 0 if event does not match sdev | |
2081 | **/ | |
2082 | static int ibmvfc_match_lun(struct ibmvfc_event *evt, void *device) | |
2083 | { | |
2084 | if (evt->cmnd && evt->cmnd->device == device) | |
2085 | return 1; | |
2086 | return 0; | |
2087 | } | |
072b91f9 | 2088 | |
d2fab5cf BK |
2089 | /** |
2090 | * ibmvfc_wait_for_ops - Wait for ops to complete | |
2091 | * @vhost: ibmvfc host struct | |
2092 | * @device: device to match (starget or sdev) | |
2093 | * @match: match function | |
2094 | * | |
2095 | * Returns: | |
2096 | * SUCCESS / FAILED | |
2097 | **/ | |
2098 | static int ibmvfc_wait_for_ops(struct ibmvfc_host *vhost, void *device, | |
2099 | int (*match) (struct ibmvfc_event *, void *)) | |
2100 | { | |
2101 | struct ibmvfc_event *evt; | |
2102 | DECLARE_COMPLETION_ONSTACK(comp); | |
2103 | int wait; | |
2104 | unsigned long flags; | |
2105 | signed long timeout = IBMVFC_ABORT_WAIT_TIMEOUT * HZ; | |
072b91f9 | 2106 | |
d2fab5cf BK |
2107 | ENTER; |
2108 | do { | |
2109 | wait = 0; | |
2110 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
2111 | list_for_each_entry(evt, &vhost->sent, queue) { | |
2112 | if (match(evt, device)) { | |
2113 | evt->eh_comp = ∁ | |
2114 | wait++; | |
2115 | } | |
2116 | } | |
2117 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
230934a6 | 2118 | |
d2fab5cf BK |
2119 | if (wait) { |
2120 | timeout = wait_for_completion_timeout(&comp, timeout); | |
072b91f9 | 2121 | |
d2fab5cf BK |
2122 | if (!timeout) { |
2123 | wait = 0; | |
2124 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
2125 | list_for_each_entry(evt, &vhost->sent, queue) { | |
2126 | if (match(evt, device)) { | |
2127 | evt->eh_comp = NULL; | |
2128 | wait++; | |
2129 | } | |
2130 | } | |
2131 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
2132 | if (wait) | |
2133 | dev_err(vhost->dev, "Timed out waiting for aborted commands\n"); | |
2134 | LEAVE; | |
2135 | return wait ? FAILED : SUCCESS; | |
2136 | } | |
2137 | } | |
2138 | } while (wait); | |
072b91f9 | 2139 | |
d2fab5cf BK |
2140 | LEAVE; |
2141 | return SUCCESS; | |
072b91f9 BK |
2142 | } |
2143 | ||
2144 | /** | |
2145 | * ibmvfc_cancel_all - Cancel all outstanding commands to the device | |
2146 | * @sdev: scsi device to cancel commands | |
2147 | * @type: type of error recovery being performed | |
2148 | * | |
2149 | * This sends a cancel to the VIOS for the specified device. This does | |
2150 | * NOT send any abort to the actual device. That must be done separately. | |
2151 | * | |
2152 | * Returns: | |
2153 | * 0 on success / other on failure | |
2154 | **/ | |
2155 | static int ibmvfc_cancel_all(struct scsi_device *sdev, int type) | |
2156 | { | |
2157 | struct ibmvfc_host *vhost = shost_priv(sdev->host); | |
ad8dcffa BK |
2158 | struct scsi_target *starget = scsi_target(sdev); |
2159 | struct fc_rport *rport = starget_to_rport(starget); | |
072b91f9 BK |
2160 | struct ibmvfc_tmf *tmf; |
2161 | struct ibmvfc_event *evt, *found_evt; | |
2162 | union ibmvfc_iu rsp; | |
2163 | int rsp_rc = -EBUSY; | |
2164 | unsigned long flags; | |
2165 | u16 status; | |
2166 | ||
2167 | ENTER; | |
2168 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
2169 | found_evt = NULL; | |
2170 | list_for_each_entry(evt, &vhost->sent, queue) { | |
2171 | if (evt->cmnd && evt->cmnd->device == sdev) { | |
2172 | found_evt = evt; | |
2173 | break; | |
2174 | } | |
2175 | } | |
2176 | ||
2177 | if (!found_evt) { | |
2178 | if (vhost->log_level > IBMVFC_DEFAULT_LOG_LEVEL) | |
2179 | sdev_printk(KERN_INFO, sdev, "No events found to cancel\n"); | |
2180 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
2181 | return 0; | |
2182 | } | |
2183 | ||
55d29bf0 | 2184 | if (vhost->logged_in) { |
072b91f9 BK |
2185 | evt = ibmvfc_get_event(vhost); |
2186 | ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_MAD_FORMAT); | |
2187 | ||
2188 | tmf = &evt->iu.tmf; | |
2189 | memset(tmf, 0, sizeof(*tmf)); | |
0aab6c3f TD |
2190 | tmf->common.version = cpu_to_be32(1); |
2191 | tmf->common.opcode = cpu_to_be32(IBMVFC_TMF_MAD); | |
2192 | tmf->common.length = cpu_to_be16(sizeof(*tmf)); | |
2193 | tmf->scsi_id = cpu_to_be64(rport->port_id); | |
072b91f9 | 2194 | int_to_scsilun(sdev->lun, &tmf->lun); |
0aab6c3f | 2195 | if (!(be64_to_cpu(vhost->login_buf->resp.capabilities) & IBMVFC_CAN_SUPPRESS_ABTS)) |
90f725db | 2196 | type &= ~IBMVFC_TMF_SUPPRESS_ABTS; |
55d29bf0 | 2197 | if (vhost->state == IBMVFC_ACTIVE) |
0aab6c3f | 2198 | tmf->flags = cpu_to_be32((type | IBMVFC_TMF_LUA_VALID)); |
55d29bf0 | 2199 | else |
0aab6c3f TD |
2200 | tmf->flags = cpu_to_be32(((type & IBMVFC_TMF_SUPPRESS_ABTS) | IBMVFC_TMF_LUA_VALID)); |
2201 | tmf->cancel_key = cpu_to_be32((unsigned long)sdev->hostdata); | |
2202 | tmf->my_cancel_key = cpu_to_be32((unsigned long)starget->hostdata); | |
072b91f9 BK |
2203 | |
2204 | evt->sync_iu = &rsp; | |
2205 | init_completion(&evt->comp); | |
2206 | rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); | |
2207 | } | |
2208 | ||
2209 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
2210 | ||
2211 | if (rsp_rc != 0) { | |
2212 | sdev_printk(KERN_ERR, sdev, "Failed to send cancel event. rc=%d\n", rsp_rc); | |
c281e32a BK |
2213 | /* If failure is received, the host adapter is most likely going |
2214 | through reset, return success so the caller will wait for the command | |
2215 | being cancelled to get returned */ | |
2216 | return 0; | |
072b91f9 BK |
2217 | } |
2218 | ||
2219 | sdev_printk(KERN_INFO, sdev, "Cancelling outstanding commands.\n"); | |
2220 | ||
2221 | wait_for_completion(&evt->comp); | |
0aab6c3f | 2222 | status = be16_to_cpu(rsp.mad_common.status); |
072b91f9 BK |
2223 | spin_lock_irqsave(vhost->host->host_lock, flags); |
2224 | ibmvfc_free_event(evt); | |
2225 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
2226 | ||
2227 | if (status != IBMVFC_MAD_SUCCESS) { | |
2228 | sdev_printk(KERN_WARNING, sdev, "Cancel failed with rc=%x\n", status); | |
c281e32a BK |
2229 | switch (status) { |
2230 | case IBMVFC_MAD_DRIVER_FAILED: | |
2231 | case IBMVFC_MAD_CRQ_ERROR: | |
2232 | /* Host adapter most likely going through reset, return success to | |
2233 | the caller will wait for the command being cancelled to get returned */ | |
2234 | return 0; | |
2235 | default: | |
2236 | return -EIO; | |
2237 | }; | |
072b91f9 BK |
2238 | } |
2239 | ||
2240 | sdev_printk(KERN_INFO, sdev, "Successfully cancelled outstanding commands\n"); | |
2241 | return 0; | |
2242 | } | |
2243 | ||
ad8dcffa | 2244 | /** |
d2fab5cf | 2245 | * ibmvfc_match_key - Match function for specified cancel key |
ad8dcffa | 2246 | * @evt: ibmvfc event struct |
d2fab5cf | 2247 | * @key: cancel key to match |
ad8dcffa BK |
2248 | * |
2249 | * Returns: | |
d2fab5cf | 2250 | * 1 if event matches key / 0 if event does not match key |
ad8dcffa | 2251 | **/ |
d2fab5cf | 2252 | static int ibmvfc_match_key(struct ibmvfc_event *evt, void *key) |
ad8dcffa | 2253 | { |
d2fab5cf | 2254 | unsigned long cancel_key = (unsigned long)key; |
ad8dcffa | 2255 | |
d2fab5cf | 2256 | if (evt->crq.format == IBMVFC_CMD_FORMAT && |
0aab6c3f | 2257 | be32_to_cpu(evt->iu.cmd.cancel_key) == cancel_key) |
ad8dcffa BK |
2258 | return 1; |
2259 | return 0; | |
2260 | } | |
2261 | ||
a077c7fa BK |
2262 | /** |
2263 | * ibmvfc_match_evt - Match function for specified event | |
2264 | * @evt: ibmvfc event struct | |
2265 | * @match: event to match | |
2266 | * | |
2267 | * Returns: | |
2268 | * 1 if event matches key / 0 if event does not match key | |
2269 | **/ | |
2270 | static int ibmvfc_match_evt(struct ibmvfc_event *evt, void *match) | |
2271 | { | |
2272 | if (evt == match) | |
2273 | return 1; | |
2274 | return 0; | |
2275 | } | |
2276 | ||
ad8dcffa | 2277 | /** |
d2fab5cf BK |
2278 | * ibmvfc_abort_task_set - Abort outstanding commands to the device |
2279 | * @sdev: scsi device to abort commands | |
2280 | * | |
2281 | * This sends an Abort Task Set to the VIOS for the specified device. This does | |
2282 | * NOT send any cancel to the VIOS. That must be done separately. | |
ad8dcffa BK |
2283 | * |
2284 | * Returns: | |
d2fab5cf | 2285 | * 0 on success / other on failure |
ad8dcffa | 2286 | **/ |
d2fab5cf | 2287 | static int ibmvfc_abort_task_set(struct scsi_device *sdev) |
ad8dcffa | 2288 | { |
d2fab5cf BK |
2289 | struct ibmvfc_host *vhost = shost_priv(sdev->host); |
2290 | struct fc_rport *rport = starget_to_rport(scsi_target(sdev)); | |
2291 | struct ibmvfc_cmd *tmf; | |
2292 | struct ibmvfc_event *evt, *found_evt; | |
2293 | union ibmvfc_iu rsp_iu; | |
2294 | struct ibmvfc_fcp_rsp *fc_rsp = &rsp_iu.cmd.rsp; | |
2295 | int rc, rsp_rc = -EBUSY; | |
2296 | unsigned long flags, timeout = IBMVFC_ABORT_TIMEOUT; | |
2297 | int rsp_code = 0; | |
ad8dcffa | 2298 | |
d2fab5cf BK |
2299 | spin_lock_irqsave(vhost->host->host_lock, flags); |
2300 | found_evt = NULL; | |
2301 | list_for_each_entry(evt, &vhost->sent, queue) { | |
2302 | if (evt->cmnd && evt->cmnd->device == sdev) { | |
2303 | found_evt = evt; | |
2304 | break; | |
ad8dcffa | 2305 | } |
d2fab5cf BK |
2306 | } |
2307 | ||
2308 | if (!found_evt) { | |
2309 | if (vhost->log_level > IBMVFC_DEFAULT_LOG_LEVEL) | |
2310 | sdev_printk(KERN_INFO, sdev, "No events found to abort\n"); | |
ad8dcffa | 2311 | spin_unlock_irqrestore(vhost->host->host_lock, flags); |
d2fab5cf BK |
2312 | return 0; |
2313 | } | |
ad8dcffa | 2314 | |
d2fab5cf BK |
2315 | if (vhost->state == IBMVFC_ACTIVE) { |
2316 | evt = ibmvfc_get_event(vhost); | |
2317 | ibmvfc_init_event(evt, ibmvfc_sync_completion, IBMVFC_CMD_FORMAT); | |
ad8dcffa | 2318 | |
d2fab5cf BK |
2319 | tmf = &evt->iu.cmd; |
2320 | memset(tmf, 0, sizeof(*tmf)); | |
0aab6c3f TD |
2321 | tmf->resp.va = cpu_to_be64(be64_to_cpu(evt->crq.ioba) + offsetof(struct ibmvfc_cmd, rsp)); |
2322 | tmf->resp.len = cpu_to_be32(sizeof(tmf->rsp)); | |
2323 | tmf->frame_type = cpu_to_be32(IBMVFC_SCSI_FCP_TYPE); | |
2324 | tmf->payload_len = cpu_to_be32(sizeof(tmf->iu)); | |
2325 | tmf->resp_len = cpu_to_be32(sizeof(tmf->rsp)); | |
2326 | tmf->cancel_key = cpu_to_be32((unsigned long)sdev->hostdata); | |
2327 | tmf->tgt_scsi_id = cpu_to_be64(rport->port_id); | |
d2fab5cf | 2328 | int_to_scsilun(sdev->lun, &tmf->iu.lun); |
0aab6c3f | 2329 | tmf->flags = cpu_to_be16((IBMVFC_NO_MEM_DESC | IBMVFC_TMF)); |
d2fab5cf BK |
2330 | tmf->iu.tmf_flags = IBMVFC_ABORT_TASK_SET; |
2331 | evt->sync_iu = &rsp_iu; | |
2332 | ||
2333 | init_completion(&evt->comp); | |
2334 | rsp_rc = ibmvfc_send_event(evt, vhost, default_timeout); | |
2335 | } | |
2336 | ||
2337 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
2338 | ||
2339 | if (rsp_rc != 0) { | |
2340 | sdev_printk(KERN_ERR, sdev, "Failed to send abort. rc=%d\n", rsp_rc); | |
2341 | return -EIO; | |
2342 | } | |
2343 | ||
2344 | sdev_printk(KERN_INFO, sdev, "Aborting outstanding commands\n"); | |
2345 | timeout = wait_for_completion_timeout(&evt->comp, timeout); | |
2346 | ||
2347 | if (!timeout) { | |
f8804b72 | 2348 | rc = ibmvfc_cancel_all(sdev, 0); |
d2fab5cf BK |
2349 | if (!rc) { |
2350 | rc = ibmvfc_wait_for_ops(vhost, sdev->hostdata, ibmvfc_match_key); | |
2351 | if (rc == SUCCESS) | |
2352 | rc = 0; | |
ad8dcffa | 2353 | } |
ad8dcffa | 2354 | |
d2fab5cf BK |
2355 | if (rc) { |
2356 | sdev_printk(KERN_INFO, sdev, "Cancel failed, resetting host\n"); | |
2357 | ibmvfc_reset_host(vhost); | |
a077c7fa BK |
2358 | rsp_rc = -EIO; |
2359 | rc = ibmvfc_wait_for_ops(vhost, sdev->hostdata, ibmvfc_match_key); | |
2360 | ||
2361 | if (rc == SUCCESS) | |
2362 | rsp_rc = 0; | |
2363 | ||
2364 | rc = ibmvfc_wait_for_ops(vhost, evt, ibmvfc_match_evt); | |
2365 | if (rc != SUCCESS) { | |
2366 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
2367 | ibmvfc_hard_reset_host(vhost); | |
2368 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
2369 | rsp_rc = 0; | |
2370 | } | |
2371 | ||
d2fab5cf BK |
2372 | goto out; |
2373 | } | |
2374 | } | |
2375 | ||
2376 | if (rsp_iu.cmd.status) | |
2377 | rsp_code = ibmvfc_get_err_result(&rsp_iu.cmd); | |
2378 | ||
2379 | if (rsp_code) { | |
2380 | if (fc_rsp->flags & FCP_RSP_LEN_VALID) | |
2381 | rsp_code = fc_rsp->data.info.rsp_code; | |
2382 | ||
2383 | sdev_printk(KERN_ERR, sdev, "Abort failed: %s (%x:%x) " | |
2384 | "flags: %x fcp_rsp: %x, scsi_status: %x\n", | |
0aab6c3f | 2385 | ibmvfc_get_cmd_error(be16_to_cpu(rsp_iu.cmd.status), be16_to_cpu(rsp_iu.cmd.error)), |
d2fab5cf BK |
2386 | rsp_iu.cmd.status, rsp_iu.cmd.error, fc_rsp->flags, rsp_code, |
2387 | fc_rsp->scsi_status); | |
2388 | rsp_rc = -EIO; | |
2389 | } else | |
2390 | sdev_printk(KERN_INFO, sdev, "Abort successful\n"); | |
2391 | ||
2392 | out: | |
2393 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
2394 | ibmvfc_free_event(evt); | |
2395 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
2396 | return rsp_rc; | |
ad8dcffa BK |
2397 | } |
2398 | ||
072b91f9 BK |
2399 | /** |
2400 | * ibmvfc_eh_abort_handler - Abort a command | |
2401 | * @cmd: scsi command to abort | |
2402 | * | |
2403 | * Returns: | |
93631b4a | 2404 | * SUCCESS / FAST_IO_FAIL / FAILED |
072b91f9 BK |
2405 | **/ |
2406 | static int ibmvfc_eh_abort_handler(struct scsi_cmnd *cmd) | |
2407 | { | |
ad8dcffa BK |
2408 | struct scsi_device *sdev = cmd->device; |
2409 | struct ibmvfc_host *vhost = shost_priv(sdev->host); | |
55d29bf0 | 2410 | int cancel_rc, block_rc; |
ad8dcffa | 2411 | int rc = FAILED; |
072b91f9 BK |
2412 | |
2413 | ENTER; | |
93631b4a | 2414 | block_rc = fc_block_scsi_eh(cmd); |
072b91f9 | 2415 | ibmvfc_wait_while_resetting(vhost); |
93631b4a BK |
2416 | if (block_rc != FAST_IO_FAIL) { |
2417 | cancel_rc = ibmvfc_cancel_all(sdev, IBMVFC_TMF_ABORT_TASK_SET); | |
55d29bf0 | 2418 | ibmvfc_abort_task_set(sdev); |
93631b4a | 2419 | } else |
90f725db | 2420 | cancel_rc = ibmvfc_cancel_all(sdev, IBMVFC_TMF_SUPPRESS_ABTS); |
072b91f9 | 2421 | |
55d29bf0 | 2422 | if (!cancel_rc) |
ad8dcffa | 2423 | rc = ibmvfc_wait_for_ops(vhost, sdev, ibmvfc_match_lun); |
072b91f9 | 2424 | |
93631b4a BK |
2425 | if (block_rc == FAST_IO_FAIL && rc != FAILED) |
2426 | rc = FAST_IO_FAIL; | |
2427 | ||
072b91f9 | 2428 | LEAVE; |
ad8dcffa | 2429 | return rc; |
072b91f9 BK |
2430 | } |
2431 | ||
2432 | /** | |
2433 | * ibmvfc_eh_device_reset_handler - Reset a single LUN | |
2434 | * @cmd: scsi command struct | |
2435 | * | |
2436 | * Returns: | |
93631b4a | 2437 | * SUCCESS / FAST_IO_FAIL / FAILED |
072b91f9 BK |
2438 | **/ |
2439 | static int ibmvfc_eh_device_reset_handler(struct scsi_cmnd *cmd) | |
2440 | { | |
ad8dcffa BK |
2441 | struct scsi_device *sdev = cmd->device; |
2442 | struct ibmvfc_host *vhost = shost_priv(sdev->host); | |
93631b4a | 2443 | int cancel_rc, block_rc, reset_rc = 0; |
ad8dcffa | 2444 | int rc = FAILED; |
072b91f9 BK |
2445 | |
2446 | ENTER; | |
93631b4a | 2447 | block_rc = fc_block_scsi_eh(cmd); |
072b91f9 | 2448 | ibmvfc_wait_while_resetting(vhost); |
93631b4a BK |
2449 | if (block_rc != FAST_IO_FAIL) { |
2450 | cancel_rc = ibmvfc_cancel_all(sdev, IBMVFC_TMF_LUN_RESET); | |
2451 | reset_rc = ibmvfc_reset_device(sdev, IBMVFC_LUN_RESET, "LUN"); | |
2452 | } else | |
90f725db | 2453 | cancel_rc = ibmvfc_cancel_all(sdev, IBMVFC_TMF_SUPPRESS_ABTS); |
072b91f9 | 2454 | |
ad8dcffa BK |
2455 | if (!cancel_rc && !reset_rc) |
2456 | rc = ibmvfc_wait_for_ops(vhost, sdev, ibmvfc_match_lun); | |
072b91f9 | 2457 | |
93631b4a BK |
2458 | if (block_rc == FAST_IO_FAIL && rc != FAILED) |
2459 | rc = FAST_IO_FAIL; | |
2460 | ||
072b91f9 | 2461 | LEAVE; |
ad8dcffa | 2462 | return rc; |
072b91f9 BK |
2463 | } |
2464 | ||
93631b4a BK |
2465 | /** |
2466 | * ibmvfc_dev_cancel_all_noreset - Device iterated cancel all function | |
2467 | * @sdev: scsi device struct | |
2468 | * @data: return code | |
2469 | * | |
2470 | **/ | |
2471 | static void ibmvfc_dev_cancel_all_noreset(struct scsi_device *sdev, void *data) | |
2472 | { | |
2473 | unsigned long *rc = data; | |
90f725db | 2474 | *rc |= ibmvfc_cancel_all(sdev, IBMVFC_TMF_SUPPRESS_ABTS); |
93631b4a BK |
2475 | } |
2476 | ||
4a5c4a5e BK |
2477 | /** |
2478 | * ibmvfc_dev_cancel_all_reset - Device iterated cancel all function | |
2479 | * @sdev: scsi device struct | |
2480 | * @data: return code | |
2481 | * | |
2482 | **/ | |
2483 | static void ibmvfc_dev_cancel_all_reset(struct scsi_device *sdev, void *data) | |
072b91f9 BK |
2484 | { |
2485 | unsigned long *rc = data; | |
2486 | *rc |= ibmvfc_cancel_all(sdev, IBMVFC_TMF_TGT_RESET); | |
2487 | } | |
2488 | ||
072b91f9 BK |
2489 | /** |
2490 | * ibmvfc_eh_target_reset_handler - Reset the target | |
2491 | * @cmd: scsi command struct | |
2492 | * | |
2493 | * Returns: | |
93631b4a | 2494 | * SUCCESS / FAST_IO_FAIL / FAILED |
072b91f9 BK |
2495 | **/ |
2496 | static int ibmvfc_eh_target_reset_handler(struct scsi_cmnd *cmd) | |
2497 | { | |
ad8dcffa BK |
2498 | struct scsi_device *sdev = cmd->device; |
2499 | struct ibmvfc_host *vhost = shost_priv(sdev->host); | |
2500 | struct scsi_target *starget = scsi_target(sdev); | |
93631b4a BK |
2501 | int block_rc; |
2502 | int reset_rc = 0; | |
ad8dcffa | 2503 | int rc = FAILED; |
072b91f9 | 2504 | unsigned long cancel_rc = 0; |
072b91f9 BK |
2505 | |
2506 | ENTER; | |
93631b4a | 2507 | block_rc = fc_block_scsi_eh(cmd); |
072b91f9 | 2508 | ibmvfc_wait_while_resetting(vhost); |
93631b4a BK |
2509 | if (block_rc != FAST_IO_FAIL) { |
2510 | starget_for_each_device(starget, &cancel_rc, ibmvfc_dev_cancel_all_reset); | |
2511 | reset_rc = ibmvfc_reset_device(sdev, IBMVFC_TARGET_RESET, "target"); | |
2512 | } else | |
2513 | starget_for_each_device(starget, &cancel_rc, ibmvfc_dev_cancel_all_noreset); | |
072b91f9 | 2514 | |
ad8dcffa BK |
2515 | if (!cancel_rc && !reset_rc) |
2516 | rc = ibmvfc_wait_for_ops(vhost, starget, ibmvfc_match_target); | |
072b91f9 | 2517 | |
93631b4a BK |
2518 | if (block_rc == FAST_IO_FAIL && rc != FAILED) |
2519 | rc = FAST_IO_FAIL; | |
2520 | ||
072b91f9 | 2521 | LEAVE; |
ad8dcffa | 2522 | return rc; |
072b91f9 BK |
2523 | } |
2524 | ||
2525 | /** | |
2526 | * ibmvfc_eh_host_reset_handler - Reset the connection to the server | |
2527 | * @cmd: struct scsi_cmnd having problems | |
2528 | * | |
2529 | **/ | |
2530 | static int ibmvfc_eh_host_reset_handler(struct scsi_cmnd *cmd) | |
2531 | { | |
93631b4a | 2532 | int rc, block_rc; |
072b91f9 BK |
2533 | struct ibmvfc_host *vhost = shost_priv(cmd->device->host); |
2534 | ||
93631b4a | 2535 | block_rc = fc_block_scsi_eh(cmd); |
072b91f9 BK |
2536 | dev_err(vhost->dev, "Resetting connection due to error recovery\n"); |
2537 | rc = ibmvfc_issue_fc_host_lip(vhost->host); | |
93631b4a BK |
2538 | |
2539 | if (block_rc == FAST_IO_FAIL) | |
2540 | return FAST_IO_FAIL; | |
2541 | ||
072b91f9 BK |
2542 | return rc ? FAILED : SUCCESS; |
2543 | } | |
2544 | ||
2545 | /** | |
2546 | * ibmvfc_terminate_rport_io - Terminate all pending I/O to the rport. | |
2547 | * @rport: rport struct | |
2548 | * | |
2549 | * Return value: | |
2550 | * none | |
2551 | **/ | |
2552 | static void ibmvfc_terminate_rport_io(struct fc_rport *rport) | |
2553 | { | |
d2fab5cf | 2554 | struct Scsi_Host *shost = rport_to_shost(rport); |
072b91f9 | 2555 | struct ibmvfc_host *vhost = shost_priv(shost); |
d2fab5cf BK |
2556 | struct fc_rport *dev_rport; |
2557 | struct scsi_device *sdev; | |
2558 | unsigned long rc; | |
072b91f9 BK |
2559 | |
2560 | ENTER; | |
d2fab5cf BK |
2561 | shost_for_each_device(sdev, shost) { |
2562 | dev_rport = starget_to_rport(scsi_target(sdev)); | |
2563 | if (dev_rport != rport) | |
2564 | continue; | |
90f725db | 2565 | ibmvfc_cancel_all(sdev, IBMVFC_TMF_SUPPRESS_ABTS); |
d2fab5cf | 2566 | } |
072b91f9 | 2567 | |
d2fab5cf | 2568 | rc = ibmvfc_wait_for_ops(vhost, rport, ibmvfc_match_rport); |
ad8dcffa BK |
2569 | |
2570 | if (rc == FAILED) | |
072b91f9 | 2571 | ibmvfc_issue_fc_host_lip(shost); |
072b91f9 BK |
2572 | LEAVE; |
2573 | } | |
2574 | ||
d99e5f48 | 2575 | static const struct ibmvfc_async_desc ae_desc [] = { |
402c6eec RJ |
2576 | { "PLOGI", IBMVFC_AE_ELS_PLOGI, IBMVFC_DEFAULT_LOG_LEVEL + 1 }, |
2577 | { "LOGO", IBMVFC_AE_ELS_LOGO, IBMVFC_DEFAULT_LOG_LEVEL + 1 }, | |
2578 | { "PRLO", IBMVFC_AE_ELS_PRLO, IBMVFC_DEFAULT_LOG_LEVEL + 1 }, | |
2579 | { "N-Port SCN", IBMVFC_AE_SCN_NPORT, IBMVFC_DEFAULT_LOG_LEVEL + 1 }, | |
2580 | { "Group SCN", IBMVFC_AE_SCN_GROUP, IBMVFC_DEFAULT_LOG_LEVEL + 1 }, | |
2581 | { "Domain SCN", IBMVFC_AE_SCN_DOMAIN, IBMVFC_DEFAULT_LOG_LEVEL }, | |
2582 | { "Fabric SCN", IBMVFC_AE_SCN_FABRIC, IBMVFC_DEFAULT_LOG_LEVEL }, | |
2583 | { "Link Up", IBMVFC_AE_LINK_UP, IBMVFC_DEFAULT_LOG_LEVEL }, | |
2584 | { "Link Down", IBMVFC_AE_LINK_DOWN, IBMVFC_DEFAULT_LOG_LEVEL }, | |
2585 | { "Link Dead", IBMVFC_AE_LINK_DEAD, IBMVFC_DEFAULT_LOG_LEVEL }, | |
2586 | { "Halt", IBMVFC_AE_HALT, IBMVFC_DEFAULT_LOG_LEVEL }, | |
2587 | { "Resume", IBMVFC_AE_RESUME, IBMVFC_DEFAULT_LOG_LEVEL }, | |
2588 | { "Adapter Failed", IBMVFC_AE_ADAPTER_FAILED, IBMVFC_DEFAULT_LOG_LEVEL }, | |
072b91f9 BK |
2589 | }; |
2590 | ||
d99e5f48 | 2591 | static const struct ibmvfc_async_desc unknown_ae = { |
402c6eec | 2592 | "Unknown async", 0, IBMVFC_DEFAULT_LOG_LEVEL |
d99e5f48 | 2593 | }; |
072b91f9 BK |
2594 | |
2595 | /** | |
2596 | * ibmvfc_get_ae_desc - Get text description for async event | |
2597 | * @ae: async event | |
2598 | * | |
2599 | **/ | |
d99e5f48 | 2600 | static const struct ibmvfc_async_desc *ibmvfc_get_ae_desc(u64 ae) |
072b91f9 BK |
2601 | { |
2602 | int i; | |
2603 | ||
2604 | for (i = 0; i < ARRAY_SIZE(ae_desc); i++) | |
2605 | if (ae_desc[i].ae == ae) | |
d99e5f48 BK |
2606 | return &ae_desc[i]; |
2607 | ||
2608 | return &unknown_ae; | |
2609 | } | |
2610 | ||
2611 | static const struct { | |
2612 | enum ibmvfc_ae_link_state state; | |
2613 | const char *desc; | |
2614 | } link_desc [] = { | |
2615 | { IBMVFC_AE_LS_LINK_UP, " link up" }, | |
2616 | { IBMVFC_AE_LS_LINK_BOUNCED, " link bounced" }, | |
2617 | { IBMVFC_AE_LS_LINK_DOWN, " link down" }, | |
2618 | { IBMVFC_AE_LS_LINK_DEAD, " link dead" }, | |
2619 | }; | |
2620 | ||
2621 | /** | |
2622 | * ibmvfc_get_link_state - Get text description for link state | |
2623 | * @state: link state | |
2624 | * | |
2625 | **/ | |
2626 | static const char *ibmvfc_get_link_state(enum ibmvfc_ae_link_state state) | |
2627 | { | |
2628 | int i; | |
2629 | ||
2630 | for (i = 0; i < ARRAY_SIZE(link_desc); i++) | |
2631 | if (link_desc[i].state == state) | |
2632 | return link_desc[i].desc; | |
072b91f9 | 2633 | |
d99e5f48 | 2634 | return ""; |
072b91f9 BK |
2635 | } |
2636 | ||
2637 | /** | |
2638 | * ibmvfc_handle_async - Handle an async event from the adapter | |
2639 | * @crq: crq to process | |
2640 | * @vhost: ibmvfc host struct | |
2641 | * | |
2642 | **/ | |
2643 | static void ibmvfc_handle_async(struct ibmvfc_async_crq *crq, | |
2644 | struct ibmvfc_host *vhost) | |
2645 | { | |
0aab6c3f | 2646 | const struct ibmvfc_async_desc *desc = ibmvfc_get_ae_desc(be64_to_cpu(crq->event)); |
6d29cc56 | 2647 | struct ibmvfc_target *tgt; |
072b91f9 | 2648 | |
d99e5f48 BK |
2649 | ibmvfc_log(vhost, desc->log_level, "%s event received. scsi_id: %llx, wwpn: %llx," |
2650 | " node_name: %llx%s\n", desc->desc, crq->scsi_id, crq->wwpn, crq->node_name, | |
2651 | ibmvfc_get_link_state(crq->link_state)); | |
072b91f9 | 2652 | |
0aab6c3f | 2653 | switch (be64_to_cpu(crq->event)) { |
072b91f9 | 2654 | case IBMVFC_AE_RESUME: |
497f9c50 BK |
2655 | switch (crq->link_state) { |
2656 | case IBMVFC_AE_LS_LINK_DOWN: | |
2657 | ibmvfc_link_down(vhost, IBMVFC_LINK_DOWN); | |
2658 | break; | |
2659 | case IBMVFC_AE_LS_LINK_DEAD: | |
2660 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
2661 | break; | |
2662 | case IBMVFC_AE_LS_LINK_UP: | |
2663 | case IBMVFC_AE_LS_LINK_BOUNCED: | |
2664 | default: | |
2665 | vhost->events_to_log |= IBMVFC_AE_LINKUP; | |
2666 | vhost->delay_init = 1; | |
2667 | __ibmvfc_reset_host(vhost); | |
2668 | break; | |
2669 | }; | |
2670 | ||
2671 | break; | |
2672 | case IBMVFC_AE_LINK_UP: | |
072b91f9 | 2673 | vhost->events_to_log |= IBMVFC_AE_LINKUP; |
d2131b33 BK |
2674 | vhost->delay_init = 1; |
2675 | __ibmvfc_reset_host(vhost); | |
072b91f9 BK |
2676 | break; |
2677 | case IBMVFC_AE_SCN_FABRIC: | |
d2131b33 | 2678 | case IBMVFC_AE_SCN_DOMAIN: |
072b91f9 | 2679 | vhost->events_to_log |= IBMVFC_AE_RSCN; |
5cdf1626 BK |
2680 | if (vhost->state < IBMVFC_HALTED) { |
2681 | vhost->delay_init = 1; | |
2682 | __ibmvfc_reset_host(vhost); | |
2683 | } | |
072b91f9 BK |
2684 | break; |
2685 | case IBMVFC_AE_SCN_NPORT: | |
2686 | case IBMVFC_AE_SCN_GROUP: | |
072b91f9 | 2687 | vhost->events_to_log |= IBMVFC_AE_RSCN; |
6d29cc56 BK |
2688 | ibmvfc_reinit_host(vhost); |
2689 | break; | |
072b91f9 BK |
2690 | case IBMVFC_AE_ELS_LOGO: |
2691 | case IBMVFC_AE_ELS_PRLO: | |
2692 | case IBMVFC_AE_ELS_PLOGI: | |
6d29cc56 BK |
2693 | list_for_each_entry(tgt, &vhost->targets, queue) { |
2694 | if (!crq->scsi_id && !crq->wwpn && !crq->node_name) | |
2695 | break; | |
0aab6c3f | 2696 | if (crq->scsi_id && cpu_to_be64(tgt->scsi_id) != crq->scsi_id) |
6d29cc56 | 2697 | continue; |
0aab6c3f | 2698 | if (crq->wwpn && cpu_to_be64(tgt->ids.port_name) != crq->wwpn) |
6d29cc56 | 2699 | continue; |
0aab6c3f | 2700 | if (crq->node_name && cpu_to_be64(tgt->ids.node_name) != crq->node_name) |
6d29cc56 | 2701 | continue; |
0aab6c3f | 2702 | if (tgt->need_login && be64_to_cpu(crq->event) == IBMVFC_AE_ELS_LOGO) |
017b2ae3 | 2703 | tgt->logo_rcvd = 1; |
0aab6c3f | 2704 | if (!tgt->need_login || be64_to_cpu(crq->event) == IBMVFC_AE_ELS_PLOGI) { |
017b2ae3 BK |
2705 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); |
2706 | ibmvfc_reinit_host(vhost); | |
2707 | } | |
6d29cc56 | 2708 | } |
072b91f9 BK |
2709 | break; |
2710 | case IBMVFC_AE_LINK_DOWN: | |
2711 | case IBMVFC_AE_ADAPTER_FAILED: | |
2712 | ibmvfc_link_down(vhost, IBMVFC_LINK_DOWN); | |
2713 | break; | |
2714 | case IBMVFC_AE_LINK_DEAD: | |
2715 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
2716 | break; | |
2717 | case IBMVFC_AE_HALT: | |
2718 | ibmvfc_link_down(vhost, IBMVFC_HALTED); | |
2719 | break; | |
2720 | default: | |
775a42ec | 2721 | dev_err(vhost->dev, "Unknown async event received: %lld\n", crq->event); |
072b91f9 BK |
2722 | break; |
2723 | }; | |
2724 | } | |
2725 | ||
2726 | /** | |
2727 | * ibmvfc_handle_crq - Handles and frees received events in the CRQ | |
2728 | * @crq: Command/Response queue | |
2729 | * @vhost: ibmvfc host struct | |
2730 | * | |
2731 | **/ | |
2732 | static void ibmvfc_handle_crq(struct ibmvfc_crq *crq, struct ibmvfc_host *vhost) | |
2733 | { | |
2734 | long rc; | |
0aab6c3f | 2735 | struct ibmvfc_event *evt = (struct ibmvfc_event *)be64_to_cpu(crq->ioba); |
072b91f9 BK |
2736 | |
2737 | switch (crq->valid) { | |
2738 | case IBMVFC_CRQ_INIT_RSP: | |
2739 | switch (crq->format) { | |
2740 | case IBMVFC_CRQ_INIT: | |
2741 | dev_info(vhost->dev, "Partner initialized\n"); | |
2742 | /* Send back a response */ | |
2743 | rc = ibmvfc_send_crq_init_complete(vhost); | |
2744 | if (rc == 0) | |
861890c6 | 2745 | ibmvfc_init_host(vhost); |
072b91f9 BK |
2746 | else |
2747 | dev_err(vhost->dev, "Unable to send init rsp. rc=%ld\n", rc); | |
2748 | break; | |
2749 | case IBMVFC_CRQ_INIT_COMPLETE: | |
2750 | dev_info(vhost->dev, "Partner initialization complete\n"); | |
861890c6 | 2751 | ibmvfc_init_host(vhost); |
072b91f9 BK |
2752 | break; |
2753 | default: | |
2754 | dev_err(vhost->dev, "Unknown crq message type: %d\n", crq->format); | |
2755 | } | |
2756 | return; | |
2757 | case IBMVFC_CRQ_XPORT_EVENT: | |
2758 | vhost->state = IBMVFC_NO_CRQ; | |
79111d08 | 2759 | vhost->logged_in = 0; |
072b91f9 BK |
2760 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_NONE); |
2761 | if (crq->format == IBMVFC_PARTITION_MIGRATED) { | |
2762 | /* We need to re-setup the interpartition connection */ | |
2763 | dev_info(vhost->dev, "Re-enabling adapter\n"); | |
2764 | vhost->client_migrated = 1; | |
2765 | ibmvfc_purge_requests(vhost, DID_REQUEUE); | |
73ee5d86 BK |
2766 | ibmvfc_link_down(vhost, IBMVFC_LINK_DOWN); |
2767 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_REENABLE); | |
072b91f9 BK |
2768 | } else { |
2769 | dev_err(vhost->dev, "Virtual adapter failed (rc=%d)\n", crq->format); | |
072b91f9 | 2770 | ibmvfc_purge_requests(vhost, DID_ERROR); |
73ee5d86 BK |
2771 | ibmvfc_link_down(vhost, IBMVFC_LINK_DOWN); |
2772 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_RESET); | |
072b91f9 BK |
2773 | } |
2774 | return; | |
2775 | case IBMVFC_CRQ_CMD_RSP: | |
2776 | break; | |
2777 | default: | |
2778 | dev_err(vhost->dev, "Got an invalid message type 0x%02x\n", crq->valid); | |
2779 | return; | |
2780 | } | |
2781 | ||
2782 | if (crq->format == IBMVFC_ASYNC_EVENT) | |
2783 | return; | |
2784 | ||
2785 | /* The only kind of payload CRQs we should get are responses to | |
2786 | * things we send. Make sure this response is to something we | |
2787 | * actually sent | |
2788 | */ | |
2789 | if (unlikely(!ibmvfc_valid_event(&vhost->pool, evt))) { | |
775a42ec | 2790 | dev_err(vhost->dev, "Returned correlation_token 0x%08llx is invalid!\n", |
072b91f9 BK |
2791 | crq->ioba); |
2792 | return; | |
2793 | } | |
2794 | ||
2795 | if (unlikely(atomic_read(&evt->free))) { | |
775a42ec | 2796 | dev_err(vhost->dev, "Received duplicate correlation_token 0x%08llx!\n", |
072b91f9 BK |
2797 | crq->ioba); |
2798 | return; | |
2799 | } | |
2800 | ||
2801 | del_timer(&evt->timer); | |
2802 | list_del(&evt->queue); | |
2803 | ibmvfc_trc_end(evt); | |
2804 | evt->done(evt); | |
2805 | } | |
2806 | ||
2807 | /** | |
2808 | * ibmvfc_scan_finished - Check if the device scan is done. | |
2809 | * @shost: scsi host struct | |
2810 | * @time: current elapsed time | |
2811 | * | |
2812 | * Returns: | |
2813 | * 0 if scan is not done / 1 if scan is done | |
2814 | **/ | |
2815 | static int ibmvfc_scan_finished(struct Scsi_Host *shost, unsigned long time) | |
2816 | { | |
2817 | unsigned long flags; | |
2818 | struct ibmvfc_host *vhost = shost_priv(shost); | |
2819 | int done = 0; | |
2820 | ||
2821 | spin_lock_irqsave(shost->host_lock, flags); | |
2822 | if (time >= (init_timeout * HZ)) { | |
2823 | dev_info(vhost->dev, "Scan taking longer than %d seconds, " | |
2824 | "continuing initialization\n", init_timeout); | |
2825 | done = 1; | |
2826 | } | |
2827 | ||
43c8da90 | 2828 | if (vhost->scan_complete) |
072b91f9 BK |
2829 | done = 1; |
2830 | spin_unlock_irqrestore(shost->host_lock, flags); | |
2831 | return done; | |
2832 | } | |
2833 | ||
2834 | /** | |
2835 | * ibmvfc_slave_alloc - Setup the device's task set value | |
2836 | * @sdev: struct scsi_device device to configure | |
2837 | * | |
2838 | * Set the device's task set value so that error handling works as | |
2839 | * expected. | |
2840 | * | |
2841 | * Returns: | |
2842 | * 0 on success / -ENXIO if device does not exist | |
2843 | **/ | |
2844 | static int ibmvfc_slave_alloc(struct scsi_device *sdev) | |
2845 | { | |
2846 | struct Scsi_Host *shost = sdev->host; | |
2847 | struct fc_rport *rport = starget_to_rport(scsi_target(sdev)); | |
2848 | struct ibmvfc_host *vhost = shost_priv(shost); | |
2849 | unsigned long flags = 0; | |
2850 | ||
2851 | if (!rport || fc_remote_port_chkready(rport)) | |
2852 | return -ENXIO; | |
2853 | ||
2854 | spin_lock_irqsave(shost->host_lock, flags); | |
2855 | sdev->hostdata = (void *)(unsigned long)vhost->task_set++; | |
2856 | spin_unlock_irqrestore(shost->host_lock, flags); | |
2857 | return 0; | |
2858 | } | |
2859 | ||
ad8dcffa BK |
2860 | /** |
2861 | * ibmvfc_target_alloc - Setup the target's task set value | |
2862 | * @starget: struct scsi_target | |
2863 | * | |
2864 | * Set the target's task set value so that error handling works as | |
2865 | * expected. | |
2866 | * | |
2867 | * Returns: | |
2868 | * 0 on success / -ENXIO if device does not exist | |
2869 | **/ | |
2870 | static int ibmvfc_target_alloc(struct scsi_target *starget) | |
2871 | { | |
2872 | struct Scsi_Host *shost = dev_to_shost(starget->dev.parent); | |
2873 | struct ibmvfc_host *vhost = shost_priv(shost); | |
2874 | unsigned long flags = 0; | |
2875 | ||
2876 | spin_lock_irqsave(shost->host_lock, flags); | |
2877 | starget->hostdata = (void *)(unsigned long)vhost->task_set++; | |
2878 | spin_unlock_irqrestore(shost->host_lock, flags); | |
2879 | return 0; | |
2880 | } | |
2881 | ||
072b91f9 BK |
2882 | /** |
2883 | * ibmvfc_slave_configure - Configure the device | |
2884 | * @sdev: struct scsi_device device to configure | |
2885 | * | |
2886 | * Enable allow_restart for a device if it is a disk. Adjust the | |
2887 | * queue_depth here also. | |
2888 | * | |
2889 | * Returns: | |
2890 | * 0 | |
2891 | **/ | |
2892 | static int ibmvfc_slave_configure(struct scsi_device *sdev) | |
2893 | { | |
2894 | struct Scsi_Host *shost = sdev->host; | |
072b91f9 BK |
2895 | unsigned long flags = 0; |
2896 | ||
2897 | spin_lock_irqsave(shost->host_lock, flags); | |
2898 | if (sdev->type == TYPE_DISK) | |
2899 | sdev->allow_restart = 1; | |
2900 | ||
2901 | if (sdev->tagged_supported) { | |
2902 | scsi_set_tag_type(sdev, MSG_SIMPLE_TAG); | |
2903 | scsi_activate_tcq(sdev, sdev->queue_depth); | |
2904 | } else | |
2905 | scsi_deactivate_tcq(sdev, sdev->queue_depth); | |
072b91f9 BK |
2906 | spin_unlock_irqrestore(shost->host_lock, flags); |
2907 | return 0; | |
2908 | } | |
2909 | ||
2910 | /** | |
2911 | * ibmvfc_change_queue_depth - Change the device's queue depth | |
2912 | * @sdev: scsi device struct | |
2913 | * @qdepth: depth to set | |
e881a172 | 2914 | * @reason: calling context |
072b91f9 BK |
2915 | * |
2916 | * Return value: | |
2917 | * actual depth set | |
2918 | **/ | |
e881a172 MC |
2919 | static int ibmvfc_change_queue_depth(struct scsi_device *sdev, int qdepth, |
2920 | int reason) | |
072b91f9 | 2921 | { |
e881a172 MC |
2922 | if (reason != SCSI_QDEPTH_DEFAULT) |
2923 | return -EOPNOTSUPP; | |
2924 | ||
072b91f9 BK |
2925 | if (qdepth > IBMVFC_MAX_CMDS_PER_LUN) |
2926 | qdepth = IBMVFC_MAX_CMDS_PER_LUN; | |
2927 | ||
2928 | scsi_adjust_queue_depth(sdev, 0, qdepth); | |
2929 | return sdev->queue_depth; | |
2930 | } | |
2931 | ||
2932 | /** | |
2933 | * ibmvfc_change_queue_type - Change the device's queue type | |
2934 | * @sdev: scsi device struct | |
2935 | * @tag_type: type of tags to use | |
2936 | * | |
2937 | * Return value: | |
2938 | * actual queue type set | |
2939 | **/ | |
2940 | static int ibmvfc_change_queue_type(struct scsi_device *sdev, int tag_type) | |
2941 | { | |
2942 | if (sdev->tagged_supported) { | |
2943 | scsi_set_tag_type(sdev, tag_type); | |
2944 | ||
2945 | if (tag_type) | |
2946 | scsi_activate_tcq(sdev, sdev->queue_depth); | |
2947 | else | |
2948 | scsi_deactivate_tcq(sdev, sdev->queue_depth); | |
2949 | } else | |
2950 | tag_type = 0; | |
2951 | ||
2952 | return tag_type; | |
2953 | } | |
2954 | ||
2955 | static ssize_t ibmvfc_show_host_partition_name(struct device *dev, | |
2956 | struct device_attribute *attr, char *buf) | |
2957 | { | |
2958 | struct Scsi_Host *shost = class_to_shost(dev); | |
2959 | struct ibmvfc_host *vhost = shost_priv(shost); | |
2960 | ||
2961 | return snprintf(buf, PAGE_SIZE, "%s\n", | |
2962 | vhost->login_buf->resp.partition_name); | |
2963 | } | |
2964 | ||
072b91f9 BK |
2965 | static ssize_t ibmvfc_show_host_device_name(struct device *dev, |
2966 | struct device_attribute *attr, char *buf) | |
2967 | { | |
2968 | struct Scsi_Host *shost = class_to_shost(dev); | |
2969 | struct ibmvfc_host *vhost = shost_priv(shost); | |
2970 | ||
2971 | return snprintf(buf, PAGE_SIZE, "%s\n", | |
2972 | vhost->login_buf->resp.device_name); | |
2973 | } | |
2974 | ||
072b91f9 BK |
2975 | static ssize_t ibmvfc_show_host_loc_code(struct device *dev, |
2976 | struct device_attribute *attr, char *buf) | |
2977 | { | |
2978 | struct Scsi_Host *shost = class_to_shost(dev); | |
2979 | struct ibmvfc_host *vhost = shost_priv(shost); | |
2980 | ||
2981 | return snprintf(buf, PAGE_SIZE, "%s\n", | |
2982 | vhost->login_buf->resp.port_loc_code); | |
2983 | } | |
2984 | ||
072b91f9 BK |
2985 | static ssize_t ibmvfc_show_host_drc_name(struct device *dev, |
2986 | struct device_attribute *attr, char *buf) | |
2987 | { | |
2988 | struct Scsi_Host *shost = class_to_shost(dev); | |
2989 | struct ibmvfc_host *vhost = shost_priv(shost); | |
2990 | ||
2991 | return snprintf(buf, PAGE_SIZE, "%s\n", | |
2992 | vhost->login_buf->resp.drc_name); | |
2993 | } | |
2994 | ||
072b91f9 BK |
2995 | static ssize_t ibmvfc_show_host_npiv_version(struct device *dev, |
2996 | struct device_attribute *attr, char *buf) | |
2997 | { | |
2998 | struct Scsi_Host *shost = class_to_shost(dev); | |
2999 | struct ibmvfc_host *vhost = shost_priv(shost); | |
3000 | return snprintf(buf, PAGE_SIZE, "%d\n", vhost->login_buf->resp.version); | |
3001 | } | |
3002 | ||
497f9c50 BK |
3003 | static ssize_t ibmvfc_show_host_capabilities(struct device *dev, |
3004 | struct device_attribute *attr, char *buf) | |
3005 | { | |
3006 | struct Scsi_Host *shost = class_to_shost(dev); | |
3007 | struct ibmvfc_host *vhost = shost_priv(shost); | |
3008 | return snprintf(buf, PAGE_SIZE, "%llx\n", vhost->login_buf->resp.capabilities); | |
3009 | } | |
3010 | ||
072b91f9 BK |
3011 | /** |
3012 | * ibmvfc_show_log_level - Show the adapter's error logging level | |
3013 | * @dev: class device struct | |
3014 | * @buf: buffer | |
3015 | * | |
3016 | * Return value: | |
3017 | * number of bytes printed to buffer | |
3018 | **/ | |
3019 | static ssize_t ibmvfc_show_log_level(struct device *dev, | |
3020 | struct device_attribute *attr, char *buf) | |
3021 | { | |
3022 | struct Scsi_Host *shost = class_to_shost(dev); | |
3023 | struct ibmvfc_host *vhost = shost_priv(shost); | |
3024 | unsigned long flags = 0; | |
3025 | int len; | |
3026 | ||
3027 | spin_lock_irqsave(shost->host_lock, flags); | |
3028 | len = snprintf(buf, PAGE_SIZE, "%d\n", vhost->log_level); | |
3029 | spin_unlock_irqrestore(shost->host_lock, flags); | |
3030 | return len; | |
3031 | } | |
3032 | ||
3033 | /** | |
3034 | * ibmvfc_store_log_level - Change the adapter's error logging level | |
3035 | * @dev: class device struct | |
3036 | * @buf: buffer | |
3037 | * | |
3038 | * Return value: | |
3039 | * number of bytes printed to buffer | |
3040 | **/ | |
3041 | static ssize_t ibmvfc_store_log_level(struct device *dev, | |
3042 | struct device_attribute *attr, | |
3043 | const char *buf, size_t count) | |
3044 | { | |
3045 | struct Scsi_Host *shost = class_to_shost(dev); | |
3046 | struct ibmvfc_host *vhost = shost_priv(shost); | |
3047 | unsigned long flags = 0; | |
3048 | ||
3049 | spin_lock_irqsave(shost->host_lock, flags); | |
3050 | vhost->log_level = simple_strtoul(buf, NULL, 10); | |
3051 | spin_unlock_irqrestore(shost->host_lock, flags); | |
3052 | return strlen(buf); | |
3053 | } | |
3054 | ||
85e2399e BK |
3055 | static DEVICE_ATTR(partition_name, S_IRUGO, ibmvfc_show_host_partition_name, NULL); |
3056 | static DEVICE_ATTR(device_name, S_IRUGO, ibmvfc_show_host_device_name, NULL); | |
3057 | static DEVICE_ATTR(port_loc_code, S_IRUGO, ibmvfc_show_host_loc_code, NULL); | |
3058 | static DEVICE_ATTR(drc_name, S_IRUGO, ibmvfc_show_host_drc_name, NULL); | |
3059 | static DEVICE_ATTR(npiv_version, S_IRUGO, ibmvfc_show_host_npiv_version, NULL); | |
497f9c50 | 3060 | static DEVICE_ATTR(capabilities, S_IRUGO, ibmvfc_show_host_capabilities, NULL); |
85e2399e BK |
3061 | static DEVICE_ATTR(log_level, S_IRUGO | S_IWUSR, |
3062 | ibmvfc_show_log_level, ibmvfc_store_log_level); | |
072b91f9 BK |
3063 | |
3064 | #ifdef CONFIG_SCSI_IBMVFC_TRACE | |
3065 | /** | |
3066 | * ibmvfc_read_trace - Dump the adapter trace | |
2c3c8bea | 3067 | * @filp: open sysfs file |
072b91f9 BK |
3068 | * @kobj: kobject struct |
3069 | * @bin_attr: bin_attribute struct | |
3070 | * @buf: buffer | |
3071 | * @off: offset | |
3072 | * @count: buffer size | |
3073 | * | |
3074 | * Return value: | |
3075 | * number of bytes printed to buffer | |
3076 | **/ | |
2c3c8bea | 3077 | static ssize_t ibmvfc_read_trace(struct file *filp, struct kobject *kobj, |
072b91f9 BK |
3078 | struct bin_attribute *bin_attr, |
3079 | char *buf, loff_t off, size_t count) | |
3080 | { | |
3081 | struct device *dev = container_of(kobj, struct device, kobj); | |
3082 | struct Scsi_Host *shost = class_to_shost(dev); | |
3083 | struct ibmvfc_host *vhost = shost_priv(shost); | |
3084 | unsigned long flags = 0; | |
3085 | int size = IBMVFC_TRACE_SIZE; | |
3086 | char *src = (char *)vhost->trace; | |
3087 | ||
3088 | if (off > size) | |
3089 | return 0; | |
3090 | if (off + count > size) { | |
3091 | size -= off; | |
3092 | count = size; | |
3093 | } | |
3094 | ||
3095 | spin_lock_irqsave(shost->host_lock, flags); | |
3096 | memcpy(buf, &src[off], count); | |
3097 | spin_unlock_irqrestore(shost->host_lock, flags); | |
3098 | return count; | |
3099 | } | |
3100 | ||
3101 | static struct bin_attribute ibmvfc_trace_attr = { | |
3102 | .attr = { | |
3103 | .name = "trace", | |
3104 | .mode = S_IRUGO, | |
3105 | }, | |
3106 | .size = 0, | |
3107 | .read = ibmvfc_read_trace, | |
3108 | }; | |
3109 | #endif | |
3110 | ||
3111 | static struct device_attribute *ibmvfc_attrs[] = { | |
85e2399e BK |
3112 | &dev_attr_partition_name, |
3113 | &dev_attr_device_name, | |
3114 | &dev_attr_port_loc_code, | |
3115 | &dev_attr_drc_name, | |
3116 | &dev_attr_npiv_version, | |
497f9c50 | 3117 | &dev_attr_capabilities, |
85e2399e | 3118 | &dev_attr_log_level, |
072b91f9 BK |
3119 | NULL |
3120 | }; | |
3121 | ||
3122 | static struct scsi_host_template driver_template = { | |
3123 | .module = THIS_MODULE, | |
3124 | .name = "IBM POWER Virtual FC Adapter", | |
3125 | .proc_name = IBMVFC_NAME, | |
3126 | .queuecommand = ibmvfc_queuecommand, | |
3127 | .eh_abort_handler = ibmvfc_eh_abort_handler, | |
3128 | .eh_device_reset_handler = ibmvfc_eh_device_reset_handler, | |
3129 | .eh_target_reset_handler = ibmvfc_eh_target_reset_handler, | |
3130 | .eh_host_reset_handler = ibmvfc_eh_host_reset_handler, | |
3131 | .slave_alloc = ibmvfc_slave_alloc, | |
3132 | .slave_configure = ibmvfc_slave_configure, | |
ad8dcffa | 3133 | .target_alloc = ibmvfc_target_alloc, |
072b91f9 BK |
3134 | .scan_finished = ibmvfc_scan_finished, |
3135 | .change_queue_depth = ibmvfc_change_queue_depth, | |
3136 | .change_queue_type = ibmvfc_change_queue_type, | |
3137 | .cmd_per_lun = 16, | |
3138 | .can_queue = IBMVFC_MAX_REQUESTS_DEFAULT, | |
3139 | .this_id = -1, | |
3140 | .sg_tablesize = SG_ALL, | |
3141 | .max_sectors = IBMVFC_MAX_SECTORS, | |
3142 | .use_clustering = ENABLE_CLUSTERING, | |
3143 | .shost_attrs = ibmvfc_attrs, | |
3144 | }; | |
3145 | ||
3146 | /** | |
3147 | * ibmvfc_next_async_crq - Returns the next entry in async queue | |
3148 | * @vhost: ibmvfc host struct | |
3149 | * | |
3150 | * Returns: | |
3151 | * Pointer to next entry in queue / NULL if empty | |
3152 | **/ | |
3153 | static struct ibmvfc_async_crq *ibmvfc_next_async_crq(struct ibmvfc_host *vhost) | |
3154 | { | |
3155 | struct ibmvfc_async_crq_queue *async_crq = &vhost->async_crq; | |
3156 | struct ibmvfc_async_crq *crq; | |
3157 | ||
3158 | crq = &async_crq->msgs[async_crq->cur]; | |
3159 | if (crq->valid & 0x80) { | |
3160 | if (++async_crq->cur == async_crq->size) | |
3161 | async_crq->cur = 0; | |
f5832fa2 | 3162 | rmb(); |
072b91f9 BK |
3163 | } else |
3164 | crq = NULL; | |
3165 | ||
3166 | return crq; | |
3167 | } | |
3168 | ||
3169 | /** | |
3170 | * ibmvfc_next_crq - Returns the next entry in message queue | |
3171 | * @vhost: ibmvfc host struct | |
3172 | * | |
3173 | * Returns: | |
3174 | * Pointer to next entry in queue / NULL if empty | |
3175 | **/ | |
3176 | static struct ibmvfc_crq *ibmvfc_next_crq(struct ibmvfc_host *vhost) | |
3177 | { | |
3178 | struct ibmvfc_crq_queue *queue = &vhost->crq; | |
3179 | struct ibmvfc_crq *crq; | |
3180 | ||
3181 | crq = &queue->msgs[queue->cur]; | |
3182 | if (crq->valid & 0x80) { | |
3183 | if (++queue->cur == queue->size) | |
3184 | queue->cur = 0; | |
f5832fa2 | 3185 | rmb(); |
072b91f9 BK |
3186 | } else |
3187 | crq = NULL; | |
3188 | ||
3189 | return crq; | |
3190 | } | |
3191 | ||
3192 | /** | |
3193 | * ibmvfc_interrupt - Interrupt handler | |
3194 | * @irq: number of irq to handle, not used | |
3195 | * @dev_instance: ibmvfc_host that received interrupt | |
3196 | * | |
3197 | * Returns: | |
3198 | * IRQ_HANDLED | |
3199 | **/ | |
3200 | static irqreturn_t ibmvfc_interrupt(int irq, void *dev_instance) | |
3201 | { | |
3202 | struct ibmvfc_host *vhost = (struct ibmvfc_host *)dev_instance; | |
039a0898 BK |
3203 | unsigned long flags; |
3204 | ||
3205 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
3206 | vio_disable_interrupts(to_vio_dev(vhost->dev)); | |
3207 | tasklet_schedule(&vhost->tasklet); | |
3208 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
3209 | return IRQ_HANDLED; | |
3210 | } | |
3211 | ||
3212 | /** | |
3213 | * ibmvfc_tasklet - Interrupt handler tasklet | |
3214 | * @data: ibmvfc host struct | |
3215 | * | |
3216 | * Returns: | |
3217 | * Nothing | |
3218 | **/ | |
3219 | static void ibmvfc_tasklet(void *data) | |
3220 | { | |
3221 | struct ibmvfc_host *vhost = data; | |
072b91f9 BK |
3222 | struct vio_dev *vdev = to_vio_dev(vhost->dev); |
3223 | struct ibmvfc_crq *crq; | |
3224 | struct ibmvfc_async_crq *async; | |
3225 | unsigned long flags; | |
3226 | int done = 0; | |
3227 | ||
3228 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
072b91f9 | 3229 | while (!done) { |
072b91f9 BK |
3230 | /* Pull all the valid messages off the async CRQ */ |
3231 | while ((async = ibmvfc_next_async_crq(vhost)) != NULL) { | |
3232 | ibmvfc_handle_async(async, vhost); | |
3233 | async->valid = 0; | |
f5832fa2 | 3234 | wmb(); |
072b91f9 BK |
3235 | } |
3236 | ||
f1d7fb7a BK |
3237 | /* Pull all the valid messages off the CRQ */ |
3238 | while ((crq = ibmvfc_next_crq(vhost)) != NULL) { | |
072b91f9 BK |
3239 | ibmvfc_handle_crq(crq, vhost); |
3240 | crq->valid = 0; | |
f5832fa2 | 3241 | wmb(); |
f1d7fb7a BK |
3242 | } |
3243 | ||
3244 | vio_enable_interrupts(vdev); | |
3245 | if ((async = ibmvfc_next_async_crq(vhost)) != NULL) { | |
072b91f9 BK |
3246 | vio_disable_interrupts(vdev); |
3247 | ibmvfc_handle_async(async, vhost); | |
4081b77c | 3248 | async->valid = 0; |
f5832fa2 | 3249 | wmb(); |
f1d7fb7a BK |
3250 | } else if ((crq = ibmvfc_next_crq(vhost)) != NULL) { |
3251 | vio_disable_interrupts(vdev); | |
3252 | ibmvfc_handle_crq(crq, vhost); | |
3253 | crq->valid = 0; | |
f5832fa2 | 3254 | wmb(); |
072b91f9 BK |
3255 | } else |
3256 | done = 1; | |
3257 | } | |
3258 | ||
3259 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
072b91f9 BK |
3260 | } |
3261 | ||
3262 | /** | |
3263 | * ibmvfc_init_tgt - Set the next init job step for the target | |
3264 | * @tgt: ibmvfc target struct | |
3265 | * @job_step: job step to perform | |
3266 | * | |
3267 | **/ | |
3268 | static void ibmvfc_init_tgt(struct ibmvfc_target *tgt, | |
3269 | void (*job_step) (struct ibmvfc_target *)) | |
3270 | { | |
3271 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_INIT); | |
3272 | tgt->job_step = job_step; | |
3273 | wake_up(&tgt->vhost->work_wait_q); | |
3274 | } | |
3275 | ||
3276 | /** | |
3277 | * ibmvfc_retry_tgt_init - Attempt to retry a step in target initialization | |
3278 | * @tgt: ibmvfc target struct | |
3279 | * @job_step: initialization job step | |
3280 | * | |
7d0e4622 BK |
3281 | * Returns: 1 if step will be retried / 0 if not |
3282 | * | |
072b91f9 | 3283 | **/ |
7d0e4622 | 3284 | static int ibmvfc_retry_tgt_init(struct ibmvfc_target *tgt, |
072b91f9 BK |
3285 | void (*job_step) (struct ibmvfc_target *)) |
3286 | { | |
1c41fa82 | 3287 | if (++tgt->init_retries > IBMVFC_MAX_TGT_INIT_RETRIES) { |
072b91f9 BK |
3288 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); |
3289 | wake_up(&tgt->vhost->work_wait_q); | |
7d0e4622 | 3290 | return 0; |
072b91f9 BK |
3291 | } else |
3292 | ibmvfc_init_tgt(tgt, job_step); | |
7d0e4622 | 3293 | return 1; |
072b91f9 BK |
3294 | } |
3295 | ||
a3b7aeab BK |
3296 | /* Defined in FC-LS */ |
3297 | static const struct { | |
3298 | int code; | |
3299 | int retry; | |
3300 | int logged_in; | |
3301 | } prli_rsp [] = { | |
3302 | { 0, 1, 0 }, | |
3303 | { 1, 0, 1 }, | |
3304 | { 2, 1, 0 }, | |
3305 | { 3, 1, 0 }, | |
3306 | { 4, 0, 0 }, | |
3307 | { 5, 0, 0 }, | |
3308 | { 6, 0, 1 }, | |
3309 | { 7, 0, 0 }, | |
3310 | { 8, 1, 0 }, | |
3311 | }; | |
3312 | ||
3313 | /** | |
3314 | * ibmvfc_get_prli_rsp - Find PRLI response index | |
3315 | * @flags: PRLI response flags | |
3316 | * | |
3317 | **/ | |
3318 | static int ibmvfc_get_prli_rsp(u16 flags) | |
3319 | { | |
3320 | int i; | |
3321 | int code = (flags & 0x0f00) >> 8; | |
3322 | ||
3323 | for (i = 0; i < ARRAY_SIZE(prli_rsp); i++) | |
3324 | if (prli_rsp[i].code == code) | |
3325 | return i; | |
3326 | ||
3327 | return 0; | |
3328 | } | |
3329 | ||
072b91f9 BK |
3330 | /** |
3331 | * ibmvfc_tgt_prli_done - Completion handler for Process Login | |
3332 | * @evt: ibmvfc event struct | |
3333 | * | |
3334 | **/ | |
3335 | static void ibmvfc_tgt_prli_done(struct ibmvfc_event *evt) | |
3336 | { | |
3337 | struct ibmvfc_target *tgt = evt->tgt; | |
3338 | struct ibmvfc_host *vhost = evt->vhost; | |
3339 | struct ibmvfc_process_login *rsp = &evt->xfer_iu->prli; | |
a3b7aeab | 3340 | struct ibmvfc_prli_svc_parms *parms = &rsp->parms; |
0aab6c3f | 3341 | u32 status = be16_to_cpu(rsp->common.status); |
7d0e4622 | 3342 | int index, level = IBMVFC_DEFAULT_LOG_LEVEL; |
072b91f9 BK |
3343 | |
3344 | vhost->discovery_threads--; | |
3345 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_NONE); | |
3346 | switch (status) { | |
3347 | case IBMVFC_MAD_SUCCESS: | |
a3b7aeab BK |
3348 | tgt_dbg(tgt, "Process Login succeeded: %X %02X %04X\n", |
3349 | parms->type, parms->flags, parms->service_parms); | |
3350 | ||
3351 | if (parms->type == IBMVFC_SCSI_FCP_TYPE) { | |
0aab6c3f | 3352 | index = ibmvfc_get_prli_rsp(be16_to_cpu(parms->flags)); |
a3b7aeab | 3353 | if (prli_rsp[index].logged_in) { |
0aab6c3f | 3354 | if (be16_to_cpu(parms->flags) & IBMVFC_PRLI_EST_IMG_PAIR) { |
a3b7aeab BK |
3355 | tgt->need_login = 0; |
3356 | tgt->ids.roles = 0; | |
0aab6c3f | 3357 | if (be32_to_cpu(parms->service_parms) & IBMVFC_PRLI_TARGET_FUNC) |
a3b7aeab | 3358 | tgt->ids.roles |= FC_PORT_ROLE_FCP_TARGET; |
0aab6c3f | 3359 | if (be32_to_cpu(parms->service_parms) & IBMVFC_PRLI_INITIATOR_FUNC) |
a3b7aeab | 3360 | tgt->ids.roles |= FC_PORT_ROLE_FCP_INITIATOR; |
43c8da90 | 3361 | tgt->add_rport = 1; |
a3b7aeab BK |
3362 | } else |
3363 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); | |
3364 | } else if (prli_rsp[index].retry) | |
3365 | ibmvfc_retry_tgt_init(tgt, ibmvfc_tgt_send_prli); | |
3366 | else | |
3367 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); | |
3368 | } else | |
3369 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); | |
072b91f9 BK |
3370 | break; |
3371 | case IBMVFC_MAD_DRIVER_FAILED: | |
3372 | break; | |
3373 | case IBMVFC_MAD_CRQ_ERROR: | |
3374 | ibmvfc_retry_tgt_init(tgt, ibmvfc_tgt_send_prli); | |
3375 | break; | |
3376 | case IBMVFC_MAD_FAILED: | |
3377 | default: | |
0aab6c3f TD |
3378 | if ((be16_to_cpu(rsp->status) & IBMVFC_VIOS_FAILURE) && |
3379 | be16_to_cpu(rsp->error) == IBMVFC_PLOGI_REQUIRED) | |
017b2ae3 BK |
3380 | level += ibmvfc_retry_tgt_init(tgt, ibmvfc_tgt_send_plogi); |
3381 | else if (tgt->logo_rcvd) | |
3382 | level += ibmvfc_retry_tgt_init(tgt, ibmvfc_tgt_send_plogi); | |
0aab6c3f | 3383 | else if (ibmvfc_retry_cmd(be16_to_cpu(rsp->status), be16_to_cpu(rsp->error))) |
7d0e4622 | 3384 | level += ibmvfc_retry_tgt_init(tgt, ibmvfc_tgt_send_prli); |
10e79499 BK |
3385 | else |
3386 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); | |
7d0e4622 BK |
3387 | |
3388 | tgt_log(tgt, level, "Process Login failed: %s (%x:%x) rc=0x%02X\n", | |
0aab6c3f | 3389 | ibmvfc_get_cmd_error(be16_to_cpu(rsp->status), be16_to_cpu(rsp->error)), |
7d0e4622 | 3390 | rsp->status, rsp->error, status); |
072b91f9 BK |
3391 | break; |
3392 | }; | |
3393 | ||
3394 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3395 | ibmvfc_free_event(evt); | |
3396 | wake_up(&vhost->work_wait_q); | |
3397 | } | |
3398 | ||
3399 | /** | |
3400 | * ibmvfc_tgt_send_prli - Send a process login | |
3401 | * @tgt: ibmvfc target struct | |
3402 | * | |
3403 | **/ | |
3404 | static void ibmvfc_tgt_send_prli(struct ibmvfc_target *tgt) | |
3405 | { | |
3406 | struct ibmvfc_process_login *prli; | |
3407 | struct ibmvfc_host *vhost = tgt->vhost; | |
3408 | struct ibmvfc_event *evt; | |
3409 | ||
3410 | if (vhost->discovery_threads >= disc_threads) | |
3411 | return; | |
3412 | ||
3413 | kref_get(&tgt->kref); | |
3414 | evt = ibmvfc_get_event(vhost); | |
3415 | vhost->discovery_threads++; | |
3416 | ibmvfc_init_event(evt, ibmvfc_tgt_prli_done, IBMVFC_MAD_FORMAT); | |
3417 | evt->tgt = tgt; | |
3418 | prli = &evt->iu.prli; | |
3419 | memset(prli, 0, sizeof(*prli)); | |
0aab6c3f TD |
3420 | prli->common.version = cpu_to_be32(1); |
3421 | prli->common.opcode = cpu_to_be32(IBMVFC_PROCESS_LOGIN); | |
3422 | prli->common.length = cpu_to_be16(sizeof(*prli)); | |
3423 | prli->scsi_id = cpu_to_be64(tgt->scsi_id); | |
072b91f9 BK |
3424 | |
3425 | prli->parms.type = IBMVFC_SCSI_FCP_TYPE; | |
0aab6c3f TD |
3426 | prli->parms.flags = cpu_to_be16(IBMVFC_PRLI_EST_IMG_PAIR); |
3427 | prli->parms.service_parms = cpu_to_be32(IBMVFC_PRLI_INITIATOR_FUNC); | |
072b91f9 BK |
3428 | |
3429 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_INIT_WAIT); | |
3430 | if (ibmvfc_send_event(evt, vhost, default_timeout)) { | |
3431 | vhost->discovery_threads--; | |
3432 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_NONE); | |
3433 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3434 | } else | |
3435 | tgt_dbg(tgt, "Sent process login\n"); | |
3436 | } | |
3437 | ||
3438 | /** | |
3439 | * ibmvfc_tgt_plogi_done - Completion handler for Port Login | |
3440 | * @evt: ibmvfc event struct | |
3441 | * | |
3442 | **/ | |
3443 | static void ibmvfc_tgt_plogi_done(struct ibmvfc_event *evt) | |
3444 | { | |
3445 | struct ibmvfc_target *tgt = evt->tgt; | |
3446 | struct ibmvfc_host *vhost = evt->vhost; | |
3447 | struct ibmvfc_port_login *rsp = &evt->xfer_iu->plogi; | |
0aab6c3f | 3448 | u32 status = be16_to_cpu(rsp->common.status); |
7d0e4622 | 3449 | int level = IBMVFC_DEFAULT_LOG_LEVEL; |
072b91f9 BK |
3450 | |
3451 | vhost->discovery_threads--; | |
3452 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_NONE); | |
3453 | switch (status) { | |
3454 | case IBMVFC_MAD_SUCCESS: | |
3455 | tgt_dbg(tgt, "Port Login succeeded\n"); | |
3456 | if (tgt->ids.port_name && | |
3457 | tgt->ids.port_name != wwn_to_u64(rsp->service_parms.port_name)) { | |
3458 | vhost->reinit = 1; | |
3459 | tgt_dbg(tgt, "Port re-init required\n"); | |
3460 | break; | |
3461 | } | |
3462 | tgt->ids.node_name = wwn_to_u64(rsp->service_parms.node_name); | |
3463 | tgt->ids.port_name = wwn_to_u64(rsp->service_parms.port_name); | |
3464 | tgt->ids.port_id = tgt->scsi_id; | |
072b91f9 BK |
3465 | memcpy(&tgt->service_parms, &rsp->service_parms, |
3466 | sizeof(tgt->service_parms)); | |
3467 | memcpy(&tgt->service_parms_change, &rsp->service_parms_change, | |
3468 | sizeof(tgt->service_parms_change)); | |
3469 | ibmvfc_init_tgt(tgt, ibmvfc_tgt_send_prli); | |
3470 | break; | |
3471 | case IBMVFC_MAD_DRIVER_FAILED: | |
3472 | break; | |
3473 | case IBMVFC_MAD_CRQ_ERROR: | |
3474 | ibmvfc_retry_tgt_init(tgt, ibmvfc_tgt_send_plogi); | |
3475 | break; | |
3476 | case IBMVFC_MAD_FAILED: | |
3477 | default: | |
0aab6c3f | 3478 | if (ibmvfc_retry_cmd(be16_to_cpu(rsp->status), be16_to_cpu(rsp->error))) |
7d0e4622 | 3479 | level += ibmvfc_retry_tgt_init(tgt, ibmvfc_tgt_send_plogi); |
10e79499 BK |
3480 | else |
3481 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); | |
7d0e4622 BK |
3482 | |
3483 | tgt_log(tgt, level, "Port Login failed: %s (%x:%x) %s (%x) %s (%x) rc=0x%02X\n", | |
0aab6c3f TD |
3484 | ibmvfc_get_cmd_error(be16_to_cpu(rsp->status), be16_to_cpu(rsp->error)), rsp->status, rsp->error, |
3485 | ibmvfc_get_fc_type(be16_to_cpu(rsp->fc_type)), rsp->fc_type, | |
3486 | ibmvfc_get_ls_explain(be16_to_cpu(rsp->fc_explain)), rsp->fc_explain, status); | |
072b91f9 BK |
3487 | break; |
3488 | }; | |
3489 | ||
3490 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3491 | ibmvfc_free_event(evt); | |
3492 | wake_up(&vhost->work_wait_q); | |
3493 | } | |
3494 | ||
3495 | /** | |
3496 | * ibmvfc_tgt_send_plogi - Send PLOGI to the specified target | |
3497 | * @tgt: ibmvfc target struct | |
3498 | * | |
3499 | **/ | |
3500 | static void ibmvfc_tgt_send_plogi(struct ibmvfc_target *tgt) | |
3501 | { | |
3502 | struct ibmvfc_port_login *plogi; | |
3503 | struct ibmvfc_host *vhost = tgt->vhost; | |
3504 | struct ibmvfc_event *evt; | |
3505 | ||
3506 | if (vhost->discovery_threads >= disc_threads) | |
3507 | return; | |
3508 | ||
3509 | kref_get(&tgt->kref); | |
017b2ae3 | 3510 | tgt->logo_rcvd = 0; |
072b91f9 BK |
3511 | evt = ibmvfc_get_event(vhost); |
3512 | vhost->discovery_threads++; | |
3513 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_INIT_WAIT); | |
3514 | ibmvfc_init_event(evt, ibmvfc_tgt_plogi_done, IBMVFC_MAD_FORMAT); | |
3515 | evt->tgt = tgt; | |
3516 | plogi = &evt->iu.plogi; | |
3517 | memset(plogi, 0, sizeof(*plogi)); | |
0aab6c3f TD |
3518 | plogi->common.version = cpu_to_be32(1); |
3519 | plogi->common.opcode = cpu_to_be32(IBMVFC_PORT_LOGIN); | |
3520 | plogi->common.length = cpu_to_be16(sizeof(*plogi)); | |
3521 | plogi->scsi_id = cpu_to_be64(tgt->scsi_id); | |
072b91f9 BK |
3522 | |
3523 | if (ibmvfc_send_event(evt, vhost, default_timeout)) { | |
3524 | vhost->discovery_threads--; | |
3525 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_NONE); | |
3526 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3527 | } else | |
3528 | tgt_dbg(tgt, "Sent port login\n"); | |
3529 | } | |
3530 | ||
3531 | /** | |
3532 | * ibmvfc_tgt_implicit_logout_done - Completion handler for Implicit Logout MAD | |
3533 | * @evt: ibmvfc event struct | |
3534 | * | |
3535 | **/ | |
3536 | static void ibmvfc_tgt_implicit_logout_done(struct ibmvfc_event *evt) | |
3537 | { | |
3538 | struct ibmvfc_target *tgt = evt->tgt; | |
3539 | struct ibmvfc_host *vhost = evt->vhost; | |
3540 | struct ibmvfc_implicit_logout *rsp = &evt->xfer_iu->implicit_logout; | |
0aab6c3f | 3541 | u32 status = be16_to_cpu(rsp->common.status); |
072b91f9 BK |
3542 | |
3543 | vhost->discovery_threads--; | |
3544 | ibmvfc_free_event(evt); | |
3545 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_NONE); | |
3546 | ||
3547 | switch (status) { | |
3548 | case IBMVFC_MAD_SUCCESS: | |
3549 | tgt_dbg(tgt, "Implicit Logout succeeded\n"); | |
3550 | break; | |
3551 | case IBMVFC_MAD_DRIVER_FAILED: | |
3552 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3553 | wake_up(&vhost->work_wait_q); | |
3554 | return; | |
3555 | case IBMVFC_MAD_FAILED: | |
3556 | default: | |
3557 | tgt_err(tgt, "Implicit Logout failed: rc=0x%02X\n", status); | |
3558 | break; | |
3559 | }; | |
3560 | ||
3561 | if (vhost->action == IBMVFC_HOST_ACTION_TGT_INIT) | |
3562 | ibmvfc_init_tgt(tgt, ibmvfc_tgt_send_plogi); | |
3563 | else if (vhost->action == IBMVFC_HOST_ACTION_QUERY_TGTS && | |
3564 | tgt->scsi_id != tgt->new_scsi_id) | |
3565 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); | |
3566 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3567 | wake_up(&vhost->work_wait_q); | |
3568 | } | |
3569 | ||
3570 | /** | |
3571 | * ibmvfc_tgt_implicit_logout - Initiate an Implicit Logout for specified target | |
3572 | * @tgt: ibmvfc target struct | |
3573 | * | |
3574 | **/ | |
3575 | static void ibmvfc_tgt_implicit_logout(struct ibmvfc_target *tgt) | |
3576 | { | |
3577 | struct ibmvfc_implicit_logout *mad; | |
3578 | struct ibmvfc_host *vhost = tgt->vhost; | |
3579 | struct ibmvfc_event *evt; | |
3580 | ||
3581 | if (vhost->discovery_threads >= disc_threads) | |
3582 | return; | |
3583 | ||
3584 | kref_get(&tgt->kref); | |
3585 | evt = ibmvfc_get_event(vhost); | |
3586 | vhost->discovery_threads++; | |
3587 | ibmvfc_init_event(evt, ibmvfc_tgt_implicit_logout_done, IBMVFC_MAD_FORMAT); | |
3588 | evt->tgt = tgt; | |
3589 | mad = &evt->iu.implicit_logout; | |
3590 | memset(mad, 0, sizeof(*mad)); | |
0aab6c3f TD |
3591 | mad->common.version = cpu_to_be32(1); |
3592 | mad->common.opcode = cpu_to_be32(IBMVFC_IMPLICIT_LOGOUT); | |
3593 | mad->common.length = cpu_to_be16(sizeof(*mad)); | |
3594 | mad->old_scsi_id = cpu_to_be64(tgt->scsi_id); | |
072b91f9 BK |
3595 | |
3596 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_INIT_WAIT); | |
3597 | if (ibmvfc_send_event(evt, vhost, default_timeout)) { | |
3598 | vhost->discovery_threads--; | |
3599 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_NONE); | |
3600 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3601 | } else | |
3602 | tgt_dbg(tgt, "Sent Implicit Logout\n"); | |
3603 | } | |
3604 | ||
989b8545 BK |
3605 | /** |
3606 | * ibmvfc_adisc_needs_plogi - Does device need PLOGI? | |
3607 | * @mad: ibmvfc passthru mad struct | |
3608 | * @tgt: ibmvfc target struct | |
3609 | * | |
3610 | * Returns: | |
3611 | * 1 if PLOGI needed / 0 if PLOGI not needed | |
3612 | **/ | |
3613 | static int ibmvfc_adisc_needs_plogi(struct ibmvfc_passthru_mad *mad, | |
3614 | struct ibmvfc_target *tgt) | |
3615 | { | |
3616 | if (memcmp(&mad->fc_iu.response[2], &tgt->ids.port_name, | |
3617 | sizeof(tgt->ids.port_name))) | |
3618 | return 1; | |
3619 | if (memcmp(&mad->fc_iu.response[4], &tgt->ids.node_name, | |
3620 | sizeof(tgt->ids.node_name))) | |
3621 | return 1; | |
0aab6c3f | 3622 | if (be32_to_cpu(mad->fc_iu.response[6]) != tgt->scsi_id) |
989b8545 BK |
3623 | return 1; |
3624 | return 0; | |
3625 | } | |
3626 | ||
3627 | /** | |
3628 | * ibmvfc_tgt_adisc_done - Completion handler for ADISC | |
3629 | * @evt: ibmvfc event struct | |
3630 | * | |
3631 | **/ | |
3632 | static void ibmvfc_tgt_adisc_done(struct ibmvfc_event *evt) | |
3633 | { | |
3634 | struct ibmvfc_target *tgt = evt->tgt; | |
3635 | struct ibmvfc_host *vhost = evt->vhost; | |
3636 | struct ibmvfc_passthru_mad *mad = &evt->xfer_iu->passthru; | |
0aab6c3f | 3637 | u32 status = be16_to_cpu(mad->common.status); |
989b8545 BK |
3638 | u8 fc_reason, fc_explain; |
3639 | ||
3640 | vhost->discovery_threads--; | |
3641 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_NONE); | |
10501e1c | 3642 | del_timer(&tgt->timer); |
989b8545 BK |
3643 | |
3644 | switch (status) { | |
3645 | case IBMVFC_MAD_SUCCESS: | |
3646 | tgt_dbg(tgt, "ADISC succeeded\n"); | |
3647 | if (ibmvfc_adisc_needs_plogi(mad, tgt)) | |
5e47167b | 3648 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); |
989b8545 BK |
3649 | break; |
3650 | case IBMVFC_MAD_DRIVER_FAILED: | |
3651 | break; | |
3652 | case IBMVFC_MAD_FAILED: | |
3653 | default: | |
5e47167b | 3654 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); |
0aab6c3f TD |
3655 | fc_reason = (be32_to_cpu(mad->fc_iu.response[1]) & 0x00ff0000) >> 16; |
3656 | fc_explain = (be32_to_cpu(mad->fc_iu.response[1]) & 0x0000ff00) >> 8; | |
989b8545 | 3657 | tgt_info(tgt, "ADISC failed: %s (%x:%x) %s (%x) %s (%x) rc=0x%02X\n", |
0aab6c3f | 3658 | ibmvfc_get_cmd_error(be16_to_cpu(mad->iu.status), be16_to_cpu(mad->iu.error)), |
989b8545 BK |
3659 | mad->iu.status, mad->iu.error, |
3660 | ibmvfc_get_fc_type(fc_reason), fc_reason, | |
3661 | ibmvfc_get_ls_explain(fc_explain), fc_explain, status); | |
3662 | break; | |
3663 | }; | |
3664 | ||
3665 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3666 | ibmvfc_free_event(evt); | |
3667 | wake_up(&vhost->work_wait_q); | |
3668 | } | |
3669 | ||
3670 | /** | |
3671 | * ibmvfc_init_passthru - Initialize an event struct for FC passthru | |
3672 | * @evt: ibmvfc event struct | |
3673 | * | |
3674 | **/ | |
3675 | static void ibmvfc_init_passthru(struct ibmvfc_event *evt) | |
3676 | { | |
3677 | struct ibmvfc_passthru_mad *mad = &evt->iu.passthru; | |
3678 | ||
3679 | memset(mad, 0, sizeof(*mad)); | |
0aab6c3f TD |
3680 | mad->common.version = cpu_to_be32(1); |
3681 | mad->common.opcode = cpu_to_be32(IBMVFC_PASSTHRU); | |
3682 | mad->common.length = cpu_to_be16(sizeof(*mad) - sizeof(mad->fc_iu) - sizeof(mad->iu)); | |
3683 | mad->cmd_ioba.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + | |
3684 | offsetof(struct ibmvfc_passthru_mad, iu)); | |
3685 | mad->cmd_ioba.len = cpu_to_be32(sizeof(mad->iu)); | |
3686 | mad->iu.cmd_len = cpu_to_be32(sizeof(mad->fc_iu.payload)); | |
3687 | mad->iu.rsp_len = cpu_to_be32(sizeof(mad->fc_iu.response)); | |
3688 | mad->iu.cmd.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + | |
989b8545 | 3689 | offsetof(struct ibmvfc_passthru_mad, fc_iu) + |
0aab6c3f TD |
3690 | offsetof(struct ibmvfc_passthru_fc_iu, payload)); |
3691 | mad->iu.cmd.len = cpu_to_be32(sizeof(mad->fc_iu.payload)); | |
3692 | mad->iu.rsp.va = cpu_to_be64((u64)be64_to_cpu(evt->crq.ioba) + | |
989b8545 | 3693 | offsetof(struct ibmvfc_passthru_mad, fc_iu) + |
0aab6c3f TD |
3694 | offsetof(struct ibmvfc_passthru_fc_iu, response)); |
3695 | mad->iu.rsp.len = cpu_to_be32(sizeof(mad->fc_iu.response)); | |
989b8545 BK |
3696 | } |
3697 | ||
10501e1c BK |
3698 | /** |
3699 | * ibmvfc_tgt_adisc_cancel_done - Completion handler when cancelling an ADISC | |
3700 | * @evt: ibmvfc event struct | |
3701 | * | |
3702 | * Just cleanup this event struct. Everything else is handled by | |
3703 | * the ADISC completion handler. If the ADISC never actually comes | |
3704 | * back, we still have the timer running on the ADISC event struct | |
3705 | * which will fire and cause the CRQ to get reset. | |
3706 | * | |
3707 | **/ | |
3708 | static void ibmvfc_tgt_adisc_cancel_done(struct ibmvfc_event *evt) | |
3709 | { | |
3710 | struct ibmvfc_host *vhost = evt->vhost; | |
3711 | struct ibmvfc_target *tgt = evt->tgt; | |
3712 | ||
3713 | tgt_dbg(tgt, "ADISC cancel complete\n"); | |
3714 | vhost->abort_threads--; | |
3715 | ibmvfc_free_event(evt); | |
3716 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3717 | wake_up(&vhost->work_wait_q); | |
3718 | } | |
3719 | ||
3720 | /** | |
3721 | * ibmvfc_adisc_timeout - Handle an ADISC timeout | |
3722 | * @tgt: ibmvfc target struct | |
3723 | * | |
3724 | * If an ADISC times out, send a cancel. If the cancel times | |
3725 | * out, reset the CRQ. When the ADISC comes back as cancelled, | |
3726 | * log back into the target. | |
3727 | **/ | |
3728 | static void ibmvfc_adisc_timeout(struct ibmvfc_target *tgt) | |
3729 | { | |
3730 | struct ibmvfc_host *vhost = tgt->vhost; | |
3731 | struct ibmvfc_event *evt; | |
3732 | struct ibmvfc_tmf *tmf; | |
3733 | unsigned long flags; | |
3734 | int rc; | |
3735 | ||
3736 | tgt_dbg(tgt, "ADISC timeout\n"); | |
3737 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
3738 | if (vhost->abort_threads >= disc_threads || | |
3739 | tgt->action != IBMVFC_TGT_ACTION_INIT_WAIT || | |
3740 | vhost->state != IBMVFC_INITIALIZING || | |
3741 | vhost->action != IBMVFC_HOST_ACTION_QUERY_TGTS) { | |
3742 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
3743 | return; | |
3744 | } | |
3745 | ||
3746 | vhost->abort_threads++; | |
3747 | kref_get(&tgt->kref); | |
3748 | evt = ibmvfc_get_event(vhost); | |
3749 | ibmvfc_init_event(evt, ibmvfc_tgt_adisc_cancel_done, IBMVFC_MAD_FORMAT); | |
3750 | ||
3751 | evt->tgt = tgt; | |
3752 | tmf = &evt->iu.tmf; | |
3753 | memset(tmf, 0, sizeof(*tmf)); | |
0aab6c3f TD |
3754 | tmf->common.version = cpu_to_be32(1); |
3755 | tmf->common.opcode = cpu_to_be32(IBMVFC_TMF_MAD); | |
3756 | tmf->common.length = cpu_to_be16(sizeof(*tmf)); | |
3757 | tmf->scsi_id = cpu_to_be64(tgt->scsi_id); | |
3758 | tmf->cancel_key = cpu_to_be32(tgt->cancel_key); | |
10501e1c BK |
3759 | |
3760 | rc = ibmvfc_send_event(evt, vhost, default_timeout); | |
3761 | ||
3762 | if (rc) { | |
3763 | tgt_err(tgt, "Failed to send cancel event for ADISC. rc=%d\n", rc); | |
3764 | vhost->abort_threads--; | |
3765 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3766 | __ibmvfc_reset_host(vhost); | |
3767 | } else | |
3768 | tgt_dbg(tgt, "Attempting to cancel ADISC\n"); | |
3769 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
3770 | } | |
3771 | ||
989b8545 BK |
3772 | /** |
3773 | * ibmvfc_tgt_adisc - Initiate an ADISC for specified target | |
3774 | * @tgt: ibmvfc target struct | |
3775 | * | |
10501e1c BK |
3776 | * When sending an ADISC we end up with two timers running. The |
3777 | * first timer is the timer in the ibmvfc target struct. If this | |
3778 | * fires, we send a cancel to the target. The second timer is the | |
3779 | * timer on the ibmvfc event for the ADISC, which is longer. If that | |
3780 | * fires, it means the ADISC timed out and our attempt to cancel it | |
3781 | * also failed, so we need to reset the CRQ. | |
989b8545 BK |
3782 | **/ |
3783 | static void ibmvfc_tgt_adisc(struct ibmvfc_target *tgt) | |
3784 | { | |
3785 | struct ibmvfc_passthru_mad *mad; | |
3786 | struct ibmvfc_host *vhost = tgt->vhost; | |
3787 | struct ibmvfc_event *evt; | |
3788 | ||
3789 | if (vhost->discovery_threads >= disc_threads) | |
3790 | return; | |
3791 | ||
3792 | kref_get(&tgt->kref); | |
3793 | evt = ibmvfc_get_event(vhost); | |
3794 | vhost->discovery_threads++; | |
3795 | ibmvfc_init_event(evt, ibmvfc_tgt_adisc_done, IBMVFC_MAD_FORMAT); | |
3796 | evt->tgt = tgt; | |
3797 | ||
3798 | ibmvfc_init_passthru(evt); | |
3799 | mad = &evt->iu.passthru; | |
0aab6c3f TD |
3800 | mad->iu.flags = cpu_to_be32(IBMVFC_FC_ELS); |
3801 | mad->iu.scsi_id = cpu_to_be64(tgt->scsi_id); | |
3802 | mad->iu.cancel_key = cpu_to_be32(tgt->cancel_key); | |
989b8545 | 3803 | |
0aab6c3f | 3804 | mad->fc_iu.payload[0] = cpu_to_be32(IBMVFC_ADISC); |
989b8545 BK |
3805 | memcpy(&mad->fc_iu.payload[2], &vhost->login_buf->resp.port_name, |
3806 | sizeof(vhost->login_buf->resp.port_name)); | |
3807 | memcpy(&mad->fc_iu.payload[4], &vhost->login_buf->resp.node_name, | |
3808 | sizeof(vhost->login_buf->resp.node_name)); | |
0aab6c3f | 3809 | mad->fc_iu.payload[6] = cpu_to_be32(be64_to_cpu(vhost->login_buf->resp.scsi_id) & 0x00ffffff); |
989b8545 | 3810 | |
10501e1c BK |
3811 | if (timer_pending(&tgt->timer)) |
3812 | mod_timer(&tgt->timer, jiffies + (IBMVFC_ADISC_TIMEOUT * HZ)); | |
3813 | else { | |
3814 | tgt->timer.data = (unsigned long) tgt; | |
3815 | tgt->timer.expires = jiffies + (IBMVFC_ADISC_TIMEOUT * HZ); | |
3816 | tgt->timer.function = (void (*)(unsigned long))ibmvfc_adisc_timeout; | |
3817 | add_timer(&tgt->timer); | |
3818 | } | |
3819 | ||
989b8545 | 3820 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_INIT_WAIT); |
10501e1c | 3821 | if (ibmvfc_send_event(evt, vhost, IBMVFC_ADISC_PLUS_CANCEL_TIMEOUT)) { |
989b8545 | 3822 | vhost->discovery_threads--; |
10501e1c | 3823 | del_timer(&tgt->timer); |
989b8545 BK |
3824 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_NONE); |
3825 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3826 | } else | |
3827 | tgt_dbg(tgt, "Sent ADISC\n"); | |
3828 | } | |
3829 | ||
072b91f9 BK |
3830 | /** |
3831 | * ibmvfc_tgt_query_target_done - Completion handler for Query Target MAD | |
3832 | * @evt: ibmvfc event struct | |
3833 | * | |
3834 | **/ | |
3835 | static void ibmvfc_tgt_query_target_done(struct ibmvfc_event *evt) | |
3836 | { | |
3837 | struct ibmvfc_target *tgt = evt->tgt; | |
3838 | struct ibmvfc_host *vhost = evt->vhost; | |
3839 | struct ibmvfc_query_tgt *rsp = &evt->xfer_iu->query_tgt; | |
0aab6c3f | 3840 | u32 status = be16_to_cpu(rsp->common.status); |
7d0e4622 | 3841 | int level = IBMVFC_DEFAULT_LOG_LEVEL; |
072b91f9 BK |
3842 | |
3843 | vhost->discovery_threads--; | |
3844 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_NONE); | |
3845 | switch (status) { | |
3846 | case IBMVFC_MAD_SUCCESS: | |
3847 | tgt_dbg(tgt, "Query Target succeeded\n"); | |
0aab6c3f TD |
3848 | tgt->new_scsi_id = be64_to_cpu(rsp->scsi_id); |
3849 | if (be64_to_cpu(rsp->scsi_id) != tgt->scsi_id) | |
072b91f9 | 3850 | ibmvfc_init_tgt(tgt, ibmvfc_tgt_implicit_logout); |
989b8545 BK |
3851 | else |
3852 | ibmvfc_init_tgt(tgt, ibmvfc_tgt_adisc); | |
072b91f9 BK |
3853 | break; |
3854 | case IBMVFC_MAD_DRIVER_FAILED: | |
3855 | break; | |
3856 | case IBMVFC_MAD_CRQ_ERROR: | |
3857 | ibmvfc_retry_tgt_init(tgt, ibmvfc_tgt_query_target); | |
3858 | break; | |
3859 | case IBMVFC_MAD_FAILED: | |
3860 | default: | |
0aab6c3f TD |
3861 | if ((be16_to_cpu(rsp->status) & IBMVFC_FABRIC_MAPPED) == IBMVFC_FABRIC_MAPPED && |
3862 | be16_to_cpu(rsp->error) == IBMVFC_UNABLE_TO_PERFORM_REQ && | |
3863 | be16_to_cpu(rsp->fc_explain) == IBMVFC_PORT_NAME_NOT_REG) | |
072b91f9 | 3864 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); |
0aab6c3f | 3865 | else if (ibmvfc_retry_cmd(be16_to_cpu(rsp->status), be16_to_cpu(rsp->error))) |
7d0e4622 | 3866 | level += ibmvfc_retry_tgt_init(tgt, ibmvfc_tgt_query_target); |
10e79499 BK |
3867 | else |
3868 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DEL_RPORT); | |
7d0e4622 BK |
3869 | |
3870 | tgt_log(tgt, level, "Query Target failed: %s (%x:%x) %s (%x) %s (%x) rc=0x%02X\n", | |
0aab6c3f TD |
3871 | ibmvfc_get_cmd_error(be16_to_cpu(rsp->status), be16_to_cpu(rsp->error)), |
3872 | rsp->status, rsp->error, ibmvfc_get_fc_type(be16_to_cpu(rsp->fc_type)), | |
3873 | rsp->fc_type, ibmvfc_get_gs_explain(be16_to_cpu(rsp->fc_explain)), | |
3874 | rsp->fc_explain, status); | |
072b91f9 BK |
3875 | break; |
3876 | }; | |
3877 | ||
3878 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3879 | ibmvfc_free_event(evt); | |
3880 | wake_up(&vhost->work_wait_q); | |
3881 | } | |
3882 | ||
3883 | /** | |
3884 | * ibmvfc_tgt_query_target - Initiate a Query Target for specified target | |
3885 | * @tgt: ibmvfc target struct | |
3886 | * | |
3887 | **/ | |
3888 | static void ibmvfc_tgt_query_target(struct ibmvfc_target *tgt) | |
3889 | { | |
3890 | struct ibmvfc_query_tgt *query_tgt; | |
3891 | struct ibmvfc_host *vhost = tgt->vhost; | |
3892 | struct ibmvfc_event *evt; | |
3893 | ||
3894 | if (vhost->discovery_threads >= disc_threads) | |
3895 | return; | |
3896 | ||
3897 | kref_get(&tgt->kref); | |
3898 | evt = ibmvfc_get_event(vhost); | |
3899 | vhost->discovery_threads++; | |
3900 | evt->tgt = tgt; | |
3901 | ibmvfc_init_event(evt, ibmvfc_tgt_query_target_done, IBMVFC_MAD_FORMAT); | |
3902 | query_tgt = &evt->iu.query_tgt; | |
3903 | memset(query_tgt, 0, sizeof(*query_tgt)); | |
0aab6c3f TD |
3904 | query_tgt->common.version = cpu_to_be32(1); |
3905 | query_tgt->common.opcode = cpu_to_be32(IBMVFC_QUERY_TARGET); | |
3906 | query_tgt->common.length = cpu_to_be16(sizeof(*query_tgt)); | |
3907 | query_tgt->wwpn = cpu_to_be64(tgt->ids.port_name); | |
072b91f9 BK |
3908 | |
3909 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_INIT_WAIT); | |
3910 | if (ibmvfc_send_event(evt, vhost, default_timeout)) { | |
3911 | vhost->discovery_threads--; | |
3912 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_NONE); | |
3913 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
3914 | } else | |
3915 | tgt_dbg(tgt, "Sent Query Target\n"); | |
3916 | } | |
3917 | ||
3918 | /** | |
3919 | * ibmvfc_alloc_target - Allocate and initialize an ibmvfc target | |
3920 | * @vhost: ibmvfc host struct | |
3921 | * @scsi_id: SCSI ID to allocate target for | |
3922 | * | |
3923 | * Returns: | |
3924 | * 0 on success / other on failure | |
3925 | **/ | |
3926 | static int ibmvfc_alloc_target(struct ibmvfc_host *vhost, u64 scsi_id) | |
3927 | { | |
3928 | struct ibmvfc_target *tgt; | |
3929 | unsigned long flags; | |
3930 | ||
3931 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
3932 | list_for_each_entry(tgt, &vhost->targets, queue) { | |
3933 | if (tgt->scsi_id == scsi_id) { | |
3934 | if (tgt->need_login) | |
3935 | ibmvfc_init_tgt(tgt, ibmvfc_tgt_implicit_logout); | |
3936 | goto unlock_out; | |
3937 | } | |
3938 | } | |
3939 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
3940 | ||
7270b9bd | 3941 | tgt = mempool_alloc(vhost->tgt_pool, GFP_NOIO); |
072b91f9 | 3942 | if (!tgt) { |
775a42ec | 3943 | dev_err(vhost->dev, "Target allocation failure for scsi id %08llx\n", |
072b91f9 BK |
3944 | scsi_id); |
3945 | return -ENOMEM; | |
3946 | } | |
3947 | ||
0883e3b3 | 3948 | memset(tgt, 0, sizeof(*tgt)); |
072b91f9 BK |
3949 | tgt->scsi_id = scsi_id; |
3950 | tgt->new_scsi_id = scsi_id; | |
3951 | tgt->vhost = vhost; | |
3952 | tgt->need_login = 1; | |
10501e1c BK |
3953 | tgt->cancel_key = vhost->task_set++; |
3954 | init_timer(&tgt->timer); | |
072b91f9 BK |
3955 | kref_init(&tgt->kref); |
3956 | ibmvfc_init_tgt(tgt, ibmvfc_tgt_implicit_logout); | |
3957 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
3958 | list_add_tail(&tgt->queue, &vhost->targets); | |
3959 | ||
3960 | unlock_out: | |
3961 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
3962 | return 0; | |
3963 | } | |
3964 | ||
3965 | /** | |
3966 | * ibmvfc_alloc_targets - Allocate and initialize ibmvfc targets | |
3967 | * @vhost: ibmvfc host struct | |
3968 | * | |
3969 | * Returns: | |
3970 | * 0 on success / other on failure | |
3971 | **/ | |
3972 | static int ibmvfc_alloc_targets(struct ibmvfc_host *vhost) | |
3973 | { | |
3974 | int i, rc; | |
3975 | ||
3976 | for (i = 0, rc = 0; !rc && i < vhost->num_targets; i++) | |
3977 | rc = ibmvfc_alloc_target(vhost, | |
0aab6c3f TD |
3978 | be32_to_cpu(vhost->disc_buf->scsi_id[i]) & |
3979 | IBMVFC_DISC_TGT_SCSI_ID_MASK); | |
072b91f9 BK |
3980 | |
3981 | return rc; | |
3982 | } | |
3983 | ||
3984 | /** | |
3985 | * ibmvfc_discover_targets_done - Completion handler for discover targets MAD | |
3986 | * @evt: ibmvfc event struct | |
3987 | * | |
3988 | **/ | |
3989 | static void ibmvfc_discover_targets_done(struct ibmvfc_event *evt) | |
3990 | { | |
3991 | struct ibmvfc_host *vhost = evt->vhost; | |
3992 | struct ibmvfc_discover_targets *rsp = &evt->xfer_iu->discover_targets; | |
0aab6c3f | 3993 | u32 mad_status = be16_to_cpu(rsp->common.status); |
7d0e4622 | 3994 | int level = IBMVFC_DEFAULT_LOG_LEVEL; |
072b91f9 BK |
3995 | |
3996 | switch (mad_status) { | |
3997 | case IBMVFC_MAD_SUCCESS: | |
3998 | ibmvfc_dbg(vhost, "Discover Targets succeeded\n"); | |
0aab6c3f | 3999 | vhost->num_targets = be32_to_cpu(rsp->num_written); |
072b91f9 BK |
4000 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_ALLOC_TGTS); |
4001 | break; | |
4002 | case IBMVFC_MAD_FAILED: | |
7d0e4622 BK |
4003 | level += ibmvfc_retry_host_init(vhost); |
4004 | ibmvfc_log(vhost, level, "Discover Targets failed: %s (%x:%x)\n", | |
0aab6c3f TD |
4005 | ibmvfc_get_cmd_error(be16_to_cpu(rsp->status), be16_to_cpu(rsp->error)), |
4006 | rsp->status, rsp->error); | |
072b91f9 BK |
4007 | break; |
4008 | case IBMVFC_MAD_DRIVER_FAILED: | |
4009 | break; | |
4010 | default: | |
4011 | dev_err(vhost->dev, "Invalid Discover Targets response: 0x%x\n", mad_status); | |
4012 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
4013 | break; | |
4014 | } | |
4015 | ||
4016 | ibmvfc_free_event(evt); | |
4017 | wake_up(&vhost->work_wait_q); | |
4018 | } | |
4019 | ||
4020 | /** | |
4021 | * ibmvfc_discover_targets - Send Discover Targets MAD | |
4022 | * @vhost: ibmvfc host struct | |
4023 | * | |
4024 | **/ | |
4025 | static void ibmvfc_discover_targets(struct ibmvfc_host *vhost) | |
4026 | { | |
4027 | struct ibmvfc_discover_targets *mad; | |
4028 | struct ibmvfc_event *evt = ibmvfc_get_event(vhost); | |
4029 | ||
4030 | ibmvfc_init_event(evt, ibmvfc_discover_targets_done, IBMVFC_MAD_FORMAT); | |
4031 | mad = &evt->iu.discover_targets; | |
4032 | memset(mad, 0, sizeof(*mad)); | |
0aab6c3f TD |
4033 | mad->common.version = cpu_to_be32(1); |
4034 | mad->common.opcode = cpu_to_be32(IBMVFC_DISC_TARGETS); | |
4035 | mad->common.length = cpu_to_be16(sizeof(*mad)); | |
4036 | mad->bufflen = cpu_to_be32(vhost->disc_buf_sz); | |
4037 | mad->buffer.va = cpu_to_be64(vhost->disc_buf_dma); | |
4038 | mad->buffer.len = cpu_to_be32(vhost->disc_buf_sz); | |
072b91f9 BK |
4039 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_INIT_WAIT); |
4040 | ||
4041 | if (!ibmvfc_send_event(evt, vhost, default_timeout)) | |
4042 | ibmvfc_dbg(vhost, "Sent discover targets\n"); | |
4043 | else | |
4044 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
4045 | } | |
4046 | ||
4047 | /** | |
4048 | * ibmvfc_npiv_login_done - Completion handler for NPIV Login | |
4049 | * @evt: ibmvfc event struct | |
4050 | * | |
4051 | **/ | |
4052 | static void ibmvfc_npiv_login_done(struct ibmvfc_event *evt) | |
4053 | { | |
4054 | struct ibmvfc_host *vhost = evt->vhost; | |
0aab6c3f | 4055 | u32 mad_status = be16_to_cpu(evt->xfer_iu->npiv_login.common.status); |
072b91f9 BK |
4056 | struct ibmvfc_npiv_login_resp *rsp = &vhost->login_buf->resp; |
4057 | unsigned int npiv_max_sectors; | |
7d0e4622 | 4058 | int level = IBMVFC_DEFAULT_LOG_LEVEL; |
072b91f9 BK |
4059 | |
4060 | switch (mad_status) { | |
4061 | case IBMVFC_MAD_SUCCESS: | |
4062 | ibmvfc_free_event(evt); | |
4063 | break; | |
4064 | case IBMVFC_MAD_FAILED: | |
0aab6c3f | 4065 | if (ibmvfc_retry_cmd(be16_to_cpu(rsp->status), be16_to_cpu(rsp->error))) |
7d0e4622 | 4066 | level += ibmvfc_retry_host_init(vhost); |
072b91f9 BK |
4067 | else |
4068 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
7d0e4622 | 4069 | ibmvfc_log(vhost, level, "NPIV Login failed: %s (%x:%x)\n", |
0aab6c3f TD |
4070 | ibmvfc_get_cmd_error(be16_to_cpu(rsp->status), be16_to_cpu(rsp->error)), |
4071 | rsp->status, rsp->error); | |
072b91f9 BK |
4072 | ibmvfc_free_event(evt); |
4073 | return; | |
4074 | case IBMVFC_MAD_CRQ_ERROR: | |
4075 | ibmvfc_retry_host_init(vhost); | |
4076 | case IBMVFC_MAD_DRIVER_FAILED: | |
4077 | ibmvfc_free_event(evt); | |
4078 | return; | |
4079 | default: | |
4080 | dev_err(vhost->dev, "Invalid NPIV Login response: 0x%x\n", mad_status); | |
4081 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
4082 | ibmvfc_free_event(evt); | |
4083 | return; | |
4084 | } | |
4085 | ||
4086 | vhost->client_migrated = 0; | |
4087 | ||
0aab6c3f | 4088 | if (!(be32_to_cpu(rsp->flags) & IBMVFC_NATIVE_FC)) { |
072b91f9 BK |
4089 | dev_err(vhost->dev, "Virtual adapter does not support FC. %x\n", |
4090 | rsp->flags); | |
4091 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
4092 | wake_up(&vhost->work_wait_q); | |
4093 | return; | |
4094 | } | |
4095 | ||
0aab6c3f | 4096 | if (be32_to_cpu(rsp->max_cmds) <= IBMVFC_NUM_INTERNAL_REQ) { |
072b91f9 BK |
4097 | dev_err(vhost->dev, "Virtual adapter supported queue depth too small: %d\n", |
4098 | rsp->max_cmds); | |
4099 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
4100 | wake_up(&vhost->work_wait_q); | |
4101 | return; | |
4102 | } | |
4103 | ||
79111d08 | 4104 | vhost->logged_in = 1; |
0aab6c3f | 4105 | npiv_max_sectors = min((uint)(be64_to_cpu(rsp->max_dma_len) >> 9), IBMVFC_MAX_SECTORS); |
072b91f9 BK |
4106 | dev_info(vhost->dev, "Host partition: %s, device: %s %s %s max sectors %u\n", |
4107 | rsp->partition_name, rsp->device_name, rsp->port_loc_code, | |
4108 | rsp->drc_name, npiv_max_sectors); | |
4109 | ||
0aab6c3f TD |
4110 | fc_host_fabric_name(vhost->host) = be64_to_cpu(rsp->node_name); |
4111 | fc_host_node_name(vhost->host) = be64_to_cpu(rsp->node_name); | |
4112 | fc_host_port_name(vhost->host) = be64_to_cpu(rsp->port_name); | |
4113 | fc_host_port_id(vhost->host) = be64_to_cpu(rsp->scsi_id); | |
072b91f9 BK |
4114 | fc_host_port_type(vhost->host) = FC_PORTTYPE_NPIV; |
4115 | fc_host_supported_classes(vhost->host) = 0; | |
0aab6c3f | 4116 | if (be32_to_cpu(rsp->service_parms.class1_parms[0]) & 0x80000000) |
072b91f9 | 4117 | fc_host_supported_classes(vhost->host) |= FC_COS_CLASS1; |
0aab6c3f | 4118 | if (be32_to_cpu(rsp->service_parms.class2_parms[0]) & 0x80000000) |
072b91f9 | 4119 | fc_host_supported_classes(vhost->host) |= FC_COS_CLASS2; |
0aab6c3f | 4120 | if (be32_to_cpu(rsp->service_parms.class3_parms[0]) & 0x80000000) |
072b91f9 BK |
4121 | fc_host_supported_classes(vhost->host) |= FC_COS_CLASS3; |
4122 | fc_host_maxframe_size(vhost->host) = | |
0aab6c3f | 4123 | be16_to_cpu(rsp->service_parms.common.bb_rcv_sz) & 0x0fff; |
072b91f9 | 4124 | |
0aab6c3f | 4125 | vhost->host->can_queue = be32_to_cpu(rsp->max_cmds) - IBMVFC_NUM_INTERNAL_REQ; |
072b91f9 BK |
4126 | vhost->host->max_sectors = npiv_max_sectors; |
4127 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_QUERY); | |
4128 | wake_up(&vhost->work_wait_q); | |
4129 | } | |
4130 | ||
4131 | /** | |
4132 | * ibmvfc_npiv_login - Sends NPIV login | |
4133 | * @vhost: ibmvfc host struct | |
4134 | * | |
4135 | **/ | |
4136 | static void ibmvfc_npiv_login(struct ibmvfc_host *vhost) | |
4137 | { | |
4138 | struct ibmvfc_npiv_login_mad *mad; | |
4139 | struct ibmvfc_event *evt = ibmvfc_get_event(vhost); | |
4140 | ||
4141 | ibmvfc_gather_partition_info(vhost); | |
4142 | ibmvfc_set_login_info(vhost); | |
4143 | ibmvfc_init_event(evt, ibmvfc_npiv_login_done, IBMVFC_MAD_FORMAT); | |
4144 | ||
4145 | memcpy(vhost->login_buf, &vhost->login_info, sizeof(vhost->login_info)); | |
4146 | mad = &evt->iu.npiv_login; | |
4147 | memset(mad, 0, sizeof(struct ibmvfc_npiv_login_mad)); | |
0aab6c3f TD |
4148 | mad->common.version = cpu_to_be32(1); |
4149 | mad->common.opcode = cpu_to_be32(IBMVFC_NPIV_LOGIN); | |
4150 | mad->common.length = cpu_to_be16(sizeof(struct ibmvfc_npiv_login_mad)); | |
4151 | mad->buffer.va = cpu_to_be64(vhost->login_buf_dma); | |
4152 | mad->buffer.len = cpu_to_be32(sizeof(*vhost->login_buf)); | |
072b91f9 | 4153 | |
072b91f9 BK |
4154 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_INIT_WAIT); |
4155 | ||
4156 | if (!ibmvfc_send_event(evt, vhost, default_timeout)) | |
4157 | ibmvfc_dbg(vhost, "Sent NPIV login\n"); | |
4158 | else | |
4159 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
4160 | }; | |
4161 | ||
79111d08 BK |
4162 | /** |
4163 | * ibmvfc_npiv_logout_done - Completion handler for NPIV Logout | |
4164 | * @vhost: ibmvfc host struct | |
4165 | * | |
4166 | **/ | |
4167 | static void ibmvfc_npiv_logout_done(struct ibmvfc_event *evt) | |
4168 | { | |
4169 | struct ibmvfc_host *vhost = evt->vhost; | |
0aab6c3f | 4170 | u32 mad_status = be16_to_cpu(evt->xfer_iu->npiv_logout.common.status); |
79111d08 BK |
4171 | |
4172 | ibmvfc_free_event(evt); | |
4173 | ||
4174 | switch (mad_status) { | |
4175 | case IBMVFC_MAD_SUCCESS: | |
4176 | if (list_empty(&vhost->sent) && | |
4177 | vhost->action == IBMVFC_HOST_ACTION_LOGO_WAIT) { | |
861890c6 | 4178 | ibmvfc_init_host(vhost); |
79111d08 BK |
4179 | return; |
4180 | } | |
4181 | break; | |
4182 | case IBMVFC_MAD_FAILED: | |
4183 | case IBMVFC_MAD_NOT_SUPPORTED: | |
4184 | case IBMVFC_MAD_CRQ_ERROR: | |
4185 | case IBMVFC_MAD_DRIVER_FAILED: | |
4186 | default: | |
4187 | ibmvfc_dbg(vhost, "NPIV Logout failed. 0x%X\n", mad_status); | |
4188 | break; | |
4189 | } | |
4190 | ||
4191 | ibmvfc_hard_reset_host(vhost); | |
4192 | } | |
4193 | ||
4194 | /** | |
4195 | * ibmvfc_npiv_logout - Issue an NPIV Logout | |
4196 | * @vhost: ibmvfc host struct | |
4197 | * | |
4198 | **/ | |
4199 | static void ibmvfc_npiv_logout(struct ibmvfc_host *vhost) | |
4200 | { | |
4201 | struct ibmvfc_npiv_logout_mad *mad; | |
4202 | struct ibmvfc_event *evt; | |
4203 | ||
4204 | evt = ibmvfc_get_event(vhost); | |
4205 | ibmvfc_init_event(evt, ibmvfc_npiv_logout_done, IBMVFC_MAD_FORMAT); | |
4206 | ||
4207 | mad = &evt->iu.npiv_logout; | |
4208 | memset(mad, 0, sizeof(*mad)); | |
0aab6c3f TD |
4209 | mad->common.version = cpu_to_be32(1); |
4210 | mad->common.opcode = cpu_to_be32(IBMVFC_NPIV_LOGOUT); | |
4211 | mad->common.length = cpu_to_be16(sizeof(struct ibmvfc_npiv_logout_mad)); | |
79111d08 BK |
4212 | |
4213 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_LOGO_WAIT); | |
4214 | ||
4215 | if (!ibmvfc_send_event(evt, vhost, default_timeout)) | |
4216 | ibmvfc_dbg(vhost, "Sent NPIV logout\n"); | |
4217 | else | |
4218 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
4219 | } | |
4220 | ||
072b91f9 BK |
4221 | /** |
4222 | * ibmvfc_dev_init_to_do - Is there target initialization work to do? | |
4223 | * @vhost: ibmvfc host struct | |
4224 | * | |
4225 | * Returns: | |
4226 | * 1 if work to do / 0 if not | |
4227 | **/ | |
4228 | static int ibmvfc_dev_init_to_do(struct ibmvfc_host *vhost) | |
4229 | { | |
4230 | struct ibmvfc_target *tgt; | |
4231 | ||
4232 | list_for_each_entry(tgt, &vhost->targets, queue) { | |
4233 | if (tgt->action == IBMVFC_TGT_ACTION_INIT || | |
4234 | tgt->action == IBMVFC_TGT_ACTION_INIT_WAIT) | |
4235 | return 1; | |
4236 | } | |
4237 | ||
4238 | return 0; | |
4239 | } | |
4240 | ||
4241 | /** | |
4242 | * __ibmvfc_work_to_do - Is there task level work to do? (no locking) | |
4243 | * @vhost: ibmvfc host struct | |
4244 | * | |
4245 | * Returns: | |
4246 | * 1 if work to do / 0 if not | |
4247 | **/ | |
4248 | static int __ibmvfc_work_to_do(struct ibmvfc_host *vhost) | |
4249 | { | |
4250 | struct ibmvfc_target *tgt; | |
4251 | ||
4252 | if (kthread_should_stop()) | |
4253 | return 1; | |
4254 | switch (vhost->action) { | |
4255 | case IBMVFC_HOST_ACTION_NONE: | |
4256 | case IBMVFC_HOST_ACTION_INIT_WAIT: | |
79111d08 | 4257 | case IBMVFC_HOST_ACTION_LOGO_WAIT: |
072b91f9 BK |
4258 | return 0; |
4259 | case IBMVFC_HOST_ACTION_TGT_INIT: | |
4260 | case IBMVFC_HOST_ACTION_QUERY_TGTS: | |
4261 | if (vhost->discovery_threads == disc_threads) | |
4262 | return 0; | |
4263 | list_for_each_entry(tgt, &vhost->targets, queue) | |
4264 | if (tgt->action == IBMVFC_TGT_ACTION_INIT) | |
4265 | return 1; | |
4266 | list_for_each_entry(tgt, &vhost->targets, queue) | |
4267 | if (tgt->action == IBMVFC_TGT_ACTION_INIT_WAIT) | |
4268 | return 0; | |
4269 | return 1; | |
79111d08 | 4270 | case IBMVFC_HOST_ACTION_LOGO: |
072b91f9 BK |
4271 | case IBMVFC_HOST_ACTION_INIT: |
4272 | case IBMVFC_HOST_ACTION_ALLOC_TGTS: | |
072b91f9 | 4273 | case IBMVFC_HOST_ACTION_TGT_DEL: |
10e79499 | 4274 | case IBMVFC_HOST_ACTION_TGT_DEL_FAILED: |
072b91f9 | 4275 | case IBMVFC_HOST_ACTION_QUERY: |
73ee5d86 BK |
4276 | case IBMVFC_HOST_ACTION_RESET: |
4277 | case IBMVFC_HOST_ACTION_REENABLE: | |
072b91f9 BK |
4278 | default: |
4279 | break; | |
4280 | }; | |
4281 | ||
4282 | return 1; | |
4283 | } | |
4284 | ||
4285 | /** | |
4286 | * ibmvfc_work_to_do - Is there task level work to do? | |
4287 | * @vhost: ibmvfc host struct | |
4288 | * | |
4289 | * Returns: | |
4290 | * 1 if work to do / 0 if not | |
4291 | **/ | |
4292 | static int ibmvfc_work_to_do(struct ibmvfc_host *vhost) | |
4293 | { | |
4294 | unsigned long flags; | |
4295 | int rc; | |
4296 | ||
4297 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
4298 | rc = __ibmvfc_work_to_do(vhost); | |
4299 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4300 | return rc; | |
4301 | } | |
4302 | ||
4303 | /** | |
4304 | * ibmvfc_log_ae - Log async events if necessary | |
4305 | * @vhost: ibmvfc host struct | |
4306 | * @events: events to log | |
4307 | * | |
4308 | **/ | |
4309 | static void ibmvfc_log_ae(struct ibmvfc_host *vhost, int events) | |
4310 | { | |
4311 | if (events & IBMVFC_AE_RSCN) | |
4312 | fc_host_post_event(vhost->host, fc_get_event_number(), FCH_EVT_RSCN, 0); | |
4313 | if ((events & IBMVFC_AE_LINKDOWN) && | |
4314 | vhost->state >= IBMVFC_HALTED) | |
4315 | fc_host_post_event(vhost->host, fc_get_event_number(), FCH_EVT_LINKDOWN, 0); | |
4316 | if ((events & IBMVFC_AE_LINKUP) && | |
4317 | vhost->state == IBMVFC_INITIALIZING) | |
4318 | fc_host_post_event(vhost->host, fc_get_event_number(), FCH_EVT_LINKUP, 0); | |
4319 | } | |
4320 | ||
4321 | /** | |
4322 | * ibmvfc_tgt_add_rport - Tell the FC transport about a new remote port | |
4323 | * @tgt: ibmvfc target struct | |
4324 | * | |
4325 | **/ | |
4326 | static void ibmvfc_tgt_add_rport(struct ibmvfc_target *tgt) | |
4327 | { | |
4328 | struct ibmvfc_host *vhost = tgt->vhost; | |
43c8da90 | 4329 | struct fc_rport *rport; |
072b91f9 BK |
4330 | unsigned long flags; |
4331 | ||
43c8da90 BK |
4332 | tgt_dbg(tgt, "Adding rport\n"); |
4333 | rport = fc_remote_port_add(vhost->host, 0, &tgt->ids); | |
4334 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
4335 | ||
4336 | if (rport && tgt->action == IBMVFC_TGT_ACTION_DEL_RPORT) { | |
4337 | tgt_dbg(tgt, "Deleting rport\n"); | |
4338 | list_del(&tgt->queue); | |
d5da3040 | 4339 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DELETED_RPORT); |
0883e3b3 | 4340 | spin_unlock_irqrestore(vhost->host->host_lock, flags); |
43c8da90 BK |
4341 | fc_remote_port_delete(rport); |
4342 | del_timer_sync(&tgt->timer); | |
4343 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
0883e3b3 | 4344 | return; |
d5da3040 BK |
4345 | } else if (rport && tgt->action == IBMVFC_TGT_ACTION_DELETED_RPORT) { |
4346 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4347 | return; | |
0883e3b3 BK |
4348 | } |
4349 | ||
072b91f9 BK |
4350 | if (rport) { |
4351 | tgt_dbg(tgt, "rport add succeeded\n"); | |
43c8da90 | 4352 | tgt->rport = rport; |
0aab6c3f | 4353 | rport->maxframe_size = be16_to_cpu(tgt->service_parms.common.bb_rcv_sz) & 0x0fff; |
072b91f9 | 4354 | rport->supported_classes = 0; |
52d7e861 | 4355 | tgt->target_id = rport->scsi_target_id; |
0aab6c3f | 4356 | if (be32_to_cpu(tgt->service_parms.class1_parms[0]) & 0x80000000) |
072b91f9 | 4357 | rport->supported_classes |= FC_COS_CLASS1; |
0aab6c3f | 4358 | if (be32_to_cpu(tgt->service_parms.class2_parms[0]) & 0x80000000) |
072b91f9 | 4359 | rport->supported_classes |= FC_COS_CLASS2; |
0aab6c3f | 4360 | if (be32_to_cpu(tgt->service_parms.class3_parms[0]) & 0x80000000) |
072b91f9 | 4361 | rport->supported_classes |= FC_COS_CLASS3; |
d31429e1 | 4362 | if (rport->rqst_q) |
8a78362c | 4363 | blk_queue_max_segments(rport->rqst_q, 1); |
072b91f9 BK |
4364 | } else |
4365 | tgt_dbg(tgt, "rport add failed\n"); | |
4366 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4367 | } | |
4368 | ||
4369 | /** | |
4370 | * ibmvfc_do_work - Do task level work | |
4371 | * @vhost: ibmvfc host struct | |
4372 | * | |
4373 | **/ | |
4374 | static void ibmvfc_do_work(struct ibmvfc_host *vhost) | |
4375 | { | |
4376 | struct ibmvfc_target *tgt; | |
4377 | unsigned long flags; | |
4378 | struct fc_rport *rport; | |
73ee5d86 | 4379 | int rc; |
072b91f9 BK |
4380 | |
4381 | ibmvfc_log_ae(vhost, vhost->events_to_log); | |
4382 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
4383 | vhost->events_to_log = 0; | |
4384 | switch (vhost->action) { | |
4385 | case IBMVFC_HOST_ACTION_NONE: | |
79111d08 | 4386 | case IBMVFC_HOST_ACTION_LOGO_WAIT: |
072b91f9 BK |
4387 | case IBMVFC_HOST_ACTION_INIT_WAIT: |
4388 | break; | |
73ee5d86 BK |
4389 | case IBMVFC_HOST_ACTION_RESET: |
4390 | vhost->action = IBMVFC_HOST_ACTION_TGT_DEL; | |
4391 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4392 | rc = ibmvfc_reset_crq(vhost); | |
4393 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
d24099df BK |
4394 | if (rc == H_CLOSED) |
4395 | vio_enable_interrupts(to_vio_dev(vhost->dev)); | |
38553564 BK |
4396 | if (rc || (rc = ibmvfc_send_crq_init(vhost)) || |
4397 | (rc = vio_enable_interrupts(to_vio_dev(vhost->dev)))) { | |
73ee5d86 BK |
4398 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); |
4399 | dev_err(vhost->dev, "Error after reset (rc=%d)\n", rc); | |
4400 | } | |
4401 | break; | |
4402 | case IBMVFC_HOST_ACTION_REENABLE: | |
4403 | vhost->action = IBMVFC_HOST_ACTION_TGT_DEL; | |
4404 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4405 | rc = ibmvfc_reenable_crq_queue(vhost); | |
4406 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
4407 | if (rc || (rc = ibmvfc_send_crq_init(vhost))) { | |
4408 | ibmvfc_link_down(vhost, IBMVFC_LINK_DEAD); | |
4409 | dev_err(vhost->dev, "Error after enable (rc=%d)\n", rc); | |
4410 | } | |
4411 | break; | |
79111d08 BK |
4412 | case IBMVFC_HOST_ACTION_LOGO: |
4413 | vhost->job_step(vhost); | |
4414 | break; | |
072b91f9 BK |
4415 | case IBMVFC_HOST_ACTION_INIT: |
4416 | BUG_ON(vhost->state != IBMVFC_INITIALIZING); | |
1c41fa82 BK |
4417 | if (vhost->delay_init) { |
4418 | vhost->delay_init = 0; | |
4419 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
d2131b33 | 4420 | ssleep(15); |
1c41fa82 BK |
4421 | return; |
4422 | } else | |
4423 | vhost->job_step(vhost); | |
072b91f9 BK |
4424 | break; |
4425 | case IBMVFC_HOST_ACTION_QUERY: | |
4426 | list_for_each_entry(tgt, &vhost->targets, queue) | |
4427 | ibmvfc_init_tgt(tgt, ibmvfc_tgt_query_target); | |
4428 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_QUERY_TGTS); | |
4429 | break; | |
4430 | case IBMVFC_HOST_ACTION_QUERY_TGTS: | |
4431 | list_for_each_entry(tgt, &vhost->targets, queue) { | |
4432 | if (tgt->action == IBMVFC_TGT_ACTION_INIT) { | |
4433 | tgt->job_step(tgt); | |
4434 | break; | |
4435 | } | |
4436 | } | |
4437 | ||
4438 | if (!ibmvfc_dev_init_to_do(vhost)) | |
4439 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_TGT_DEL); | |
4440 | break; | |
4441 | case IBMVFC_HOST_ACTION_TGT_DEL: | |
10e79499 | 4442 | case IBMVFC_HOST_ACTION_TGT_DEL_FAILED: |
072b91f9 BK |
4443 | list_for_each_entry(tgt, &vhost->targets, queue) { |
4444 | if (tgt->action == IBMVFC_TGT_ACTION_DEL_RPORT) { | |
4445 | tgt_dbg(tgt, "Deleting rport\n"); | |
4446 | rport = tgt->rport; | |
4447 | tgt->rport = NULL; | |
4448 | list_del(&tgt->queue); | |
d5da3040 | 4449 | ibmvfc_set_tgt_action(tgt, IBMVFC_TGT_ACTION_DELETED_RPORT); |
072b91f9 BK |
4450 | spin_unlock_irqrestore(vhost->host->host_lock, flags); |
4451 | if (rport) | |
4452 | fc_remote_port_delete(rport); | |
10501e1c | 4453 | del_timer_sync(&tgt->timer); |
072b91f9 BK |
4454 | kref_put(&tgt->kref, ibmvfc_release_tgt); |
4455 | return; | |
4456 | } | |
4457 | } | |
4458 | ||
4459 | if (vhost->state == IBMVFC_INITIALIZING) { | |
10e79499 | 4460 | if (vhost->action == IBMVFC_HOST_ACTION_TGT_DEL_FAILED) { |
43c8da90 BK |
4461 | if (vhost->reinit) { |
4462 | vhost->reinit = 0; | |
4463 | scsi_block_requests(vhost->host); | |
4464 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_QUERY); | |
4465 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4466 | } else { | |
4467 | ibmvfc_set_host_state(vhost, IBMVFC_ACTIVE); | |
4468 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_NONE); | |
4469 | wake_up(&vhost->init_wait_q); | |
4470 | schedule_work(&vhost->rport_add_work_q); | |
4471 | vhost->init_retries = 0; | |
4472 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4473 | scsi_unblock_requests(vhost->host); | |
4474 | } | |
4475 | ||
10e79499 BK |
4476 | return; |
4477 | } else { | |
4478 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_INIT); | |
4479 | vhost->job_step = ibmvfc_discover_targets; | |
4480 | } | |
072b91f9 BK |
4481 | } else { |
4482 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_NONE); | |
4483 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4484 | scsi_unblock_requests(vhost->host); | |
4485 | wake_up(&vhost->init_wait_q); | |
4486 | return; | |
4487 | } | |
4488 | break; | |
4489 | case IBMVFC_HOST_ACTION_ALLOC_TGTS: | |
4490 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_TGT_INIT); | |
4491 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4492 | ibmvfc_alloc_targets(vhost); | |
4493 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
4494 | break; | |
4495 | case IBMVFC_HOST_ACTION_TGT_INIT: | |
4496 | list_for_each_entry(tgt, &vhost->targets, queue) { | |
4497 | if (tgt->action == IBMVFC_TGT_ACTION_INIT) { | |
4498 | tgt->job_step(tgt); | |
4499 | break; | |
4500 | } | |
4501 | } | |
4502 | ||
10e79499 BK |
4503 | if (!ibmvfc_dev_init_to_do(vhost)) |
4504 | ibmvfc_set_host_action(vhost, IBMVFC_HOST_ACTION_TGT_DEL_FAILED); | |
072b91f9 | 4505 | break; |
072b91f9 BK |
4506 | default: |
4507 | break; | |
4508 | }; | |
4509 | ||
4510 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4511 | } | |
4512 | ||
4513 | /** | |
4514 | * ibmvfc_work - Do task level work | |
4515 | * @data: ibmvfc host struct | |
4516 | * | |
4517 | * Returns: | |
4518 | * zero | |
4519 | **/ | |
4520 | static int ibmvfc_work(void *data) | |
4521 | { | |
4522 | struct ibmvfc_host *vhost = data; | |
4523 | int rc; | |
4524 | ||
8698a745 | 4525 | set_user_nice(current, MIN_NICE); |
072b91f9 BK |
4526 | |
4527 | while (1) { | |
4528 | rc = wait_event_interruptible(vhost->work_wait_q, | |
4529 | ibmvfc_work_to_do(vhost)); | |
4530 | ||
4531 | BUG_ON(rc); | |
4532 | ||
4533 | if (kthread_should_stop()) | |
4534 | break; | |
4535 | ||
4536 | ibmvfc_do_work(vhost); | |
4537 | } | |
4538 | ||
4539 | ibmvfc_dbg(vhost, "ibmvfc kthread exiting...\n"); | |
4540 | return 0; | |
4541 | } | |
4542 | ||
4543 | /** | |
4544 | * ibmvfc_init_crq - Initializes and registers CRQ with hypervisor | |
4545 | * @vhost: ibmvfc host struct | |
4546 | * | |
4547 | * Allocates a page for messages, maps it for dma, and registers | |
4548 | * the crq with the hypervisor. | |
4549 | * | |
4550 | * Return value: | |
4551 | * zero on success / other on failure | |
4552 | **/ | |
4553 | static int ibmvfc_init_crq(struct ibmvfc_host *vhost) | |
4554 | { | |
4555 | int rc, retrc = -ENOMEM; | |
4556 | struct device *dev = vhost->dev; | |
4557 | struct vio_dev *vdev = to_vio_dev(dev); | |
4558 | struct ibmvfc_crq_queue *crq = &vhost->crq; | |
4559 | ||
4560 | ENTER; | |
4561 | crq->msgs = (struct ibmvfc_crq *)get_zeroed_page(GFP_KERNEL); | |
4562 | ||
4563 | if (!crq->msgs) | |
4564 | return -ENOMEM; | |
4565 | ||
4566 | crq->size = PAGE_SIZE / sizeof(*crq->msgs); | |
4567 | crq->msg_token = dma_map_single(dev, crq->msgs, | |
4568 | PAGE_SIZE, DMA_BIDIRECTIONAL); | |
4569 | ||
8d8bb39b | 4570 | if (dma_mapping_error(dev, crq->msg_token)) |
072b91f9 BK |
4571 | goto map_failed; |
4572 | ||
4573 | retrc = rc = plpar_hcall_norets(H_REG_CRQ, vdev->unit_address, | |
4574 | crq->msg_token, PAGE_SIZE); | |
4575 | ||
4576 | if (rc == H_RESOURCE) | |
4577 | /* maybe kexecing and resource is busy. try a reset */ | |
4578 | retrc = rc = ibmvfc_reset_crq(vhost); | |
4579 | ||
4580 | if (rc == H_CLOSED) | |
4581 | dev_warn(dev, "Partner adapter not ready\n"); | |
4582 | else if (rc) { | |
4583 | dev_warn(dev, "Error %d opening adapter\n", rc); | |
4584 | goto reg_crq_failed; | |
4585 | } | |
4586 | ||
4587 | retrc = 0; | |
4588 | ||
039a0898 BK |
4589 | tasklet_init(&vhost->tasklet, (void *)ibmvfc_tasklet, (unsigned long)vhost); |
4590 | ||
072b91f9 BK |
4591 | if ((rc = request_irq(vdev->irq, ibmvfc_interrupt, 0, IBMVFC_NAME, vhost))) { |
4592 | dev_err(dev, "Couldn't register irq 0x%x. rc=%d\n", vdev->irq, rc); | |
4593 | goto req_irq_failed; | |
4594 | } | |
4595 | ||
4596 | if ((rc = vio_enable_interrupts(vdev))) { | |
4597 | dev_err(dev, "Error %d enabling interrupts\n", rc); | |
4598 | goto req_irq_failed; | |
4599 | } | |
4600 | ||
4601 | crq->cur = 0; | |
4602 | LEAVE; | |
4603 | return retrc; | |
4604 | ||
4605 | req_irq_failed: | |
039a0898 | 4606 | tasklet_kill(&vhost->tasklet); |
072b91f9 BK |
4607 | do { |
4608 | rc = plpar_hcall_norets(H_FREE_CRQ, vdev->unit_address); | |
4609 | } while (rc == H_BUSY || H_IS_LONG_BUSY(rc)); | |
4610 | reg_crq_failed: | |
4611 | dma_unmap_single(dev, crq->msg_token, PAGE_SIZE, DMA_BIDIRECTIONAL); | |
4612 | map_failed: | |
4613 | free_page((unsigned long)crq->msgs); | |
4614 | return retrc; | |
4615 | } | |
4616 | ||
4617 | /** | |
4618 | * ibmvfc_free_mem - Free memory for vhost | |
4619 | * @vhost: ibmvfc host struct | |
4620 | * | |
4621 | * Return value: | |
4622 | * none | |
4623 | **/ | |
4624 | static void ibmvfc_free_mem(struct ibmvfc_host *vhost) | |
4625 | { | |
4626 | struct ibmvfc_async_crq_queue *async_q = &vhost->async_crq; | |
4627 | ||
4628 | ENTER; | |
4629 | mempool_destroy(vhost->tgt_pool); | |
4630 | kfree(vhost->trace); | |
4631 | dma_free_coherent(vhost->dev, vhost->disc_buf_sz, vhost->disc_buf, | |
4632 | vhost->disc_buf_dma); | |
4633 | dma_free_coherent(vhost->dev, sizeof(*vhost->login_buf), | |
4634 | vhost->login_buf, vhost->login_buf_dma); | |
4635 | dma_pool_destroy(vhost->sg_pool); | |
4636 | dma_unmap_single(vhost->dev, async_q->msg_token, | |
4637 | async_q->size * sizeof(*async_q->msgs), DMA_BIDIRECTIONAL); | |
4638 | free_page((unsigned long)async_q->msgs); | |
4639 | LEAVE; | |
4640 | } | |
4641 | ||
4642 | /** | |
4643 | * ibmvfc_alloc_mem - Allocate memory for vhost | |
4644 | * @vhost: ibmvfc host struct | |
4645 | * | |
4646 | * Return value: | |
4647 | * 0 on success / non-zero on failure | |
4648 | **/ | |
4649 | static int ibmvfc_alloc_mem(struct ibmvfc_host *vhost) | |
4650 | { | |
4651 | struct ibmvfc_async_crq_queue *async_q = &vhost->async_crq; | |
4652 | struct device *dev = vhost->dev; | |
4653 | ||
4654 | ENTER; | |
4655 | async_q->msgs = (struct ibmvfc_async_crq *)get_zeroed_page(GFP_KERNEL); | |
4656 | if (!async_q->msgs) { | |
4657 | dev_err(dev, "Couldn't allocate async queue.\n"); | |
4658 | goto nomem; | |
4659 | } | |
4660 | ||
4661 | async_q->size = PAGE_SIZE / sizeof(struct ibmvfc_async_crq); | |
4662 | async_q->msg_token = dma_map_single(dev, async_q->msgs, | |
4663 | async_q->size * sizeof(*async_q->msgs), | |
4664 | DMA_BIDIRECTIONAL); | |
4665 | ||
8d8bb39b | 4666 | if (dma_mapping_error(dev, async_q->msg_token)) { |
072b91f9 BK |
4667 | dev_err(dev, "Failed to map async queue\n"); |
4668 | goto free_async_crq; | |
4669 | } | |
4670 | ||
4671 | vhost->sg_pool = dma_pool_create(IBMVFC_NAME, dev, | |
4672 | SG_ALL * sizeof(struct srp_direct_buf), | |
4673 | sizeof(struct srp_direct_buf), 0); | |
4674 | ||
4675 | if (!vhost->sg_pool) { | |
4676 | dev_err(dev, "Failed to allocate sg pool\n"); | |
4677 | goto unmap_async_crq; | |
4678 | } | |
4679 | ||
4680 | vhost->login_buf = dma_alloc_coherent(dev, sizeof(*vhost->login_buf), | |
4681 | &vhost->login_buf_dma, GFP_KERNEL); | |
4682 | ||
4683 | if (!vhost->login_buf) { | |
4684 | dev_err(dev, "Couldn't allocate NPIV login buffer\n"); | |
4685 | goto free_sg_pool; | |
4686 | } | |
4687 | ||
4688 | vhost->disc_buf_sz = sizeof(vhost->disc_buf->scsi_id[0]) * max_targets; | |
4689 | vhost->disc_buf = dma_alloc_coherent(dev, vhost->disc_buf_sz, | |
4690 | &vhost->disc_buf_dma, GFP_KERNEL); | |
4691 | ||
4692 | if (!vhost->disc_buf) { | |
4693 | dev_err(dev, "Couldn't allocate Discover Targets buffer\n"); | |
4694 | goto free_login_buffer; | |
4695 | } | |
4696 | ||
4697 | vhost->trace = kcalloc(IBMVFC_NUM_TRACE_ENTRIES, | |
4698 | sizeof(struct ibmvfc_trace_entry), GFP_KERNEL); | |
4699 | ||
4700 | if (!vhost->trace) | |
4701 | goto free_disc_buffer; | |
4702 | ||
d6886692 | 4703 | vhost->tgt_pool = mempool_create_kmalloc_pool(IBMVFC_TGT_MEMPOOL_SZ, |
072b91f9 BK |
4704 | sizeof(struct ibmvfc_target)); |
4705 | ||
4706 | if (!vhost->tgt_pool) { | |
4707 | dev_err(dev, "Couldn't allocate target memory pool\n"); | |
4708 | goto free_trace; | |
4709 | } | |
4710 | ||
4711 | LEAVE; | |
4712 | return 0; | |
4713 | ||
4714 | free_trace: | |
4715 | kfree(vhost->trace); | |
4716 | free_disc_buffer: | |
4717 | dma_free_coherent(dev, vhost->disc_buf_sz, vhost->disc_buf, | |
4718 | vhost->disc_buf_dma); | |
4719 | free_login_buffer: | |
4720 | dma_free_coherent(dev, sizeof(*vhost->login_buf), | |
4721 | vhost->login_buf, vhost->login_buf_dma); | |
4722 | free_sg_pool: | |
4723 | dma_pool_destroy(vhost->sg_pool); | |
4724 | unmap_async_crq: | |
4725 | dma_unmap_single(dev, async_q->msg_token, | |
4726 | async_q->size * sizeof(*async_q->msgs), DMA_BIDIRECTIONAL); | |
4727 | free_async_crq: | |
4728 | free_page((unsigned long)async_q->msgs); | |
4729 | nomem: | |
4730 | LEAVE; | |
4731 | return -ENOMEM; | |
4732 | } | |
4733 | ||
43c8da90 BK |
4734 | /** |
4735 | * ibmvfc_rport_add_thread - Worker thread for rport adds | |
4736 | * @work: work struct | |
4737 | * | |
4738 | **/ | |
4739 | static void ibmvfc_rport_add_thread(struct work_struct *work) | |
4740 | { | |
4741 | struct ibmvfc_host *vhost = container_of(work, struct ibmvfc_host, | |
4742 | rport_add_work_q); | |
4743 | struct ibmvfc_target *tgt; | |
4744 | struct fc_rport *rport; | |
4745 | unsigned long flags; | |
4746 | int did_work; | |
4747 | ||
4748 | ENTER; | |
4749 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
4750 | do { | |
4751 | did_work = 0; | |
4752 | if (vhost->state != IBMVFC_ACTIVE) | |
4753 | break; | |
4754 | ||
4755 | list_for_each_entry(tgt, &vhost->targets, queue) { | |
4756 | if (tgt->add_rport) { | |
4757 | did_work = 1; | |
4758 | tgt->add_rport = 0; | |
4759 | kref_get(&tgt->kref); | |
4760 | rport = tgt->rport; | |
4761 | if (!rport) { | |
4762 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4763 | ibmvfc_tgt_add_rport(tgt); | |
4764 | } else if (get_device(&rport->dev)) { | |
4765 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4766 | tgt_dbg(tgt, "Setting rport roles\n"); | |
4767 | fc_remote_port_rolechg(rport, tgt->ids.roles); | |
4768 | put_device(&rport->dev); | |
4769 | } | |
4770 | ||
4771 | kref_put(&tgt->kref, ibmvfc_release_tgt); | |
4772 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
4773 | break; | |
4774 | } | |
4775 | } | |
4776 | } while(did_work); | |
4777 | ||
4778 | if (vhost->state == IBMVFC_ACTIVE) | |
4779 | vhost->scan_complete = 1; | |
4780 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4781 | LEAVE; | |
4782 | } | |
4783 | ||
072b91f9 BK |
4784 | /** |
4785 | * ibmvfc_probe - Adapter hot plug add entry point | |
4786 | * @vdev: vio device struct | |
4787 | * @id: vio device id struct | |
4788 | * | |
4789 | * Return value: | |
4790 | * 0 on success / non-zero on failure | |
4791 | **/ | |
4792 | static int ibmvfc_probe(struct vio_dev *vdev, const struct vio_device_id *id) | |
4793 | { | |
4794 | struct ibmvfc_host *vhost; | |
4795 | struct Scsi_Host *shost; | |
4796 | struct device *dev = &vdev->dev; | |
4797 | int rc = -ENOMEM; | |
4798 | ||
4799 | ENTER; | |
4800 | shost = scsi_host_alloc(&driver_template, sizeof(*vhost)); | |
4801 | if (!shost) { | |
4802 | dev_err(dev, "Couldn't allocate host data\n"); | |
4803 | goto out; | |
4804 | } | |
4805 | ||
4806 | shost->transportt = ibmvfc_transport_template; | |
4807 | shost->can_queue = max_requests; | |
4808 | shost->max_lun = max_lun; | |
4809 | shost->max_id = max_targets; | |
4810 | shost->max_sectors = IBMVFC_MAX_SECTORS; | |
4811 | shost->max_cmd_len = IBMVFC_MAX_CDB_LEN; | |
4812 | shost->unique_id = shost->host_no; | |
4813 | ||
4814 | vhost = shost_priv(shost); | |
4815 | INIT_LIST_HEAD(&vhost->sent); | |
4816 | INIT_LIST_HEAD(&vhost->free); | |
4817 | INIT_LIST_HEAD(&vhost->targets); | |
4818 | sprintf(vhost->name, IBMVFC_NAME); | |
4819 | vhost->host = shost; | |
4820 | vhost->dev = dev; | |
4821 | vhost->partition_number = -1; | |
4822 | vhost->log_level = log_level; | |
10501e1c | 4823 | vhost->task_set = 1; |
072b91f9 BK |
4824 | strcpy(vhost->partition_name, "UNKNOWN"); |
4825 | init_waitqueue_head(&vhost->work_wait_q); | |
4826 | init_waitqueue_head(&vhost->init_wait_q); | |
43c8da90 | 4827 | INIT_WORK(&vhost->rport_add_work_q, ibmvfc_rport_add_thread); |
d31429e1 | 4828 | mutex_init(&vhost->passthru_mutex); |
072b91f9 BK |
4829 | |
4830 | if ((rc = ibmvfc_alloc_mem(vhost))) | |
4831 | goto free_scsi_host; | |
4832 | ||
4833 | vhost->work_thread = kthread_run(ibmvfc_work, vhost, "%s_%d", IBMVFC_NAME, | |
4834 | shost->host_no); | |
4835 | ||
4836 | if (IS_ERR(vhost->work_thread)) { | |
4837 | dev_err(dev, "Couldn't create kernel thread: %ld\n", | |
4838 | PTR_ERR(vhost->work_thread)); | |
4839 | goto free_host_mem; | |
4840 | } | |
4841 | ||
4842 | if ((rc = ibmvfc_init_crq(vhost))) { | |
4843 | dev_err(dev, "Couldn't initialize crq. rc=%d\n", rc); | |
4844 | goto kill_kthread; | |
4845 | } | |
4846 | ||
4847 | if ((rc = ibmvfc_init_event_pool(vhost))) { | |
4848 | dev_err(dev, "Couldn't initialize event pool. rc=%d\n", rc); | |
4849 | goto release_crq; | |
4850 | } | |
4851 | ||
4852 | if ((rc = scsi_add_host(shost, dev))) | |
4853 | goto release_event_pool; | |
4854 | ||
a5110f29 MC |
4855 | fc_host_dev_loss_tmo(shost) = IBMVFC_DEV_LOSS_TMO; |
4856 | ||
072b91f9 BK |
4857 | if ((rc = ibmvfc_create_trace_file(&shost->shost_dev.kobj, |
4858 | &ibmvfc_trace_attr))) { | |
4859 | dev_err(dev, "Failed to create trace file. rc=%d\n", rc); | |
4860 | goto remove_shost; | |
4861 | } | |
4862 | ||
d31429e1 | 4863 | if (shost_to_fc_host(shost)->rqst_q) |
8a78362c | 4864 | blk_queue_max_segments(shost_to_fc_host(shost)->rqst_q, 1); |
072b91f9 BK |
4865 | dev_set_drvdata(dev, vhost); |
4866 | spin_lock(&ibmvfc_driver_lock); | |
4867 | list_add_tail(&vhost->queue, &ibmvfc_head); | |
4868 | spin_unlock(&ibmvfc_driver_lock); | |
4869 | ||
4870 | ibmvfc_send_crq_init(vhost); | |
4871 | scsi_scan_host(shost); | |
4872 | return 0; | |
4873 | ||
4874 | remove_shost: | |
4875 | scsi_remove_host(shost); | |
4876 | release_event_pool: | |
4877 | ibmvfc_free_event_pool(vhost); | |
4878 | release_crq: | |
4879 | ibmvfc_release_crq_queue(vhost); | |
4880 | kill_kthread: | |
4881 | kthread_stop(vhost->work_thread); | |
4882 | free_host_mem: | |
4883 | ibmvfc_free_mem(vhost); | |
4884 | free_scsi_host: | |
4885 | scsi_host_put(shost); | |
4886 | out: | |
4887 | LEAVE; | |
4888 | return rc; | |
4889 | } | |
4890 | ||
4891 | /** | |
4892 | * ibmvfc_remove - Adapter hot plug remove entry point | |
4893 | * @vdev: vio device struct | |
4894 | * | |
4895 | * Return value: | |
4896 | * 0 | |
4897 | **/ | |
4898 | static int ibmvfc_remove(struct vio_dev *vdev) | |
4899 | { | |
4900 | struct ibmvfc_host *vhost = dev_get_drvdata(&vdev->dev); | |
4901 | unsigned long flags; | |
4902 | ||
4903 | ENTER; | |
4904 | ibmvfc_remove_trace_file(&vhost->host->shost_dev.kobj, &ibmvfc_trace_attr); | |
70431105 BK |
4905 | |
4906 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
2d0da2a4 | 4907 | ibmvfc_link_down(vhost, IBMVFC_HOST_OFFLINE); |
70431105 BK |
4908 | spin_unlock_irqrestore(vhost->host->host_lock, flags); |
4909 | ||
2d0da2a4 BK |
4910 | ibmvfc_wait_while_resetting(vhost); |
4911 | ibmvfc_release_crq_queue(vhost); | |
072b91f9 BK |
4912 | kthread_stop(vhost->work_thread); |
4913 | fc_remove_host(vhost->host); | |
4914 | scsi_remove_host(vhost->host); | |
072b91f9 BK |
4915 | |
4916 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
4917 | ibmvfc_purge_requests(vhost, DID_ERROR); | |
4918 | ibmvfc_free_event_pool(vhost); | |
4919 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4920 | ||
4921 | ibmvfc_free_mem(vhost); | |
4922 | spin_lock(&ibmvfc_driver_lock); | |
4923 | list_del(&vhost->queue); | |
4924 | spin_unlock(&ibmvfc_driver_lock); | |
4925 | scsi_host_put(vhost->host); | |
4926 | LEAVE; | |
4927 | return 0; | |
4928 | } | |
4929 | ||
b0f4d4cf BK |
4930 | /** |
4931 | * ibmvfc_resume - Resume from suspend | |
4932 | * @dev: device struct | |
4933 | * | |
4934 | * We may have lost an interrupt across suspend/resume, so kick the | |
4935 | * interrupt handler | |
4936 | * | |
4937 | */ | |
4938 | static int ibmvfc_resume(struct device *dev) | |
4939 | { | |
4940 | unsigned long flags; | |
4941 | struct ibmvfc_host *vhost = dev_get_drvdata(dev); | |
4942 | struct vio_dev *vdev = to_vio_dev(dev); | |
4943 | ||
4944 | spin_lock_irqsave(vhost->host->host_lock, flags); | |
4945 | vio_disable_interrupts(vdev); | |
4946 | tasklet_schedule(&vhost->tasklet); | |
4947 | spin_unlock_irqrestore(vhost->host->host_lock, flags); | |
4948 | return 0; | |
4949 | } | |
4950 | ||
39c1ffec BK |
4951 | /** |
4952 | * ibmvfc_get_desired_dma - Calculate DMA resources needed by the driver | |
4953 | * @vdev: vio device struct | |
4954 | * | |
4955 | * Return value: | |
4956 | * Number of bytes the driver will need to DMA map at the same time in | |
4957 | * order to perform well. | |
4958 | */ | |
4959 | static unsigned long ibmvfc_get_desired_dma(struct vio_dev *vdev) | |
4960 | { | |
4961 | unsigned long pool_dma = max_requests * sizeof(union ibmvfc_iu); | |
4962 | return pool_dma + ((512 * 1024) * driver_template.cmd_per_lun); | |
4963 | } | |
4964 | ||
6f039790 | 4965 | static struct vio_device_id ibmvfc_device_table[] = { |
072b91f9 BK |
4966 | {"fcp", "IBM,vfc-client"}, |
4967 | { "", "" } | |
4968 | }; | |
4969 | MODULE_DEVICE_TABLE(vio, ibmvfc_device_table); | |
4970 | ||
b0f4d4cf BK |
4971 | static struct dev_pm_ops ibmvfc_pm_ops = { |
4972 | .resume = ibmvfc_resume | |
4973 | }; | |
4974 | ||
072b91f9 BK |
4975 | static struct vio_driver ibmvfc_driver = { |
4976 | .id_table = ibmvfc_device_table, | |
4977 | .probe = ibmvfc_probe, | |
4978 | .remove = ibmvfc_remove, | |
39c1ffec | 4979 | .get_desired_dma = ibmvfc_get_desired_dma, |
cb52d897 BH |
4980 | .name = IBMVFC_NAME, |
4981 | .pm = &ibmvfc_pm_ops, | |
072b91f9 BK |
4982 | }; |
4983 | ||
4984 | static struct fc_function_template ibmvfc_transport_functions = { | |
4985 | .show_host_fabric_name = 1, | |
4986 | .show_host_node_name = 1, | |
4987 | .show_host_port_name = 1, | |
4988 | .show_host_supported_classes = 1, | |
4989 | .show_host_port_type = 1, | |
4990 | .show_host_port_id = 1, | |
9ab3610f | 4991 | .show_host_maxframe_size = 1, |
072b91f9 BK |
4992 | |
4993 | .get_host_port_state = ibmvfc_get_host_port_state, | |
4994 | .show_host_port_state = 1, | |
4995 | ||
4996 | .get_host_speed = ibmvfc_get_host_speed, | |
4997 | .show_host_speed = 1, | |
4998 | ||
4999 | .issue_fc_host_lip = ibmvfc_issue_fc_host_lip, | |
5000 | .terminate_rport_io = ibmvfc_terminate_rport_io, | |
5001 | ||
5002 | .show_rport_maxframe_size = 1, | |
5003 | .show_rport_supported_classes = 1, | |
5004 | ||
5005 | .set_rport_dev_loss_tmo = ibmvfc_set_rport_dev_loss_tmo, | |
5006 | .show_rport_dev_loss_tmo = 1, | |
5007 | ||
5008 | .get_starget_node_name = ibmvfc_get_starget_node_name, | |
5009 | .show_starget_node_name = 1, | |
5010 | ||
5011 | .get_starget_port_name = ibmvfc_get_starget_port_name, | |
5012 | .show_starget_port_name = 1, | |
5013 | ||
5014 | .get_starget_port_id = ibmvfc_get_starget_port_id, | |
5015 | .show_starget_port_id = 1, | |
d31429e1 BK |
5016 | |
5017 | .bsg_request = ibmvfc_bsg_request, | |
5018 | .bsg_timeout = ibmvfc_bsg_timeout, | |
072b91f9 BK |
5019 | }; |
5020 | ||
5021 | /** | |
5022 | * ibmvfc_module_init - Initialize the ibmvfc module | |
5023 | * | |
5024 | * Return value: | |
5025 | * 0 on success / other on failure | |
5026 | **/ | |
5027 | static int __init ibmvfc_module_init(void) | |
5028 | { | |
5029 | int rc; | |
5030 | ||
5031 | if (!firmware_has_feature(FW_FEATURE_VIO)) | |
5032 | return -ENODEV; | |
5033 | ||
5034 | printk(KERN_INFO IBMVFC_NAME": IBM Virtual Fibre Channel Driver version: %s %s\n", | |
5035 | IBMVFC_DRIVER_VERSION, IBMVFC_DRIVER_DATE); | |
5036 | ||
5037 | ibmvfc_transport_template = fc_attach_transport(&ibmvfc_transport_functions); | |
5038 | if (!ibmvfc_transport_template) | |
5039 | return -ENOMEM; | |
5040 | ||
5041 | rc = vio_register_driver(&ibmvfc_driver); | |
5042 | if (rc) | |
5043 | fc_release_transport(ibmvfc_transport_template); | |
5044 | return rc; | |
5045 | } | |
5046 | ||
5047 | /** | |
5048 | * ibmvfc_module_exit - Teardown the ibmvfc module | |
5049 | * | |
5050 | * Return value: | |
5051 | * nothing | |
5052 | **/ | |
5053 | static void __exit ibmvfc_module_exit(void) | |
5054 | { | |
5055 | vio_unregister_driver(&ibmvfc_driver); | |
5056 | fc_release_transport(ibmvfc_transport_template); | |
5057 | } | |
5058 | ||
5059 | module_init(ibmvfc_module_init); | |
5060 | module_exit(ibmvfc_module_exit); |