Merge branch 'core-urgent-for-linus' of git://git.kernel.org/pub/scm/linux/kernel...
[deliverable/linux.git] / drivers / scsi / bnx2fc / bnx2fc_io.c
1 /* bnx2fc_io.c: QLogic Linux FCoE offload driver.
2 * IO manager and SCSI IO processing.
3 *
4 * Copyright (c) 2008-2013 Broadcom Corporation
5 * Copyright (c) 2014-2015 QLogic Corporation
6 *
7 * This program is free software; you can redistribute it and/or modify
8 * it under the terms of the GNU General Public License as published by
9 * the Free Software Foundation.
10 *
11 * Written by: Bhanu Prakash Gollapudi (bprakash@broadcom.com)
12 */
13
14 #include "bnx2fc.h"
15
16 #define RESERVE_FREE_LIST_INDEX num_possible_cpus()
17
18 static int bnx2fc_split_bd(struct bnx2fc_cmd *io_req, u64 addr, int sg_len,
19 int bd_index);
20 static int bnx2fc_map_sg(struct bnx2fc_cmd *io_req);
21 static int bnx2fc_build_bd_list_from_sg(struct bnx2fc_cmd *io_req);
22 static void bnx2fc_unmap_sg_list(struct bnx2fc_cmd *io_req);
23 static void bnx2fc_free_mp_resc(struct bnx2fc_cmd *io_req);
24 static void bnx2fc_parse_fcp_rsp(struct bnx2fc_cmd *io_req,
25 struct fcoe_fcp_rsp_payload *fcp_rsp,
26 u8 num_rq);
27
28 void bnx2fc_cmd_timer_set(struct bnx2fc_cmd *io_req,
29 unsigned int timer_msec)
30 {
31 struct bnx2fc_interface *interface = io_req->port->priv;
32
33 if (queue_delayed_work(interface->timer_work_queue,
34 &io_req->timeout_work,
35 msecs_to_jiffies(timer_msec)))
36 kref_get(&io_req->refcount);
37 }
38
39 static void bnx2fc_cmd_timeout(struct work_struct *work)
40 {
41 struct bnx2fc_cmd *io_req = container_of(work, struct bnx2fc_cmd,
42 timeout_work.work);
43 u8 cmd_type = io_req->cmd_type;
44 struct bnx2fc_rport *tgt = io_req->tgt;
45 int rc;
46
47 BNX2FC_IO_DBG(io_req, "cmd_timeout, cmd_type = %d,"
48 "req_flags = %lx\n", cmd_type, io_req->req_flags);
49
50 spin_lock_bh(&tgt->tgt_lock);
51 if (test_and_clear_bit(BNX2FC_FLAG_ISSUE_RRQ, &io_req->req_flags)) {
52 clear_bit(BNX2FC_FLAG_RETIRE_OXID, &io_req->req_flags);
53 /*
54 * ideally we should hold the io_req until RRQ complets,
55 * and release io_req from timeout hold.
56 */
57 spin_unlock_bh(&tgt->tgt_lock);
58 bnx2fc_send_rrq(io_req);
59 return;
60 }
61 if (test_and_clear_bit(BNX2FC_FLAG_RETIRE_OXID, &io_req->req_flags)) {
62 BNX2FC_IO_DBG(io_req, "IO ready for reuse now\n");
63 goto done;
64 }
65
66 switch (cmd_type) {
67 case BNX2FC_SCSI_CMD:
68 if (test_and_clear_bit(BNX2FC_FLAG_EH_ABORT,
69 &io_req->req_flags)) {
70 /* Handle eh_abort timeout */
71 BNX2FC_IO_DBG(io_req, "eh_abort timed out\n");
72 complete(&io_req->tm_done);
73 } else if (test_bit(BNX2FC_FLAG_ISSUE_ABTS,
74 &io_req->req_flags)) {
75 /* Handle internally generated ABTS timeout */
76 BNX2FC_IO_DBG(io_req, "ABTS timed out refcnt = %d\n",
77 io_req->refcount.refcount.counter);
78 if (!(test_and_set_bit(BNX2FC_FLAG_ABTS_DONE,
79 &io_req->req_flags))) {
80 /*
81 * Cleanup and return original command to
82 * mid-layer.
83 */
84 bnx2fc_initiate_cleanup(io_req);
85 kref_put(&io_req->refcount, bnx2fc_cmd_release);
86 spin_unlock_bh(&tgt->tgt_lock);
87
88 return;
89 }
90 } else {
91 /* Hanlde IO timeout */
92 BNX2FC_IO_DBG(io_req, "IO timed out. issue ABTS\n");
93 if (test_and_set_bit(BNX2FC_FLAG_IO_COMPL,
94 &io_req->req_flags)) {
95 BNX2FC_IO_DBG(io_req, "IO completed before "
96 " timer expiry\n");
97 goto done;
98 }
99
100 if (!test_and_set_bit(BNX2FC_FLAG_ISSUE_ABTS,
101 &io_req->req_flags)) {
102 rc = bnx2fc_initiate_abts(io_req);
103 if (rc == SUCCESS)
104 goto done;
105
106 kref_put(&io_req->refcount, bnx2fc_cmd_release);
107 spin_unlock_bh(&tgt->tgt_lock);
108
109 return;
110 } else {
111 BNX2FC_IO_DBG(io_req, "IO already in "
112 "ABTS processing\n");
113 }
114 }
115 break;
116 case BNX2FC_ELS:
117
118 if (test_bit(BNX2FC_FLAG_ISSUE_ABTS, &io_req->req_flags)) {
119 BNX2FC_IO_DBG(io_req, "ABTS for ELS timed out\n");
120
121 if (!test_and_set_bit(BNX2FC_FLAG_ABTS_DONE,
122 &io_req->req_flags)) {
123 kref_put(&io_req->refcount, bnx2fc_cmd_release);
124 spin_unlock_bh(&tgt->tgt_lock);
125
126 return;
127 }
128 } else {
129 /*
130 * Handle ELS timeout.
131 * tgt_lock is used to sync compl path and timeout
132 * path. If els compl path is processing this IO, we
133 * have nothing to do here, just release the timer hold
134 */
135 BNX2FC_IO_DBG(io_req, "ELS timed out\n");
136 if (test_and_set_bit(BNX2FC_FLAG_ELS_DONE,
137 &io_req->req_flags))
138 goto done;
139
140 /* Indicate the cb_func that this ELS is timed out */
141 set_bit(BNX2FC_FLAG_ELS_TIMEOUT, &io_req->req_flags);
142
143 if ((io_req->cb_func) && (io_req->cb_arg)) {
144 io_req->cb_func(io_req->cb_arg);
145 io_req->cb_arg = NULL;
146 }
147 }
148 break;
149 default:
150 printk(KERN_ERR PFX "cmd_timeout: invalid cmd_type %d\n",
151 cmd_type);
152 break;
153 }
154
155 done:
156 /* release the cmd that was held when timer was set */
157 kref_put(&io_req->refcount, bnx2fc_cmd_release);
158 spin_unlock_bh(&tgt->tgt_lock);
159 }
160
161 static void bnx2fc_scsi_done(struct bnx2fc_cmd *io_req, int err_code)
162 {
163 /* Called with host lock held */
164 struct scsi_cmnd *sc_cmd = io_req->sc_cmd;
165
166 /*
167 * active_cmd_queue may have other command types as well,
168 * and during flush operation, we want to error back only
169 * scsi commands.
170 */
171 if (io_req->cmd_type != BNX2FC_SCSI_CMD)
172 return;
173
174 BNX2FC_IO_DBG(io_req, "scsi_done. err_code = 0x%x\n", err_code);
175 if (test_bit(BNX2FC_FLAG_CMD_LOST, &io_req->req_flags)) {
176 /* Do not call scsi done for this IO */
177 return;
178 }
179
180 bnx2fc_unmap_sg_list(io_req);
181 io_req->sc_cmd = NULL;
182
183 /* Sanity checks before returning command to mid-layer */
184 if (!sc_cmd) {
185 printk(KERN_ERR PFX "scsi_done - sc_cmd NULL. "
186 "IO(0x%x) already cleaned up\n",
187 io_req->xid);
188 return;
189 }
190 if (!sc_cmd->device) {
191 pr_err(PFX "0x%x: sc_cmd->device is NULL.\n", io_req->xid);
192 return;
193 }
194 if (!sc_cmd->device->host) {
195 pr_err(PFX "0x%x: sc_cmd->device->host is NULL.\n",
196 io_req->xid);
197 return;
198 }
199
200 sc_cmd->result = err_code << 16;
201
202 BNX2FC_IO_DBG(io_req, "sc=%p, result=0x%x, retries=%d, allowed=%d\n",
203 sc_cmd, host_byte(sc_cmd->result), sc_cmd->retries,
204 sc_cmd->allowed);
205 scsi_set_resid(sc_cmd, scsi_bufflen(sc_cmd));
206 sc_cmd->SCp.ptr = NULL;
207 sc_cmd->scsi_done(sc_cmd);
208 }
209
210 struct bnx2fc_cmd_mgr *bnx2fc_cmd_mgr_alloc(struct bnx2fc_hba *hba)
211 {
212 struct bnx2fc_cmd_mgr *cmgr;
213 struct io_bdt *bdt_info;
214 struct bnx2fc_cmd *io_req;
215 size_t len;
216 u32 mem_size;
217 u16 xid;
218 int i;
219 int num_ios, num_pri_ios;
220 size_t bd_tbl_sz;
221 int arr_sz = num_possible_cpus() + 1;
222 u16 min_xid = BNX2FC_MIN_XID;
223 u16 max_xid = hba->max_xid;
224
225 if (max_xid <= min_xid || max_xid == FC_XID_UNKNOWN) {
226 printk(KERN_ERR PFX "cmd_mgr_alloc: Invalid min_xid 0x%x \
227 and max_xid 0x%x\n", min_xid, max_xid);
228 return NULL;
229 }
230 BNX2FC_MISC_DBG("min xid 0x%x, max xid 0x%x\n", min_xid, max_xid);
231
232 num_ios = max_xid - min_xid + 1;
233 len = (num_ios * (sizeof(struct bnx2fc_cmd *)));
234 len += sizeof(struct bnx2fc_cmd_mgr);
235
236 cmgr = kzalloc(len, GFP_KERNEL);
237 if (!cmgr) {
238 printk(KERN_ERR PFX "failed to alloc cmgr\n");
239 return NULL;
240 }
241
242 cmgr->free_list = kzalloc(sizeof(*cmgr->free_list) *
243 arr_sz, GFP_KERNEL);
244 if (!cmgr->free_list) {
245 printk(KERN_ERR PFX "failed to alloc free_list\n");
246 goto mem_err;
247 }
248
249 cmgr->free_list_lock = kzalloc(sizeof(*cmgr->free_list_lock) *
250 arr_sz, GFP_KERNEL);
251 if (!cmgr->free_list_lock) {
252 printk(KERN_ERR PFX "failed to alloc free_list_lock\n");
253 kfree(cmgr->free_list);
254 cmgr->free_list = NULL;
255 goto mem_err;
256 }
257
258 cmgr->hba = hba;
259 cmgr->cmds = (struct bnx2fc_cmd **)(cmgr + 1);
260
261 for (i = 0; i < arr_sz; i++) {
262 INIT_LIST_HEAD(&cmgr->free_list[i]);
263 spin_lock_init(&cmgr->free_list_lock[i]);
264 }
265
266 /*
267 * Pre-allocated pool of bnx2fc_cmds.
268 * Last entry in the free list array is the free list
269 * of slow path requests.
270 */
271 xid = BNX2FC_MIN_XID;
272 num_pri_ios = num_ios - hba->elstm_xids;
273 for (i = 0; i < num_ios; i++) {
274 io_req = kzalloc(sizeof(*io_req), GFP_KERNEL);
275
276 if (!io_req) {
277 printk(KERN_ERR PFX "failed to alloc io_req\n");
278 goto mem_err;
279 }
280
281 INIT_LIST_HEAD(&io_req->link);
282 INIT_DELAYED_WORK(&io_req->timeout_work, bnx2fc_cmd_timeout);
283
284 io_req->xid = xid++;
285 if (i < num_pri_ios)
286 list_add_tail(&io_req->link,
287 &cmgr->free_list[io_req->xid %
288 num_possible_cpus()]);
289 else
290 list_add_tail(&io_req->link,
291 &cmgr->free_list[num_possible_cpus()]);
292 io_req++;
293 }
294
295 /* Allocate pool of io_bdts - one for each bnx2fc_cmd */
296 mem_size = num_ios * sizeof(struct io_bdt *);
297 cmgr->io_bdt_pool = kmalloc(mem_size, GFP_KERNEL);
298 if (!cmgr->io_bdt_pool) {
299 printk(KERN_ERR PFX "failed to alloc io_bdt_pool\n");
300 goto mem_err;
301 }
302
303 mem_size = sizeof(struct io_bdt);
304 for (i = 0; i < num_ios; i++) {
305 cmgr->io_bdt_pool[i] = kmalloc(mem_size, GFP_KERNEL);
306 if (!cmgr->io_bdt_pool[i]) {
307 printk(KERN_ERR PFX "failed to alloc "
308 "io_bdt_pool[%d]\n", i);
309 goto mem_err;
310 }
311 }
312
313 /* Allocate an map fcoe_bdt_ctx structures */
314 bd_tbl_sz = BNX2FC_MAX_BDS_PER_CMD * sizeof(struct fcoe_bd_ctx);
315 for (i = 0; i < num_ios; i++) {
316 bdt_info = cmgr->io_bdt_pool[i];
317 bdt_info->bd_tbl = dma_alloc_coherent(&hba->pcidev->dev,
318 bd_tbl_sz,
319 &bdt_info->bd_tbl_dma,
320 GFP_KERNEL);
321 if (!bdt_info->bd_tbl) {
322 printk(KERN_ERR PFX "failed to alloc "
323 "bdt_tbl[%d]\n", i);
324 goto mem_err;
325 }
326 }
327
328 return cmgr;
329
330 mem_err:
331 bnx2fc_cmd_mgr_free(cmgr);
332 return NULL;
333 }
334
335 void bnx2fc_cmd_mgr_free(struct bnx2fc_cmd_mgr *cmgr)
336 {
337 struct io_bdt *bdt_info;
338 struct bnx2fc_hba *hba = cmgr->hba;
339 size_t bd_tbl_sz;
340 u16 min_xid = BNX2FC_MIN_XID;
341 u16 max_xid = hba->max_xid;
342 int num_ios;
343 int i;
344
345 num_ios = max_xid - min_xid + 1;
346
347 /* Free fcoe_bdt_ctx structures */
348 if (!cmgr->io_bdt_pool)
349 goto free_cmd_pool;
350
351 bd_tbl_sz = BNX2FC_MAX_BDS_PER_CMD * sizeof(struct fcoe_bd_ctx);
352 for (i = 0; i < num_ios; i++) {
353 bdt_info = cmgr->io_bdt_pool[i];
354 if (bdt_info->bd_tbl) {
355 dma_free_coherent(&hba->pcidev->dev, bd_tbl_sz,
356 bdt_info->bd_tbl,
357 bdt_info->bd_tbl_dma);
358 bdt_info->bd_tbl = NULL;
359 }
360 }
361
362 /* Destroy io_bdt pool */
363 for (i = 0; i < num_ios; i++) {
364 kfree(cmgr->io_bdt_pool[i]);
365 cmgr->io_bdt_pool[i] = NULL;
366 }
367
368 kfree(cmgr->io_bdt_pool);
369 cmgr->io_bdt_pool = NULL;
370
371 free_cmd_pool:
372 kfree(cmgr->free_list_lock);
373
374 /* Destroy cmd pool */
375 if (!cmgr->free_list)
376 goto free_cmgr;
377
378 for (i = 0; i < num_possible_cpus() + 1; i++) {
379 struct bnx2fc_cmd *tmp, *io_req;
380
381 list_for_each_entry_safe(io_req, tmp,
382 &cmgr->free_list[i], link) {
383 list_del(&io_req->link);
384 kfree(io_req);
385 }
386 }
387 kfree(cmgr->free_list);
388 free_cmgr:
389 /* Free command manager itself */
390 kfree(cmgr);
391 }
392
393 struct bnx2fc_cmd *bnx2fc_elstm_alloc(struct bnx2fc_rport *tgt, int type)
394 {
395 struct fcoe_port *port = tgt->port;
396 struct bnx2fc_interface *interface = port->priv;
397 struct bnx2fc_cmd_mgr *cmd_mgr = interface->hba->cmd_mgr;
398 struct bnx2fc_cmd *io_req;
399 struct list_head *listp;
400 struct io_bdt *bd_tbl;
401 int index = RESERVE_FREE_LIST_INDEX;
402 u32 free_sqes;
403 u32 max_sqes;
404 u16 xid;
405
406 max_sqes = tgt->max_sqes;
407 switch (type) {
408 case BNX2FC_TASK_MGMT_CMD:
409 max_sqes = BNX2FC_TM_MAX_SQES;
410 break;
411 case BNX2FC_ELS:
412 max_sqes = BNX2FC_ELS_MAX_SQES;
413 break;
414 default:
415 break;
416 }
417
418 /*
419 * NOTE: Free list insertions and deletions are protected with
420 * cmgr lock
421 */
422 spin_lock_bh(&cmd_mgr->free_list_lock[index]);
423 free_sqes = atomic_read(&tgt->free_sqes);
424 if ((list_empty(&(cmd_mgr->free_list[index]))) ||
425 (tgt->num_active_ios.counter >= max_sqes) ||
426 (free_sqes + max_sqes <= BNX2FC_SQ_WQES_MAX)) {
427 BNX2FC_TGT_DBG(tgt, "No free els_tm cmds available "
428 "ios(%d):sqes(%d)\n",
429 tgt->num_active_ios.counter, tgt->max_sqes);
430 if (list_empty(&(cmd_mgr->free_list[index])))
431 printk(KERN_ERR PFX "elstm_alloc: list_empty\n");
432 spin_unlock_bh(&cmd_mgr->free_list_lock[index]);
433 return NULL;
434 }
435
436 listp = (struct list_head *)
437 cmd_mgr->free_list[index].next;
438 list_del_init(listp);
439 io_req = (struct bnx2fc_cmd *) listp;
440 xid = io_req->xid;
441 cmd_mgr->cmds[xid] = io_req;
442 atomic_inc(&tgt->num_active_ios);
443 atomic_dec(&tgt->free_sqes);
444 spin_unlock_bh(&cmd_mgr->free_list_lock[index]);
445
446 INIT_LIST_HEAD(&io_req->link);
447
448 io_req->port = port;
449 io_req->cmd_mgr = cmd_mgr;
450 io_req->req_flags = 0;
451 io_req->cmd_type = type;
452
453 /* Bind io_bdt for this io_req */
454 /* Have a static link between io_req and io_bdt_pool */
455 bd_tbl = io_req->bd_tbl = cmd_mgr->io_bdt_pool[xid];
456 bd_tbl->io_req = io_req;
457
458 /* Hold the io_req against deletion */
459 kref_init(&io_req->refcount);
460 return io_req;
461 }
462
463 struct bnx2fc_cmd *bnx2fc_cmd_alloc(struct bnx2fc_rport *tgt)
464 {
465 struct fcoe_port *port = tgt->port;
466 struct bnx2fc_interface *interface = port->priv;
467 struct bnx2fc_cmd_mgr *cmd_mgr = interface->hba->cmd_mgr;
468 struct bnx2fc_cmd *io_req;
469 struct list_head *listp;
470 struct io_bdt *bd_tbl;
471 u32 free_sqes;
472 u32 max_sqes;
473 u16 xid;
474 int index = get_cpu();
475
476 max_sqes = BNX2FC_SCSI_MAX_SQES;
477 /*
478 * NOTE: Free list insertions and deletions are protected with
479 * cmgr lock
480 */
481 spin_lock_bh(&cmd_mgr->free_list_lock[index]);
482 free_sqes = atomic_read(&tgt->free_sqes);
483 if ((list_empty(&cmd_mgr->free_list[index])) ||
484 (tgt->num_active_ios.counter >= max_sqes) ||
485 (free_sqes + max_sqes <= BNX2FC_SQ_WQES_MAX)) {
486 spin_unlock_bh(&cmd_mgr->free_list_lock[index]);
487 put_cpu();
488 return NULL;
489 }
490
491 listp = (struct list_head *)
492 cmd_mgr->free_list[index].next;
493 list_del_init(listp);
494 io_req = (struct bnx2fc_cmd *) listp;
495 xid = io_req->xid;
496 cmd_mgr->cmds[xid] = io_req;
497 atomic_inc(&tgt->num_active_ios);
498 atomic_dec(&tgt->free_sqes);
499 spin_unlock_bh(&cmd_mgr->free_list_lock[index]);
500 put_cpu();
501
502 INIT_LIST_HEAD(&io_req->link);
503
504 io_req->port = port;
505 io_req->cmd_mgr = cmd_mgr;
506 io_req->req_flags = 0;
507
508 /* Bind io_bdt for this io_req */
509 /* Have a static link between io_req and io_bdt_pool */
510 bd_tbl = io_req->bd_tbl = cmd_mgr->io_bdt_pool[xid];
511 bd_tbl->io_req = io_req;
512
513 /* Hold the io_req against deletion */
514 kref_init(&io_req->refcount);
515 return io_req;
516 }
517
518 void bnx2fc_cmd_release(struct kref *ref)
519 {
520 struct bnx2fc_cmd *io_req = container_of(ref,
521 struct bnx2fc_cmd, refcount);
522 struct bnx2fc_cmd_mgr *cmd_mgr = io_req->cmd_mgr;
523 int index;
524
525 if (io_req->cmd_type == BNX2FC_SCSI_CMD)
526 index = io_req->xid % num_possible_cpus();
527 else
528 index = RESERVE_FREE_LIST_INDEX;
529
530
531 spin_lock_bh(&cmd_mgr->free_list_lock[index]);
532 if (io_req->cmd_type != BNX2FC_SCSI_CMD)
533 bnx2fc_free_mp_resc(io_req);
534 cmd_mgr->cmds[io_req->xid] = NULL;
535 /* Delete IO from retire queue */
536 list_del_init(&io_req->link);
537 /* Add it to the free list */
538 list_add(&io_req->link,
539 &cmd_mgr->free_list[index]);
540 atomic_dec(&io_req->tgt->num_active_ios);
541 spin_unlock_bh(&cmd_mgr->free_list_lock[index]);
542
543 }
544
545 static void bnx2fc_free_mp_resc(struct bnx2fc_cmd *io_req)
546 {
547 struct bnx2fc_mp_req *mp_req = &(io_req->mp_req);
548 struct bnx2fc_interface *interface = io_req->port->priv;
549 struct bnx2fc_hba *hba = interface->hba;
550 size_t sz = sizeof(struct fcoe_bd_ctx);
551
552 /* clear tm flags */
553 mp_req->tm_flags = 0;
554 if (mp_req->mp_req_bd) {
555 dma_free_coherent(&hba->pcidev->dev, sz,
556 mp_req->mp_req_bd,
557 mp_req->mp_req_bd_dma);
558 mp_req->mp_req_bd = NULL;
559 }
560 if (mp_req->mp_resp_bd) {
561 dma_free_coherent(&hba->pcidev->dev, sz,
562 mp_req->mp_resp_bd,
563 mp_req->mp_resp_bd_dma);
564 mp_req->mp_resp_bd = NULL;
565 }
566 if (mp_req->req_buf) {
567 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
568 mp_req->req_buf,
569 mp_req->req_buf_dma);
570 mp_req->req_buf = NULL;
571 }
572 if (mp_req->resp_buf) {
573 dma_free_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
574 mp_req->resp_buf,
575 mp_req->resp_buf_dma);
576 mp_req->resp_buf = NULL;
577 }
578 }
579
580 int bnx2fc_init_mp_req(struct bnx2fc_cmd *io_req)
581 {
582 struct bnx2fc_mp_req *mp_req;
583 struct fcoe_bd_ctx *mp_req_bd;
584 struct fcoe_bd_ctx *mp_resp_bd;
585 struct bnx2fc_interface *interface = io_req->port->priv;
586 struct bnx2fc_hba *hba = interface->hba;
587 dma_addr_t addr;
588 size_t sz;
589
590 mp_req = (struct bnx2fc_mp_req *)&(io_req->mp_req);
591 memset(mp_req, 0, sizeof(struct bnx2fc_mp_req));
592
593 if (io_req->cmd_type != BNX2FC_ELS) {
594 mp_req->req_len = sizeof(struct fcp_cmnd);
595 io_req->data_xfer_len = mp_req->req_len;
596 } else
597 mp_req->req_len = io_req->data_xfer_len;
598
599 mp_req->req_buf = dma_alloc_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
600 &mp_req->req_buf_dma,
601 GFP_ATOMIC);
602 if (!mp_req->req_buf) {
603 printk(KERN_ERR PFX "unable to alloc MP req buffer\n");
604 bnx2fc_free_mp_resc(io_req);
605 return FAILED;
606 }
607
608 mp_req->resp_buf = dma_alloc_coherent(&hba->pcidev->dev, CNIC_PAGE_SIZE,
609 &mp_req->resp_buf_dma,
610 GFP_ATOMIC);
611 if (!mp_req->resp_buf) {
612 printk(KERN_ERR PFX "unable to alloc TM resp buffer\n");
613 bnx2fc_free_mp_resc(io_req);
614 return FAILED;
615 }
616 memset(mp_req->req_buf, 0, CNIC_PAGE_SIZE);
617 memset(mp_req->resp_buf, 0, CNIC_PAGE_SIZE);
618
619 /* Allocate and map mp_req_bd and mp_resp_bd */
620 sz = sizeof(struct fcoe_bd_ctx);
621 mp_req->mp_req_bd = dma_alloc_coherent(&hba->pcidev->dev, sz,
622 &mp_req->mp_req_bd_dma,
623 GFP_ATOMIC);
624 if (!mp_req->mp_req_bd) {
625 printk(KERN_ERR PFX "unable to alloc MP req bd\n");
626 bnx2fc_free_mp_resc(io_req);
627 return FAILED;
628 }
629 mp_req->mp_resp_bd = dma_alloc_coherent(&hba->pcidev->dev, sz,
630 &mp_req->mp_resp_bd_dma,
631 GFP_ATOMIC);
632 if (!mp_req->mp_resp_bd) {
633 printk(KERN_ERR PFX "unable to alloc MP resp bd\n");
634 bnx2fc_free_mp_resc(io_req);
635 return FAILED;
636 }
637 /* Fill bd table */
638 addr = mp_req->req_buf_dma;
639 mp_req_bd = mp_req->mp_req_bd;
640 mp_req_bd->buf_addr_lo = (u32)addr & 0xffffffff;
641 mp_req_bd->buf_addr_hi = (u32)((u64)addr >> 32);
642 mp_req_bd->buf_len = CNIC_PAGE_SIZE;
643 mp_req_bd->flags = 0;
644
645 /*
646 * MP buffer is either a task mgmt command or an ELS.
647 * So the assumption is that it consumes a single bd
648 * entry in the bd table
649 */
650 mp_resp_bd = mp_req->mp_resp_bd;
651 addr = mp_req->resp_buf_dma;
652 mp_resp_bd->buf_addr_lo = (u32)addr & 0xffffffff;
653 mp_resp_bd->buf_addr_hi = (u32)((u64)addr >> 32);
654 mp_resp_bd->buf_len = CNIC_PAGE_SIZE;
655 mp_resp_bd->flags = 0;
656
657 return SUCCESS;
658 }
659
660 static int bnx2fc_initiate_tmf(struct scsi_cmnd *sc_cmd, u8 tm_flags)
661 {
662 struct fc_lport *lport;
663 struct fc_rport *rport;
664 struct fc_rport_libfc_priv *rp;
665 struct fcoe_port *port;
666 struct bnx2fc_interface *interface;
667 struct bnx2fc_rport *tgt;
668 struct bnx2fc_cmd *io_req;
669 struct bnx2fc_mp_req *tm_req;
670 struct fcoe_task_ctx_entry *task;
671 struct fcoe_task_ctx_entry *task_page;
672 struct Scsi_Host *host = sc_cmd->device->host;
673 struct fc_frame_header *fc_hdr;
674 struct fcp_cmnd *fcp_cmnd;
675 int task_idx, index;
676 int rc = SUCCESS;
677 u16 xid;
678 u32 sid, did;
679 unsigned long start = jiffies;
680
681 lport = shost_priv(host);
682 rport = starget_to_rport(scsi_target(sc_cmd->device));
683 port = lport_priv(lport);
684 interface = port->priv;
685
686 if (rport == NULL) {
687 printk(KERN_ERR PFX "device_reset: rport is NULL\n");
688 rc = FAILED;
689 goto tmf_err;
690 }
691 rp = rport->dd_data;
692
693 rc = fc_block_scsi_eh(sc_cmd);
694 if (rc)
695 return rc;
696
697 if (lport->state != LPORT_ST_READY || !(lport->link_up)) {
698 printk(KERN_ERR PFX "device_reset: link is not ready\n");
699 rc = FAILED;
700 goto tmf_err;
701 }
702 /* rport and tgt are allocated together, so tgt should be non-NULL */
703 tgt = (struct bnx2fc_rport *)&rp[1];
704
705 if (!(test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags))) {
706 printk(KERN_ERR PFX "device_reset: tgt not offloaded\n");
707 rc = FAILED;
708 goto tmf_err;
709 }
710 retry_tmf:
711 io_req = bnx2fc_elstm_alloc(tgt, BNX2FC_TASK_MGMT_CMD);
712 if (!io_req) {
713 if (time_after(jiffies, start + HZ)) {
714 printk(KERN_ERR PFX "tmf: Failed TMF");
715 rc = FAILED;
716 goto tmf_err;
717 }
718 msleep(20);
719 goto retry_tmf;
720 }
721 /* Initialize rest of io_req fields */
722 io_req->sc_cmd = sc_cmd;
723 io_req->port = port;
724 io_req->tgt = tgt;
725
726 tm_req = (struct bnx2fc_mp_req *)&(io_req->mp_req);
727
728 rc = bnx2fc_init_mp_req(io_req);
729 if (rc == FAILED) {
730 printk(KERN_ERR PFX "Task mgmt MP request init failed\n");
731 spin_lock_bh(&tgt->tgt_lock);
732 kref_put(&io_req->refcount, bnx2fc_cmd_release);
733 spin_unlock_bh(&tgt->tgt_lock);
734 goto tmf_err;
735 }
736
737 /* Set TM flags */
738 io_req->io_req_flags = 0;
739 tm_req->tm_flags = tm_flags;
740
741 /* Fill FCP_CMND */
742 bnx2fc_build_fcp_cmnd(io_req, (struct fcp_cmnd *)tm_req->req_buf);
743 fcp_cmnd = (struct fcp_cmnd *)tm_req->req_buf;
744 memset(fcp_cmnd->fc_cdb, 0, sc_cmd->cmd_len);
745 fcp_cmnd->fc_dl = 0;
746
747 /* Fill FC header */
748 fc_hdr = &(tm_req->req_fc_hdr);
749 sid = tgt->sid;
750 did = rport->port_id;
751 __fc_fill_fc_hdr(fc_hdr, FC_RCTL_DD_UNSOL_CMD, did, sid,
752 FC_TYPE_FCP, FC_FC_FIRST_SEQ | FC_FC_END_SEQ |
753 FC_FC_SEQ_INIT, 0);
754 /* Obtain exchange id */
755 xid = io_req->xid;
756
757 BNX2FC_TGT_DBG(tgt, "Initiate TMF - xid = 0x%x\n", xid);
758 task_idx = xid/BNX2FC_TASKS_PER_PAGE;
759 index = xid % BNX2FC_TASKS_PER_PAGE;
760
761 /* Initialize task context for this IO request */
762 task_page = (struct fcoe_task_ctx_entry *)
763 interface->hba->task_ctx[task_idx];
764 task = &(task_page[index]);
765 bnx2fc_init_mp_task(io_req, task);
766
767 sc_cmd->SCp.ptr = (char *)io_req;
768
769 /* Obtain free SQ entry */
770 spin_lock_bh(&tgt->tgt_lock);
771 bnx2fc_add_2_sq(tgt, xid);
772
773 /* Enqueue the io_req to active_tm_queue */
774 io_req->on_tmf_queue = 1;
775 list_add_tail(&io_req->link, &tgt->active_tm_queue);
776
777 init_completion(&io_req->tm_done);
778 io_req->wait_for_comp = 1;
779
780 /* Ring doorbell */
781 bnx2fc_ring_doorbell(tgt);
782 spin_unlock_bh(&tgt->tgt_lock);
783
784 rc = wait_for_completion_timeout(&io_req->tm_done,
785 interface->tm_timeout * HZ);
786 spin_lock_bh(&tgt->tgt_lock);
787
788 io_req->wait_for_comp = 0;
789 if (!(test_bit(BNX2FC_FLAG_TM_COMPL, &io_req->req_flags))) {
790 set_bit(BNX2FC_FLAG_TM_TIMEOUT, &io_req->req_flags);
791 if (io_req->on_tmf_queue) {
792 list_del_init(&io_req->link);
793 io_req->on_tmf_queue = 0;
794 }
795 io_req->wait_for_comp = 1;
796 bnx2fc_initiate_cleanup(io_req);
797 spin_unlock_bh(&tgt->tgt_lock);
798 rc = wait_for_completion_timeout(&io_req->tm_done,
799 BNX2FC_FW_TIMEOUT);
800 spin_lock_bh(&tgt->tgt_lock);
801 io_req->wait_for_comp = 0;
802 if (!rc)
803 kref_put(&io_req->refcount, bnx2fc_cmd_release);
804 }
805
806 spin_unlock_bh(&tgt->tgt_lock);
807
808 if (!rc) {
809 BNX2FC_TGT_DBG(tgt, "task mgmt command failed...\n");
810 rc = FAILED;
811 } else {
812 BNX2FC_TGT_DBG(tgt, "task mgmt command success...\n");
813 rc = SUCCESS;
814 }
815 tmf_err:
816 return rc;
817 }
818
819 int bnx2fc_initiate_abts(struct bnx2fc_cmd *io_req)
820 {
821 struct fc_lport *lport;
822 struct bnx2fc_rport *tgt = io_req->tgt;
823 struct fc_rport *rport = tgt->rport;
824 struct fc_rport_priv *rdata = tgt->rdata;
825 struct bnx2fc_interface *interface;
826 struct fcoe_port *port;
827 struct bnx2fc_cmd *abts_io_req;
828 struct fcoe_task_ctx_entry *task;
829 struct fcoe_task_ctx_entry *task_page;
830 struct fc_frame_header *fc_hdr;
831 struct bnx2fc_mp_req *abts_req;
832 int task_idx, index;
833 u32 sid, did;
834 u16 xid;
835 int rc = SUCCESS;
836 u32 r_a_tov = rdata->r_a_tov;
837
838 /* called with tgt_lock held */
839 BNX2FC_IO_DBG(io_req, "Entered bnx2fc_initiate_abts\n");
840
841 port = io_req->port;
842 interface = port->priv;
843 lport = port->lport;
844
845 if (!test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags)) {
846 printk(KERN_ERR PFX "initiate_abts: tgt not offloaded\n");
847 rc = FAILED;
848 goto abts_err;
849 }
850
851 if (rport == NULL) {
852 printk(KERN_ERR PFX "initiate_abts: rport is NULL\n");
853 rc = FAILED;
854 goto abts_err;
855 }
856
857 if (lport->state != LPORT_ST_READY || !(lport->link_up)) {
858 printk(KERN_ERR PFX "initiate_abts: link is not ready\n");
859 rc = FAILED;
860 goto abts_err;
861 }
862
863 abts_io_req = bnx2fc_elstm_alloc(tgt, BNX2FC_ABTS);
864 if (!abts_io_req) {
865 printk(KERN_ERR PFX "abts: couldnt allocate cmd\n");
866 rc = FAILED;
867 goto abts_err;
868 }
869
870 /* Initialize rest of io_req fields */
871 abts_io_req->sc_cmd = NULL;
872 abts_io_req->port = port;
873 abts_io_req->tgt = tgt;
874 abts_io_req->data_xfer_len = 0; /* No data transfer for ABTS */
875
876 abts_req = (struct bnx2fc_mp_req *)&(abts_io_req->mp_req);
877 memset(abts_req, 0, sizeof(struct bnx2fc_mp_req));
878
879 /* Fill FC header */
880 fc_hdr = &(abts_req->req_fc_hdr);
881
882 /* Obtain oxid and rxid for the original exchange to be aborted */
883 fc_hdr->fh_ox_id = htons(io_req->xid);
884 fc_hdr->fh_rx_id = htons(io_req->task->rxwr_txrd.var_ctx.rx_id);
885
886 sid = tgt->sid;
887 did = rport->port_id;
888
889 __fc_fill_fc_hdr(fc_hdr, FC_RCTL_BA_ABTS, did, sid,
890 FC_TYPE_BLS, FC_FC_FIRST_SEQ | FC_FC_END_SEQ |
891 FC_FC_SEQ_INIT, 0);
892
893 xid = abts_io_req->xid;
894 BNX2FC_IO_DBG(abts_io_req, "ABTS io_req\n");
895 task_idx = xid/BNX2FC_TASKS_PER_PAGE;
896 index = xid % BNX2FC_TASKS_PER_PAGE;
897
898 /* Initialize task context for this IO request */
899 task_page = (struct fcoe_task_ctx_entry *)
900 interface->hba->task_ctx[task_idx];
901 task = &(task_page[index]);
902 bnx2fc_init_mp_task(abts_io_req, task);
903
904 /*
905 * ABTS task is a temporary task that will be cleaned up
906 * irrespective of ABTS response. We need to start the timer
907 * for the original exchange, as the CQE is posted for the original
908 * IO request.
909 *
910 * Timer for ABTS is started only when it is originated by a
911 * TM request. For the ABTS issued as part of ULP timeout,
912 * scsi-ml maintains the timers.
913 */
914
915 /* if (test_bit(BNX2FC_FLAG_ISSUE_ABTS, &io_req->req_flags))*/
916 bnx2fc_cmd_timer_set(io_req, 2 * r_a_tov);
917
918 /* Obtain free SQ entry */
919 bnx2fc_add_2_sq(tgt, xid);
920
921 /* Ring doorbell */
922 bnx2fc_ring_doorbell(tgt);
923
924 abts_err:
925 return rc;
926 }
927
928 int bnx2fc_initiate_seq_cleanup(struct bnx2fc_cmd *orig_io_req, u32 offset,
929 enum fc_rctl r_ctl)
930 {
931 struct fc_lport *lport;
932 struct bnx2fc_rport *tgt = orig_io_req->tgt;
933 struct bnx2fc_interface *interface;
934 struct fcoe_port *port;
935 struct bnx2fc_cmd *seq_clnp_req;
936 struct fcoe_task_ctx_entry *task;
937 struct fcoe_task_ctx_entry *task_page;
938 struct bnx2fc_els_cb_arg *cb_arg = NULL;
939 int task_idx, index;
940 u16 xid;
941 int rc = 0;
942
943 BNX2FC_IO_DBG(orig_io_req, "bnx2fc_initiate_seq_cleanup xid = 0x%x\n",
944 orig_io_req->xid);
945 kref_get(&orig_io_req->refcount);
946
947 port = orig_io_req->port;
948 interface = port->priv;
949 lport = port->lport;
950
951 cb_arg = kzalloc(sizeof(struct bnx2fc_els_cb_arg), GFP_ATOMIC);
952 if (!cb_arg) {
953 printk(KERN_ERR PFX "Unable to alloc cb_arg for seq clnup\n");
954 rc = -ENOMEM;
955 goto cleanup_err;
956 }
957
958 seq_clnp_req = bnx2fc_elstm_alloc(tgt, BNX2FC_SEQ_CLEANUP);
959 if (!seq_clnp_req) {
960 printk(KERN_ERR PFX "cleanup: couldnt allocate cmd\n");
961 rc = -ENOMEM;
962 kfree(cb_arg);
963 goto cleanup_err;
964 }
965 /* Initialize rest of io_req fields */
966 seq_clnp_req->sc_cmd = NULL;
967 seq_clnp_req->port = port;
968 seq_clnp_req->tgt = tgt;
969 seq_clnp_req->data_xfer_len = 0; /* No data transfer for cleanup */
970
971 xid = seq_clnp_req->xid;
972
973 task_idx = xid/BNX2FC_TASKS_PER_PAGE;
974 index = xid % BNX2FC_TASKS_PER_PAGE;
975
976 /* Initialize task context for this IO request */
977 task_page = (struct fcoe_task_ctx_entry *)
978 interface->hba->task_ctx[task_idx];
979 task = &(task_page[index]);
980 cb_arg->aborted_io_req = orig_io_req;
981 cb_arg->io_req = seq_clnp_req;
982 cb_arg->r_ctl = r_ctl;
983 cb_arg->offset = offset;
984 seq_clnp_req->cb_arg = cb_arg;
985
986 printk(KERN_ERR PFX "call init_seq_cleanup_task\n");
987 bnx2fc_init_seq_cleanup_task(seq_clnp_req, task, orig_io_req, offset);
988
989 /* Obtain free SQ entry */
990 bnx2fc_add_2_sq(tgt, xid);
991
992 /* Ring doorbell */
993 bnx2fc_ring_doorbell(tgt);
994 cleanup_err:
995 return rc;
996 }
997
998 int bnx2fc_initiate_cleanup(struct bnx2fc_cmd *io_req)
999 {
1000 struct fc_lport *lport;
1001 struct bnx2fc_rport *tgt = io_req->tgt;
1002 struct bnx2fc_interface *interface;
1003 struct fcoe_port *port;
1004 struct bnx2fc_cmd *cleanup_io_req;
1005 struct fcoe_task_ctx_entry *task;
1006 struct fcoe_task_ctx_entry *task_page;
1007 int task_idx, index;
1008 u16 xid, orig_xid;
1009 int rc = 0;
1010
1011 /* ASSUMPTION: called with tgt_lock held */
1012 BNX2FC_IO_DBG(io_req, "Entered bnx2fc_initiate_cleanup\n");
1013
1014 port = io_req->port;
1015 interface = port->priv;
1016 lport = port->lport;
1017
1018 cleanup_io_req = bnx2fc_elstm_alloc(tgt, BNX2FC_CLEANUP);
1019 if (!cleanup_io_req) {
1020 printk(KERN_ERR PFX "cleanup: couldnt allocate cmd\n");
1021 rc = -1;
1022 goto cleanup_err;
1023 }
1024
1025 /* Initialize rest of io_req fields */
1026 cleanup_io_req->sc_cmd = NULL;
1027 cleanup_io_req->port = port;
1028 cleanup_io_req->tgt = tgt;
1029 cleanup_io_req->data_xfer_len = 0; /* No data transfer for cleanup */
1030
1031 xid = cleanup_io_req->xid;
1032
1033 task_idx = xid/BNX2FC_TASKS_PER_PAGE;
1034 index = xid % BNX2FC_TASKS_PER_PAGE;
1035
1036 /* Initialize task context for this IO request */
1037 task_page = (struct fcoe_task_ctx_entry *)
1038 interface->hba->task_ctx[task_idx];
1039 task = &(task_page[index]);
1040 orig_xid = io_req->xid;
1041
1042 BNX2FC_IO_DBG(io_req, "CLEANUP io_req xid = 0x%x\n", xid);
1043
1044 bnx2fc_init_cleanup_task(cleanup_io_req, task, orig_xid);
1045
1046 /* Obtain free SQ entry */
1047 bnx2fc_add_2_sq(tgt, xid);
1048
1049 /* Ring doorbell */
1050 bnx2fc_ring_doorbell(tgt);
1051
1052 cleanup_err:
1053 return rc;
1054 }
1055
1056 /**
1057 * bnx2fc_eh_target_reset: Reset a target
1058 *
1059 * @sc_cmd: SCSI command
1060 *
1061 * Set from SCSI host template to send task mgmt command to the target
1062 * and wait for the response
1063 */
1064 int bnx2fc_eh_target_reset(struct scsi_cmnd *sc_cmd)
1065 {
1066 return bnx2fc_initiate_tmf(sc_cmd, FCP_TMF_TGT_RESET);
1067 }
1068
1069 /**
1070 * bnx2fc_eh_device_reset - Reset a single LUN
1071 *
1072 * @sc_cmd: SCSI command
1073 *
1074 * Set from SCSI host template to send task mgmt command to the target
1075 * and wait for the response
1076 */
1077 int bnx2fc_eh_device_reset(struct scsi_cmnd *sc_cmd)
1078 {
1079 return bnx2fc_initiate_tmf(sc_cmd, FCP_TMF_LUN_RESET);
1080 }
1081
1082 int bnx2fc_abts_cleanup(struct bnx2fc_cmd *io_req)
1083 {
1084 struct bnx2fc_rport *tgt = io_req->tgt;
1085 int rc = SUCCESS;
1086
1087 io_req->wait_for_comp = 1;
1088 bnx2fc_initiate_cleanup(io_req);
1089
1090 spin_unlock_bh(&tgt->tgt_lock);
1091
1092 wait_for_completion(&io_req->tm_done);
1093
1094 io_req->wait_for_comp = 0;
1095 /*
1096 * release the reference taken in eh_abort to allow the
1097 * target to re-login after flushing IOs
1098 */
1099 kref_put(&io_req->refcount, bnx2fc_cmd_release);
1100
1101 spin_lock_bh(&tgt->tgt_lock);
1102 return rc;
1103 }
1104 /**
1105 * bnx2fc_eh_abort - eh_abort_handler api to abort an outstanding
1106 * SCSI command
1107 *
1108 * @sc_cmd: SCSI_ML command pointer
1109 *
1110 * SCSI abort request handler
1111 */
1112 int bnx2fc_eh_abort(struct scsi_cmnd *sc_cmd)
1113 {
1114 struct fc_rport *rport = starget_to_rport(scsi_target(sc_cmd->device));
1115 struct fc_rport_libfc_priv *rp = rport->dd_data;
1116 struct bnx2fc_cmd *io_req;
1117 struct fc_lport *lport;
1118 struct bnx2fc_rport *tgt;
1119 int rc;
1120
1121 rc = fc_block_scsi_eh(sc_cmd);
1122 if (rc)
1123 return rc;
1124
1125 lport = shost_priv(sc_cmd->device->host);
1126 if ((lport->state != LPORT_ST_READY) || !(lport->link_up)) {
1127 printk(KERN_ERR PFX "eh_abort: link not ready\n");
1128 return FAILED;
1129 }
1130
1131 tgt = (struct bnx2fc_rport *)&rp[1];
1132
1133 BNX2FC_TGT_DBG(tgt, "Entered bnx2fc_eh_abort\n");
1134
1135 spin_lock_bh(&tgt->tgt_lock);
1136 io_req = (struct bnx2fc_cmd *)sc_cmd->SCp.ptr;
1137 if (!io_req) {
1138 /* Command might have just completed */
1139 printk(KERN_ERR PFX "eh_abort: io_req is NULL\n");
1140 spin_unlock_bh(&tgt->tgt_lock);
1141 return SUCCESS;
1142 }
1143 BNX2FC_IO_DBG(io_req, "eh_abort - refcnt = %d\n",
1144 io_req->refcount.refcount.counter);
1145
1146 /* Hold IO request across abort processing */
1147 kref_get(&io_req->refcount);
1148
1149 BUG_ON(tgt != io_req->tgt);
1150
1151 /* Remove the io_req from the active_q. */
1152 /*
1153 * Task Mgmt functions (LUN RESET & TGT RESET) will not
1154 * issue an ABTS on this particular IO req, as the
1155 * io_req is no longer in the active_q.
1156 */
1157 if (tgt->flush_in_prog) {
1158 printk(KERN_ERR PFX "eh_abort: io_req (xid = 0x%x) "
1159 "flush in progress\n", io_req->xid);
1160 kref_put(&io_req->refcount, bnx2fc_cmd_release);
1161 spin_unlock_bh(&tgt->tgt_lock);
1162 return SUCCESS;
1163 }
1164
1165 if (io_req->on_active_queue == 0) {
1166 printk(KERN_ERR PFX "eh_abort: io_req (xid = 0x%x) "
1167 "not on active_q\n", io_req->xid);
1168 /*
1169 * This condition can happen only due to the FW bug,
1170 * where we do not receive cleanup response from
1171 * the FW. Handle this case gracefully by erroring
1172 * back the IO request to SCSI-ml
1173 */
1174 bnx2fc_scsi_done(io_req, DID_ABORT);
1175
1176 kref_put(&io_req->refcount, bnx2fc_cmd_release);
1177 spin_unlock_bh(&tgt->tgt_lock);
1178 return SUCCESS;
1179 }
1180
1181 /*
1182 * Only eh_abort processing will remove the IO from
1183 * active_cmd_q before processing the request. this is
1184 * done to avoid race conditions between IOs aborted
1185 * as part of task management completion and eh_abort
1186 * processing
1187 */
1188 list_del_init(&io_req->link);
1189 io_req->on_active_queue = 0;
1190 /* Move IO req to retire queue */
1191 list_add_tail(&io_req->link, &tgt->io_retire_queue);
1192
1193 init_completion(&io_req->tm_done);
1194
1195 if (test_and_set_bit(BNX2FC_FLAG_ISSUE_ABTS, &io_req->req_flags)) {
1196 printk(KERN_ERR PFX "eh_abort: io_req (xid = 0x%x) "
1197 "already in abts processing\n", io_req->xid);
1198 if (cancel_delayed_work(&io_req->timeout_work))
1199 kref_put(&io_req->refcount,
1200 bnx2fc_cmd_release); /* drop timer hold */
1201 rc = bnx2fc_abts_cleanup(io_req);
1202 /* This only occurs when an task abort was requested while ABTS
1203 is in progress. Setting the IO_CLEANUP flag will skip the
1204 RRQ process in the case when the fw generated SCSI_CMD cmpl
1205 was a result from the ABTS request rather than the CLEANUP
1206 request */
1207 set_bit(BNX2FC_FLAG_IO_CLEANUP, &io_req->req_flags);
1208 goto out;
1209 }
1210
1211 /* Cancel the current timer running on this io_req */
1212 if (cancel_delayed_work(&io_req->timeout_work))
1213 kref_put(&io_req->refcount,
1214 bnx2fc_cmd_release); /* drop timer hold */
1215 set_bit(BNX2FC_FLAG_EH_ABORT, &io_req->req_flags);
1216 io_req->wait_for_comp = 1;
1217 rc = bnx2fc_initiate_abts(io_req);
1218 if (rc == FAILED) {
1219 bnx2fc_initiate_cleanup(io_req);
1220 spin_unlock_bh(&tgt->tgt_lock);
1221 wait_for_completion(&io_req->tm_done);
1222 spin_lock_bh(&tgt->tgt_lock);
1223 io_req->wait_for_comp = 0;
1224 goto done;
1225 }
1226 spin_unlock_bh(&tgt->tgt_lock);
1227
1228 wait_for_completion(&io_req->tm_done);
1229
1230 spin_lock_bh(&tgt->tgt_lock);
1231 io_req->wait_for_comp = 0;
1232 if (test_bit(BNX2FC_FLAG_IO_COMPL, &io_req->req_flags)) {
1233 BNX2FC_IO_DBG(io_req, "IO completed in a different context\n");
1234 rc = SUCCESS;
1235 } else if (!(test_and_set_bit(BNX2FC_FLAG_ABTS_DONE,
1236 &io_req->req_flags))) {
1237 /* Let the scsi-ml try to recover this command */
1238 printk(KERN_ERR PFX "abort failed, xid = 0x%x\n",
1239 io_req->xid);
1240 rc = bnx2fc_abts_cleanup(io_req);
1241 goto out;
1242 } else {
1243 /*
1244 * We come here even when there was a race condition
1245 * between timeout and abts completion, and abts
1246 * completion happens just in time.
1247 */
1248 BNX2FC_IO_DBG(io_req, "abort succeeded\n");
1249 rc = SUCCESS;
1250 bnx2fc_scsi_done(io_req, DID_ABORT);
1251 kref_put(&io_req->refcount, bnx2fc_cmd_release);
1252 }
1253 done:
1254 /* release the reference taken in eh_abort */
1255 kref_put(&io_req->refcount, bnx2fc_cmd_release);
1256 out:
1257 spin_unlock_bh(&tgt->tgt_lock);
1258 return rc;
1259 }
1260
1261 void bnx2fc_process_seq_cleanup_compl(struct bnx2fc_cmd *seq_clnp_req,
1262 struct fcoe_task_ctx_entry *task,
1263 u8 rx_state)
1264 {
1265 struct bnx2fc_els_cb_arg *cb_arg = seq_clnp_req->cb_arg;
1266 struct bnx2fc_cmd *orig_io_req = cb_arg->aborted_io_req;
1267 u32 offset = cb_arg->offset;
1268 enum fc_rctl r_ctl = cb_arg->r_ctl;
1269 int rc = 0;
1270 struct bnx2fc_rport *tgt = orig_io_req->tgt;
1271
1272 BNX2FC_IO_DBG(orig_io_req, "Entered process_cleanup_compl xid = 0x%x"
1273 "cmd_type = %d\n",
1274 seq_clnp_req->xid, seq_clnp_req->cmd_type);
1275
1276 if (rx_state == FCOE_TASK_RX_STATE_IGNORED_SEQUENCE_CLEANUP) {
1277 printk(KERN_ERR PFX "seq cleanup ignored - xid = 0x%x\n",
1278 seq_clnp_req->xid);
1279 goto free_cb_arg;
1280 }
1281
1282 spin_unlock_bh(&tgt->tgt_lock);
1283 rc = bnx2fc_send_srr(orig_io_req, offset, r_ctl);
1284 spin_lock_bh(&tgt->tgt_lock);
1285
1286 if (rc)
1287 printk(KERN_ERR PFX "clnup_compl: Unable to send SRR"
1288 " IO will abort\n");
1289 seq_clnp_req->cb_arg = NULL;
1290 kref_put(&orig_io_req->refcount, bnx2fc_cmd_release);
1291 free_cb_arg:
1292 kfree(cb_arg);
1293 return;
1294 }
1295
1296 void bnx2fc_process_cleanup_compl(struct bnx2fc_cmd *io_req,
1297 struct fcoe_task_ctx_entry *task,
1298 u8 num_rq)
1299 {
1300 BNX2FC_IO_DBG(io_req, "Entered process_cleanup_compl "
1301 "refcnt = %d, cmd_type = %d\n",
1302 io_req->refcount.refcount.counter, io_req->cmd_type);
1303 bnx2fc_scsi_done(io_req, DID_ERROR);
1304 kref_put(&io_req->refcount, bnx2fc_cmd_release);
1305 if (io_req->wait_for_comp)
1306 complete(&io_req->tm_done);
1307 }
1308
1309 void bnx2fc_process_abts_compl(struct bnx2fc_cmd *io_req,
1310 struct fcoe_task_ctx_entry *task,
1311 u8 num_rq)
1312 {
1313 u32 r_ctl;
1314 u32 r_a_tov = FC_DEF_R_A_TOV;
1315 u8 issue_rrq = 0;
1316 struct bnx2fc_rport *tgt = io_req->tgt;
1317
1318 BNX2FC_IO_DBG(io_req, "Entered process_abts_compl xid = 0x%x"
1319 "refcnt = %d, cmd_type = %d\n",
1320 io_req->xid,
1321 io_req->refcount.refcount.counter, io_req->cmd_type);
1322
1323 if (test_and_set_bit(BNX2FC_FLAG_ABTS_DONE,
1324 &io_req->req_flags)) {
1325 BNX2FC_IO_DBG(io_req, "Timer context finished processing"
1326 " this io\n");
1327 return;
1328 }
1329
1330 /* Do not issue RRQ as this IO is already cleanedup */
1331 if (test_and_set_bit(BNX2FC_FLAG_IO_CLEANUP,
1332 &io_req->req_flags))
1333 goto io_compl;
1334
1335 /*
1336 * For ABTS issued due to SCSI eh_abort_handler, timeout
1337 * values are maintained by scsi-ml itself. Cancel timeout
1338 * in case ABTS issued as part of task management function
1339 * or due to FW error.
1340 */
1341 if (test_bit(BNX2FC_FLAG_ISSUE_ABTS, &io_req->req_flags))
1342 if (cancel_delayed_work(&io_req->timeout_work))
1343 kref_put(&io_req->refcount,
1344 bnx2fc_cmd_release); /* drop timer hold */
1345
1346 r_ctl = (u8)task->rxwr_only.union_ctx.comp_info.abts_rsp.r_ctl;
1347
1348 switch (r_ctl) {
1349 case FC_RCTL_BA_ACC:
1350 /*
1351 * Dont release this cmd yet. It will be relesed
1352 * after we get RRQ response
1353 */
1354 BNX2FC_IO_DBG(io_req, "ABTS response - ACC Send RRQ\n");
1355 issue_rrq = 1;
1356 break;
1357
1358 case FC_RCTL_BA_RJT:
1359 BNX2FC_IO_DBG(io_req, "ABTS response - RJT\n");
1360 break;
1361 default:
1362 printk(KERN_ERR PFX "Unknown ABTS response\n");
1363 break;
1364 }
1365
1366 if (issue_rrq) {
1367 BNX2FC_IO_DBG(io_req, "Issue RRQ after R_A_TOV\n");
1368 set_bit(BNX2FC_FLAG_ISSUE_RRQ, &io_req->req_flags);
1369 }
1370 set_bit(BNX2FC_FLAG_RETIRE_OXID, &io_req->req_flags);
1371 bnx2fc_cmd_timer_set(io_req, r_a_tov);
1372
1373 io_compl:
1374 if (io_req->wait_for_comp) {
1375 if (test_and_clear_bit(BNX2FC_FLAG_EH_ABORT,
1376 &io_req->req_flags))
1377 complete(&io_req->tm_done);
1378 } else {
1379 /*
1380 * We end up here when ABTS is issued as
1381 * in asynchronous context, i.e., as part
1382 * of task management completion, or
1383 * when FW error is received or when the
1384 * ABTS is issued when the IO is timed
1385 * out.
1386 */
1387
1388 if (io_req->on_active_queue) {
1389 list_del_init(&io_req->link);
1390 io_req->on_active_queue = 0;
1391 /* Move IO req to retire queue */
1392 list_add_tail(&io_req->link, &tgt->io_retire_queue);
1393 }
1394 bnx2fc_scsi_done(io_req, DID_ERROR);
1395 kref_put(&io_req->refcount, bnx2fc_cmd_release);
1396 }
1397 }
1398
1399 static void bnx2fc_lun_reset_cmpl(struct bnx2fc_cmd *io_req)
1400 {
1401 struct scsi_cmnd *sc_cmd = io_req->sc_cmd;
1402 struct bnx2fc_rport *tgt = io_req->tgt;
1403 struct bnx2fc_cmd *cmd, *tmp;
1404 u64 tm_lun = sc_cmd->device->lun;
1405 u64 lun;
1406 int rc = 0;
1407
1408 /* called with tgt_lock held */
1409 BNX2FC_IO_DBG(io_req, "Entered bnx2fc_lun_reset_cmpl\n");
1410 /*
1411 * Walk thru the active_ios queue and ABORT the IO
1412 * that matches with the LUN that was reset
1413 */
1414 list_for_each_entry_safe(cmd, tmp, &tgt->active_cmd_queue, link) {
1415 BNX2FC_TGT_DBG(tgt, "LUN RST cmpl: scan for pending IOs\n");
1416 lun = cmd->sc_cmd->device->lun;
1417 if (lun == tm_lun) {
1418 /* Initiate ABTS on this cmd */
1419 if (!test_and_set_bit(BNX2FC_FLAG_ISSUE_ABTS,
1420 &cmd->req_flags)) {
1421 /* cancel the IO timeout */
1422 if (cancel_delayed_work(&io_req->timeout_work))
1423 kref_put(&io_req->refcount,
1424 bnx2fc_cmd_release);
1425 /* timer hold */
1426 rc = bnx2fc_initiate_abts(cmd);
1427 /* abts shouldn't fail in this context */
1428 WARN_ON(rc != SUCCESS);
1429 } else
1430 printk(KERN_ERR PFX "lun_rst: abts already in"
1431 " progress for this IO 0x%x\n",
1432 cmd->xid);
1433 }
1434 }
1435 }
1436
1437 static void bnx2fc_tgt_reset_cmpl(struct bnx2fc_cmd *io_req)
1438 {
1439 struct bnx2fc_rport *tgt = io_req->tgt;
1440 struct bnx2fc_cmd *cmd, *tmp;
1441 int rc = 0;
1442
1443 /* called with tgt_lock held */
1444 BNX2FC_IO_DBG(io_req, "Entered bnx2fc_tgt_reset_cmpl\n");
1445 /*
1446 * Walk thru the active_ios queue and ABORT the IO
1447 * that matches with the LUN that was reset
1448 */
1449 list_for_each_entry_safe(cmd, tmp, &tgt->active_cmd_queue, link) {
1450 BNX2FC_TGT_DBG(tgt, "TGT RST cmpl: scan for pending IOs\n");
1451 /* Initiate ABTS */
1452 if (!test_and_set_bit(BNX2FC_FLAG_ISSUE_ABTS,
1453 &cmd->req_flags)) {
1454 /* cancel the IO timeout */
1455 if (cancel_delayed_work(&io_req->timeout_work))
1456 kref_put(&io_req->refcount,
1457 bnx2fc_cmd_release); /* timer hold */
1458 rc = bnx2fc_initiate_abts(cmd);
1459 /* abts shouldn't fail in this context */
1460 WARN_ON(rc != SUCCESS);
1461
1462 } else
1463 printk(KERN_ERR PFX "tgt_rst: abts already in progress"
1464 " for this IO 0x%x\n", cmd->xid);
1465 }
1466 }
1467
1468 void bnx2fc_process_tm_compl(struct bnx2fc_cmd *io_req,
1469 struct fcoe_task_ctx_entry *task, u8 num_rq)
1470 {
1471 struct bnx2fc_mp_req *tm_req;
1472 struct fc_frame_header *fc_hdr;
1473 struct scsi_cmnd *sc_cmd = io_req->sc_cmd;
1474 u64 *hdr;
1475 u64 *temp_hdr;
1476 void *rsp_buf;
1477
1478 /* Called with tgt_lock held */
1479 BNX2FC_IO_DBG(io_req, "Entered process_tm_compl\n");
1480
1481 if (!(test_bit(BNX2FC_FLAG_TM_TIMEOUT, &io_req->req_flags)))
1482 set_bit(BNX2FC_FLAG_TM_COMPL, &io_req->req_flags);
1483 else {
1484 /* TM has already timed out and we got
1485 * delayed completion. Ignore completion
1486 * processing.
1487 */
1488 return;
1489 }
1490
1491 tm_req = &(io_req->mp_req);
1492 fc_hdr = &(tm_req->resp_fc_hdr);
1493 hdr = (u64 *)fc_hdr;
1494 temp_hdr = (u64 *)
1495 &task->rxwr_only.union_ctx.comp_info.mp_rsp.fc_hdr;
1496 hdr[0] = cpu_to_be64(temp_hdr[0]);
1497 hdr[1] = cpu_to_be64(temp_hdr[1]);
1498 hdr[2] = cpu_to_be64(temp_hdr[2]);
1499
1500 tm_req->resp_len =
1501 task->rxwr_only.union_ctx.comp_info.mp_rsp.mp_payload_len;
1502
1503 rsp_buf = tm_req->resp_buf;
1504
1505 if (fc_hdr->fh_r_ctl == FC_RCTL_DD_CMD_STATUS) {
1506 bnx2fc_parse_fcp_rsp(io_req,
1507 (struct fcoe_fcp_rsp_payload *)
1508 rsp_buf, num_rq);
1509 if (io_req->fcp_rsp_code == 0) {
1510 /* TM successful */
1511 if (tm_req->tm_flags & FCP_TMF_LUN_RESET)
1512 bnx2fc_lun_reset_cmpl(io_req);
1513 else if (tm_req->tm_flags & FCP_TMF_TGT_RESET)
1514 bnx2fc_tgt_reset_cmpl(io_req);
1515 }
1516 } else {
1517 printk(KERN_ERR PFX "tmf's fc_hdr r_ctl = 0x%x\n",
1518 fc_hdr->fh_r_ctl);
1519 }
1520 if (!sc_cmd->SCp.ptr) {
1521 printk(KERN_ERR PFX "tm_compl: SCp.ptr is NULL\n");
1522 return;
1523 }
1524 switch (io_req->fcp_status) {
1525 case FC_GOOD:
1526 if (io_req->cdb_status == 0) {
1527 /* Good IO completion */
1528 sc_cmd->result = DID_OK << 16;
1529 } else {
1530 /* Transport status is good, SCSI status not good */
1531 sc_cmd->result = (DID_OK << 16) | io_req->cdb_status;
1532 }
1533 if (io_req->fcp_resid)
1534 scsi_set_resid(sc_cmd, io_req->fcp_resid);
1535 break;
1536
1537 default:
1538 BNX2FC_IO_DBG(io_req, "process_tm_compl: fcp_status = %d\n",
1539 io_req->fcp_status);
1540 break;
1541 }
1542
1543 sc_cmd = io_req->sc_cmd;
1544 io_req->sc_cmd = NULL;
1545
1546 /* check if the io_req exists in tgt's tmf_q */
1547 if (io_req->on_tmf_queue) {
1548
1549 list_del_init(&io_req->link);
1550 io_req->on_tmf_queue = 0;
1551 } else {
1552
1553 printk(KERN_ERR PFX "Command not on active_cmd_queue!\n");
1554 return;
1555 }
1556
1557 sc_cmd->SCp.ptr = NULL;
1558 sc_cmd->scsi_done(sc_cmd);
1559
1560 kref_put(&io_req->refcount, bnx2fc_cmd_release);
1561 if (io_req->wait_for_comp) {
1562 BNX2FC_IO_DBG(io_req, "tm_compl - wake up the waiter\n");
1563 complete(&io_req->tm_done);
1564 }
1565 }
1566
1567 static int bnx2fc_split_bd(struct bnx2fc_cmd *io_req, u64 addr, int sg_len,
1568 int bd_index)
1569 {
1570 struct fcoe_bd_ctx *bd = io_req->bd_tbl->bd_tbl;
1571 int frag_size, sg_frags;
1572
1573 sg_frags = 0;
1574 while (sg_len) {
1575 if (sg_len >= BNX2FC_BD_SPLIT_SZ)
1576 frag_size = BNX2FC_BD_SPLIT_SZ;
1577 else
1578 frag_size = sg_len;
1579 bd[bd_index + sg_frags].buf_addr_lo = addr & 0xffffffff;
1580 bd[bd_index + sg_frags].buf_addr_hi = addr >> 32;
1581 bd[bd_index + sg_frags].buf_len = (u16)frag_size;
1582 bd[bd_index + sg_frags].flags = 0;
1583
1584 addr += (u64) frag_size;
1585 sg_frags++;
1586 sg_len -= frag_size;
1587 }
1588 return sg_frags;
1589
1590 }
1591
1592 static int bnx2fc_map_sg(struct bnx2fc_cmd *io_req)
1593 {
1594 struct bnx2fc_interface *interface = io_req->port->priv;
1595 struct bnx2fc_hba *hba = interface->hba;
1596 struct scsi_cmnd *sc = io_req->sc_cmd;
1597 struct fcoe_bd_ctx *bd = io_req->bd_tbl->bd_tbl;
1598 struct scatterlist *sg;
1599 int byte_count = 0;
1600 int sg_count = 0;
1601 int bd_count = 0;
1602 int sg_frags;
1603 unsigned int sg_len;
1604 u64 addr;
1605 int i;
1606
1607 /*
1608 * Use dma_map_sg directly to ensure we're using the correct
1609 * dev struct off of pcidev.
1610 */
1611 sg_count = dma_map_sg(&hba->pcidev->dev, scsi_sglist(sc),
1612 scsi_sg_count(sc), sc->sc_data_direction);
1613 scsi_for_each_sg(sc, sg, sg_count, i) {
1614 sg_len = sg_dma_len(sg);
1615 addr = sg_dma_address(sg);
1616 if (sg_len > BNX2FC_MAX_BD_LEN) {
1617 sg_frags = bnx2fc_split_bd(io_req, addr, sg_len,
1618 bd_count);
1619 } else {
1620
1621 sg_frags = 1;
1622 bd[bd_count].buf_addr_lo = addr & 0xffffffff;
1623 bd[bd_count].buf_addr_hi = addr >> 32;
1624 bd[bd_count].buf_len = (u16)sg_len;
1625 bd[bd_count].flags = 0;
1626 }
1627 bd_count += sg_frags;
1628 byte_count += sg_len;
1629 }
1630 if (byte_count != scsi_bufflen(sc))
1631 printk(KERN_ERR PFX "byte_count = %d != scsi_bufflen = %d, "
1632 "task_id = 0x%x\n", byte_count, scsi_bufflen(sc),
1633 io_req->xid);
1634 return bd_count;
1635 }
1636
1637 static int bnx2fc_build_bd_list_from_sg(struct bnx2fc_cmd *io_req)
1638 {
1639 struct scsi_cmnd *sc = io_req->sc_cmd;
1640 struct fcoe_bd_ctx *bd = io_req->bd_tbl->bd_tbl;
1641 int bd_count;
1642
1643 if (scsi_sg_count(sc)) {
1644 bd_count = bnx2fc_map_sg(io_req);
1645 if (bd_count == 0)
1646 return -ENOMEM;
1647 } else {
1648 bd_count = 0;
1649 bd[0].buf_addr_lo = bd[0].buf_addr_hi = 0;
1650 bd[0].buf_len = bd[0].flags = 0;
1651 }
1652 io_req->bd_tbl->bd_valid = bd_count;
1653
1654 return 0;
1655 }
1656
1657 static void bnx2fc_unmap_sg_list(struct bnx2fc_cmd *io_req)
1658 {
1659 struct scsi_cmnd *sc = io_req->sc_cmd;
1660 struct bnx2fc_interface *interface = io_req->port->priv;
1661 struct bnx2fc_hba *hba = interface->hba;
1662
1663 /*
1664 * Use dma_unmap_sg directly to ensure we're using the correct
1665 * dev struct off of pcidev.
1666 */
1667 if (io_req->bd_tbl->bd_valid && sc && scsi_sg_count(sc)) {
1668 dma_unmap_sg(&hba->pcidev->dev, scsi_sglist(sc),
1669 scsi_sg_count(sc), sc->sc_data_direction);
1670 io_req->bd_tbl->bd_valid = 0;
1671 }
1672 }
1673
1674 void bnx2fc_build_fcp_cmnd(struct bnx2fc_cmd *io_req,
1675 struct fcp_cmnd *fcp_cmnd)
1676 {
1677 struct scsi_cmnd *sc_cmd = io_req->sc_cmd;
1678
1679 memset(fcp_cmnd, 0, sizeof(struct fcp_cmnd));
1680
1681 int_to_scsilun(sc_cmd->device->lun, &fcp_cmnd->fc_lun);
1682
1683 fcp_cmnd->fc_dl = htonl(io_req->data_xfer_len);
1684 memcpy(fcp_cmnd->fc_cdb, sc_cmd->cmnd, sc_cmd->cmd_len);
1685
1686 fcp_cmnd->fc_cmdref = 0;
1687 fcp_cmnd->fc_pri_ta = 0;
1688 fcp_cmnd->fc_tm_flags = io_req->mp_req.tm_flags;
1689 fcp_cmnd->fc_flags = io_req->io_req_flags;
1690 fcp_cmnd->fc_pri_ta = FCP_PTA_SIMPLE;
1691 }
1692
1693 static void bnx2fc_parse_fcp_rsp(struct bnx2fc_cmd *io_req,
1694 struct fcoe_fcp_rsp_payload *fcp_rsp,
1695 u8 num_rq)
1696 {
1697 struct scsi_cmnd *sc_cmd = io_req->sc_cmd;
1698 struct bnx2fc_rport *tgt = io_req->tgt;
1699 u8 rsp_flags = fcp_rsp->fcp_flags.flags;
1700 u32 rq_buff_len = 0;
1701 int i;
1702 unsigned char *rq_data;
1703 unsigned char *dummy;
1704 int fcp_sns_len = 0;
1705 int fcp_rsp_len = 0;
1706
1707 io_req->fcp_status = FC_GOOD;
1708 io_req->fcp_resid = 0;
1709 if (rsp_flags & (FCOE_FCP_RSP_FLAGS_FCP_RESID_OVER |
1710 FCOE_FCP_RSP_FLAGS_FCP_RESID_UNDER))
1711 io_req->fcp_resid = fcp_rsp->fcp_resid;
1712
1713 io_req->scsi_comp_flags = rsp_flags;
1714 CMD_SCSI_STATUS(sc_cmd) = io_req->cdb_status =
1715 fcp_rsp->scsi_status_code;
1716
1717 /* Fetch fcp_rsp_info and fcp_sns_info if available */
1718 if (num_rq) {
1719
1720 /*
1721 * We do not anticipate num_rq >1, as the linux defined
1722 * SCSI_SENSE_BUFFERSIZE is 96 bytes + 8 bytes of FCP_RSP_INFO
1723 * 256 bytes of single rq buffer is good enough to hold this.
1724 */
1725
1726 if (rsp_flags &
1727 FCOE_FCP_RSP_FLAGS_FCP_RSP_LEN_VALID) {
1728 fcp_rsp_len = rq_buff_len
1729 = fcp_rsp->fcp_rsp_len;
1730 }
1731
1732 if (rsp_flags &
1733 FCOE_FCP_RSP_FLAGS_FCP_SNS_LEN_VALID) {
1734 fcp_sns_len = fcp_rsp->fcp_sns_len;
1735 rq_buff_len += fcp_rsp->fcp_sns_len;
1736 }
1737
1738 io_req->fcp_rsp_len = fcp_rsp_len;
1739 io_req->fcp_sns_len = fcp_sns_len;
1740
1741 if (rq_buff_len > num_rq * BNX2FC_RQ_BUF_SZ) {
1742 /* Invalid sense sense length. */
1743 printk(KERN_ERR PFX "invalid sns length %d\n",
1744 rq_buff_len);
1745 /* reset rq_buff_len */
1746 rq_buff_len = num_rq * BNX2FC_RQ_BUF_SZ;
1747 }
1748
1749 rq_data = bnx2fc_get_next_rqe(tgt, 1);
1750
1751 if (num_rq > 1) {
1752 /* We do not need extra sense data */
1753 for (i = 1; i < num_rq; i++)
1754 dummy = bnx2fc_get_next_rqe(tgt, 1);
1755 }
1756
1757 /* fetch fcp_rsp_code */
1758 if ((fcp_rsp_len == 4) || (fcp_rsp_len == 8)) {
1759 /* Only for task management function */
1760 io_req->fcp_rsp_code = rq_data[3];
1761 BNX2FC_IO_DBG(io_req, "fcp_rsp_code = %d\n",
1762 io_req->fcp_rsp_code);
1763 }
1764
1765 /* fetch sense data */
1766 rq_data += fcp_rsp_len;
1767
1768 if (fcp_sns_len > SCSI_SENSE_BUFFERSIZE) {
1769 printk(KERN_ERR PFX "Truncating sense buffer\n");
1770 fcp_sns_len = SCSI_SENSE_BUFFERSIZE;
1771 }
1772
1773 memset(sc_cmd->sense_buffer, 0, SCSI_SENSE_BUFFERSIZE);
1774 if (fcp_sns_len)
1775 memcpy(sc_cmd->sense_buffer, rq_data, fcp_sns_len);
1776
1777 /* return RQ entries */
1778 for (i = 0; i < num_rq; i++)
1779 bnx2fc_return_rqe(tgt, 1);
1780 }
1781 }
1782
1783 /**
1784 * bnx2fc_queuecommand - Queuecommand function of the scsi template
1785 *
1786 * @host: The Scsi_Host the command was issued to
1787 * @sc_cmd: struct scsi_cmnd to be executed
1788 *
1789 * This is the IO strategy routine, called by SCSI-ML
1790 **/
1791 int bnx2fc_queuecommand(struct Scsi_Host *host,
1792 struct scsi_cmnd *sc_cmd)
1793 {
1794 struct fc_lport *lport = shost_priv(host);
1795 struct fc_rport *rport = starget_to_rport(scsi_target(sc_cmd->device));
1796 struct fc_rport_libfc_priv *rp = rport->dd_data;
1797 struct bnx2fc_rport *tgt;
1798 struct bnx2fc_cmd *io_req;
1799 int rc = 0;
1800 int rval;
1801
1802 rval = fc_remote_port_chkready(rport);
1803 if (rval) {
1804 sc_cmd->result = rval;
1805 sc_cmd->scsi_done(sc_cmd);
1806 return 0;
1807 }
1808
1809 if ((lport->state != LPORT_ST_READY) || !(lport->link_up)) {
1810 rc = SCSI_MLQUEUE_HOST_BUSY;
1811 goto exit_qcmd;
1812 }
1813
1814 /* rport and tgt are allocated together, so tgt should be non-NULL */
1815 tgt = (struct bnx2fc_rport *)&rp[1];
1816
1817 if (!test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags)) {
1818 /*
1819 * Session is not offloaded yet. Let SCSI-ml retry
1820 * the command.
1821 */
1822 rc = SCSI_MLQUEUE_TARGET_BUSY;
1823 goto exit_qcmd;
1824 }
1825 if (tgt->retry_delay_timestamp) {
1826 if (time_after(jiffies, tgt->retry_delay_timestamp)) {
1827 tgt->retry_delay_timestamp = 0;
1828 } else {
1829 /* If retry_delay timer is active, flow off the ML */
1830 rc = SCSI_MLQUEUE_TARGET_BUSY;
1831 goto exit_qcmd;
1832 }
1833 }
1834
1835 spin_lock_bh(&tgt->tgt_lock);
1836
1837 io_req = bnx2fc_cmd_alloc(tgt);
1838 if (!io_req) {
1839 rc = SCSI_MLQUEUE_HOST_BUSY;
1840 goto exit_qcmd_tgtlock;
1841 }
1842 io_req->sc_cmd = sc_cmd;
1843
1844 if (bnx2fc_post_io_req(tgt, io_req)) {
1845 printk(KERN_ERR PFX "Unable to post io_req\n");
1846 rc = SCSI_MLQUEUE_HOST_BUSY;
1847 goto exit_qcmd_tgtlock;
1848 }
1849
1850 exit_qcmd_tgtlock:
1851 spin_unlock_bh(&tgt->tgt_lock);
1852 exit_qcmd:
1853 return rc;
1854 }
1855
1856 void bnx2fc_process_scsi_cmd_compl(struct bnx2fc_cmd *io_req,
1857 struct fcoe_task_ctx_entry *task,
1858 u8 num_rq)
1859 {
1860 struct fcoe_fcp_rsp_payload *fcp_rsp;
1861 struct bnx2fc_rport *tgt = io_req->tgt;
1862 struct scsi_cmnd *sc_cmd;
1863 struct Scsi_Host *host;
1864
1865
1866 /* scsi_cmd_cmpl is called with tgt lock held */
1867
1868 if (test_and_set_bit(BNX2FC_FLAG_IO_COMPL, &io_req->req_flags)) {
1869 /* we will not receive ABTS response for this IO */
1870 BNX2FC_IO_DBG(io_req, "Timer context finished processing "
1871 "this scsi cmd\n");
1872 }
1873
1874 /* Cancel the timeout_work, as we received IO completion */
1875 if (cancel_delayed_work(&io_req->timeout_work))
1876 kref_put(&io_req->refcount,
1877 bnx2fc_cmd_release); /* drop timer hold */
1878
1879 sc_cmd = io_req->sc_cmd;
1880 if (sc_cmd == NULL) {
1881 printk(KERN_ERR PFX "scsi_cmd_compl - sc_cmd is NULL\n");
1882 return;
1883 }
1884
1885 /* Fetch fcp_rsp from task context and perform cmd completion */
1886 fcp_rsp = (struct fcoe_fcp_rsp_payload *)
1887 &(task->rxwr_only.union_ctx.comp_info.fcp_rsp.payload);
1888
1889 /* parse fcp_rsp and obtain sense data from RQ if available */
1890 bnx2fc_parse_fcp_rsp(io_req, fcp_rsp, num_rq);
1891
1892 host = sc_cmd->device->host;
1893 if (!sc_cmd->SCp.ptr) {
1894 printk(KERN_ERR PFX "SCp.ptr is NULL\n");
1895 return;
1896 }
1897
1898 if (io_req->on_active_queue) {
1899 list_del_init(&io_req->link);
1900 io_req->on_active_queue = 0;
1901 /* Move IO req to retire queue */
1902 list_add_tail(&io_req->link, &tgt->io_retire_queue);
1903 } else {
1904 /* This should not happen, but could have been pulled
1905 * by bnx2fc_flush_active_ios(), or during a race
1906 * between command abort and (late) completion.
1907 */
1908 BNX2FC_IO_DBG(io_req, "xid not on active_cmd_queue\n");
1909 if (io_req->wait_for_comp)
1910 if (test_and_clear_bit(BNX2FC_FLAG_EH_ABORT,
1911 &io_req->req_flags))
1912 complete(&io_req->tm_done);
1913 }
1914
1915 bnx2fc_unmap_sg_list(io_req);
1916 io_req->sc_cmd = NULL;
1917
1918 switch (io_req->fcp_status) {
1919 case FC_GOOD:
1920 if (io_req->cdb_status == 0) {
1921 /* Good IO completion */
1922 sc_cmd->result = DID_OK << 16;
1923 } else {
1924 /* Transport status is good, SCSI status not good */
1925 BNX2FC_IO_DBG(io_req, "scsi_cmpl: cdb_status = %d"
1926 " fcp_resid = 0x%x\n",
1927 io_req->cdb_status, io_req->fcp_resid);
1928 sc_cmd->result = (DID_OK << 16) | io_req->cdb_status;
1929
1930 if (io_req->cdb_status == SAM_STAT_TASK_SET_FULL ||
1931 io_req->cdb_status == SAM_STAT_BUSY) {
1932 /* Set the jiffies + retry_delay_timer * 100ms
1933 for the rport/tgt */
1934 tgt->retry_delay_timestamp = jiffies +
1935 fcp_rsp->retry_delay_timer * HZ / 10;
1936 }
1937
1938 }
1939 if (io_req->fcp_resid)
1940 scsi_set_resid(sc_cmd, io_req->fcp_resid);
1941 break;
1942 default:
1943 printk(KERN_ERR PFX "scsi_cmd_compl: fcp_status = %d\n",
1944 io_req->fcp_status);
1945 break;
1946 }
1947 sc_cmd->SCp.ptr = NULL;
1948 sc_cmd->scsi_done(sc_cmd);
1949 kref_put(&io_req->refcount, bnx2fc_cmd_release);
1950 }
1951
1952 int bnx2fc_post_io_req(struct bnx2fc_rport *tgt,
1953 struct bnx2fc_cmd *io_req)
1954 {
1955 struct fcoe_task_ctx_entry *task;
1956 struct fcoe_task_ctx_entry *task_page;
1957 struct scsi_cmnd *sc_cmd = io_req->sc_cmd;
1958 struct fcoe_port *port = tgt->port;
1959 struct bnx2fc_interface *interface = port->priv;
1960 struct bnx2fc_hba *hba = interface->hba;
1961 struct fc_lport *lport = port->lport;
1962 struct fc_stats *stats;
1963 int task_idx, index;
1964 u16 xid;
1965
1966 /* bnx2fc_post_io_req() is called with the tgt_lock held */
1967
1968 /* Initialize rest of io_req fields */
1969 io_req->cmd_type = BNX2FC_SCSI_CMD;
1970 io_req->port = port;
1971 io_req->tgt = tgt;
1972 io_req->data_xfer_len = scsi_bufflen(sc_cmd);
1973 sc_cmd->SCp.ptr = (char *)io_req;
1974
1975 stats = per_cpu_ptr(lport->stats, get_cpu());
1976 if (sc_cmd->sc_data_direction == DMA_FROM_DEVICE) {
1977 io_req->io_req_flags = BNX2FC_READ;
1978 stats->InputRequests++;
1979 stats->InputBytes += io_req->data_xfer_len;
1980 } else if (sc_cmd->sc_data_direction == DMA_TO_DEVICE) {
1981 io_req->io_req_flags = BNX2FC_WRITE;
1982 stats->OutputRequests++;
1983 stats->OutputBytes += io_req->data_xfer_len;
1984 } else {
1985 io_req->io_req_flags = 0;
1986 stats->ControlRequests++;
1987 }
1988 put_cpu();
1989
1990 xid = io_req->xid;
1991
1992 /* Build buffer descriptor list for firmware from sg list */
1993 if (bnx2fc_build_bd_list_from_sg(io_req)) {
1994 printk(KERN_ERR PFX "BD list creation failed\n");
1995 kref_put(&io_req->refcount, bnx2fc_cmd_release);
1996 return -EAGAIN;
1997 }
1998
1999 task_idx = xid / BNX2FC_TASKS_PER_PAGE;
2000 index = xid % BNX2FC_TASKS_PER_PAGE;
2001
2002 /* Initialize task context for this IO request */
2003 task_page = (struct fcoe_task_ctx_entry *) hba->task_ctx[task_idx];
2004 task = &(task_page[index]);
2005 bnx2fc_init_task(io_req, task);
2006
2007 if (tgt->flush_in_prog) {
2008 printk(KERN_ERR PFX "Flush in progress..Host Busy\n");
2009 kref_put(&io_req->refcount, bnx2fc_cmd_release);
2010 return -EAGAIN;
2011 }
2012
2013 if (!test_bit(BNX2FC_FLAG_SESSION_READY, &tgt->flags)) {
2014 printk(KERN_ERR PFX "Session not ready...post_io\n");
2015 kref_put(&io_req->refcount, bnx2fc_cmd_release);
2016 return -EAGAIN;
2017 }
2018
2019 /* Time IO req */
2020 if (tgt->io_timeout)
2021 bnx2fc_cmd_timer_set(io_req, BNX2FC_IO_TIMEOUT);
2022 /* Obtain free SQ entry */
2023 bnx2fc_add_2_sq(tgt, xid);
2024
2025 /* Enqueue the io_req to active_cmd_queue */
2026
2027 io_req->on_active_queue = 1;
2028 /* move io_req from pending_queue to active_queue */
2029 list_add_tail(&io_req->link, &tgt->active_cmd_queue);
2030
2031 /* Ring doorbell */
2032 bnx2fc_ring_doorbell(tgt);
2033 return 0;
2034 }
This page took 0.074312 seconds and 5 git commands to generate.