Commit | Line | Data |
---|---|---|
85b4aa49 RL |
1 | /* |
2 | * Copyright(c) 2007 - 2008 Intel Corporation. All rights reserved. | |
3 | * | |
4 | * This program is free software; you can redistribute it and/or modify it | |
5 | * under the terms and conditions of the GNU General Public License, | |
6 | * version 2, as published by the Free Software Foundation. | |
7 | * | |
8 | * This program is distributed in the hope it will be useful, but WITHOUT | |
9 | * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or | |
10 | * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for | |
11 | * more details. | |
12 | * | |
13 | * You should have received a copy of the GNU General Public License along with | |
14 | * this program; if not, write to the Free Software Foundation, Inc., | |
15 | * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA. | |
16 | * | |
17 | * Maintained at www.Open-FCoE.org | |
18 | */ | |
19 | ||
20 | #include <linux/module.h> | |
21 | #include <linux/version.h> | |
22 | #include <linux/kernel.h> | |
23 | #include <linux/spinlock.h> | |
24 | #include <linux/skbuff.h> | |
25 | #include <linux/netdevice.h> | |
26 | #include <linux/etherdevice.h> | |
27 | #include <linux/ethtool.h> | |
28 | #include <linux/if_ether.h> | |
29 | #include <linux/if_vlan.h> | |
30 | #include <linux/kthread.h> | |
31 | #include <linux/crc32.h> | |
32 | #include <linux/cpu.h> | |
33 | #include <linux/fs.h> | |
34 | #include <linux/sysfs.h> | |
35 | #include <linux/ctype.h> | |
36 | #include <scsi/scsi_tcq.h> | |
37 | #include <scsi/scsicam.h> | |
38 | #include <scsi/scsi_transport.h> | |
39 | #include <scsi/scsi_transport_fc.h> | |
40 | #include <net/rtnetlink.h> | |
41 | ||
42 | #include <scsi/fc/fc_encaps.h> | |
43 | ||
44 | #include <scsi/libfc.h> | |
45 | #include <scsi/fc_frame.h> | |
46 | #include <scsi/libfcoe.h> | |
47 | #include <scsi/fc_transport_fcoe.h> | |
48 | ||
49 | static int debug_fcoe; | |
50 | ||
51 | #define FCOE_MAX_QUEUE_DEPTH 256 | |
52 | ||
53 | /* destination address mode */ | |
54 | #define FCOE_GW_ADDR_MODE 0x00 | |
55 | #define FCOE_FCOUI_ADDR_MODE 0x01 | |
56 | ||
57 | #define FCOE_WORD_TO_BYTE 4 | |
58 | ||
59 | MODULE_AUTHOR("Open-FCoE.org"); | |
60 | MODULE_DESCRIPTION("FCoE"); | |
61 | MODULE_LICENSE("GPL"); | |
62 | ||
63 | /* fcoe host list */ | |
64 | LIST_HEAD(fcoe_hostlist); | |
65 | DEFINE_RWLOCK(fcoe_hostlist_lock); | |
66 | DEFINE_TIMER(fcoe_timer, NULL, 0, 0); | |
67 | struct fcoe_percpu_s *fcoe_percpu[NR_CPUS]; | |
68 | ||
69 | ||
70 | /* Function Prototyes */ | |
71 | static int fcoe_check_wait_queue(struct fc_lport *); | |
72 | static void fcoe_insert_wait_queue_head(struct fc_lport *, struct sk_buff *); | |
73 | static void fcoe_insert_wait_queue(struct fc_lport *, struct sk_buff *); | |
74 | static void fcoe_recv_flogi(struct fcoe_softc *, struct fc_frame *, u8 *); | |
75 | #ifdef CONFIG_HOTPLUG_CPU | |
76 | static int fcoe_cpu_callback(struct notifier_block *, ulong, void *); | |
77 | #endif /* CONFIG_HOTPLUG_CPU */ | |
78 | static int fcoe_device_notification(struct notifier_block *, ulong, void *); | |
79 | static void fcoe_dev_setup(void); | |
80 | static void fcoe_dev_cleanup(void); | |
81 | ||
82 | /* notification function from net device */ | |
83 | static struct notifier_block fcoe_notifier = { | |
84 | .notifier_call = fcoe_device_notification, | |
85 | }; | |
86 | ||
87 | ||
88 | #ifdef CONFIG_HOTPLUG_CPU | |
89 | static struct notifier_block fcoe_cpu_notifier = { | |
90 | .notifier_call = fcoe_cpu_callback, | |
91 | }; | |
92 | ||
93 | /** | |
94 | * fcoe_create_percpu_data - creates the associated cpu data | |
95 | * @cpu: index for the cpu where fcoe cpu data will be created | |
96 | * | |
97 | * create percpu stats block, from cpu add notifier | |
98 | * | |
99 | * Returns: none | |
100 | **/ | |
101 | static void fcoe_create_percpu_data(int cpu) | |
102 | { | |
103 | struct fc_lport *lp; | |
104 | struct fcoe_softc *fc; | |
105 | ||
106 | write_lock_bh(&fcoe_hostlist_lock); | |
107 | list_for_each_entry(fc, &fcoe_hostlist, list) { | |
108 | lp = fc->lp; | |
109 | if (lp->dev_stats[cpu] == NULL) | |
110 | lp->dev_stats[cpu] = | |
111 | kzalloc(sizeof(struct fcoe_dev_stats), | |
112 | GFP_KERNEL); | |
113 | } | |
114 | write_unlock_bh(&fcoe_hostlist_lock); | |
115 | } | |
116 | ||
117 | /** | |
118 | * fcoe_destroy_percpu_data - destroys the associated cpu data | |
119 | * @cpu: index for the cpu where fcoe cpu data will destroyed | |
120 | * | |
121 | * destroy percpu stats block called by cpu add/remove notifier | |
122 | * | |
123 | * Retuns: none | |
124 | **/ | |
125 | static void fcoe_destroy_percpu_data(int cpu) | |
126 | { | |
127 | struct fc_lport *lp; | |
128 | struct fcoe_softc *fc; | |
129 | ||
130 | write_lock_bh(&fcoe_hostlist_lock); | |
131 | list_for_each_entry(fc, &fcoe_hostlist, list) { | |
132 | lp = fc->lp; | |
133 | kfree(lp->dev_stats[cpu]); | |
134 | lp->dev_stats[cpu] = NULL; | |
135 | } | |
136 | write_unlock_bh(&fcoe_hostlist_lock); | |
137 | } | |
138 | ||
139 | /** | |
140 | * fcoe_cpu_callback - fcoe cpu hotplug event callback | |
141 | * @nfb: callback data block | |
142 | * @action: event triggering the callback | |
143 | * @hcpu: index for the cpu of this event | |
144 | * | |
145 | * this creates or destroys per cpu data for fcoe | |
146 | * | |
147 | * Returns NOTIFY_OK always. | |
148 | **/ | |
149 | static int fcoe_cpu_callback(struct notifier_block *nfb, unsigned long action, | |
150 | void *hcpu) | |
151 | { | |
152 | unsigned int cpu = (unsigned long)hcpu; | |
153 | ||
154 | switch (action) { | |
155 | case CPU_ONLINE: | |
156 | fcoe_create_percpu_data(cpu); | |
157 | break; | |
158 | case CPU_DEAD: | |
159 | fcoe_destroy_percpu_data(cpu); | |
160 | break; | |
161 | default: | |
162 | break; | |
163 | } | |
164 | return NOTIFY_OK; | |
165 | } | |
166 | #endif /* CONFIG_HOTPLUG_CPU */ | |
167 | ||
168 | /** | |
56b854bb | 169 | * fcoe_rcv - this is the fcoe receive function called by NET_RX_SOFTIRQ |
85b4aa49 RL |
170 | * @skb: the receive skb |
171 | * @dev: associated net device | |
172 | * @ptype: context | |
173 | * @odldev: last device | |
174 | * | |
175 | * this function will receive the packet and build fc frame and pass it up | |
176 | * | |
177 | * Returns: 0 for success | |
178 | **/ | |
179 | int fcoe_rcv(struct sk_buff *skb, struct net_device *dev, | |
180 | struct packet_type *ptype, struct net_device *olddev) | |
181 | { | |
182 | struct fc_lport *lp; | |
183 | struct fcoe_rcv_info *fr; | |
184 | struct fcoe_softc *fc; | |
185 | struct fcoe_dev_stats *stats; | |
186 | struct fc_frame_header *fh; | |
187 | unsigned short oxid; | |
188 | int cpu_idx; | |
189 | struct fcoe_percpu_s *fps; | |
190 | ||
191 | fc = container_of(ptype, struct fcoe_softc, fcoe_packet_type); | |
192 | lp = fc->lp; | |
193 | if (unlikely(lp == NULL)) { | |
194 | FC_DBG("cannot find hba structure"); | |
195 | goto err2; | |
196 | } | |
197 | ||
198 | if (unlikely(debug_fcoe)) { | |
199 | FC_DBG("skb_info: len:%d data_len:%d head:%p data:%p tail:%p " | |
200 | "end:%p sum:%d dev:%s", skb->len, skb->data_len, | |
201 | skb->head, skb->data, skb_tail_pointer(skb), | |
202 | skb_end_pointer(skb), skb->csum, | |
203 | skb->dev ? skb->dev->name : "<NULL>"); | |
204 | ||
205 | } | |
206 | ||
207 | /* check for FCOE packet type */ | |
208 | if (unlikely(eth_hdr(skb)->h_proto != htons(ETH_P_FCOE))) { | |
209 | FC_DBG("wrong FC type frame"); | |
210 | goto err; | |
211 | } | |
212 | ||
213 | /* | |
214 | * Check for minimum frame length, and make sure required FCoE | |
215 | * and FC headers are pulled into the linear data area. | |
216 | */ | |
217 | if (unlikely((skb->len < FCOE_MIN_FRAME) || | |
218 | !pskb_may_pull(skb, FCOE_HEADER_LEN))) | |
219 | goto err; | |
220 | ||
221 | skb_set_transport_header(skb, sizeof(struct fcoe_hdr)); | |
222 | fh = (struct fc_frame_header *) skb_transport_header(skb); | |
223 | ||
224 | oxid = ntohs(fh->fh_ox_id); | |
225 | ||
226 | fr = fcoe_dev_from_skb(skb); | |
227 | fr->fr_dev = lp; | |
228 | fr->ptype = ptype; | |
229 | cpu_idx = 0; | |
230 | #ifdef CONFIG_SMP | |
231 | /* | |
232 | * The incoming frame exchange id(oxid) is ANDed with num of online | |
233 | * cpu bits to get cpu_idx and then this cpu_idx is used for selecting | |
234 | * a per cpu kernel thread from fcoe_percpu. In case the cpu is | |
235 | * offline or no kernel thread for derived cpu_idx then cpu_idx is | |
236 | * initialize to first online cpu index. | |
237 | */ | |
238 | cpu_idx = oxid & (num_online_cpus() - 1); | |
239 | if (!fcoe_percpu[cpu_idx] || !cpu_online(cpu_idx)) | |
240 | cpu_idx = first_cpu(cpu_online_map); | |
241 | #endif | |
242 | fps = fcoe_percpu[cpu_idx]; | |
243 | ||
244 | spin_lock_bh(&fps->fcoe_rx_list.lock); | |
245 | __skb_queue_tail(&fps->fcoe_rx_list, skb); | |
246 | if (fps->fcoe_rx_list.qlen == 1) | |
247 | wake_up_process(fps->thread); | |
248 | ||
249 | spin_unlock_bh(&fps->fcoe_rx_list.lock); | |
250 | ||
251 | return 0; | |
252 | err: | |
253 | #ifdef CONFIG_SMP | |
254 | stats = lp->dev_stats[smp_processor_id()]; | |
255 | #else | |
256 | stats = lp->dev_stats[0]; | |
257 | #endif | |
258 | if (stats) | |
259 | stats->ErrorFrames++; | |
260 | ||
261 | err2: | |
262 | kfree_skb(skb); | |
263 | return -1; | |
264 | } | |
265 | EXPORT_SYMBOL_GPL(fcoe_rcv); | |
266 | ||
267 | /** | |
268 | * fcoe_start_io - pass to netdev to start xmit for fcoe | |
269 | * @skb: the skb to be xmitted | |
270 | * | |
271 | * Returns: 0 for success | |
272 | **/ | |
273 | static inline int fcoe_start_io(struct sk_buff *skb) | |
274 | { | |
275 | int rc; | |
276 | ||
277 | skb_get(skb); | |
278 | rc = dev_queue_xmit(skb); | |
279 | if (rc != 0) | |
280 | return rc; | |
281 | kfree_skb(skb); | |
282 | return 0; | |
283 | } | |
284 | ||
285 | /** | |
286 | * fcoe_get_paged_crc_eof - in case we need alloc a page for crc_eof | |
287 | * @skb: the skb to be xmitted | |
288 | * @tlen: total len | |
289 | * | |
290 | * Returns: 0 for success | |
291 | **/ | |
292 | static int fcoe_get_paged_crc_eof(struct sk_buff *skb, int tlen) | |
293 | { | |
294 | struct fcoe_percpu_s *fps; | |
295 | struct page *page; | |
296 | int cpu_idx; | |
297 | ||
298 | cpu_idx = get_cpu(); | |
299 | fps = fcoe_percpu[cpu_idx]; | |
300 | page = fps->crc_eof_page; | |
301 | if (!page) { | |
302 | page = alloc_page(GFP_ATOMIC); | |
303 | if (!page) { | |
304 | put_cpu(); | |
305 | return -ENOMEM; | |
306 | } | |
307 | fps->crc_eof_page = page; | |
308 | WARN_ON(fps->crc_eof_offset != 0); | |
309 | } | |
310 | ||
311 | get_page(page); | |
312 | skb_fill_page_desc(skb, skb_shinfo(skb)->nr_frags, page, | |
313 | fps->crc_eof_offset, tlen); | |
314 | skb->len += tlen; | |
315 | skb->data_len += tlen; | |
316 | skb->truesize += tlen; | |
317 | fps->crc_eof_offset += sizeof(struct fcoe_crc_eof); | |
318 | ||
319 | if (fps->crc_eof_offset >= PAGE_SIZE) { | |
320 | fps->crc_eof_page = NULL; | |
321 | fps->crc_eof_offset = 0; | |
322 | put_page(page); | |
323 | } | |
324 | put_cpu(); | |
325 | return 0; | |
326 | } | |
327 | ||
328 | /** | |
329 | * fcoe_fc_crc - calculates FC CRC in this fcoe skb | |
330 | * @fp: the fc_frame containg data to be checksummed | |
331 | * | |
332 | * This uses crc32() to calculate the crc for fc frame | |
333 | * Return : 32 bit crc | |
334 | * | |
335 | **/ | |
336 | u32 fcoe_fc_crc(struct fc_frame *fp) | |
337 | { | |
338 | struct sk_buff *skb = fp_skb(fp); | |
339 | struct skb_frag_struct *frag; | |
340 | unsigned char *data; | |
341 | unsigned long off, len, clen; | |
342 | u32 crc; | |
343 | unsigned i; | |
344 | ||
345 | crc = crc32(~0, skb->data, skb_headlen(skb)); | |
346 | ||
347 | for (i = 0; i < skb_shinfo(skb)->nr_frags; i++) { | |
348 | frag = &skb_shinfo(skb)->frags[i]; | |
349 | off = frag->page_offset; | |
350 | len = frag->size; | |
351 | while (len > 0) { | |
352 | clen = min(len, PAGE_SIZE - (off & ~PAGE_MASK)); | |
353 | data = kmap_atomic(frag->page + (off >> PAGE_SHIFT), | |
354 | KM_SKB_DATA_SOFTIRQ); | |
355 | crc = crc32(crc, data + (off & ~PAGE_MASK), clen); | |
356 | kunmap_atomic(data, KM_SKB_DATA_SOFTIRQ); | |
357 | off += clen; | |
358 | len -= clen; | |
359 | } | |
360 | } | |
361 | return crc; | |
362 | } | |
363 | EXPORT_SYMBOL_GPL(fcoe_fc_crc); | |
364 | ||
365 | /** | |
366 | * fcoe_xmit - FCoE frame transmit function | |
367 | * @lp: the associated local port | |
368 | * @fp: the fc_frame to be transmitted | |
369 | * | |
370 | * Return : 0 for success | |
371 | * | |
372 | **/ | |
373 | int fcoe_xmit(struct fc_lport *lp, struct fc_frame *fp) | |
374 | { | |
375 | int wlen, rc = 0; | |
376 | u32 crc; | |
377 | struct ethhdr *eh; | |
378 | struct fcoe_crc_eof *cp; | |
379 | struct sk_buff *skb; | |
380 | struct fcoe_dev_stats *stats; | |
381 | struct fc_frame_header *fh; | |
382 | unsigned int hlen; /* header length implies the version */ | |
383 | unsigned int tlen; /* trailer length */ | |
384 | unsigned int elen; /* eth header, may include vlan */ | |
385 | int flogi_in_progress = 0; | |
386 | struct fcoe_softc *fc; | |
387 | u8 sof, eof; | |
388 | struct fcoe_hdr *hp; | |
389 | ||
390 | WARN_ON((fr_len(fp) % sizeof(u32)) != 0); | |
391 | ||
392 | fc = fcoe_softc(lp); | |
393 | /* | |
394 | * if it is a flogi then we need to learn gw-addr | |
395 | * and my own fcid | |
396 | */ | |
397 | fh = fc_frame_header_get(fp); | |
398 | if (unlikely(fh->fh_r_ctl == FC_RCTL_ELS_REQ)) { | |
399 | if (fc_frame_payload_op(fp) == ELS_FLOGI) { | |
400 | fc->flogi_oxid = ntohs(fh->fh_ox_id); | |
401 | fc->address_mode = FCOE_FCOUI_ADDR_MODE; | |
402 | fc->flogi_progress = 1; | |
403 | flogi_in_progress = 1; | |
404 | } else if (fc->flogi_progress && ntoh24(fh->fh_s_id) != 0) { | |
405 | /* | |
406 | * Here we must've gotten an SID by accepting an FLOGI | |
407 | * from a point-to-point connection. Switch to using | |
408 | * the source mac based on the SID. The destination | |
409 | * MAC in this case would have been set by receving the | |
410 | * FLOGI. | |
411 | */ | |
412 | fc_fcoe_set_mac(fc->data_src_addr, fh->fh_s_id); | |
413 | fc->flogi_progress = 0; | |
414 | } | |
415 | } | |
416 | ||
417 | skb = fp_skb(fp); | |
418 | sof = fr_sof(fp); | |
419 | eof = fr_eof(fp); | |
420 | ||
421 | elen = (fc->real_dev->priv_flags & IFF_802_1Q_VLAN) ? | |
422 | sizeof(struct vlan_ethhdr) : sizeof(struct ethhdr); | |
423 | hlen = sizeof(struct fcoe_hdr); | |
424 | tlen = sizeof(struct fcoe_crc_eof); | |
425 | wlen = (skb->len - tlen + sizeof(crc)) / FCOE_WORD_TO_BYTE; | |
426 | ||
427 | /* crc offload */ | |
428 | if (likely(lp->crc_offload)) { | |
429 | skb->ip_summed = CHECKSUM_COMPLETE; | |
430 | skb->csum_start = skb_headroom(skb); | |
431 | skb->csum_offset = skb->len; | |
432 | crc = 0; | |
433 | } else { | |
434 | skb->ip_summed = CHECKSUM_NONE; | |
435 | crc = fcoe_fc_crc(fp); | |
436 | } | |
437 | ||
438 | /* copy fc crc and eof to the skb buff */ | |
439 | if (skb_is_nonlinear(skb)) { | |
440 | skb_frag_t *frag; | |
441 | if (fcoe_get_paged_crc_eof(skb, tlen)) { | |
442 | kfree(skb); | |
443 | return -ENOMEM; | |
444 | } | |
445 | frag = &skb_shinfo(skb)->frags[skb_shinfo(skb)->nr_frags - 1]; | |
446 | cp = kmap_atomic(frag->page, KM_SKB_DATA_SOFTIRQ) | |
447 | + frag->page_offset; | |
448 | } else { | |
449 | cp = (struct fcoe_crc_eof *)skb_put(skb, tlen); | |
450 | } | |
451 | ||
452 | memset(cp, 0, sizeof(*cp)); | |
453 | cp->fcoe_eof = eof; | |
454 | cp->fcoe_crc32 = cpu_to_le32(~crc); | |
455 | ||
456 | if (skb_is_nonlinear(skb)) { | |
457 | kunmap_atomic(cp, KM_SKB_DATA_SOFTIRQ); | |
458 | cp = NULL; | |
459 | } | |
460 | ||
461 | /* adjust skb netowrk/transport offsets to match mac/fcoe/fc */ | |
462 | skb_push(skb, elen + hlen); | |
463 | skb_reset_mac_header(skb); | |
464 | skb_reset_network_header(skb); | |
465 | skb->mac_len = elen; | |
466 | skb->protocol = htons(ETH_P_802_3); | |
467 | skb->dev = fc->real_dev; | |
468 | ||
469 | /* fill up mac and fcoe headers */ | |
470 | eh = eth_hdr(skb); | |
471 | eh->h_proto = htons(ETH_P_FCOE); | |
472 | if (fc->address_mode == FCOE_FCOUI_ADDR_MODE) | |
473 | fc_fcoe_set_mac(eh->h_dest, fh->fh_d_id); | |
474 | else | |
475 | /* insert GW address */ | |
476 | memcpy(eh->h_dest, fc->dest_addr, ETH_ALEN); | |
477 | ||
478 | if (unlikely(flogi_in_progress)) | |
479 | memcpy(eh->h_source, fc->ctl_src_addr, ETH_ALEN); | |
480 | else | |
481 | memcpy(eh->h_source, fc->data_src_addr, ETH_ALEN); | |
482 | ||
483 | hp = (struct fcoe_hdr *)(eh + 1); | |
484 | memset(hp, 0, sizeof(*hp)); | |
485 | if (FC_FCOE_VER) | |
486 | FC_FCOE_ENCAPS_VER(hp, FC_FCOE_VER); | |
487 | hp->fcoe_sof = sof; | |
488 | ||
489 | /* update tx stats: regardless if LLD fails */ | |
490 | stats = lp->dev_stats[smp_processor_id()]; | |
491 | if (stats) { | |
492 | stats->TxFrames++; | |
493 | stats->TxWords += wlen; | |
494 | } | |
495 | ||
496 | /* send down to lld */ | |
497 | fr_dev(fp) = lp; | |
498 | if (fc->fcoe_pending_queue.qlen) | |
499 | rc = fcoe_check_wait_queue(lp); | |
500 | ||
501 | if (rc == 0) | |
502 | rc = fcoe_start_io(skb); | |
503 | ||
504 | if (rc) { | |
505 | fcoe_insert_wait_queue(lp, skb); | |
506 | if (fc->fcoe_pending_queue.qlen > FCOE_MAX_QUEUE_DEPTH) | |
bc0e17f6 | 507 | lp->qfull = 1; |
85b4aa49 RL |
508 | } |
509 | ||
510 | return 0; | |
511 | } | |
512 | EXPORT_SYMBOL_GPL(fcoe_xmit); | |
513 | ||
514 | /* | |
515 | * fcoe_percpu_receive_thread - recv thread per cpu | |
516 | * @arg: ptr to the fcoe per cpu struct | |
517 | * | |
518 | * Return: 0 for success | |
519 | * | |
520 | */ | |
521 | int fcoe_percpu_receive_thread(void *arg) | |
522 | { | |
523 | struct fcoe_percpu_s *p = arg; | |
524 | u32 fr_len; | |
525 | struct fc_lport *lp; | |
526 | struct fcoe_rcv_info *fr; | |
527 | struct fcoe_dev_stats *stats; | |
528 | struct fc_frame_header *fh; | |
529 | struct sk_buff *skb; | |
530 | struct fcoe_crc_eof crc_eof; | |
531 | struct fc_frame *fp; | |
532 | u8 *mac = NULL; | |
533 | struct fcoe_softc *fc; | |
534 | struct fcoe_hdr *hp; | |
535 | ||
536 | set_user_nice(current, 19); | |
537 | ||
538 | while (!kthread_should_stop()) { | |
539 | ||
540 | spin_lock_bh(&p->fcoe_rx_list.lock); | |
541 | while ((skb = __skb_dequeue(&p->fcoe_rx_list)) == NULL) { | |
542 | set_current_state(TASK_INTERRUPTIBLE); | |
543 | spin_unlock_bh(&p->fcoe_rx_list.lock); | |
544 | schedule(); | |
545 | set_current_state(TASK_RUNNING); | |
546 | if (kthread_should_stop()) | |
547 | return 0; | |
548 | spin_lock_bh(&p->fcoe_rx_list.lock); | |
549 | } | |
550 | spin_unlock_bh(&p->fcoe_rx_list.lock); | |
551 | fr = fcoe_dev_from_skb(skb); | |
552 | lp = fr->fr_dev; | |
553 | if (unlikely(lp == NULL)) { | |
554 | FC_DBG("invalid HBA Structure"); | |
555 | kfree_skb(skb); | |
556 | continue; | |
557 | } | |
558 | ||
559 | stats = lp->dev_stats[smp_processor_id()]; | |
560 | ||
561 | if (unlikely(debug_fcoe)) { | |
562 | FC_DBG("skb_info: len:%d data_len:%d head:%p data:%p " | |
563 | "tail:%p end:%p sum:%d dev:%s", | |
564 | skb->len, skb->data_len, | |
565 | skb->head, skb->data, skb_tail_pointer(skb), | |
566 | skb_end_pointer(skb), skb->csum, | |
567 | skb->dev ? skb->dev->name : "<NULL>"); | |
568 | } | |
569 | ||
570 | /* | |
571 | * Save source MAC address before discarding header. | |
572 | */ | |
573 | fc = lport_priv(lp); | |
574 | if (unlikely(fc->flogi_progress)) | |
575 | mac = eth_hdr(skb)->h_source; | |
576 | ||
577 | if (skb_is_nonlinear(skb)) | |
578 | skb_linearize(skb); /* not ideal */ | |
579 | ||
580 | /* | |
581 | * Frame length checks and setting up the header pointers | |
582 | * was done in fcoe_rcv already. | |
583 | */ | |
584 | hp = (struct fcoe_hdr *) skb_network_header(skb); | |
585 | fh = (struct fc_frame_header *) skb_transport_header(skb); | |
586 | ||
587 | if (unlikely(FC_FCOE_DECAPS_VER(hp) != FC_FCOE_VER)) { | |
588 | if (stats) { | |
589 | if (stats->ErrorFrames < 5) | |
590 | FC_DBG("unknown FCoE version %x", | |
591 | FC_FCOE_DECAPS_VER(hp)); | |
592 | stats->ErrorFrames++; | |
593 | } | |
594 | kfree_skb(skb); | |
595 | continue; | |
596 | } | |
597 | ||
598 | skb_pull(skb, sizeof(struct fcoe_hdr)); | |
599 | fr_len = skb->len - sizeof(struct fcoe_crc_eof); | |
600 | ||
601 | if (stats) { | |
602 | stats->RxFrames++; | |
603 | stats->RxWords += fr_len / FCOE_WORD_TO_BYTE; | |
604 | } | |
605 | ||
606 | fp = (struct fc_frame *)skb; | |
607 | fc_frame_init(fp); | |
608 | fr_dev(fp) = lp; | |
609 | fr_sof(fp) = hp->fcoe_sof; | |
610 | ||
611 | /* Copy out the CRC and EOF trailer for access */ | |
612 | if (skb_copy_bits(skb, fr_len, &crc_eof, sizeof(crc_eof))) { | |
613 | kfree_skb(skb); | |
614 | continue; | |
615 | } | |
616 | fr_eof(fp) = crc_eof.fcoe_eof; | |
617 | fr_crc(fp) = crc_eof.fcoe_crc32; | |
618 | if (pskb_trim(skb, fr_len)) { | |
619 | kfree_skb(skb); | |
620 | continue; | |
621 | } | |
622 | ||
623 | /* | |
624 | * We only check CRC if no offload is available and if it is | |
625 | * it's solicited data, in which case, the FCP layer would | |
626 | * check it during the copy. | |
627 | */ | |
628 | if (lp->crc_offload) | |
629 | fr_flags(fp) &= ~FCPHF_CRC_UNCHECKED; | |
630 | else | |
631 | fr_flags(fp) |= FCPHF_CRC_UNCHECKED; | |
632 | ||
633 | fh = fc_frame_header_get(fp); | |
634 | if (fh->fh_r_ctl == FC_RCTL_DD_SOL_DATA && | |
635 | fh->fh_type == FC_TYPE_FCP) { | |
636 | fc_exch_recv(lp, lp->emp, fp); | |
637 | continue; | |
638 | } | |
639 | if (fr_flags(fp) & FCPHF_CRC_UNCHECKED) { | |
640 | if (le32_to_cpu(fr_crc(fp)) != | |
641 | ~crc32(~0, skb->data, fr_len)) { | |
642 | if (debug_fcoe || stats->InvalidCRCCount < 5) | |
643 | printk(KERN_WARNING "fcoe: dropping " | |
644 | "frame with CRC error\n"); | |
645 | stats->InvalidCRCCount++; | |
646 | stats->ErrorFrames++; | |
647 | fc_frame_free(fp); | |
648 | continue; | |
649 | } | |
650 | fr_flags(fp) &= ~FCPHF_CRC_UNCHECKED; | |
651 | } | |
652 | /* non flogi and non data exchanges are handled here */ | |
653 | if (unlikely(fc->flogi_progress)) | |
654 | fcoe_recv_flogi(fc, fp, mac); | |
655 | fc_exch_recv(lp, lp->emp, fp); | |
656 | } | |
657 | return 0; | |
658 | } | |
659 | ||
660 | /** | |
661 | * fcoe_recv_flogi - flogi receive function | |
662 | * @fc: associated fcoe_softc | |
663 | * @fp: the recieved frame | |
664 | * @sa: the source address of this flogi | |
665 | * | |
666 | * This is responsible to parse the flogi response and sets the corresponding | |
667 | * mac address for the initiator, eitehr OUI based or GW based. | |
668 | * | |
669 | * Returns: none | |
670 | **/ | |
671 | static void fcoe_recv_flogi(struct fcoe_softc *fc, struct fc_frame *fp, u8 *sa) | |
672 | { | |
673 | struct fc_frame_header *fh; | |
674 | u8 op; | |
675 | ||
676 | fh = fc_frame_header_get(fp); | |
677 | if (fh->fh_type != FC_TYPE_ELS) | |
678 | return; | |
679 | op = fc_frame_payload_op(fp); | |
680 | if (op == ELS_LS_ACC && fh->fh_r_ctl == FC_RCTL_ELS_REP && | |
681 | fc->flogi_oxid == ntohs(fh->fh_ox_id)) { | |
682 | /* | |
683 | * FLOGI accepted. | |
684 | * If the src mac addr is FC_OUI-based, then we mark the | |
685 | * address_mode flag to use FC_OUI-based Ethernet DA. | |
686 | * Otherwise we use the FCoE gateway addr | |
687 | */ | |
688 | if (!compare_ether_addr(sa, (u8[6]) FC_FCOE_FLOGI_MAC)) { | |
689 | fc->address_mode = FCOE_FCOUI_ADDR_MODE; | |
690 | } else { | |
691 | memcpy(fc->dest_addr, sa, ETH_ALEN); | |
692 | fc->address_mode = FCOE_GW_ADDR_MODE; | |
693 | } | |
694 | ||
695 | /* | |
696 | * Remove any previously-set unicast MAC filter. | |
697 | * Add secondary FCoE MAC address filter for our OUI. | |
698 | */ | |
699 | rtnl_lock(); | |
700 | if (compare_ether_addr(fc->data_src_addr, (u8[6]) { 0 })) | |
701 | dev_unicast_delete(fc->real_dev, fc->data_src_addr, | |
702 | ETH_ALEN); | |
703 | fc_fcoe_set_mac(fc->data_src_addr, fh->fh_d_id); | |
704 | dev_unicast_add(fc->real_dev, fc->data_src_addr, ETH_ALEN); | |
705 | rtnl_unlock(); | |
706 | ||
707 | fc->flogi_progress = 0; | |
708 | } else if (op == ELS_FLOGI && fh->fh_r_ctl == FC_RCTL_ELS_REQ && sa) { | |
709 | /* | |
710 | * Save source MAC for point-to-point responses. | |
711 | */ | |
712 | memcpy(fc->dest_addr, sa, ETH_ALEN); | |
713 | fc->address_mode = FCOE_GW_ADDR_MODE; | |
714 | } | |
715 | } | |
716 | ||
717 | /** | |
718 | * fcoe_watchdog - fcoe timer callback | |
719 | * @vp: | |
720 | * | |
bc0e17f6 | 721 | * This checks the pending queue length for fcoe and set lport qfull |
85b4aa49 RL |
722 | * if the FCOE_MAX_QUEUE_DEPTH is reached. This is done for all fc_lport on the |
723 | * fcoe_hostlist. | |
724 | * | |
725 | * Returns: 0 for success | |
726 | **/ | |
727 | void fcoe_watchdog(ulong vp) | |
728 | { | |
729 | struct fc_lport *lp; | |
730 | struct fcoe_softc *fc; | |
bc0e17f6 | 731 | int qfilled = 0; |
85b4aa49 RL |
732 | |
733 | read_lock(&fcoe_hostlist_lock); | |
734 | list_for_each_entry(fc, &fcoe_hostlist, list) { | |
735 | lp = fc->lp; | |
736 | if (lp) { | |
737 | if (fc->fcoe_pending_queue.qlen > FCOE_MAX_QUEUE_DEPTH) | |
bc0e17f6 | 738 | qfilled = 1; |
85b4aa49 | 739 | if (fcoe_check_wait_queue(lp) < FCOE_MAX_QUEUE_DEPTH) { |
bc0e17f6 VD |
740 | if (qfilled) |
741 | lp->qfull = 0; | |
85b4aa49 RL |
742 | } |
743 | } | |
744 | } | |
745 | read_unlock(&fcoe_hostlist_lock); | |
746 | ||
747 | fcoe_timer.expires = jiffies + (1 * HZ); | |
748 | add_timer(&fcoe_timer); | |
749 | } | |
750 | ||
751 | ||
752 | /** | |
753 | * fcoe_check_wait_queue - put the skb into fcoe pending xmit queue | |
754 | * @lp: the fc_port for this skb | |
755 | * @skb: the associated skb to be xmitted | |
756 | * | |
757 | * This empties the wait_queue, dequeue the head of the wait_queue queue | |
758 | * and calls fcoe_start_io() for each packet, if all skb have been | |
759 | * transmitted, return 0 if a error occurs, then restore wait_queue and | |
760 | * try again later. | |
761 | * | |
762 | * The wait_queue is used when the skb transmit fails. skb will go | |
763 | * in the wait_queue which will be emptied by the time function OR | |
764 | * by the next skb transmit. | |
765 | * | |
766 | * Returns: 0 for success | |
767 | **/ | |
768 | static int fcoe_check_wait_queue(struct fc_lport *lp) | |
769 | { | |
bc0e17f6 | 770 | int rc; |
85b4aa49 RL |
771 | struct sk_buff *skb; |
772 | struct fcoe_softc *fc; | |
773 | ||
774 | fc = fcoe_softc(lp); | |
775 | spin_lock_bh(&fc->fcoe_pending_queue.lock); | |
776 | ||
777 | /* | |
bc0e17f6 | 778 | * if interface pending queue full then set qfull in lport. |
85b4aa49 RL |
779 | */ |
780 | if (fc->fcoe_pending_queue.qlen > FCOE_MAX_QUEUE_DEPTH) | |
bc0e17f6 | 781 | lp->qfull = 1; |
85b4aa49 RL |
782 | if (fc->fcoe_pending_queue.qlen) { |
783 | while ((skb = __skb_dequeue(&fc->fcoe_pending_queue)) != NULL) { | |
784 | spin_unlock_bh(&fc->fcoe_pending_queue.lock); | |
785 | rc = fcoe_start_io(skb); | |
786 | if (rc) { | |
787 | fcoe_insert_wait_queue_head(lp, skb); | |
788 | return rc; | |
789 | } | |
790 | spin_lock_bh(&fc->fcoe_pending_queue.lock); | |
791 | } | |
792 | if (fc->fcoe_pending_queue.qlen < FCOE_MAX_QUEUE_DEPTH) | |
bc0e17f6 | 793 | lp->qfull = 0; |
85b4aa49 RL |
794 | } |
795 | spin_unlock_bh(&fc->fcoe_pending_queue.lock); | |
85b4aa49 RL |
796 | return fc->fcoe_pending_queue.qlen; |
797 | } | |
798 | ||
799 | /** | |
800 | * fcoe_insert_wait_queue_head - puts skb to fcoe pending queue head | |
801 | * @lp: the fc_port for this skb | |
802 | * @skb: the associated skb to be xmitted | |
803 | * | |
804 | * Returns: none | |
805 | **/ | |
806 | static void fcoe_insert_wait_queue_head(struct fc_lport *lp, | |
807 | struct sk_buff *skb) | |
808 | { | |
809 | struct fcoe_softc *fc; | |
810 | ||
811 | fc = fcoe_softc(lp); | |
812 | spin_lock_bh(&fc->fcoe_pending_queue.lock); | |
813 | __skb_queue_head(&fc->fcoe_pending_queue, skb); | |
814 | spin_unlock_bh(&fc->fcoe_pending_queue.lock); | |
815 | } | |
816 | ||
817 | /** | |
818 | * fcoe_insert_wait_queue - put the skb into fcoe pending queue tail | |
819 | * @lp: the fc_port for this skb | |
820 | * @skb: the associated skb to be xmitted | |
821 | * | |
822 | * Returns: none | |
823 | **/ | |
824 | static void fcoe_insert_wait_queue(struct fc_lport *lp, | |
825 | struct sk_buff *skb) | |
826 | { | |
827 | struct fcoe_softc *fc; | |
828 | ||
829 | fc = fcoe_softc(lp); | |
830 | spin_lock_bh(&fc->fcoe_pending_queue.lock); | |
831 | __skb_queue_tail(&fc->fcoe_pending_queue, skb); | |
832 | spin_unlock_bh(&fc->fcoe_pending_queue.lock); | |
833 | } | |
834 | ||
835 | /** | |
836 | * fcoe_dev_setup - setup link change notification interface | |
837 | * | |
838 | **/ | |
839 | static void fcoe_dev_setup(void) | |
840 | { | |
841 | /* | |
842 | * here setup a interface specific wd time to | |
843 | * monitor the link state | |
844 | */ | |
845 | register_netdevice_notifier(&fcoe_notifier); | |
846 | } | |
847 | ||
848 | /** | |
849 | * fcoe_dev_setup - cleanup link change notification interface | |
850 | **/ | |
851 | static void fcoe_dev_cleanup(void) | |
852 | { | |
853 | unregister_netdevice_notifier(&fcoe_notifier); | |
854 | } | |
855 | ||
856 | /** | |
857 | * fcoe_device_notification - netdev event notification callback | |
858 | * @notifier: context of the notification | |
859 | * @event: type of event | |
860 | * @ptr: fixed array for output parsed ifname | |
861 | * | |
862 | * This function is called by the ethernet driver in case of link change event | |
863 | * | |
864 | * Returns: 0 for success | |
865 | **/ | |
866 | static int fcoe_device_notification(struct notifier_block *notifier, | |
867 | ulong event, void *ptr) | |
868 | { | |
869 | struct fc_lport *lp = NULL; | |
870 | struct net_device *real_dev = ptr; | |
871 | struct fcoe_softc *fc; | |
872 | struct fcoe_dev_stats *stats; | |
bc0e17f6 | 873 | u32 new_link_up; |
85b4aa49 RL |
874 | u32 mfs; |
875 | int rc = NOTIFY_OK; | |
876 | ||
877 | read_lock(&fcoe_hostlist_lock); | |
878 | list_for_each_entry(fc, &fcoe_hostlist, list) { | |
879 | if (fc->real_dev == real_dev) { | |
880 | lp = fc->lp; | |
881 | break; | |
882 | } | |
883 | } | |
884 | read_unlock(&fcoe_hostlist_lock); | |
885 | if (lp == NULL) { | |
886 | rc = NOTIFY_DONE; | |
887 | goto out; | |
888 | } | |
889 | ||
bc0e17f6 | 890 | new_link_up = lp->link_up; |
85b4aa49 RL |
891 | switch (event) { |
892 | case NETDEV_DOWN: | |
893 | case NETDEV_GOING_DOWN: | |
bc0e17f6 | 894 | new_link_up = 0; |
85b4aa49 RL |
895 | break; |
896 | case NETDEV_UP: | |
897 | case NETDEV_CHANGE: | |
bc0e17f6 | 898 | new_link_up = !fcoe_link_ok(lp); |
85b4aa49 RL |
899 | break; |
900 | case NETDEV_CHANGEMTU: | |
901 | mfs = fc->real_dev->mtu - | |
902 | (sizeof(struct fcoe_hdr) + | |
903 | sizeof(struct fcoe_crc_eof)); | |
904 | if (mfs >= FC_MIN_MAX_FRAME) | |
905 | fc_set_mfs(lp, mfs); | |
bc0e17f6 | 906 | new_link_up = !fcoe_link_ok(lp); |
85b4aa49 RL |
907 | break; |
908 | case NETDEV_REGISTER: | |
909 | break; | |
910 | default: | |
911 | FC_DBG("unknown event %ld call", event); | |
912 | } | |
bc0e17f6 VD |
913 | if (lp->link_up != new_link_up) { |
914 | if (new_link_up) | |
85b4aa49 RL |
915 | fc_linkup(lp); |
916 | else { | |
917 | stats = lp->dev_stats[smp_processor_id()]; | |
918 | if (stats) | |
919 | stats->LinkFailureCount++; | |
920 | fc_linkdown(lp); | |
921 | fcoe_clean_pending_queue(lp); | |
922 | } | |
923 | } | |
924 | out: | |
925 | return rc; | |
926 | } | |
927 | ||
928 | /** | |
929 | * fcoe_if_to_netdev - parse a name buffer to get netdev | |
930 | * @ifname: fixed array for output parsed ifname | |
931 | * @buffer: incoming buffer to be copied | |
932 | * | |
933 | * Returns: NULL or ptr to netdeive | |
934 | **/ | |
935 | static struct net_device *fcoe_if_to_netdev(const char *buffer) | |
936 | { | |
937 | char *cp; | |
938 | char ifname[IFNAMSIZ + 2]; | |
939 | ||
940 | if (buffer) { | |
941 | strlcpy(ifname, buffer, IFNAMSIZ); | |
942 | cp = ifname + strlen(ifname); | |
943 | while (--cp >= ifname && *cp == '\n') | |
944 | *cp = '\0'; | |
945 | return dev_get_by_name(&init_net, ifname); | |
946 | } | |
947 | return NULL; | |
948 | } | |
949 | ||
950 | /** | |
951 | * fcoe_netdev_to_module_owner - finds out the nic drive moddule of the netdev | |
952 | * @netdev: the target netdev | |
953 | * | |
954 | * Returns: ptr to the struct module, NULL for failure | |
955 | **/ | |
956 | static struct module *fcoe_netdev_to_module_owner( | |
957 | const struct net_device *netdev) | |
958 | { | |
959 | struct device *dev; | |
960 | ||
961 | if (!netdev) | |
962 | return NULL; | |
963 | ||
964 | dev = netdev->dev.parent; | |
965 | if (!dev) | |
966 | return NULL; | |
967 | ||
968 | if (!dev->driver) | |
969 | return NULL; | |
970 | ||
971 | return dev->driver->owner; | |
972 | } | |
973 | ||
974 | /** | |
975 | * fcoe_ethdrv_get - holds the nic driver module by try_module_get() for | |
976 | * the corresponding netdev. | |
977 | * @netdev: the target netdev | |
978 | * | |
979 | * Returns: 0 for succsss | |
980 | **/ | |
981 | static int fcoe_ethdrv_get(const struct net_device *netdev) | |
982 | { | |
983 | struct module *owner; | |
984 | ||
985 | owner = fcoe_netdev_to_module_owner(netdev); | |
986 | if (owner) { | |
56b854bb JB |
987 | printk(KERN_DEBUG "fcoe:hold driver module %s for %s\n", |
988 | module_name(owner), netdev->name); | |
85b4aa49 RL |
989 | return try_module_get(owner); |
990 | } | |
991 | return -ENODEV; | |
992 | } | |
993 | ||
994 | /** | |
995 | * fcoe_ethdrv_get - releases the nic driver module by module_put for | |
996 | * the corresponding netdev. | |
997 | * @netdev: the target netdev | |
998 | * | |
999 | * Returns: 0 for succsss | |
1000 | **/ | |
1001 | static int fcoe_ethdrv_put(const struct net_device *netdev) | |
1002 | { | |
1003 | struct module *owner; | |
1004 | ||
1005 | owner = fcoe_netdev_to_module_owner(netdev); | |
1006 | if (owner) { | |
56b854bb JB |
1007 | printk(KERN_DEBUG "fcoe:release driver module %s for %s\n", |
1008 | module_name(owner), netdev->name); | |
85b4aa49 RL |
1009 | module_put(owner); |
1010 | return 0; | |
1011 | } | |
1012 | return -ENODEV; | |
1013 | } | |
1014 | ||
1015 | /** | |
1016 | * fcoe_destroy- handles the destroy from sysfs | |
1017 | * @buffer: expcted to be a eth if name | |
1018 | * @kp: associated kernel param | |
1019 | * | |
1020 | * Returns: 0 for success | |
1021 | **/ | |
1022 | static int fcoe_destroy(const char *buffer, struct kernel_param *kp) | |
1023 | { | |
1024 | int rc; | |
1025 | struct net_device *netdev; | |
1026 | ||
1027 | netdev = fcoe_if_to_netdev(buffer); | |
1028 | if (!netdev) { | |
1029 | rc = -ENODEV; | |
1030 | goto out_nodev; | |
1031 | } | |
1032 | /* look for existing lport */ | |
1033 | if (!fcoe_hostlist_lookup(netdev)) { | |
1034 | rc = -ENODEV; | |
1035 | goto out_putdev; | |
1036 | } | |
1037 | /* pass to transport */ | |
1038 | rc = fcoe_transport_release(netdev); | |
1039 | if (rc) { | |
1040 | printk(KERN_ERR "fcoe: fcoe_transport_release(%s) failed\n", | |
1041 | netdev->name); | |
1042 | rc = -EIO; | |
1043 | goto out_putdev; | |
1044 | } | |
1045 | fcoe_ethdrv_put(netdev); | |
1046 | rc = 0; | |
1047 | out_putdev: | |
1048 | dev_put(netdev); | |
1049 | out_nodev: | |
1050 | return rc; | |
1051 | } | |
1052 | ||
1053 | /** | |
1054 | * fcoe_create - handles the create call from sysfs | |
1055 | * @buffer: expcted to be a eth if name | |
1056 | * @kp: associated kernel param | |
1057 | * | |
1058 | * Returns: 0 for success | |
1059 | **/ | |
1060 | static int fcoe_create(const char *buffer, struct kernel_param *kp) | |
1061 | { | |
1062 | int rc; | |
1063 | struct net_device *netdev; | |
1064 | ||
1065 | netdev = fcoe_if_to_netdev(buffer); | |
1066 | if (!netdev) { | |
1067 | rc = -ENODEV; | |
1068 | goto out_nodev; | |
1069 | } | |
1070 | /* look for existing lport */ | |
1071 | if (fcoe_hostlist_lookup(netdev)) { | |
1072 | rc = -EEXIST; | |
1073 | goto out_putdev; | |
1074 | } | |
1075 | fcoe_ethdrv_get(netdev); | |
1076 | ||
1077 | /* pass to transport */ | |
1078 | rc = fcoe_transport_attach(netdev); | |
1079 | if (rc) { | |
1080 | printk(KERN_ERR "fcoe: fcoe_transport_attach(%s) failed\n", | |
1081 | netdev->name); | |
1082 | fcoe_ethdrv_put(netdev); | |
1083 | rc = -EIO; | |
1084 | goto out_putdev; | |
1085 | } | |
1086 | rc = 0; | |
1087 | out_putdev: | |
1088 | dev_put(netdev); | |
1089 | out_nodev: | |
1090 | return rc; | |
1091 | } | |
1092 | ||
1093 | module_param_call(create, fcoe_create, NULL, NULL, S_IWUSR); | |
1094 | __MODULE_PARM_TYPE(create, "string"); | |
1095 | MODULE_PARM_DESC(create, "Create fcoe port using net device passed in."); | |
1096 | module_param_call(destroy, fcoe_destroy, NULL, NULL, S_IWUSR); | |
1097 | __MODULE_PARM_TYPE(destroy, "string"); | |
1098 | MODULE_PARM_DESC(destroy, "Destroy fcoe port"); | |
1099 | ||
1100 | /* | |
1101 | * fcoe_link_ok - check if link is ok for the fc_lport | |
1102 | * @lp: ptr to the fc_lport | |
1103 | * | |
1104 | * Any permanently-disqualifying conditions have been previously checked. | |
1105 | * This also updates the speed setting, which may change with link for 100/1000. | |
1106 | * | |
1107 | * This function should probably be checking for PAUSE support at some point | |
1108 | * in the future. Currently Per-priority-pause is not determinable using | |
1109 | * ethtool, so we shouldn't be restrictive until that problem is resolved. | |
1110 | * | |
1111 | * Returns: 0 if link is OK for use by FCoE. | |
1112 | * | |
1113 | */ | |
1114 | int fcoe_link_ok(struct fc_lport *lp) | |
1115 | { | |
1116 | struct fcoe_softc *fc = fcoe_softc(lp); | |
1117 | struct net_device *dev = fc->real_dev; | |
1118 | struct ethtool_cmd ecmd = { ETHTOOL_GSET }; | |
1119 | int rc = 0; | |
1120 | ||
1121 | if ((dev->flags & IFF_UP) && netif_carrier_ok(dev)) { | |
1122 | dev = fc->phys_dev; | |
1123 | if (dev->ethtool_ops->get_settings) { | |
1124 | dev->ethtool_ops->get_settings(dev, &ecmd); | |
1125 | lp->link_supported_speeds &= | |
1126 | ~(FC_PORTSPEED_1GBIT | FC_PORTSPEED_10GBIT); | |
1127 | if (ecmd.supported & (SUPPORTED_1000baseT_Half | | |
1128 | SUPPORTED_1000baseT_Full)) | |
1129 | lp->link_supported_speeds |= FC_PORTSPEED_1GBIT; | |
1130 | if (ecmd.supported & SUPPORTED_10000baseT_Full) | |
1131 | lp->link_supported_speeds |= | |
1132 | FC_PORTSPEED_10GBIT; | |
1133 | if (ecmd.speed == SPEED_1000) | |
1134 | lp->link_speed = FC_PORTSPEED_1GBIT; | |
1135 | if (ecmd.speed == SPEED_10000) | |
1136 | lp->link_speed = FC_PORTSPEED_10GBIT; | |
1137 | } | |
1138 | } else | |
1139 | rc = -1; | |
1140 | ||
1141 | return rc; | |
1142 | } | |
1143 | EXPORT_SYMBOL_GPL(fcoe_link_ok); | |
1144 | ||
1145 | /* | |
1146 | * fcoe_percpu_clean - frees skb of the corresponding lport from the per | |
1147 | * cpu queue. | |
1148 | * @lp: the fc_lport | |
1149 | */ | |
1150 | void fcoe_percpu_clean(struct fc_lport *lp) | |
1151 | { | |
1152 | int idx; | |
1153 | struct fcoe_percpu_s *pp; | |
1154 | struct fcoe_rcv_info *fr; | |
1155 | struct sk_buff_head *list; | |
1156 | struct sk_buff *skb, *next; | |
1157 | struct sk_buff *head; | |
1158 | ||
1159 | for (idx = 0; idx < NR_CPUS; idx++) { | |
1160 | if (fcoe_percpu[idx]) { | |
1161 | pp = fcoe_percpu[idx]; | |
1162 | spin_lock_bh(&pp->fcoe_rx_list.lock); | |
1163 | list = &pp->fcoe_rx_list; | |
1164 | head = list->next; | |
1165 | for (skb = head; skb != (struct sk_buff *)list; | |
1166 | skb = next) { | |
1167 | next = skb->next; | |
1168 | fr = fcoe_dev_from_skb(skb); | |
1169 | if (fr->fr_dev == lp) { | |
1170 | __skb_unlink(skb, list); | |
1171 | kfree_skb(skb); | |
1172 | } | |
1173 | } | |
1174 | spin_unlock_bh(&pp->fcoe_rx_list.lock); | |
1175 | } | |
1176 | } | |
1177 | } | |
1178 | EXPORT_SYMBOL_GPL(fcoe_percpu_clean); | |
1179 | ||
1180 | /** | |
1181 | * fcoe_clean_pending_queue - dequeue skb and free it | |
1182 | * @lp: the corresponding fc_lport | |
1183 | * | |
1184 | * Returns: none | |
1185 | **/ | |
1186 | void fcoe_clean_pending_queue(struct fc_lport *lp) | |
1187 | { | |
1188 | struct fcoe_softc *fc = lport_priv(lp); | |
1189 | struct sk_buff *skb; | |
1190 | ||
1191 | spin_lock_bh(&fc->fcoe_pending_queue.lock); | |
1192 | while ((skb = __skb_dequeue(&fc->fcoe_pending_queue)) != NULL) { | |
1193 | spin_unlock_bh(&fc->fcoe_pending_queue.lock); | |
1194 | kfree_skb(skb); | |
1195 | spin_lock_bh(&fc->fcoe_pending_queue.lock); | |
1196 | } | |
1197 | spin_unlock_bh(&fc->fcoe_pending_queue.lock); | |
1198 | } | |
1199 | EXPORT_SYMBOL_GPL(fcoe_clean_pending_queue); | |
1200 | ||
1201 | /** | |
1202 | * libfc_host_alloc - allocate a Scsi_Host with room for the fc_lport | |
1203 | * @sht: ptr to the scsi host templ | |
1204 | * @priv_size: size of private data after fc_lport | |
1205 | * | |
1206 | * Returns: ptr to Scsi_Host | |
1207 | * TODO - to libfc? | |
1208 | */ | |
1209 | static inline struct Scsi_Host *libfc_host_alloc( | |
1210 | struct scsi_host_template *sht, int priv_size) | |
1211 | { | |
1212 | return scsi_host_alloc(sht, sizeof(struct fc_lport) + priv_size); | |
1213 | } | |
1214 | ||
1215 | /** | |
1216 | * fcoe_host_alloc - allocate a Scsi_Host with room for the fcoe_softc | |
1217 | * @sht: ptr to the scsi host templ | |
1218 | * @priv_size: size of private data after fc_lport | |
1219 | * | |
1220 | * Returns: ptr to Scsi_Host | |
1221 | */ | |
1222 | struct Scsi_Host *fcoe_host_alloc(struct scsi_host_template *sht, int priv_size) | |
1223 | { | |
1224 | return libfc_host_alloc(sht, sizeof(struct fcoe_softc) + priv_size); | |
1225 | } | |
1226 | EXPORT_SYMBOL_GPL(fcoe_host_alloc); | |
1227 | ||
1228 | /* | |
1229 | * fcoe_reset - resets the fcoe | |
1230 | * @shost: shost the reset is from | |
1231 | * | |
1232 | * Returns: always 0 | |
1233 | */ | |
1234 | int fcoe_reset(struct Scsi_Host *shost) | |
1235 | { | |
1236 | struct fc_lport *lport = shost_priv(shost); | |
1237 | fc_lport_reset(lport); | |
1238 | return 0; | |
1239 | } | |
1240 | EXPORT_SYMBOL_GPL(fcoe_reset); | |
1241 | ||
1242 | /* | |
1243 | * fcoe_wwn_from_mac - converts 48-bit IEEE MAC address to 64-bit FC WWN. | |
1244 | * @mac: mac address | |
1245 | * @scheme: check port | |
1246 | * @port: port indicator for converting | |
1247 | * | |
1248 | * Returns: u64 fc world wide name | |
1249 | */ | |
1250 | u64 fcoe_wwn_from_mac(unsigned char mac[MAX_ADDR_LEN], | |
1251 | unsigned int scheme, unsigned int port) | |
1252 | { | |
1253 | u64 wwn; | |
1254 | u64 host_mac; | |
1255 | ||
1256 | /* The MAC is in NO, so flip only the low 48 bits */ | |
1257 | host_mac = ((u64) mac[0] << 40) | | |
1258 | ((u64) mac[1] << 32) | | |
1259 | ((u64) mac[2] << 24) | | |
1260 | ((u64) mac[3] << 16) | | |
1261 | ((u64) mac[4] << 8) | | |
1262 | (u64) mac[5]; | |
1263 | ||
1264 | WARN_ON(host_mac >= (1ULL << 48)); | |
1265 | wwn = host_mac | ((u64) scheme << 60); | |
1266 | switch (scheme) { | |
1267 | case 1: | |
1268 | WARN_ON(port != 0); | |
1269 | break; | |
1270 | case 2: | |
1271 | WARN_ON(port >= 0xfff); | |
1272 | wwn |= (u64) port << 48; | |
1273 | break; | |
1274 | default: | |
1275 | WARN_ON(1); | |
1276 | break; | |
1277 | } | |
1278 | ||
1279 | return wwn; | |
1280 | } | |
1281 | EXPORT_SYMBOL_GPL(fcoe_wwn_from_mac); | |
1282 | /* | |
1283 | * fcoe_hostlist_lookup_softc - find the corresponding lport by a given device | |
1284 | * @device: this is currently ptr to net_device | |
1285 | * | |
1286 | * Returns: NULL or the located fcoe_softc | |
1287 | */ | |
1288 | static struct fcoe_softc *fcoe_hostlist_lookup_softc( | |
1289 | const struct net_device *dev) | |
1290 | { | |
1291 | struct fcoe_softc *fc; | |
1292 | ||
1293 | read_lock(&fcoe_hostlist_lock); | |
1294 | list_for_each_entry(fc, &fcoe_hostlist, list) { | |
1295 | if (fc->real_dev == dev) { | |
1296 | read_unlock(&fcoe_hostlist_lock); | |
1297 | return fc; | |
1298 | } | |
1299 | } | |
1300 | read_unlock(&fcoe_hostlist_lock); | |
1301 | return NULL; | |
1302 | } | |
1303 | ||
1304 | /* | |
1305 | * fcoe_hostlist_lookup - find the corresponding lport by netdev | |
1306 | * @netdev: ptr to net_device | |
1307 | * | |
1308 | * Returns: 0 for success | |
1309 | */ | |
1310 | struct fc_lport *fcoe_hostlist_lookup(const struct net_device *netdev) | |
1311 | { | |
1312 | struct fcoe_softc *fc; | |
1313 | ||
1314 | fc = fcoe_hostlist_lookup_softc(netdev); | |
1315 | ||
1316 | return (fc) ? fc->lp : NULL; | |
1317 | } | |
1318 | EXPORT_SYMBOL_GPL(fcoe_hostlist_lookup); | |
1319 | ||
1320 | /* | |
1321 | * fcoe_hostlist_add - add a lport to lports list | |
1322 | * @lp: ptr to the fc_lport to badded | |
1323 | * | |
1324 | * Returns: 0 for success | |
1325 | */ | |
1326 | int fcoe_hostlist_add(const struct fc_lport *lp) | |
1327 | { | |
1328 | struct fcoe_softc *fc; | |
1329 | ||
1330 | fc = fcoe_hostlist_lookup_softc(fcoe_netdev(lp)); | |
1331 | if (!fc) { | |
1332 | fc = fcoe_softc(lp); | |
1333 | write_lock_bh(&fcoe_hostlist_lock); | |
1334 | list_add_tail(&fc->list, &fcoe_hostlist); | |
1335 | write_unlock_bh(&fcoe_hostlist_lock); | |
1336 | } | |
1337 | return 0; | |
1338 | } | |
1339 | EXPORT_SYMBOL_GPL(fcoe_hostlist_add); | |
1340 | ||
1341 | /* | |
1342 | * fcoe_hostlist_remove - remove a lport from lports list | |
1343 | * @lp: ptr to the fc_lport to badded | |
1344 | * | |
1345 | * Returns: 0 for success | |
1346 | */ | |
1347 | int fcoe_hostlist_remove(const struct fc_lport *lp) | |
1348 | { | |
1349 | struct fcoe_softc *fc; | |
1350 | ||
1351 | fc = fcoe_hostlist_lookup_softc(fcoe_netdev(lp)); | |
1352 | BUG_ON(!fc); | |
1353 | write_lock_bh(&fcoe_hostlist_lock); | |
1354 | list_del(&fc->list); | |
1355 | write_unlock_bh(&fcoe_hostlist_lock); | |
1356 | ||
1357 | return 0; | |
1358 | } | |
1359 | EXPORT_SYMBOL_GPL(fcoe_hostlist_remove); | |
1360 | ||
1361 | /** | |
1362 | * fcoe_libfc_config - sets up libfc related properties for lport | |
1363 | * @lp: ptr to the fc_lport | |
1364 | * @tt: libfc function template | |
1365 | * | |
1366 | * Returns : 0 for success | |
1367 | **/ | |
1368 | int fcoe_libfc_config(struct fc_lport *lp, struct libfc_function_template *tt) | |
1369 | { | |
1370 | /* Set the function pointers set by the LLDD */ | |
1371 | memcpy(&lp->tt, tt, sizeof(*tt)); | |
1372 | if (fc_fcp_init(lp)) | |
1373 | return -ENOMEM; | |
1374 | fc_exch_init(lp); | |
1375 | fc_elsct_init(lp); | |
1376 | fc_lport_init(lp); | |
1377 | fc_rport_init(lp); | |
1378 | fc_disc_init(lp); | |
1379 | ||
1380 | return 0; | |
1381 | } | |
1382 | EXPORT_SYMBOL_GPL(fcoe_libfc_config); | |
1383 | ||
1384 | /** | |
1385 | * fcoe_init - fcoe module loading initialization | |
1386 | * | |
1387 | * Initialization routine | |
1388 | * 1. Will create fc transport software structure | |
1389 | * 2. initialize the link list of port information structure | |
1390 | * | |
1391 | * Returns 0 on success, negative on failure | |
1392 | **/ | |
1393 | static int __init fcoe_init(void) | |
1394 | { | |
1395 | int cpu; | |
1396 | struct fcoe_percpu_s *p; | |
1397 | ||
1398 | ||
1399 | INIT_LIST_HEAD(&fcoe_hostlist); | |
1400 | rwlock_init(&fcoe_hostlist_lock); | |
1401 | ||
1402 | #ifdef CONFIG_HOTPLUG_CPU | |
1403 | register_cpu_notifier(&fcoe_cpu_notifier); | |
1404 | #endif /* CONFIG_HOTPLUG_CPU */ | |
1405 | ||
1406 | /* | |
1407 | * initialize per CPU interrupt thread | |
1408 | */ | |
1409 | for_each_online_cpu(cpu) { | |
1410 | p = kzalloc(sizeof(struct fcoe_percpu_s), GFP_KERNEL); | |
1411 | if (p) { | |
1412 | p->thread = kthread_create(fcoe_percpu_receive_thread, | |
1413 | (void *)p, | |
1414 | "fcoethread/%d", cpu); | |
1415 | ||
1416 | /* | |
1417 | * if there is no error then bind the thread to the cpu | |
1418 | * initialize the semaphore and skb queue head | |
1419 | */ | |
1420 | if (likely(!IS_ERR(p->thread))) { | |
1421 | p->cpu = cpu; | |
1422 | fcoe_percpu[cpu] = p; | |
1423 | skb_queue_head_init(&p->fcoe_rx_list); | |
1424 | kthread_bind(p->thread, cpu); | |
1425 | wake_up_process(p->thread); | |
1426 | } else { | |
1427 | fcoe_percpu[cpu] = NULL; | |
1428 | kfree(p); | |
1429 | ||
1430 | } | |
1431 | } | |
1432 | } | |
1433 | ||
1434 | /* | |
1435 | * setup link change notification | |
1436 | */ | |
1437 | fcoe_dev_setup(); | |
1438 | ||
1439 | init_timer(&fcoe_timer); | |
1440 | fcoe_timer.data = 0; | |
1441 | fcoe_timer.function = fcoe_watchdog; | |
1442 | fcoe_timer.expires = (jiffies + (10 * HZ)); | |
1443 | add_timer(&fcoe_timer); | |
1444 | ||
1445 | /* initiatlize the fcoe transport */ | |
1446 | fcoe_transport_init(); | |
1447 | ||
1448 | fcoe_sw_init(); | |
1449 | ||
1450 | return 0; | |
1451 | } | |
1452 | module_init(fcoe_init); | |
1453 | ||
1454 | /** | |
1455 | * fcoe_exit - fcoe module unloading cleanup | |
1456 | * | |
1457 | * Returns 0 on success, negative on failure | |
1458 | **/ | |
1459 | static void __exit fcoe_exit(void) | |
1460 | { | |
1461 | u32 idx; | |
1462 | struct fcoe_softc *fc, *tmp; | |
1463 | struct fcoe_percpu_s *p; | |
1464 | struct sk_buff *skb; | |
1465 | ||
1466 | /* | |
1467 | * Stop all call back interfaces | |
1468 | */ | |
1469 | #ifdef CONFIG_HOTPLUG_CPU | |
1470 | unregister_cpu_notifier(&fcoe_cpu_notifier); | |
1471 | #endif /* CONFIG_HOTPLUG_CPU */ | |
1472 | fcoe_dev_cleanup(); | |
1473 | ||
1474 | /* | |
1475 | * stop timer | |
1476 | */ | |
1477 | del_timer_sync(&fcoe_timer); | |
1478 | ||
1479 | /* releases the assocaited fcoe transport for each lport */ | |
1480 | list_for_each_entry_safe(fc, tmp, &fcoe_hostlist, list) | |
1481 | fcoe_transport_release(fc->real_dev); | |
1482 | ||
1483 | for (idx = 0; idx < NR_CPUS; idx++) { | |
1484 | if (fcoe_percpu[idx]) { | |
1485 | kthread_stop(fcoe_percpu[idx]->thread); | |
1486 | p = fcoe_percpu[idx]; | |
1487 | spin_lock_bh(&p->fcoe_rx_list.lock); | |
1488 | while ((skb = __skb_dequeue(&p->fcoe_rx_list)) != NULL) | |
1489 | kfree_skb(skb); | |
1490 | spin_unlock_bh(&p->fcoe_rx_list.lock); | |
1491 | if (fcoe_percpu[idx]->crc_eof_page) | |
1492 | put_page(fcoe_percpu[idx]->crc_eof_page); | |
1493 | kfree(fcoe_percpu[idx]); | |
1494 | } | |
1495 | } | |
1496 | ||
1497 | /* remove sw trasnport */ | |
1498 | fcoe_sw_exit(); | |
1499 | ||
1500 | /* detach the transport */ | |
1501 | fcoe_transport_exit(); | |
1502 | } | |
1503 | module_exit(fcoe_exit); |