]> git.proxmox.com Git - mirror_ubuntu-artful-kernel.git/blame - drivers/scsi/libfc/fc_lport.c
[SCSI] libfc: add FLOGI state to rport for VN2VN
[mirror_ubuntu-artful-kernel.git] / drivers / scsi / libfc / fc_lport.c
CommitLineData
42e9a92f
RL
1/*
2 * Copyright(c) 2007 Intel Corporation. All rights reserved.
3 *
4 * This program is free software; you can redistribute it and/or modify it
5 * under the terms and conditions of the GNU General Public License,
6 * version 2, as published by the Free Software Foundation.
7 *
8 * This program is distributed in the hope it will be useful, but WITHOUT
9 * ANY WARRANTY; without even the implied warranty of MERCHANTABILITY or
10 * FITNESS FOR A PARTICULAR PURPOSE. See the GNU General Public License for
11 * more details.
12 *
13 * You should have received a copy of the GNU General Public License along with
14 * this program; if not, write to the Free Software Foundation, Inc.,
15 * 51 Franklin St - Fifth Floor, Boston, MA 02110-1301 USA.
16 *
17 * Maintained at www.Open-FCoE.org
18 */
19
20/*
21 * PORT LOCKING NOTES
22 *
23 * These comments only apply to the 'port code' which consists of the lport,
24 * disc and rport blocks.
25 *
26 * MOTIVATION
27 *
28 * The lport, disc and rport blocks all have mutexes that are used to protect
29 * those objects. The main motivation for these locks is to prevent from
30 * having an lport reset just before we send a frame. In that scenario the
31 * lport's FID would get set to zero and then we'd send a frame with an
32 * invalid SID. We also need to ensure that states don't change unexpectedly
33 * while processing another state.
34 *
35 * HEIRARCHY
36 *
37 * The following heirarchy defines the locking rules. A greater lock
38 * may be held before acquiring a lesser lock, but a lesser lock should never
39 * be held while attempting to acquire a greater lock. Here is the heirarchy-
40 *
41 * lport > disc, lport > rport, disc > rport
42 *
43 * CALLBACKS
44 *
45 * The callbacks cause complications with this scheme. There is a callback
46 * from the rport (to either lport or disc) and a callback from disc
47 * (to the lport).
48 *
49 * As rports exit the rport state machine a callback is made to the owner of
50 * the rport to notify success or failure. Since the callback is likely to
51 * cause the lport or disc to grab its lock we cannot hold the rport lock
52 * while making the callback. To ensure that the rport is not free'd while
53 * processing the callback the rport callbacks are serialized through a
54 * single-threaded workqueue. An rport would never be free'd while in a
55 * callback handler becuase no other rport work in this queue can be executed
56 * at the same time.
57 *
58 * When discovery succeeds or fails a callback is made to the lport as
af901ca1 59 * notification. Currently, successful discovery causes the lport to take no
42e9a92f
RL
60 * action. A failure will cause the lport to reset. There is likely a circular
61 * locking problem with this implementation.
62 */
63
64/*
65 * LPORT LOCKING
66 *
67 * The critical sections protected by the lport's mutex are quite broad and
68 * may be improved upon in the future. The lport code and its locking doesn't
69 * influence the I/O path, so excessive locking doesn't penalize I/O
70 * performance.
71 *
72 * The strategy is to lock whenever processing a request or response. Note
73 * that every _enter_* function corresponds to a state change. They generally
74 * change the lports state and then send a request out on the wire. We lock
75 * before calling any of these functions to protect that state change. This
76 * means that the entry points into the lport block manage the locks while
77 * the state machine can transition between states (i.e. _enter_* functions)
78 * while always staying protected.
79 *
80 * When handling responses we also hold the lport mutex broadly. When the
81 * lport receives the response frame it locks the mutex and then calls the
82 * appropriate handler for the particuar response. Generally a response will
83 * trigger a state change and so the lock must already be held.
84 *
85 * Retries also have to consider the locking. The retries occur from a work
86 * context and the work function will lock the lport and then retry the state
87 * (i.e. _enter_* function).
88 */
89
90#include <linux/timer.h>
5a0e3ad6 91#include <linux/slab.h>
42e9a92f
RL
92#include <asm/unaligned.h>
93
94#include <scsi/fc/fc_gs.h>
95
96#include <scsi/libfc.h>
97#include <scsi/fc_encode.h>
a51ab396 98#include <linux/scatterlist.h>
42e9a92f 99
8866a5d9
RL
100#include "fc_libfc.h"
101
42e9a92f
RL
102/* Fabric IDs to use for point-to-point mode, chosen on whims. */
103#define FC_LOCAL_PTP_FID_LO 0x010101
104#define FC_LOCAL_PTP_FID_HI 0x010102
105
106#define DNS_DELAY 3 /* Discovery delay after RSCN (in seconds)*/
107
42e9a92f
RL
108static void fc_lport_error(struct fc_lport *, struct fc_frame *);
109
110static void fc_lport_enter_reset(struct fc_lport *);
111static void fc_lport_enter_flogi(struct fc_lport *);
112static void fc_lport_enter_dns(struct fc_lport *);
c914f7d1 113static void fc_lport_enter_ns(struct fc_lport *, enum fc_lport_state);
42e9a92f
RL
114static void fc_lport_enter_scr(struct fc_lport *);
115static void fc_lport_enter_ready(struct fc_lport *);
116static void fc_lport_enter_logo(struct fc_lport *);
117
118static const char *fc_lport_state_names[] = {
b1d9fd55 119 [LPORT_ST_DISABLED] = "disabled",
42e9a92f
RL
120 [LPORT_ST_FLOGI] = "FLOGI",
121 [LPORT_ST_DNS] = "dNS",
c9c7bd7a 122 [LPORT_ST_RNN_ID] = "RNN_ID",
5baa17c3 123 [LPORT_ST_RSNN_NN] = "RSNN_NN",
c9866a54 124 [LPORT_ST_RSPN_ID] = "RSPN_ID",
42e9a92f 125 [LPORT_ST_RFT_ID] = "RFT_ID",
ab593b18 126 [LPORT_ST_RFF_ID] = "RFF_ID",
42e9a92f
RL
127 [LPORT_ST_SCR] = "SCR",
128 [LPORT_ST_READY] = "Ready",
129 [LPORT_ST_LOGO] = "LOGO",
130 [LPORT_ST_RESET] = "reset",
131};
132
a51ab396
SM
133/**
134 * struct fc_bsg_info - FC Passthrough managemet structure
135 * @job: The passthrough job
136 * @lport: The local port to pass through a command
137 * @rsp_code: The expected response code
3a3b42bf 138 * @sg: job->reply_payload.sg_list
a51ab396
SM
139 * @nents: job->reply_payload.sg_cnt
140 * @offset: The offset into the response data
141 */
142struct fc_bsg_info {
143 struct fc_bsg_job *job;
144 struct fc_lport *lport;
145 u16 rsp_code;
146 struct scatterlist *sg;
147 u32 nents;
148 size_t offset;
149};
150
3a3b42bf
RL
151/**
152 * fc_frame_drop() - Dummy frame handler
153 * @lport: The local port the frame was received on
154 * @fp: The received frame
155 */
42e9a92f
RL
156static int fc_frame_drop(struct fc_lport *lport, struct fc_frame *fp)
157{
158 fc_frame_free(fp);
159 return 0;
160}
161
162/**
34f42a07 163 * fc_lport_rport_callback() - Event handler for rport events
42e9a92f 164 * @lport: The lport which is receiving the event
9fb9d328 165 * @rdata: private remote port data
42e9a92f
RL
166 * @event: The event that occured
167 *
168 * Locking Note: The rport lock should not be held when calling
169 * this function.
170 */
171static void fc_lport_rport_callback(struct fc_lport *lport,
9fb9d328 172 struct fc_rport_priv *rdata,
42e9a92f
RL
173 enum fc_rport_event event)
174{
ce8b5df0 175 FC_LPORT_DBG(lport, "Received a %d event for port (%6.6x)\n", event,
f211fa51 176 rdata->ids.port_id);
42e9a92f 177
b5cbf083 178 mutex_lock(&lport->lp_mutex);
42e9a92f 179 switch (event) {
4c0f62b5 180 case RPORT_EV_READY:
b5cbf083 181 if (lport->state == LPORT_ST_DNS) {
3a3b42bf 182 lport->dns_rdata = rdata;
c914f7d1 183 fc_lport_enter_ns(lport, LPORT_ST_RNN_ID);
b5cbf083
JE
184 } else {
185 FC_LPORT_DBG(lport, "Received an READY event "
ce8b5df0 186 "on port (%6.6x) for the directory "
b5cbf083
JE
187 "server, but the lport is not "
188 "in the DNS state, it's in the "
189 "%d state", rdata->ids.port_id,
190 lport->state);
191 lport->tt.rport_logoff(rdata);
192 }
42e9a92f
RL
193 break;
194 case RPORT_EV_LOGO:
195 case RPORT_EV_FAILED:
196 case RPORT_EV_STOP:
3a3b42bf 197 lport->dns_rdata = NULL;
42e9a92f
RL
198 break;
199 case RPORT_EV_NONE:
200 break;
201 }
b5cbf083 202 mutex_unlock(&lport->lp_mutex);
42e9a92f
RL
203}
204
205/**
34f42a07 206 * fc_lport_state() - Return a string which represents the lport's state
42e9a92f
RL
207 * @lport: The lport whose state is to converted to a string
208 */
209static const char *fc_lport_state(struct fc_lport *lport)
210{
211 const char *cp;
212
213 cp = fc_lport_state_names[lport->state];
214 if (!cp)
215 cp = "unknown";
216 return cp;
217}
218
219/**
34f42a07 220 * fc_lport_ptp_setup() - Create an rport for point-to-point mode
3a3b42bf
RL
221 * @lport: The lport to attach the ptp rport to
222 * @remote_fid: The FID of the ptp rport
42e9a92f
RL
223 * @remote_wwpn: The WWPN of the ptp rport
224 * @remote_wwnn: The WWNN of the ptp rport
225 */
226static void fc_lport_ptp_setup(struct fc_lport *lport,
227 u32 remote_fid, u64 remote_wwpn,
228 u64 remote_wwnn)
229{
48f00902 230 mutex_lock(&lport->disc.disc_mutex);
2f2ac4a0 231 if (lport->ptp_rdata) {
3a3b42bf 232 lport->tt.rport_logoff(lport->ptp_rdata);
2f2ac4a0
JE
233 kref_put(&lport->ptp_rdata->kref, lport->tt.rport_destroy);
234 }
3a3b42bf 235 lport->ptp_rdata = lport->tt.rport_create(lport, remote_fid);
2f2ac4a0 236 kref_get(&lport->ptp_rdata->kref);
3a3b42bf
RL
237 lport->ptp_rdata->ids.port_name = remote_wwpn;
238 lport->ptp_rdata->ids.node_name = remote_wwnn;
48f00902 239 mutex_unlock(&lport->disc.disc_mutex);
42e9a92f 240
3a3b42bf 241 lport->tt.rport_login(lport->ptp_rdata);
42e9a92f
RL
242
243 fc_lport_enter_ready(lport);
244}
245
3a3b42bf
RL
246/**
247 * fc_get_host_port_state() - Return the port state of the given Scsi_Host
248 * @shost: The SCSI host whose port state is to be determined
249 */
42e9a92f
RL
250void fc_get_host_port_state(struct Scsi_Host *shost)
251{
3a3b42bf 252 struct fc_lport *lport = shost_priv(shost);
42e9a92f 253
3a3b42bf
RL
254 mutex_lock(&lport->lp_mutex);
255 if (!lport->link_up)
8faecddb 256 fc_host_port_state(shost) = FC_PORTSTATE_LINKDOWN;
42e9a92f 257 else
3a3b42bf 258 switch (lport->state) {
8faecddb
CL
259 case LPORT_ST_READY:
260 fc_host_port_state(shost) = FC_PORTSTATE_ONLINE;
261 break;
262 default:
263 fc_host_port_state(shost) = FC_PORTSTATE_OFFLINE;
264 }
3a3b42bf 265 mutex_unlock(&lport->lp_mutex);
42e9a92f
RL
266}
267EXPORT_SYMBOL(fc_get_host_port_state);
268
3a3b42bf
RL
269/**
270 * fc_get_host_speed() - Return the speed of the given Scsi_Host
271 * @shost: The SCSI host whose port speed is to be determined
272 */
42e9a92f
RL
273void fc_get_host_speed(struct Scsi_Host *shost)
274{
275 struct fc_lport *lport = shost_priv(shost);
276
277 fc_host_speed(shost) = lport->link_speed;
278}
279EXPORT_SYMBOL(fc_get_host_speed);
280
3a3b42bf
RL
281/**
282 * fc_get_host_stats() - Return the Scsi_Host's statistics
283 * @shost: The SCSI host whose statistics are to be returned
284 */
42e9a92f
RL
285struct fc_host_statistics *fc_get_host_stats(struct Scsi_Host *shost)
286{
42e9a92f 287 struct fc_host_statistics *fcoe_stats;
3a3b42bf 288 struct fc_lport *lport = shost_priv(shost);
42e9a92f 289 struct timespec v0, v1;
582b45bc 290 unsigned int cpu;
42e9a92f 291
3a3b42bf 292 fcoe_stats = &lport->host_stats;
42e9a92f
RL
293 memset(fcoe_stats, 0, sizeof(struct fc_host_statistics));
294
295 jiffies_to_timespec(jiffies, &v0);
3a3b42bf 296 jiffies_to_timespec(lport->boot_time, &v1);
42e9a92f
RL
297 fcoe_stats->seconds_since_last_reset = (v0.tv_sec - v1.tv_sec);
298
582b45bc
RL
299 for_each_possible_cpu(cpu) {
300 struct fcoe_dev_stats *stats;
301
3a3b42bf 302 stats = per_cpu_ptr(lport->dev_stats, cpu);
582b45bc 303
42e9a92f
RL
304 fcoe_stats->tx_frames += stats->TxFrames;
305 fcoe_stats->tx_words += stats->TxWords;
306 fcoe_stats->rx_frames += stats->RxFrames;
307 fcoe_stats->rx_words += stats->RxWords;
308 fcoe_stats->error_frames += stats->ErrorFrames;
309 fcoe_stats->invalid_crc_count += stats->InvalidCRCCount;
310 fcoe_stats->fcp_input_requests += stats->InputRequests;
311 fcoe_stats->fcp_output_requests += stats->OutputRequests;
312 fcoe_stats->fcp_control_requests += stats->ControlRequests;
313 fcoe_stats->fcp_input_megabytes += stats->InputMegabytes;
314 fcoe_stats->fcp_output_megabytes += stats->OutputMegabytes;
315 fcoe_stats->link_failure_count += stats->LinkFailureCount;
316 }
317 fcoe_stats->lip_count = -1;
318 fcoe_stats->nos_count = -1;
319 fcoe_stats->loss_of_sync_count = -1;
320 fcoe_stats->loss_of_signal_count = -1;
321 fcoe_stats->prim_seq_protocol_err_count = -1;
322 fcoe_stats->dumped_frames = -1;
323 return fcoe_stats;
324}
325EXPORT_SYMBOL(fc_get_host_stats);
326
3a3b42bf
RL
327/**
328 * fc_lport_flogi_fill() - Fill in FLOGI command for request
329 * @lport: The local port the FLOGI is for
330 * @flogi: The FLOGI command
331 * @op: The opcode
42e9a92f 332 */
3a3b42bf
RL
333static void fc_lport_flogi_fill(struct fc_lport *lport,
334 struct fc_els_flogi *flogi,
335 unsigned int op)
42e9a92f
RL
336{
337 struct fc_els_csp *sp;
338 struct fc_els_cssp *cp;
339
340 memset(flogi, 0, sizeof(*flogi));
341 flogi->fl_cmd = (u8) op;
342 put_unaligned_be64(lport->wwpn, &flogi->fl_wwpn);
343 put_unaligned_be64(lport->wwnn, &flogi->fl_wwnn);
344 sp = &flogi->fl_csp;
345 sp->sp_hi_ver = 0x20;
346 sp->sp_lo_ver = 0x20;
347 sp->sp_bb_cred = htons(10); /* this gets set by gateway */
348 sp->sp_bb_data = htons((u16) lport->mfs);
349 cp = &flogi->fl_cssp[3 - 1]; /* class 3 parameters */
350 cp->cp_class = htons(FC_CPC_VALID | FC_CPC_SEQ);
351 if (op != ELS_FLOGI) {
352 sp->sp_features = htons(FC_SP_FT_CIRO);
353 sp->sp_tot_seq = htons(255); /* seq. we accept */
354 sp->sp_rel_off = htons(0x1f);
355 sp->sp_e_d_tov = htonl(lport->e_d_tov);
356
357 cp->cp_rdfs = htons((u16) lport->mfs);
358 cp->cp_con_seq = htons(255);
359 cp->cp_open_seq = 1;
360 }
361}
362
3a3b42bf
RL
363/**
364 * fc_lport_add_fc4_type() - Add a supported FC-4 type to a local port
365 * @lport: The local port to add a new FC-4 type to
366 * @type: The new FC-4 type
42e9a92f
RL
367 */
368static void fc_lport_add_fc4_type(struct fc_lport *lport, enum fc_fh_type type)
369{
370 __be32 *mp;
371
372 mp = &lport->fcts.ff_type_map[type / FC_NS_BPW];
373 *mp = htonl(ntohl(*mp) | 1UL << (type % FC_NS_BPW));
374}
375
376/**
34f42a07 377 * fc_lport_recv_rlir_req() - Handle received Registered Link Incident Report.
3a3b42bf
RL
378 * @sp: The sequence in the RLIR exchange
379 * @fp: The RLIR request frame
42e9a92f 380 * @lport: Fibre Channel local port recieving the RLIR
42e9a92f 381 *
1b69bc06 382 * Locking Note: The lport lock is expected to be held before calling
42e9a92f
RL
383 * this function.
384 */
385static void fc_lport_recv_rlir_req(struct fc_seq *sp, struct fc_frame *fp,
386 struct fc_lport *lport)
387{
7414705e
RL
388 FC_LPORT_DBG(lport, "Received RLIR request while in state %s\n",
389 fc_lport_state(lport));
42e9a92f
RL
390
391 lport->tt.seq_els_rsp_send(sp, ELS_LS_ACC, NULL);
392 fc_frame_free(fp);
393}
394
395/**
34f42a07 396 * fc_lport_recv_echo_req() - Handle received ECHO request
3a3b42bf
RL
397 * @sp: The sequence in the ECHO exchange
398 * @fp: ECHO request frame
399 * @lport: The local port recieving the ECHO
42e9a92f 400 *
1b69bc06 401 * Locking Note: The lport lock is expected to be held before calling
42e9a92f
RL
402 * this function.
403 */
404static void fc_lport_recv_echo_req(struct fc_seq *sp, struct fc_frame *in_fp,
405 struct fc_lport *lport)
406{
407 struct fc_frame *fp;
408 struct fc_exch *ep = fc_seq_exch(sp);
409 unsigned int len;
410 void *pp;
411 void *dp;
412 u32 f_ctl;
413
1b69bc06 414 FC_LPORT_DBG(lport, "Received ECHO request while in state %s\n",
7414705e 415 fc_lport_state(lport));
42e9a92f
RL
416
417 len = fr_len(in_fp) - sizeof(struct fc_frame_header);
418 pp = fc_frame_payload_get(in_fp, len);
419
420 if (len < sizeof(__be32))
421 len = sizeof(__be32);
422
423 fp = fc_frame_alloc(lport, len);
424 if (fp) {
425 dp = fc_frame_payload_get(fp, len);
426 memcpy(dp, pp, len);
1b69bc06 427 *((__be32 *)dp) = htonl(ELS_LS_ACC << 24);
42e9a92f
RL
428 sp = lport->tt.seq_start_next(sp);
429 f_ctl = FC_FC_EX_CTX | FC_FC_LAST_SEQ | FC_FC_END_SEQ;
430 fc_fill_fc_hdr(fp, FC_RCTL_ELS_REP, ep->did, ep->sid,
431 FC_TYPE_ELS, f_ctl, 0);
432 lport->tt.seq_send(lport, sp, fp);
433 }
434 fc_frame_free(in_fp);
435}
436
437/**
1b69bc06
JE
438 * fc_lport_recv_rnid_req() - Handle received Request Node ID data request
439 * @sp: The sequence in the RNID exchange
440 * @fp: The RNID request frame
441 * @lport: The local port recieving the RNID
42e9a92f 442 *
1b69bc06 443 * Locking Note: The lport lock is expected to be held before calling
42e9a92f
RL
444 * this function.
445 */
446static void fc_lport_recv_rnid_req(struct fc_seq *sp, struct fc_frame *in_fp,
447 struct fc_lport *lport)
448{
449 struct fc_frame *fp;
450 struct fc_exch *ep = fc_seq_exch(sp);
451 struct fc_els_rnid *req;
452 struct {
453 struct fc_els_rnid_resp rnid;
454 struct fc_els_rnid_cid cid;
455 struct fc_els_rnid_gen gen;
456 } *rp;
457 struct fc_seq_els_data rjt_data;
458 u8 fmt;
459 size_t len;
460 u32 f_ctl;
461
7414705e
RL
462 FC_LPORT_DBG(lport, "Received RNID request while in state %s\n",
463 fc_lport_state(lport));
42e9a92f
RL
464
465 req = fc_frame_payload_get(in_fp, sizeof(*req));
466 if (!req) {
467 rjt_data.fp = NULL;
468 rjt_data.reason = ELS_RJT_LOGIC;
469 rjt_data.explan = ELS_EXPL_NONE;
470 lport->tt.seq_els_rsp_send(sp, ELS_LS_RJT, &rjt_data);
471 } else {
472 fmt = req->rnid_fmt;
473 len = sizeof(*rp);
474 if (fmt != ELS_RNIDF_GEN ||
475 ntohl(lport->rnid_gen.rnid_atype) == 0) {
476 fmt = ELS_RNIDF_NONE; /* nothing to provide */
477 len -= sizeof(rp->gen);
478 }
479 fp = fc_frame_alloc(lport, len);
480 if (fp) {
481 rp = fc_frame_payload_get(fp, len);
482 memset(rp, 0, len);
483 rp->rnid.rnid_cmd = ELS_LS_ACC;
484 rp->rnid.rnid_fmt = fmt;
485 rp->rnid.rnid_cid_len = sizeof(rp->cid);
486 rp->cid.rnid_wwpn = htonll(lport->wwpn);
487 rp->cid.rnid_wwnn = htonll(lport->wwnn);
488 if (fmt == ELS_RNIDF_GEN) {
489 rp->rnid.rnid_sid_len = sizeof(rp->gen);
490 memcpy(&rp->gen, &lport->rnid_gen,
491 sizeof(rp->gen));
492 }
493 sp = lport->tt.seq_start_next(sp);
494 f_ctl = FC_FC_EX_CTX | FC_FC_LAST_SEQ;
495 f_ctl |= FC_FC_END_SEQ | FC_FC_SEQ_INIT;
496 fc_fill_fc_hdr(fp, FC_RCTL_ELS_REP, ep->did, ep->sid,
497 FC_TYPE_ELS, f_ctl, 0);
498 lport->tt.seq_send(lport, sp, fp);
499 }
500 }
501 fc_frame_free(in_fp);
502}
503
42e9a92f 504/**
34f42a07 505 * fc_lport_recv_logo_req() - Handle received fabric LOGO request
3a3b42bf
RL
506 * @sp: The sequence in the LOGO exchange
507 * @fp: The LOGO request frame
508 * @lport: The local port recieving the LOGO
42e9a92f
RL
509 *
510 * Locking Note: The lport lock is exected to be held before calling
511 * this function.
512 */
513static void fc_lport_recv_logo_req(struct fc_seq *sp, struct fc_frame *fp,
514 struct fc_lport *lport)
515{
516 lport->tt.seq_els_rsp_send(sp, ELS_LS_ACC, NULL);
517 fc_lport_enter_reset(lport);
518 fc_frame_free(fp);
519}
520
521/**
34f42a07 522 * fc_fabric_login() - Start the lport state machine
3a3b42bf 523 * @lport: The local port that should log into the fabric
42e9a92f
RL
524 *
525 * Locking Note: This function should not be called
526 * with the lport lock held.
527 */
528int fc_fabric_login(struct fc_lport *lport)
529{
530 int rc = -1;
531
532 mutex_lock(&lport->lp_mutex);
55a66d3c
VD
533 if (lport->state == LPORT_ST_DISABLED ||
534 lport->state == LPORT_ST_LOGO) {
535 fc_lport_state_enter(lport, LPORT_ST_RESET);
42e9a92f
RL
536 fc_lport_enter_reset(lport);
537 rc = 0;
538 }
539 mutex_unlock(&lport->lp_mutex);
540
541 return rc;
542}
543EXPORT_SYMBOL(fc_fabric_login);
544
545/**
8faecddb 546 * __fc_linkup() - Handler for transport linkup events
42e9a92f 547 * @lport: The lport whose link is up
8faecddb
CL
548 *
549 * Locking: must be called with the lp_mutex held
42e9a92f 550 */
8faecddb 551void __fc_linkup(struct fc_lport *lport)
42e9a92f 552{
bc0e17f6
VD
553 if (!lport->link_up) {
554 lport->link_up = 1;
42e9a92f
RL
555
556 if (lport->state == LPORT_ST_RESET)
557 fc_lport_enter_flogi(lport);
558 }
8faecddb
CL
559}
560
561/**
562 * fc_linkup() - Handler for transport linkup events
3a3b42bf 563 * @lport: The local port whose link is up
8faecddb
CL
564 */
565void fc_linkup(struct fc_lport *lport)
566{
ce8b5df0 567 printk(KERN_INFO "host%d: libfc: Link up on port (%6.6x)\n",
7b2787ec 568 lport->host->host_no, lport->port_id);
8faecddb
CL
569
570 mutex_lock(&lport->lp_mutex);
571 __fc_linkup(lport);
42e9a92f
RL
572 mutex_unlock(&lport->lp_mutex);
573}
574EXPORT_SYMBOL(fc_linkup);
575
576/**
8faecddb 577 * __fc_linkdown() - Handler for transport linkdown events
42e9a92f 578 * @lport: The lport whose link is down
8faecddb
CL
579 *
580 * Locking: must be called with the lp_mutex held
42e9a92f 581 */
8faecddb 582void __fc_linkdown(struct fc_lport *lport)
42e9a92f 583{
bc0e17f6
VD
584 if (lport->link_up) {
585 lport->link_up = 0;
42e9a92f
RL
586 fc_lport_enter_reset(lport);
587 lport->tt.fcp_cleanup(lport);
588 }
8faecddb
CL
589}
590
591/**
592 * fc_linkdown() - Handler for transport linkdown events
3a3b42bf 593 * @lport: The local port whose link is down
8faecddb
CL
594 */
595void fc_linkdown(struct fc_lport *lport)
596{
ce8b5df0 597 printk(KERN_INFO "host%d: libfc: Link down on port (%6.6x)\n",
7b2787ec 598 lport->host->host_no, lport->port_id);
8faecddb
CL
599
600 mutex_lock(&lport->lp_mutex);
601 __fc_linkdown(lport);
42e9a92f
RL
602 mutex_unlock(&lport->lp_mutex);
603}
604EXPORT_SYMBOL(fc_linkdown);
605
42e9a92f 606/**
34f42a07 607 * fc_fabric_logoff() - Logout of the fabric
3a3b42bf 608 * @lport: The local port to logoff the fabric
42e9a92f
RL
609 *
610 * Return value:
611 * 0 for success, -1 for failure
34f42a07 612 */
42e9a92f
RL
613int fc_fabric_logoff(struct fc_lport *lport)
614{
615 lport->tt.disc_stop_final(lport);
616 mutex_lock(&lport->lp_mutex);
3a3b42bf
RL
617 if (lport->dns_rdata)
618 lport->tt.rport_logoff(lport->dns_rdata);
a0fd2e49
AJ
619 mutex_unlock(&lport->lp_mutex);
620 lport->tt.rport_flush_queue();
621 mutex_lock(&lport->lp_mutex);
42e9a92f
RL
622 fc_lport_enter_logo(lport);
623 mutex_unlock(&lport->lp_mutex);
f7db2c15 624 cancel_delayed_work_sync(&lport->retry_work);
42e9a92f
RL
625 return 0;
626}
627EXPORT_SYMBOL(fc_fabric_logoff);
628
629/**
3a3b42bf
RL
630 * fc_lport_destroy() - Unregister a fc_lport
631 * @lport: The local port to unregister
42e9a92f 632 *
42e9a92f
RL
633 * Note:
634 * exit routine for fc_lport instance
635 * clean-up all the allocated memory
636 * and free up other system resources.
637 *
34f42a07 638 */
42e9a92f
RL
639int fc_lport_destroy(struct fc_lport *lport)
640{
bbf15669 641 mutex_lock(&lport->lp_mutex);
b1d9fd55 642 lport->state = LPORT_ST_DISABLED;
bbf15669 643 lport->link_up = 0;
42e9a92f 644 lport->tt.frame_send = fc_frame_drop;
bbf15669
AJ
645 mutex_unlock(&lport->lp_mutex);
646
42e9a92f 647 lport->tt.fcp_abort_io(lport);
e9ba8b42 648 lport->tt.disc_stop_final(lport);
1f6ff364 649 lport->tt.exch_mgr_reset(lport, 0, 0);
42e9a92f
RL
650 return 0;
651}
652EXPORT_SYMBOL(fc_lport_destroy);
653
654/**
3a3b42bf
RL
655 * fc_set_mfs() - Set the maximum frame size for a local port
656 * @lport: The local port to set the MFS for
657 * @mfs: The new MFS
34f42a07 658 */
42e9a92f
RL
659int fc_set_mfs(struct fc_lport *lport, u32 mfs)
660{
661 unsigned int old_mfs;
662 int rc = -EINVAL;
663
664 mutex_lock(&lport->lp_mutex);
665
666 old_mfs = lport->mfs;
667
668 if (mfs >= FC_MIN_MAX_FRAME) {
669 mfs &= ~3;
670 if (mfs > FC_MAX_FRAME)
671 mfs = FC_MAX_FRAME;
672 mfs -= sizeof(struct fc_frame_header);
673 lport->mfs = mfs;
674 rc = 0;
675 }
676
677 if (!rc && mfs < old_mfs)
678 fc_lport_enter_reset(lport);
679
680 mutex_unlock(&lport->lp_mutex);
681
682 return rc;
683}
684EXPORT_SYMBOL(fc_set_mfs);
685
686/**
34f42a07 687 * fc_lport_disc_callback() - Callback for discovery events
3a3b42bf 688 * @lport: The local port receiving the event
42e9a92f
RL
689 * @event: The discovery event
690 */
691void fc_lport_disc_callback(struct fc_lport *lport, enum fc_disc_event event)
692{
693 switch (event) {
694 case DISC_EV_SUCCESS:
7414705e 695 FC_LPORT_DBG(lport, "Discovery succeeded\n");
42e9a92f
RL
696 break;
697 case DISC_EV_FAILED:
e6d8a1b0 698 printk(KERN_ERR "host%d: libfc: "
ce8b5df0 699 "Discovery failed for port (%6.6x)\n",
7b2787ec 700 lport->host->host_no, lport->port_id);
42e9a92f
RL
701 mutex_lock(&lport->lp_mutex);
702 fc_lport_enter_reset(lport);
703 mutex_unlock(&lport->lp_mutex);
704 break;
705 case DISC_EV_NONE:
706 WARN_ON(1);
707 break;
708 }
709}
710
711/**
34f42a07 712 * fc_rport_enter_ready() - Enter the ready state and start discovery
3a3b42bf 713 * @lport: The local port that is ready
42e9a92f
RL
714 *
715 * Locking Note: The lport lock is expected to be held before calling
716 * this routine.
717 */
718static void fc_lport_enter_ready(struct fc_lport *lport)
719{
7414705e
RL
720 FC_LPORT_DBG(lport, "Entered READY from state %s\n",
721 fc_lport_state(lport));
42e9a92f
RL
722
723 fc_lport_state_enter(lport, LPORT_ST_READY);
8faecddb
CL
724 if (lport->vport)
725 fc_vport_set_state(lport->vport, FC_VPORT_ACTIVE);
726 fc_vports_linkchange(lport);
42e9a92f 727
3a3b42bf 728 if (!lport->ptp_rdata)
29d898e9 729 lport->tt.disc_start(fc_lport_disc_callback, lport);
42e9a92f
RL
730}
731
093bb6a2
JE
732/**
733 * fc_lport_set_port_id() - set the local port Port ID
734 * @lport: The local port which will have its Port ID set.
735 * @port_id: The new port ID.
736 * @fp: The frame containing the incoming request, or NULL.
737 *
738 * Locking Note: The lport lock is expected to be held before calling
739 * this function.
740 */
741static void fc_lport_set_port_id(struct fc_lport *lport, u32 port_id,
742 struct fc_frame *fp)
743{
744 if (port_id)
ce8b5df0 745 printk(KERN_INFO "host%d: Assigned Port ID %6.6x\n",
093bb6a2
JE
746 lport->host->host_no, port_id);
747
7b2787ec
RL
748 lport->port_id = port_id;
749
750 /* Update the fc_host */
093bb6a2 751 fc_host_port_id(lport->host) = port_id;
7b2787ec 752
093bb6a2
JE
753 if (lport->tt.lport_set_port_id)
754 lport->tt.lport_set_port_id(lport, port_id, fp);
755}
756
3726f358
JE
757/**
758 * fc_lport_set_port_id() - set the local port Port ID for point-to-multipoint
759 * @lport: The local port which will have its Port ID set.
760 * @port_id: The new port ID.
761 *
762 * Called by the lower-level driver when transport sets the local port_id.
763 * This is used in VN_port to VN_port mode for FCoE, and causes FLOGI and
764 * discovery to be skipped.
765 */
766void fc_lport_set_local_id(struct fc_lport *lport, u32 port_id)
767{
768 mutex_lock(&lport->lp_mutex);
769
770 fc_lport_set_port_id(lport, port_id, NULL);
771
772 switch (lport->state) {
773 case LPORT_ST_RESET:
774 case LPORT_ST_FLOGI:
775 if (port_id)
776 fc_lport_enter_ready(lport);
777 break;
778 default:
779 break;
780 }
781 mutex_unlock(&lport->lp_mutex);
782}
783EXPORT_SYMBOL(fc_lport_set_local_id);
784
42e9a92f 785/**
34f42a07 786 * fc_lport_recv_flogi_req() - Receive a FLOGI request
42e9a92f 787 * @sp_in: The sequence the FLOGI is on
3a3b42bf
RL
788 * @rx_fp: The FLOGI frame
789 * @lport: The local port that recieved the request
42e9a92f
RL
790 *
791 * A received FLOGI request indicates a point-to-point connection.
792 * Accept it with the common service parameters indicating our N port.
793 * Set up to do a PLOGI if we have the higher-number WWPN.
794 *
1b69bc06 795 * Locking Note: The lport lock is expected to be held before calling
42e9a92f
RL
796 * this function.
797 */
798static void fc_lport_recv_flogi_req(struct fc_seq *sp_in,
799 struct fc_frame *rx_fp,
800 struct fc_lport *lport)
801{
802 struct fc_frame *fp;
803 struct fc_frame_header *fh;
804 struct fc_seq *sp;
805 struct fc_exch *ep;
806 struct fc_els_flogi *flp;
807 struct fc_els_flogi *new_flp;
808 u64 remote_wwpn;
809 u32 remote_fid;
810 u32 local_fid;
811 u32 f_ctl;
812
7414705e
RL
813 FC_LPORT_DBG(lport, "Received FLOGI request while in state %s\n",
814 fc_lport_state(lport));
42e9a92f
RL
815
816 fh = fc_frame_header_get(rx_fp);
817 remote_fid = ntoh24(fh->fh_s_id);
818 flp = fc_frame_payload_get(rx_fp, sizeof(*flp));
819 if (!flp)
820 goto out;
821 remote_wwpn = get_unaligned_be64(&flp->fl_wwpn);
822 if (remote_wwpn == lport->wwpn) {
e6d8a1b0 823 printk(KERN_WARNING "host%d: libfc: Received FLOGI from port "
9f8f3aa6 824 "with same WWPN %16.16llx\n",
e6d8a1b0 825 lport->host->host_no, remote_wwpn);
42e9a92f
RL
826 goto out;
827 }
9f8f3aa6 828 FC_LPORT_DBG(lport, "FLOGI from port WWPN %16.16llx\n", remote_wwpn);
42e9a92f
RL
829
830 /*
831 * XXX what is the right thing to do for FIDs?
832 * The originator might expect our S_ID to be 0xfffffe.
833 * But if so, both of us could end up with the same FID.
834 */
835 local_fid = FC_LOCAL_PTP_FID_LO;
836 if (remote_wwpn < lport->wwpn) {
837 local_fid = FC_LOCAL_PTP_FID_HI;
838 if (!remote_fid || remote_fid == local_fid)
839 remote_fid = FC_LOCAL_PTP_FID_LO;
840 } else if (!remote_fid) {
841 remote_fid = FC_LOCAL_PTP_FID_HI;
842 }
843
093bb6a2 844 fc_lport_set_port_id(lport, local_fid, rx_fp);
42e9a92f
RL
845
846 fp = fc_frame_alloc(lport, sizeof(*flp));
847 if (fp) {
848 sp = lport->tt.seq_start_next(fr_seq(rx_fp));
849 new_flp = fc_frame_payload_get(fp, sizeof(*flp));
850 fc_lport_flogi_fill(lport, new_flp, ELS_FLOGI);
851 new_flp->fl_cmd = (u8) ELS_LS_ACC;
852
853 /*
854 * Send the response. If this fails, the originator should
855 * repeat the sequence.
856 */
857 f_ctl = FC_FC_EX_CTX | FC_FC_LAST_SEQ | FC_FC_END_SEQ;
858 ep = fc_seq_exch(sp);
ccfc3098 859 fc_fill_fc_hdr(fp, FC_RCTL_ELS_REP, remote_fid, local_fid,
42e9a92f
RL
860 FC_TYPE_ELS, f_ctl, 0);
861 lport->tt.seq_send(lport, sp, fp);
862
863 } else {
864 fc_lport_error(lport, fp);
865 }
866 fc_lport_ptp_setup(lport, remote_fid, remote_wwpn,
867 get_unaligned_be64(&flp->fl_wwnn));
868
42e9a92f
RL
869out:
870 sp = fr_seq(rx_fp);
871 fc_frame_free(rx_fp);
872}
873
874/**
34f42a07 875 * fc_lport_recv_req() - The generic lport request handler
3a3b42bf
RL
876 * @lport: The local port that received the request
877 * @sp: The sequence the request is on
878 * @fp: The request frame
42e9a92f
RL
879 *
880 * This function will see if the lport handles the request or
881 * if an rport should handle the request.
882 *
883 * Locking Note: This function should not be called with the lport
884 * lock held becuase it will grab the lock.
885 */
886static void fc_lport_recv_req(struct fc_lport *lport, struct fc_seq *sp,
887 struct fc_frame *fp)
888{
889 struct fc_frame_header *fh = fc_frame_header_get(fp);
890 void (*recv) (struct fc_seq *, struct fc_frame *, struct fc_lport *);
42e9a92f
RL
891
892 mutex_lock(&lport->lp_mutex);
893
894 /*
895 * Handle special ELS cases like FLOGI, LOGO, and
896 * RSCN here. These don't require a session.
897 * Even if we had a session, it might not be ready.
898 */
e9ba8b42
JE
899 if (!lport->link_up)
900 fc_frame_free(fp);
901 else if (fh->fh_type == FC_TYPE_ELS &&
902 fh->fh_r_ctl == FC_RCTL_ELS_REQ) {
42e9a92f
RL
903 /*
904 * Check opcode.
905 */
131203a1 906 recv = lport->tt.rport_recv_req;
42e9a92f
RL
907 switch (fc_frame_payload_op(fp)) {
908 case ELS_FLOGI:
a7b12a27
JE
909 if (!lport->point_to_multipoint)
910 recv = fc_lport_recv_flogi_req;
42e9a92f
RL
911 break;
912 case ELS_LOGO:
42e9a92f
RL
913 if (ntoh24(fh->fh_s_id) == FC_FID_FLOGI)
914 recv = fc_lport_recv_logo_req;
915 break;
916 case ELS_RSCN:
917 recv = lport->tt.disc_recv_req;
918 break;
919 case ELS_ECHO:
920 recv = fc_lport_recv_echo_req;
921 break;
922 case ELS_RLIR:
923 recv = fc_lport_recv_rlir_req;
924 break;
925 case ELS_RNID:
926 recv = fc_lport_recv_rnid_req;
927 break;
42e9a92f
RL
928 }
929
131203a1 930 recv(sp, fp, lport);
42e9a92f 931 } else {
7414705e
RL
932 FC_LPORT_DBG(lport, "dropping invalid frame (eof %x)\n",
933 fr_eof(fp));
42e9a92f
RL
934 fc_frame_free(fp);
935 }
936 mutex_unlock(&lport->lp_mutex);
937
938 /*
939 * The common exch_done for all request may not be good
940 * if any request requires longer hold on exhange. XXX
941 */
942 lport->tt.exch_done(sp);
943}
944
945/**
3a3b42bf
RL
946 * fc_lport_reset() - Reset a local port
947 * @lport: The local port which should be reset
42e9a92f
RL
948 *
949 * Locking Note: This functions should not be called with the
950 * lport lock held.
951 */
952int fc_lport_reset(struct fc_lport *lport)
953{
f7db2c15 954 cancel_delayed_work_sync(&lport->retry_work);
42e9a92f
RL
955 mutex_lock(&lport->lp_mutex);
956 fc_lport_enter_reset(lport);
957 mutex_unlock(&lport->lp_mutex);
958 return 0;
959}
960EXPORT_SYMBOL(fc_lport_reset);
961
962/**
3a3b42bf
RL
963 * fc_lport_reset_locked() - Reset the local port w/ the lport lock held
964 * @lport: The local port to be reset
42e9a92f
RL
965 *
966 * Locking Note: The lport lock is expected to be held before calling
967 * this routine.
968 */
1190d925 969static void fc_lport_reset_locked(struct fc_lport *lport)
42e9a92f 970{
3a3b42bf
RL
971 if (lport->dns_rdata)
972 lport->tt.rport_logoff(lport->dns_rdata);
42e9a92f 973
2f2ac4a0
JE
974 if (lport->ptp_rdata) {
975 lport->tt.rport_logoff(lport->ptp_rdata);
976 kref_put(&lport->ptp_rdata->kref, lport->tt.rport_destroy);
977 lport->ptp_rdata = NULL;
978 }
42e9a92f
RL
979
980 lport->tt.disc_stop(lport);
981
1f6ff364 982 lport->tt.exch_mgr_reset(lport, 0, 0);
42e9a92f 983 fc_host_fabric_name(lport->host) = 0;
093bb6a2 984
3726f358 985 if (lport->port_id && (!lport->point_to_multipoint || !lport->link_up))
093bb6a2 986 fc_lport_set_port_id(lport, 0, NULL);
1190d925 987}
42e9a92f 988
1190d925
JE
989/**
990 * fc_lport_enter_reset() - Reset the local port
3a3b42bf 991 * @lport: The local port to be reset
1190d925
JE
992 *
993 * Locking Note: The lport lock is expected to be held before calling
994 * this routine.
995 */
996static void fc_lport_enter_reset(struct fc_lport *lport)
997{
998 FC_LPORT_DBG(lport, "Entered RESET state from %s state\n",
999 fc_lport_state(lport));
1000
55a66d3c
VD
1001 if (lport->state == LPORT_ST_DISABLED || lport->state == LPORT_ST_LOGO)
1002 return;
1003
8faecddb
CL
1004 if (lport->vport) {
1005 if (lport->link_up)
1006 fc_vport_set_state(lport->vport, FC_VPORT_INITIALIZING);
1007 else
1008 fc_vport_set_state(lport->vport, FC_VPORT_LINKDOWN);
1009 }
1190d925 1010 fc_lport_state_enter(lport, LPORT_ST_RESET);
8faecddb 1011 fc_vports_linkchange(lport);
1190d925 1012 fc_lport_reset_locked(lport);
bc0e17f6 1013 if (lport->link_up)
42e9a92f
RL
1014 fc_lport_enter_flogi(lport);
1015}
1016
1190d925 1017/**
3a3b42bf
RL
1018 * fc_lport_enter_disabled() - Disable the local port
1019 * @lport: The local port to be reset
1190d925
JE
1020 *
1021 * Locking Note: The lport lock is expected to be held before calling
1022 * this routine.
1023 */
1024static void fc_lport_enter_disabled(struct fc_lport *lport)
1025{
1026 FC_LPORT_DBG(lport, "Entered disabled state from %s state\n",
1027 fc_lport_state(lport));
1028
1029 fc_lport_state_enter(lport, LPORT_ST_DISABLED);
8faecddb 1030 fc_vports_linkchange(lport);
1190d925
JE
1031 fc_lport_reset_locked(lport);
1032}
1033
42e9a92f 1034/**
34f42a07 1035 * fc_lport_error() - Handler for any errors
3a3b42bf
RL
1036 * @lport: The local port that the error was on
1037 * @fp: The error code encoded in a frame pointer
42e9a92f
RL
1038 *
1039 * If the error was caused by a resource allocation failure
1040 * then wait for half a second and retry, otherwise retry
1041 * after the e_d_tov time.
1042 */
1043static void fc_lport_error(struct fc_lport *lport, struct fc_frame *fp)
1044{
1045 unsigned long delay = 0;
7414705e
RL
1046 FC_LPORT_DBG(lport, "Error %ld in state %s, retries %d\n",
1047 PTR_ERR(fp), fc_lport_state(lport),
1048 lport->retry_count);
42e9a92f
RL
1049
1050 if (!fp || PTR_ERR(fp) == -FC_EX_TIMEOUT) {
1051 /*
1052 * Memory allocation failure, or the exchange timed out.
1053 * Retry after delay
1054 */
1055 if (lport->retry_count < lport->max_retry_count) {
1056 lport->retry_count++;
1057 if (!fp)
1058 delay = msecs_to_jiffies(500);
1059 else
1060 delay = msecs_to_jiffies(lport->e_d_tov);
1061
1062 schedule_delayed_work(&lport->retry_work, delay);
1063 } else {
1064 switch (lport->state) {
b1d9fd55 1065 case LPORT_ST_DISABLED:
42e9a92f
RL
1066 case LPORT_ST_READY:
1067 case LPORT_ST_RESET:
c9c7bd7a 1068 case LPORT_ST_RNN_ID:
5baa17c3 1069 case LPORT_ST_RSNN_NN:
c9866a54 1070 case LPORT_ST_RSPN_ID:
42e9a92f 1071 case LPORT_ST_RFT_ID:
ab593b18 1072 case LPORT_ST_RFF_ID:
42e9a92f
RL
1073 case LPORT_ST_SCR:
1074 case LPORT_ST_DNS:
1075 case LPORT_ST_FLOGI:
1076 case LPORT_ST_LOGO:
1077 fc_lport_enter_reset(lport);
1078 break;
1079 }
1080 }
1081 }
1082}
1083
1084/**
7cccc157 1085 * fc_lport_ns_resp() - Handle response to a name server
3a3b42bf
RL
1086 * registration exchange
1087 * @sp: current sequence in exchange
1088 * @fp: response frame
42e9a92f
RL
1089 * @lp_arg: Fibre Channel host port instance
1090 *
1091 * Locking Note: This function will be called without the lport lock
3a3b42bf 1092 * held, but it will lock, call an _enter_* function or fc_lport_error()
42e9a92f
RL
1093 * and then unlock the lport.
1094 */
7cccc157
CL
1095static void fc_lport_ns_resp(struct fc_seq *sp, struct fc_frame *fp,
1096 void *lp_arg)
42e9a92f
RL
1097{
1098 struct fc_lport *lport = lp_arg;
1099 struct fc_frame_header *fh;
1100 struct fc_ct_hdr *ct;
1101
7cccc157 1102 FC_LPORT_DBG(lport, "Received a ns %s\n", fc_els_resp_type(fp));
f657d299 1103
42e9a92f
RL
1104 if (fp == ERR_PTR(-FC_EX_CLOSED))
1105 return;
1106
1107 mutex_lock(&lport->lp_mutex);
1108
ab593b18 1109 if (lport->state < LPORT_ST_RNN_ID || lport->state > LPORT_ST_RFF_ID) {
7cccc157 1110 FC_LPORT_DBG(lport, "Received a name server response, "
3a3b42bf 1111 "but in state %s\n", fc_lport_state(lport));
76f6804e
AJ
1112 if (IS_ERR(fp))
1113 goto err;
42e9a92f
RL
1114 goto out;
1115 }
1116
76f6804e
AJ
1117 if (IS_ERR(fp)) {
1118 fc_lport_error(lport, fp);
1119 goto err;
1120 }
1121
42e9a92f
RL
1122 fh = fc_frame_header_get(fp);
1123 ct = fc_frame_payload_get(fp, sizeof(*ct));
1124
1125 if (fh && ct && fh->fh_type == FC_TYPE_CT &&
1126 ct->ct_fs_type == FC_FST_DIR &&
1127 ct->ct_fs_subtype == FC_NS_SUBTYPE &&
1128 ntohs(ct->ct_cmd) == FC_FS_ACC)
7cccc157
CL
1129 switch (lport->state) {
1130 case LPORT_ST_RNN_ID:
c914f7d1 1131 fc_lport_enter_ns(lport, LPORT_ST_RSNN_NN);
7cccc157
CL
1132 break;
1133 case LPORT_ST_RSNN_NN:
c914f7d1 1134 fc_lport_enter_ns(lport, LPORT_ST_RSPN_ID);
7cccc157
CL
1135 break;
1136 case LPORT_ST_RSPN_ID:
c914f7d1 1137 fc_lport_enter_ns(lport, LPORT_ST_RFT_ID);
7cccc157
CL
1138 break;
1139 case LPORT_ST_RFT_ID:
ab593b18
JE
1140 fc_lport_enter_ns(lport, LPORT_ST_RFF_ID);
1141 break;
1142 case LPORT_ST_RFF_ID:
7cccc157
CL
1143 fc_lport_enter_scr(lport);
1144 break;
1145 default:
1146 /* should have already been caught by state checks */
1147 break;
1148 }
c9c7bd7a
CL
1149 else
1150 fc_lport_error(lport, fp);
c9c7bd7a
CL
1151out:
1152 fc_frame_free(fp);
1153err:
1154 mutex_unlock(&lport->lp_mutex);
1155}
1156
42e9a92f 1157/**
34f42a07 1158 * fc_lport_scr_resp() - Handle response to State Change Register (SCR) request
3a3b42bf
RL
1159 * @sp: current sequence in SCR exchange
1160 * @fp: response frame
42e9a92f
RL
1161 * @lp_arg: Fibre Channel lport port instance that sent the registration request
1162 *
1163 * Locking Note: This function will be called without the lport lock
1164 * held, but it will lock, call an _enter_* function or fc_lport_error
1165 * and then unlock the lport.
1166 */
1167static void fc_lport_scr_resp(struct fc_seq *sp, struct fc_frame *fp,
1168 void *lp_arg)
1169{
1170 struct fc_lport *lport = lp_arg;
1171 u8 op;
1172
f657d299
JE
1173 FC_LPORT_DBG(lport, "Received a SCR %s\n", fc_els_resp_type(fp));
1174
42e9a92f
RL
1175 if (fp == ERR_PTR(-FC_EX_CLOSED))
1176 return;
1177
1178 mutex_lock(&lport->lp_mutex);
1179
42e9a92f 1180 if (lport->state != LPORT_ST_SCR) {
7414705e
RL
1181 FC_LPORT_DBG(lport, "Received a SCR response, but in state "
1182 "%s\n", fc_lport_state(lport));
76f6804e
AJ
1183 if (IS_ERR(fp))
1184 goto err;
42e9a92f
RL
1185 goto out;
1186 }
1187
76f6804e
AJ
1188 if (IS_ERR(fp)) {
1189 fc_lport_error(lport, fp);
1190 goto err;
1191 }
1192
42e9a92f
RL
1193 op = fc_frame_payload_op(fp);
1194 if (op == ELS_LS_ACC)
1195 fc_lport_enter_ready(lport);
1196 else
1197 fc_lport_error(lport, fp);
1198
1199out:
1200 fc_frame_free(fp);
1201err:
1202 mutex_unlock(&lport->lp_mutex);
1203}
1204
1205/**
3a3b42bf
RL
1206 * fc_lport_enter_scr() - Send a SCR (State Change Register) request
1207 * @lport: The local port to register for state changes
42e9a92f
RL
1208 *
1209 * Locking Note: The lport lock is expected to be held before calling
1210 * this routine.
1211 */
1212static void fc_lport_enter_scr(struct fc_lport *lport)
1213{
1214 struct fc_frame *fp;
1215
7414705e
RL
1216 FC_LPORT_DBG(lport, "Entered SCR state from %s state\n",
1217 fc_lport_state(lport));
42e9a92f
RL
1218
1219 fc_lport_state_enter(lport, LPORT_ST_SCR);
1220
1221 fp = fc_frame_alloc(lport, sizeof(struct fc_els_scr));
1222 if (!fp) {
1223 fc_lport_error(lport, fp);
1224 return;
1225 }
1226
a46f327a 1227 if (!lport->tt.elsct_send(lport, FC_FID_FCTRL, fp, ELS_SCR,
b94f8951
JE
1228 fc_lport_scr_resp, lport,
1229 2 * lport->r_a_tov))
8f550f93 1230 fc_lport_error(lport, NULL);
42e9a92f
RL
1231}
1232
1233/**
c914f7d1 1234 * fc_lport_enter_ns() - register some object with the name server
42e9a92f
RL
1235 * @lport: Fibre Channel local port to register
1236 *
1237 * Locking Note: The lport lock is expected to be held before calling
1238 * this routine.
1239 */
c914f7d1 1240static void fc_lport_enter_ns(struct fc_lport *lport, enum fc_lport_state state)
c9866a54
CL
1241{
1242 struct fc_frame *fp;
c914f7d1
CL
1243 enum fc_ns_req cmd;
1244 int size = sizeof(struct fc_ct_hdr);
c9866a54
CL
1245 size_t len;
1246
c914f7d1
CL
1247 FC_LPORT_DBG(lport, "Entered %s state from %s state\n",
1248 fc_lport_state_names[state],
c9866a54
CL
1249 fc_lport_state(lport));
1250
c914f7d1 1251 fc_lport_state_enter(lport, state);
c9866a54 1252
c914f7d1
CL
1253 switch (state) {
1254 case LPORT_ST_RNN_ID:
1255 cmd = FC_NS_RNN_ID;
1256 size += sizeof(struct fc_ns_rn_id);
1257 break;
1258 case LPORT_ST_RSNN_NN:
1259 len = strnlen(fc_host_symbolic_name(lport->host), 255);
1260 /* if there is no symbolic name, skip to RFT_ID */
1261 if (!len)
1262 return fc_lport_enter_ns(lport, LPORT_ST_RFT_ID);
1263 cmd = FC_NS_RSNN_NN;
1264 size += sizeof(struct fc_ns_rsnn) + len;
1265 break;
1266 case LPORT_ST_RSPN_ID:
1267 len = strnlen(fc_host_symbolic_name(lport->host), 255);
1268 /* if there is no symbolic name, skip to RFT_ID */
1269 if (!len)
1270 return fc_lport_enter_ns(lport, LPORT_ST_RFT_ID);
1271 cmd = FC_NS_RSPN_ID;
1272 size += sizeof(struct fc_ns_rspn) + len;
1273 break;
1274 case LPORT_ST_RFT_ID:
1275 cmd = FC_NS_RFT_ID;
1276 size += sizeof(struct fc_ns_rft);
1277 break;
ab593b18
JE
1278 case LPORT_ST_RFF_ID:
1279 cmd = FC_NS_RFF_ID;
1280 size += sizeof(struct fc_ns_rff_id);
1281 break;
c914f7d1
CL
1282 default:
1283 fc_lport_error(lport, NULL);
5baa17c3
CL
1284 return;
1285 }
1286
c914f7d1 1287 fp = fc_frame_alloc(lport, size);
c9c7bd7a
CL
1288 if (!fp) {
1289 fc_lport_error(lport, fp);
1290 return;
1291 }
1292
c914f7d1 1293 if (!lport->tt.elsct_send(lport, FC_FID_DIR_SERV, fp, cmd,
7cccc157 1294 fc_lport_ns_resp,
b94f8951 1295 lport, 3 * lport->r_a_tov))
c9c7bd7a
CL
1296 fc_lport_error(lport, fp);
1297}
1298
42e9a92f
RL
1299static struct fc_rport_operations fc_lport_rport_ops = {
1300 .event_callback = fc_lport_rport_callback,
1301};
1302
1303/**
3a3b42bf
RL
1304 * fc_rport_enter_dns() - Create a fc_rport for the name server
1305 * @lport: The local port requesting a remote port for the name server
42e9a92f
RL
1306 *
1307 * Locking Note: The lport lock is expected to be held before calling
1308 * this routine.
1309 */
1310static void fc_lport_enter_dns(struct fc_lport *lport)
1311{
ab28f1fd 1312 struct fc_rport_priv *rdata;
42e9a92f 1313
7414705e
RL
1314 FC_LPORT_DBG(lport, "Entered DNS state from %s state\n",
1315 fc_lport_state(lport));
42e9a92f
RL
1316
1317 fc_lport_state_enter(lport, LPORT_ST_DNS);
1318
48f00902 1319 mutex_lock(&lport->disc.disc_mutex);
9737e6a7 1320 rdata = lport->tt.rport_create(lport, FC_FID_DIR_SERV);
48f00902 1321 mutex_unlock(&lport->disc.disc_mutex);
9fb9d328 1322 if (!rdata)
42e9a92f
RL
1323 goto err;
1324
42e9a92f 1325 rdata->ops = &fc_lport_rport_ops;
9fb9d328 1326 lport->tt.rport_login(rdata);
42e9a92f
RL
1327 return;
1328
1329err:
1330 fc_lport_error(lport, NULL);
1331}
1332
1333/**
3a3b42bf
RL
1334 * fc_lport_timeout() - Handler for the retry_work timer
1335 * @work: The work struct of the local port
42e9a92f
RL
1336 */
1337static void fc_lport_timeout(struct work_struct *work)
1338{
1339 struct fc_lport *lport =
1340 container_of(work, struct fc_lport,
1341 retry_work.work);
1342
1343 mutex_lock(&lport->lp_mutex);
1344
1345 switch (lport->state) {
b1d9fd55 1346 case LPORT_ST_DISABLED:
22655ac2
JE
1347 WARN_ON(1);
1348 break;
42e9a92f 1349 case LPORT_ST_READY:
42e9a92f
RL
1350 WARN_ON(1);
1351 break;
22655ac2
JE
1352 case LPORT_ST_RESET:
1353 break;
42e9a92f
RL
1354 case LPORT_ST_FLOGI:
1355 fc_lport_enter_flogi(lport);
1356 break;
1357 case LPORT_ST_DNS:
1358 fc_lport_enter_dns(lport);
1359 break;
c9c7bd7a 1360 case LPORT_ST_RNN_ID:
5baa17c3 1361 case LPORT_ST_RSNN_NN:
c9866a54 1362 case LPORT_ST_RSPN_ID:
42e9a92f 1363 case LPORT_ST_RFT_ID:
ab593b18 1364 case LPORT_ST_RFF_ID:
c914f7d1 1365 fc_lport_enter_ns(lport, lport->state);
42e9a92f
RL
1366 break;
1367 case LPORT_ST_SCR:
1368 fc_lport_enter_scr(lport);
1369 break;
1370 case LPORT_ST_LOGO:
1371 fc_lport_enter_logo(lport);
1372 break;
1373 }
1374
1375 mutex_unlock(&lport->lp_mutex);
1376}
1377
1378/**
34f42a07 1379 * fc_lport_logo_resp() - Handle response to LOGO request
3a3b42bf
RL
1380 * @sp: The sequence that the LOGO was on
1381 * @fp: The LOGO frame
1382 * @lp_arg: The lport port that received the LOGO request
42e9a92f
RL
1383 *
1384 * Locking Note: This function will be called without the lport lock
3a3b42bf 1385 * held, but it will lock, call an _enter_* function or fc_lport_error()
42e9a92f
RL
1386 * and then unlock the lport.
1387 */
11b56188 1388void fc_lport_logo_resp(struct fc_seq *sp, struct fc_frame *fp,
3a3b42bf 1389 void *lp_arg)
42e9a92f
RL
1390{
1391 struct fc_lport *lport = lp_arg;
1392 u8 op;
1393
f657d299
JE
1394 FC_LPORT_DBG(lport, "Received a LOGO %s\n", fc_els_resp_type(fp));
1395
42e9a92f
RL
1396 if (fp == ERR_PTR(-FC_EX_CLOSED))
1397 return;
1398
1399 mutex_lock(&lport->lp_mutex);
1400
42e9a92f 1401 if (lport->state != LPORT_ST_LOGO) {
7414705e
RL
1402 FC_LPORT_DBG(lport, "Received a LOGO response, but in state "
1403 "%s\n", fc_lport_state(lport));
76f6804e
AJ
1404 if (IS_ERR(fp))
1405 goto err;
42e9a92f
RL
1406 goto out;
1407 }
1408
76f6804e
AJ
1409 if (IS_ERR(fp)) {
1410 fc_lport_error(lport, fp);
1411 goto err;
1412 }
1413
42e9a92f
RL
1414 op = fc_frame_payload_op(fp);
1415 if (op == ELS_LS_ACC)
1190d925 1416 fc_lport_enter_disabled(lport);
42e9a92f
RL
1417 else
1418 fc_lport_error(lport, fp);
1419
1420out:
1421 fc_frame_free(fp);
1422err:
1423 mutex_unlock(&lport->lp_mutex);
1424}
11b56188 1425EXPORT_SYMBOL(fc_lport_logo_resp);
42e9a92f
RL
1426
1427/**
34f42a07 1428 * fc_rport_enter_logo() - Logout of the fabric
3a3b42bf 1429 * @lport: The local port to be logged out
42e9a92f
RL
1430 *
1431 * Locking Note: The lport lock is expected to be held before calling
1432 * this routine.
1433 */
1434static void fc_lport_enter_logo(struct fc_lport *lport)
1435{
1436 struct fc_frame *fp;
1437 struct fc_els_logo *logo;
1438
7414705e
RL
1439 FC_LPORT_DBG(lport, "Entered LOGO state from %s state\n",
1440 fc_lport_state(lport));
42e9a92f
RL
1441
1442 fc_lport_state_enter(lport, LPORT_ST_LOGO);
8faecddb 1443 fc_vports_linkchange(lport);
42e9a92f 1444
42e9a92f
RL
1445 fp = fc_frame_alloc(lport, sizeof(*logo));
1446 if (!fp) {
1447 fc_lport_error(lport, fp);
1448 return;
1449 }
1450
a46f327a 1451 if (!lport->tt.elsct_send(lport, FC_FID_FLOGI, fp, ELS_LOGO,
b94f8951
JE
1452 fc_lport_logo_resp, lport,
1453 2 * lport->r_a_tov))
8f550f93 1454 fc_lport_error(lport, NULL);
42e9a92f
RL
1455}
1456
1457/**
34f42a07 1458 * fc_lport_flogi_resp() - Handle response to FLOGI request
3a3b42bf
RL
1459 * @sp: The sequence that the FLOGI was on
1460 * @fp: The FLOGI response frame
1461 * @lp_arg: The lport port that received the FLOGI response
42e9a92f
RL
1462 *
1463 * Locking Note: This function will be called without the lport lock
3a3b42bf 1464 * held, but it will lock, call an _enter_* function or fc_lport_error()
42e9a92f
RL
1465 * and then unlock the lport.
1466 */
11b56188 1467void fc_lport_flogi_resp(struct fc_seq *sp, struct fc_frame *fp,
3a3b42bf 1468 void *lp_arg)
42e9a92f
RL
1469{
1470 struct fc_lport *lport = lp_arg;
1471 struct fc_frame_header *fh;
1472 struct fc_els_flogi *flp;
1473 u32 did;
1474 u16 csp_flags;
1475 unsigned int r_a_tov;
1476 unsigned int e_d_tov;
1477 u16 mfs;
1478
f657d299
JE
1479 FC_LPORT_DBG(lport, "Received a FLOGI %s\n", fc_els_resp_type(fp));
1480
42e9a92f
RL
1481 if (fp == ERR_PTR(-FC_EX_CLOSED))
1482 return;
1483
1484 mutex_lock(&lport->lp_mutex);
1485
42e9a92f 1486 if (lport->state != LPORT_ST_FLOGI) {
7414705e
RL
1487 FC_LPORT_DBG(lport, "Received a FLOGI response, but in state "
1488 "%s\n", fc_lport_state(lport));
76f6804e
AJ
1489 if (IS_ERR(fp))
1490 goto err;
42e9a92f
RL
1491 goto out;
1492 }
1493
76f6804e
AJ
1494 if (IS_ERR(fp)) {
1495 fc_lport_error(lport, fp);
1496 goto err;
1497 }
1498
42e9a92f
RL
1499 fh = fc_frame_header_get(fp);
1500 did = ntoh24(fh->fh_d_id);
1501 if (fc_frame_payload_op(fp) == ELS_LS_ACC && did != 0) {
42e9a92f
RL
1502 flp = fc_frame_payload_get(fp, sizeof(*flp));
1503 if (flp) {
1504 mfs = ntohs(flp->fl_csp.sp_bb_data) &
1505 FC_SP_BB_DATA_MASK;
1506 if (mfs >= FC_SP_MIN_MAX_PAYLOAD &&
1507 mfs < lport->mfs)
1508 lport->mfs = mfs;
1509 csp_flags = ntohs(flp->fl_csp.sp_features);
1510 r_a_tov = ntohl(flp->fl_csp.sp_r_a_tov);
1511 e_d_tov = ntohl(flp->fl_csp.sp_e_d_tov);
1512 if (csp_flags & FC_SP_FT_EDTR)
1513 e_d_tov /= 1000000;
db36c06c
CL
1514
1515 lport->npiv_enabled = !!(csp_flags & FC_SP_FT_NPIV_ACC);
1516
42e9a92f
RL
1517 if ((csp_flags & FC_SP_FT_FPORT) == 0) {
1518 if (e_d_tov > lport->e_d_tov)
1519 lport->e_d_tov = e_d_tov;
1520 lport->r_a_tov = 2 * e_d_tov;
093bb6a2 1521 fc_lport_set_port_id(lport, did, fp);
e6d8a1b0 1522 printk(KERN_INFO "host%d: libfc: "
ce8b5df0 1523 "Port (%6.6x) entered "
e6d8a1b0
JE
1524 "point-to-point mode\n",
1525 lport->host->host_no, did);
42e9a92f
RL
1526 fc_lport_ptp_setup(lport, ntoh24(fh->fh_s_id),
1527 get_unaligned_be64(
1528 &flp->fl_wwpn),
1529 get_unaligned_be64(
1530 &flp->fl_wwnn));
1531 } else {
1532 lport->e_d_tov = e_d_tov;
1533 lport->r_a_tov = r_a_tov;
1534 fc_host_fabric_name(lport->host) =
1535 get_unaligned_be64(&flp->fl_wwnn);
093bb6a2 1536 fc_lport_set_port_id(lport, did, fp);
42e9a92f
RL
1537 fc_lport_enter_dns(lport);
1538 }
1539 }
42e9a92f 1540 } else {
7414705e 1541 FC_LPORT_DBG(lport, "Bad FLOGI response\n");
42e9a92f
RL
1542 }
1543
1544out:
1545 fc_frame_free(fp);
1546err:
1547 mutex_unlock(&lport->lp_mutex);
1548}
11b56188 1549EXPORT_SYMBOL(fc_lport_flogi_resp);
42e9a92f
RL
1550
1551/**
34f42a07 1552 * fc_rport_enter_flogi() - Send a FLOGI request to the fabric manager
42e9a92f
RL
1553 * @lport: Fibre Channel local port to be logged in to the fabric
1554 *
1555 * Locking Note: The lport lock is expected to be held before calling
1556 * this routine.
1557 */
1558void fc_lport_enter_flogi(struct fc_lport *lport)
1559{
1560 struct fc_frame *fp;
1561
7414705e
RL
1562 FC_LPORT_DBG(lport, "Entered FLOGI state from %s state\n",
1563 fc_lport_state(lport));
42e9a92f
RL
1564
1565 fc_lport_state_enter(lport, LPORT_ST_FLOGI);
1566
3726f358
JE
1567 if (lport->point_to_multipoint) {
1568 if (lport->port_id)
1569 fc_lport_enter_ready(lport);
1570 return;
1571 }
1572
42e9a92f
RL
1573 fp = fc_frame_alloc(lport, sizeof(struct fc_els_flogi));
1574 if (!fp)
1575 return fc_lport_error(lport, fp);
1576
db36c06c
CL
1577 if (!lport->tt.elsct_send(lport, FC_FID_FLOGI, fp,
1578 lport->vport ? ELS_FDISC : ELS_FLOGI,
b94f8951
JE
1579 fc_lport_flogi_resp, lport,
1580 lport->vport ? 2 * lport->r_a_tov :
1581 lport->e_d_tov))
8f550f93 1582 fc_lport_error(lport, NULL);
42e9a92f
RL
1583}
1584
3a3b42bf
RL
1585/**
1586 * fc_lport_config() - Configure a fc_lport
1587 * @lport: The local port to be configured
1588 */
42e9a92f
RL
1589int fc_lport_config(struct fc_lport *lport)
1590{
1591 INIT_DELAYED_WORK(&lport->retry_work, fc_lport_timeout);
1592 mutex_init(&lport->lp_mutex);
1593
b1d9fd55 1594 fc_lport_state_enter(lport, LPORT_ST_DISABLED);
42e9a92f
RL
1595
1596 fc_lport_add_fc4_type(lport, FC_TYPE_FCP);
1597 fc_lport_add_fc4_type(lport, FC_TYPE_CT);
1598
1599 return 0;
1600}
1601EXPORT_SYMBOL(fc_lport_config);
1602
3a3b42bf
RL
1603/**
1604 * fc_lport_init() - Initialize the lport layer for a local port
1605 * @lport: The local port to initialize the exchange layer for
1606 */
42e9a92f
RL
1607int fc_lport_init(struct fc_lport *lport)
1608{
1609 if (!lport->tt.lport_recv)
1610 lport->tt.lport_recv = fc_lport_recv_req;
1611
1612 if (!lport->tt.lport_reset)
1613 lport->tt.lport_reset = fc_lport_reset;
1614
1615 fc_host_port_type(lport->host) = FC_PORTTYPE_NPORT;
1616 fc_host_node_name(lport->host) = lport->wwnn;
1617 fc_host_port_name(lport->host) = lport->wwpn;
1618 fc_host_supported_classes(lport->host) = FC_COS_CLASS3;
1619 memset(fc_host_supported_fc4s(lport->host), 0,
1620 sizeof(fc_host_supported_fc4s(lport->host)));
1621 fc_host_supported_fc4s(lport->host)[2] = 1;
1622 fc_host_supported_fc4s(lport->host)[7] = 1;
1623
1624 /* This value is also unchanging */
1625 memset(fc_host_active_fc4s(lport->host), 0,
1626 sizeof(fc_host_active_fc4s(lport->host)));
1627 fc_host_active_fc4s(lport->host)[2] = 1;
1628 fc_host_active_fc4s(lport->host)[7] = 1;
1629 fc_host_maxframe_size(lport->host) = lport->mfs;
1630 fc_host_supported_speeds(lport->host) = 0;
1631 if (lport->link_supported_speeds & FC_PORTSPEED_1GBIT)
1632 fc_host_supported_speeds(lport->host) |= FC_PORTSPEED_1GBIT;
1633 if (lport->link_supported_speeds & FC_PORTSPEED_10GBIT)
1634 fc_host_supported_speeds(lport->host) |= FC_PORTSPEED_10GBIT;
1635
1636 return 0;
1637}
1638EXPORT_SYMBOL(fc_lport_init);
a51ab396
SM
1639
1640/**
3a3b42bf
RL
1641 * fc_lport_bsg_resp() - The common response handler for FC Passthrough requests
1642 * @sp: The sequence for the FC Passthrough response
1643 * @fp: The response frame
1644 * @info_arg: The BSG info that the response is for
a51ab396
SM
1645 */
1646static void fc_lport_bsg_resp(struct fc_seq *sp, struct fc_frame *fp,
1647 void *info_arg)
1648{
1649 struct fc_bsg_info *info = info_arg;
1650 struct fc_bsg_job *job = info->job;
1651 struct fc_lport *lport = info->lport;
1652 struct fc_frame_header *fh;
1653 size_t len;
1654 void *buf;
1655
1656 if (IS_ERR(fp)) {
1657 job->reply->result = (PTR_ERR(fp) == -FC_EX_CLOSED) ?
1658 -ECONNABORTED : -ETIMEDOUT;
1659 job->reply_len = sizeof(uint32_t);
1660 job->state_flags |= FC_RQST_STATE_DONE;
1661 job->job_done(job);
1662 kfree(info);
1663 return;
1664 }
1665
1666 mutex_lock(&lport->lp_mutex);
1667 fh = fc_frame_header_get(fp);
1668 len = fr_len(fp) - sizeof(*fh);
1669 buf = fc_frame_payload_get(fp, 0);
1670
1671 if (fr_sof(fp) == FC_SOF_I3 && !ntohs(fh->fh_seq_cnt)) {
1672 /* Get the response code from the first frame payload */
1673 unsigned short cmd = (info->rsp_code == FC_FS_ACC) ?
1674 ntohs(((struct fc_ct_hdr *)buf)->ct_cmd) :
1675 (unsigned short)fc_frame_payload_op(fp);
1676
1677 /* Save the reply status of the job */
1678 job->reply->reply_data.ctels_reply.status =
1679 (cmd == info->rsp_code) ?
1680 FC_CTELS_STATUS_OK : FC_CTELS_STATUS_REJECT;
1681 }
1682
1683 job->reply->reply_payload_rcv_len +=
1684 fc_copy_buffer_to_sglist(buf, len, info->sg, &info->nents,
1685 &info->offset, KM_BIO_SRC_IRQ, NULL);
1686
1687 if (fr_eof(fp) == FC_EOF_T &&
1688 (ntoh24(fh->fh_f_ctl) & (FC_FC_LAST_SEQ | FC_FC_END_SEQ)) ==
1689 (FC_FC_LAST_SEQ | FC_FC_END_SEQ)) {
1690 if (job->reply->reply_payload_rcv_len >
1691 job->reply_payload.payload_len)
1692 job->reply->reply_payload_rcv_len =
1693 job->reply_payload.payload_len;
1694 job->reply->result = 0;
1695 job->state_flags |= FC_RQST_STATE_DONE;
1696 job->job_done(job);
1697 kfree(info);
1698 }
1699 fc_frame_free(fp);
1700 mutex_unlock(&lport->lp_mutex);
1701}
1702
1703/**
3a3b42bf
RL
1704 * fc_lport_els_request() - Send ELS passthrough request
1705 * @job: The BSG Passthrough job
a51ab396 1706 * @lport: The local port sending the request
3a3b42bf 1707 * @did: The destination port id
a51ab396
SM
1708 *
1709 * Locking Note: The lport lock is expected to be held before calling
1710 * this routine.
1711 */
1712static int fc_lport_els_request(struct fc_bsg_job *job,
1713 struct fc_lport *lport,
1714 u32 did, u32 tov)
1715{
1716 struct fc_bsg_info *info;
1717 struct fc_frame *fp;
1718 struct fc_frame_header *fh;
1719 char *pp;
1720 int len;
1721
70d919fb 1722 fp = fc_frame_alloc(lport, job->request_payload.payload_len);
a51ab396
SM
1723 if (!fp)
1724 return -ENOMEM;
1725
1726 len = job->request_payload.payload_len;
1727 pp = fc_frame_payload_get(fp, len);
1728
1729 sg_copy_to_buffer(job->request_payload.sg_list,
1730 job->request_payload.sg_cnt,
1731 pp, len);
1732
1733 fh = fc_frame_header_get(fp);
1734 fh->fh_r_ctl = FC_RCTL_ELS_REQ;
1735 hton24(fh->fh_d_id, did);
7b2787ec 1736 hton24(fh->fh_s_id, lport->port_id);
a51ab396
SM
1737 fh->fh_type = FC_TYPE_ELS;
1738 hton24(fh->fh_f_ctl, FC_FC_FIRST_SEQ |
1739 FC_FC_END_SEQ | FC_FC_SEQ_INIT);
1740 fh->fh_cs_ctl = 0;
1741 fh->fh_df_ctl = 0;
1742 fh->fh_parm_offset = 0;
1743
1744 info = kzalloc(sizeof(struct fc_bsg_info), GFP_KERNEL);
1745 if (!info) {
1746 fc_frame_free(fp);
1747 return -ENOMEM;
1748 }
1749
1750 info->job = job;
1751 info->lport = lport;
1752 info->rsp_code = ELS_LS_ACC;
1753 info->nents = job->reply_payload.sg_cnt;
1754 info->sg = job->reply_payload.sg_list;
1755
1756 if (!lport->tt.exch_seq_send(lport, fp, fc_lport_bsg_resp,
1757 NULL, info, tov))
1758 return -ECOMM;
1759 return 0;
1760}
1761
1762/**
3a3b42bf
RL
1763 * fc_lport_ct_request() - Send CT Passthrough request
1764 * @job: The BSG Passthrough job
a51ab396
SM
1765 * @lport: The local port sending the request
1766 * @did: The destination FC-ID
3a3b42bf 1767 * @tov: The timeout period to wait for the response
a51ab396
SM
1768 *
1769 * Locking Note: The lport lock is expected to be held before calling
1770 * this routine.
1771 */
1772static int fc_lport_ct_request(struct fc_bsg_job *job,
1773 struct fc_lport *lport, u32 did, u32 tov)
1774{
1775 struct fc_bsg_info *info;
1776 struct fc_frame *fp;
1777 struct fc_frame_header *fh;
1778 struct fc_ct_req *ct;
1779 size_t len;
1780
1781 fp = fc_frame_alloc(lport, sizeof(struct fc_ct_hdr) +
1782 job->request_payload.payload_len);
1783 if (!fp)
1784 return -ENOMEM;
1785
1786 len = job->request_payload.payload_len;
1787 ct = fc_frame_payload_get(fp, len);
1788
1789 sg_copy_to_buffer(job->request_payload.sg_list,
1790 job->request_payload.sg_cnt,
1791 ct, len);
1792
1793 fh = fc_frame_header_get(fp);
1794 fh->fh_r_ctl = FC_RCTL_DD_UNSOL_CTL;
1795 hton24(fh->fh_d_id, did);
7b2787ec 1796 hton24(fh->fh_s_id, lport->port_id);
a51ab396
SM
1797 fh->fh_type = FC_TYPE_CT;
1798 hton24(fh->fh_f_ctl, FC_FC_FIRST_SEQ |
1799 FC_FC_END_SEQ | FC_FC_SEQ_INIT);
1800 fh->fh_cs_ctl = 0;
1801 fh->fh_df_ctl = 0;
1802 fh->fh_parm_offset = 0;
1803
1804 info = kzalloc(sizeof(struct fc_bsg_info), GFP_KERNEL);
1805 if (!info) {
1806 fc_frame_free(fp);
1807 return -ENOMEM;
1808 }
1809
1810 info->job = job;
1811 info->lport = lport;
1812 info->rsp_code = FC_FS_ACC;
1813 info->nents = job->reply_payload.sg_cnt;
1814 info->sg = job->reply_payload.sg_list;
1815
1816 if (!lport->tt.exch_seq_send(lport, fp, fc_lport_bsg_resp,
1817 NULL, info, tov))
1818 return -ECOMM;
1819 return 0;
1820}
1821
1822/**
1823 * fc_lport_bsg_request() - The common entry point for sending
3a3b42bf
RL
1824 * FC Passthrough requests
1825 * @job: The BSG passthrough job
a51ab396
SM
1826 */
1827int fc_lport_bsg_request(struct fc_bsg_job *job)
1828{
1829 struct request *rsp = job->req->next_rq;
1830 struct Scsi_Host *shost = job->shost;
1831 struct fc_lport *lport = shost_priv(shost);
1832 struct fc_rport *rport;
1833 struct fc_rport_priv *rdata;
1834 int rc = -EINVAL;
1835 u32 did;
1836
1837 job->reply->reply_payload_rcv_len = 0;
b248df30
HD
1838 if (rsp)
1839 rsp->resid_len = job->reply_payload.payload_len;
a51ab396
SM
1840
1841 mutex_lock(&lport->lp_mutex);
1842
1843 switch (job->request->msgcode) {
1844 case FC_BSG_RPT_ELS:
1845 rport = job->rport;
1846 if (!rport)
1847 break;
1848
1849 rdata = rport->dd_data;
1850 rc = fc_lport_els_request(job, lport, rport->port_id,
1851 rdata->e_d_tov);
1852 break;
1853
1854 case FC_BSG_RPT_CT:
1855 rport = job->rport;
1856 if (!rport)
1857 break;
1858
1859 rdata = rport->dd_data;
1860 rc = fc_lport_ct_request(job, lport, rport->port_id,
1861 rdata->e_d_tov);
1862 break;
1863
1864 case FC_BSG_HST_CT:
1865 did = ntoh24(job->request->rqst_data.h_ct.port_id);
1866 if (did == FC_FID_DIR_SERV)
3a3b42bf 1867 rdata = lport->dns_rdata;
a51ab396
SM
1868 else
1869 rdata = lport->tt.rport_lookup(lport, did);
1870
1871 if (!rdata)
1872 break;
1873
1874 rc = fc_lport_ct_request(job, lport, did, rdata->e_d_tov);
1875 break;
1876
1877 case FC_BSG_HST_ELS_NOLOGIN:
1878 did = ntoh24(job->request->rqst_data.h_els.port_id);
1879 rc = fc_lport_els_request(job, lport, did, lport->e_d_tov);
1880 break;
1881 }
1882
1883 mutex_unlock(&lport->lp_mutex);
1884 return rc;
1885}
1886EXPORT_SYMBOL(fc_lport_bsg_request);