]>
Commit | Line | Data |
---|---|---|
e48354ce NB |
1 | /******************************************************************************* |
2 | * This file contains the iSCSI Target specific utility functions. | |
3 | * | |
4 | * \u00a9 Copyright 2007-2011 RisingTide Systems LLC. | |
5 | * | |
6 | * Licensed to the Linux Foundation under the General Public License (GPL) version 2. | |
7 | * | |
8 | * Author: Nicholas A. Bellinger <nab@linux-iscsi.org> | |
9 | * | |
10 | * This program is free software; you can redistribute it and/or modify | |
11 | * it under the terms of the GNU General Public License as published by | |
12 | * the Free Software Foundation; either version 2 of the License, or | |
13 | * (at your option) any later version. | |
14 | * | |
15 | * This program is distributed in the hope that it will be useful, | |
16 | * but WITHOUT ANY WARRANTY; without even the implied warranty of | |
17 | * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the | |
18 | * GNU General Public License for more details. | |
19 | ******************************************************************************/ | |
20 | ||
21 | #include <linux/list.h> | |
22 | #include <scsi/scsi_tcq.h> | |
23 | #include <scsi/iscsi_proto.h> | |
24 | #include <target/target_core_base.h> | |
25 | #include <target/target_core_transport.h> | |
26 | #include <target/target_core_tmr.h> | |
27 | #include <target/target_core_fabric_ops.h> | |
28 | #include <target/target_core_configfs.h> | |
29 | ||
30 | #include "iscsi_target_core.h" | |
31 | #include "iscsi_target_parameters.h" | |
32 | #include "iscsi_target_seq_pdu_list.h" | |
33 | #include "iscsi_target_datain_values.h" | |
34 | #include "iscsi_target_erl0.h" | |
35 | #include "iscsi_target_erl1.h" | |
36 | #include "iscsi_target_erl2.h" | |
37 | #include "iscsi_target_tpg.h" | |
38 | #include "iscsi_target_tq.h" | |
39 | #include "iscsi_target_util.h" | |
40 | #include "iscsi_target.h" | |
41 | ||
42 | #define PRINT_BUFF(buff, len) \ | |
43 | { \ | |
44 | int zzz; \ | |
45 | \ | |
46 | pr_debug("%d:\n", __LINE__); \ | |
47 | for (zzz = 0; zzz < len; zzz++) { \ | |
48 | if (zzz % 16 == 0) { \ | |
49 | if (zzz) \ | |
50 | pr_debug("\n"); \ | |
51 | pr_debug("%4i: ", zzz); \ | |
52 | } \ | |
53 | pr_debug("%02x ", (unsigned char) (buff)[zzz]); \ | |
54 | } \ | |
55 | if ((len + 1) % 16) \ | |
56 | pr_debug("\n"); \ | |
57 | } | |
58 | ||
59 | extern struct list_head g_tiqn_list; | |
60 | extern spinlock_t tiqn_lock; | |
61 | ||
62 | /* | |
63 | * Called with cmd->r2t_lock held. | |
64 | */ | |
65 | int iscsit_add_r2t_to_list( | |
66 | struct iscsi_cmd *cmd, | |
67 | u32 offset, | |
68 | u32 xfer_len, | |
69 | int recovery, | |
70 | u32 r2t_sn) | |
71 | { | |
72 | struct iscsi_r2t *r2t; | |
73 | ||
74 | r2t = kmem_cache_zalloc(lio_r2t_cache, GFP_ATOMIC); | |
75 | if (!r2t) { | |
76 | pr_err("Unable to allocate memory for struct iscsi_r2t.\n"); | |
77 | return -1; | |
78 | } | |
79 | INIT_LIST_HEAD(&r2t->r2t_list); | |
80 | ||
81 | r2t->recovery_r2t = recovery; | |
82 | r2t->r2t_sn = (!r2t_sn) ? cmd->r2t_sn++ : r2t_sn; | |
83 | r2t->offset = offset; | |
84 | r2t->xfer_len = xfer_len; | |
85 | list_add_tail(&r2t->r2t_list, &cmd->cmd_r2t_list); | |
86 | spin_unlock_bh(&cmd->r2t_lock); | |
87 | ||
88 | iscsit_add_cmd_to_immediate_queue(cmd, cmd->conn, ISTATE_SEND_R2T); | |
89 | ||
90 | spin_lock_bh(&cmd->r2t_lock); | |
91 | return 0; | |
92 | } | |
93 | ||
94 | struct iscsi_r2t *iscsit_get_r2t_for_eos( | |
95 | struct iscsi_cmd *cmd, | |
96 | u32 offset, | |
97 | u32 length) | |
98 | { | |
99 | struct iscsi_r2t *r2t; | |
100 | ||
101 | spin_lock_bh(&cmd->r2t_lock); | |
102 | list_for_each_entry(r2t, &cmd->cmd_r2t_list, r2t_list) { | |
103 | if ((r2t->offset <= offset) && | |
104 | (r2t->offset + r2t->xfer_len) >= (offset + length)) { | |
105 | spin_unlock_bh(&cmd->r2t_lock); | |
106 | return r2t; | |
107 | } | |
108 | } | |
109 | spin_unlock_bh(&cmd->r2t_lock); | |
110 | ||
111 | pr_err("Unable to locate R2T for Offset: %u, Length:" | |
112 | " %u\n", offset, length); | |
113 | return NULL; | |
114 | } | |
115 | ||
116 | struct iscsi_r2t *iscsit_get_r2t_from_list(struct iscsi_cmd *cmd) | |
117 | { | |
118 | struct iscsi_r2t *r2t; | |
119 | ||
120 | spin_lock_bh(&cmd->r2t_lock); | |
121 | list_for_each_entry(r2t, &cmd->cmd_r2t_list, r2t_list) { | |
122 | if (!r2t->sent_r2t) { | |
123 | spin_unlock_bh(&cmd->r2t_lock); | |
124 | return r2t; | |
125 | } | |
126 | } | |
127 | spin_unlock_bh(&cmd->r2t_lock); | |
128 | ||
129 | pr_err("Unable to locate next R2T to send for ITT:" | |
130 | " 0x%08x.\n", cmd->init_task_tag); | |
131 | return NULL; | |
132 | } | |
133 | ||
134 | /* | |
135 | * Called with cmd->r2t_lock held. | |
136 | */ | |
137 | void iscsit_free_r2t(struct iscsi_r2t *r2t, struct iscsi_cmd *cmd) | |
138 | { | |
139 | list_del(&r2t->r2t_list); | |
140 | kmem_cache_free(lio_r2t_cache, r2t); | |
141 | } | |
142 | ||
143 | void iscsit_free_r2ts_from_list(struct iscsi_cmd *cmd) | |
144 | { | |
145 | struct iscsi_r2t *r2t, *r2t_tmp; | |
146 | ||
147 | spin_lock_bh(&cmd->r2t_lock); | |
148 | list_for_each_entry_safe(r2t, r2t_tmp, &cmd->cmd_r2t_list, r2t_list) | |
149 | iscsit_free_r2t(r2t, cmd); | |
150 | spin_unlock_bh(&cmd->r2t_lock); | |
151 | } | |
152 | ||
153 | /* | |
154 | * May be called from software interrupt (timer) context for allocating | |
155 | * iSCSI NopINs. | |
156 | */ | |
157 | struct iscsi_cmd *iscsit_allocate_cmd(struct iscsi_conn *conn, gfp_t gfp_mask) | |
158 | { | |
159 | struct iscsi_cmd *cmd; | |
160 | ||
161 | cmd = kmem_cache_zalloc(lio_cmd_cache, gfp_mask); | |
162 | if (!cmd) { | |
163 | pr_err("Unable to allocate memory for struct iscsi_cmd.\n"); | |
164 | return NULL; | |
165 | } | |
166 | ||
167 | cmd->conn = conn; | |
168 | INIT_LIST_HEAD(&cmd->i_list); | |
169 | INIT_LIST_HEAD(&cmd->datain_list); | |
170 | INIT_LIST_HEAD(&cmd->cmd_r2t_list); | |
171 | init_completion(&cmd->reject_comp); | |
172 | spin_lock_init(&cmd->datain_lock); | |
173 | spin_lock_init(&cmd->dataout_timeout_lock); | |
174 | spin_lock_init(&cmd->istate_lock); | |
175 | spin_lock_init(&cmd->error_lock); | |
176 | spin_lock_init(&cmd->r2t_lock); | |
177 | ||
178 | return cmd; | |
179 | } | |
180 | ||
181 | /* | |
182 | * Called from iscsi_handle_scsi_cmd() | |
183 | */ | |
184 | struct iscsi_cmd *iscsit_allocate_se_cmd( | |
185 | struct iscsi_conn *conn, | |
186 | u32 data_length, | |
187 | int data_direction, | |
188 | int iscsi_task_attr) | |
189 | { | |
190 | struct iscsi_cmd *cmd; | |
191 | struct se_cmd *se_cmd; | |
192 | int sam_task_attr; | |
193 | ||
194 | cmd = iscsit_allocate_cmd(conn, GFP_KERNEL); | |
195 | if (!cmd) | |
196 | return NULL; | |
197 | ||
198 | cmd->data_direction = data_direction; | |
199 | cmd->data_length = data_length; | |
200 | /* | |
201 | * Figure out the SAM Task Attribute for the incoming SCSI CDB | |
202 | */ | |
203 | if ((iscsi_task_attr == ISCSI_ATTR_UNTAGGED) || | |
204 | (iscsi_task_attr == ISCSI_ATTR_SIMPLE)) | |
205 | sam_task_attr = MSG_SIMPLE_TAG; | |
206 | else if (iscsi_task_attr == ISCSI_ATTR_ORDERED) | |
207 | sam_task_attr = MSG_ORDERED_TAG; | |
208 | else if (iscsi_task_attr == ISCSI_ATTR_HEAD_OF_QUEUE) | |
209 | sam_task_attr = MSG_HEAD_TAG; | |
210 | else if (iscsi_task_attr == ISCSI_ATTR_ACA) | |
211 | sam_task_attr = MSG_ACA_TAG; | |
212 | else { | |
213 | pr_debug("Unknown iSCSI Task Attribute: 0x%02x, using" | |
214 | " MSG_SIMPLE_TAG\n", iscsi_task_attr); | |
215 | sam_task_attr = MSG_SIMPLE_TAG; | |
216 | } | |
217 | ||
218 | se_cmd = &cmd->se_cmd; | |
219 | /* | |
220 | * Initialize struct se_cmd descriptor from target_core_mod infrastructure | |
221 | */ | |
222 | transport_init_se_cmd(se_cmd, &lio_target_fabric_configfs->tf_ops, | |
223 | conn->sess->se_sess, data_length, data_direction, | |
224 | sam_task_attr, &cmd->sense_buffer[0]); | |
225 | return cmd; | |
226 | } | |
227 | ||
228 | struct iscsi_cmd *iscsit_allocate_se_cmd_for_tmr( | |
229 | struct iscsi_conn *conn, | |
230 | u8 function) | |
231 | { | |
232 | struct iscsi_cmd *cmd; | |
233 | struct se_cmd *se_cmd; | |
234 | u8 tcm_function; | |
235 | ||
236 | cmd = iscsit_allocate_cmd(conn, GFP_KERNEL); | |
237 | if (!cmd) | |
238 | return NULL; | |
239 | ||
240 | cmd->data_direction = DMA_NONE; | |
241 | ||
242 | cmd->tmr_req = kzalloc(sizeof(struct iscsi_tmr_req), GFP_KERNEL); | |
243 | if (!cmd->tmr_req) { | |
244 | pr_err("Unable to allocate memory for" | |
245 | " Task Management command!\n"); | |
ba773669 | 246 | goto out; |
e48354ce NB |
247 | } |
248 | /* | |
249 | * TASK_REASSIGN for ERL=2 / connection stays inside of | |
250 | * LIO-Target $FABRIC_MOD | |
251 | */ | |
252 | if (function == ISCSI_TM_FUNC_TASK_REASSIGN) | |
253 | return cmd; | |
254 | ||
255 | se_cmd = &cmd->se_cmd; | |
256 | /* | |
257 | * Initialize struct se_cmd descriptor from target_core_mod infrastructure | |
258 | */ | |
259 | transport_init_se_cmd(se_cmd, &lio_target_fabric_configfs->tf_ops, | |
260 | conn->sess->se_sess, 0, DMA_NONE, | |
261 | MSG_SIMPLE_TAG, &cmd->sense_buffer[0]); | |
262 | ||
263 | switch (function) { | |
264 | case ISCSI_TM_FUNC_ABORT_TASK: | |
265 | tcm_function = TMR_ABORT_TASK; | |
266 | break; | |
267 | case ISCSI_TM_FUNC_ABORT_TASK_SET: | |
268 | tcm_function = TMR_ABORT_TASK_SET; | |
269 | break; | |
270 | case ISCSI_TM_FUNC_CLEAR_ACA: | |
271 | tcm_function = TMR_CLEAR_ACA; | |
272 | break; | |
273 | case ISCSI_TM_FUNC_CLEAR_TASK_SET: | |
274 | tcm_function = TMR_CLEAR_TASK_SET; | |
275 | break; | |
276 | case ISCSI_TM_FUNC_LOGICAL_UNIT_RESET: | |
277 | tcm_function = TMR_LUN_RESET; | |
278 | break; | |
279 | case ISCSI_TM_FUNC_TARGET_WARM_RESET: | |
280 | tcm_function = TMR_TARGET_WARM_RESET; | |
281 | break; | |
282 | case ISCSI_TM_FUNC_TARGET_COLD_RESET: | |
283 | tcm_function = TMR_TARGET_COLD_RESET; | |
284 | break; | |
285 | default: | |
286 | pr_err("Unknown iSCSI TMR Function:" | |
287 | " 0x%02x\n", function); | |
288 | goto out; | |
289 | } | |
290 | ||
291 | se_cmd->se_tmr_req = core_tmr_alloc_req(se_cmd, | |
292 | (void *)cmd->tmr_req, tcm_function); | |
293 | if (!se_cmd->se_tmr_req) | |
294 | goto out; | |
295 | ||
296 | cmd->tmr_req->se_tmr_req = se_cmd->se_tmr_req; | |
297 | ||
298 | return cmd; | |
299 | out: | |
300 | iscsit_release_cmd(cmd); | |
e48354ce NB |
301 | return NULL; |
302 | } | |
303 | ||
304 | int iscsit_decide_list_to_build( | |
305 | struct iscsi_cmd *cmd, | |
306 | u32 immediate_data_length) | |
307 | { | |
308 | struct iscsi_build_list bl; | |
309 | struct iscsi_conn *conn = cmd->conn; | |
310 | struct iscsi_session *sess = conn->sess; | |
311 | struct iscsi_node_attrib *na; | |
312 | ||
313 | if (sess->sess_ops->DataSequenceInOrder && | |
314 | sess->sess_ops->DataPDUInOrder) | |
315 | return 0; | |
316 | ||
317 | if (cmd->data_direction == DMA_NONE) | |
318 | return 0; | |
319 | ||
320 | na = iscsit_tpg_get_node_attrib(sess); | |
321 | memset(&bl, 0, sizeof(struct iscsi_build_list)); | |
322 | ||
323 | if (cmd->data_direction == DMA_FROM_DEVICE) { | |
324 | bl.data_direction = ISCSI_PDU_READ; | |
325 | bl.type = PDULIST_NORMAL; | |
326 | if (na->random_datain_pdu_offsets) | |
327 | bl.randomize |= RANDOM_DATAIN_PDU_OFFSETS; | |
328 | if (na->random_datain_seq_offsets) | |
329 | bl.randomize |= RANDOM_DATAIN_SEQ_OFFSETS; | |
330 | } else { | |
331 | bl.data_direction = ISCSI_PDU_WRITE; | |
332 | bl.immediate_data_length = immediate_data_length; | |
333 | if (na->random_r2t_offsets) | |
334 | bl.randomize |= RANDOM_R2T_OFFSETS; | |
335 | ||
336 | if (!cmd->immediate_data && !cmd->unsolicited_data) | |
337 | bl.type = PDULIST_NORMAL; | |
338 | else if (cmd->immediate_data && !cmd->unsolicited_data) | |
339 | bl.type = PDULIST_IMMEDIATE; | |
340 | else if (!cmd->immediate_data && cmd->unsolicited_data) | |
341 | bl.type = PDULIST_UNSOLICITED; | |
342 | else if (cmd->immediate_data && cmd->unsolicited_data) | |
343 | bl.type = PDULIST_IMMEDIATE_AND_UNSOLICITED; | |
344 | } | |
345 | ||
346 | return iscsit_do_build_list(cmd, &bl); | |
347 | } | |
348 | ||
349 | struct iscsi_seq *iscsit_get_seq_holder_for_datain( | |
350 | struct iscsi_cmd *cmd, | |
351 | u32 seq_send_order) | |
352 | { | |
353 | u32 i; | |
354 | ||
355 | for (i = 0; i < cmd->seq_count; i++) | |
356 | if (cmd->seq_list[i].seq_send_order == seq_send_order) | |
357 | return &cmd->seq_list[i]; | |
358 | ||
359 | return NULL; | |
360 | } | |
361 | ||
362 | struct iscsi_seq *iscsit_get_seq_holder_for_r2t(struct iscsi_cmd *cmd) | |
363 | { | |
364 | u32 i; | |
365 | ||
366 | if (!cmd->seq_list) { | |
367 | pr_err("struct iscsi_cmd->seq_list is NULL!\n"); | |
368 | return NULL; | |
369 | } | |
370 | ||
371 | for (i = 0; i < cmd->seq_count; i++) { | |
372 | if (cmd->seq_list[i].type != SEQTYPE_NORMAL) | |
373 | continue; | |
374 | if (cmd->seq_list[i].seq_send_order == cmd->seq_send_order) { | |
375 | cmd->seq_send_order++; | |
376 | return &cmd->seq_list[i]; | |
377 | } | |
378 | } | |
379 | ||
380 | return NULL; | |
381 | } | |
382 | ||
383 | struct iscsi_r2t *iscsit_get_holder_for_r2tsn( | |
384 | struct iscsi_cmd *cmd, | |
385 | u32 r2t_sn) | |
386 | { | |
387 | struct iscsi_r2t *r2t; | |
388 | ||
389 | spin_lock_bh(&cmd->r2t_lock); | |
390 | list_for_each_entry(r2t, &cmd->cmd_r2t_list, r2t_list) { | |
391 | if (r2t->r2t_sn == r2t_sn) { | |
392 | spin_unlock_bh(&cmd->r2t_lock); | |
393 | return r2t; | |
394 | } | |
395 | } | |
396 | spin_unlock_bh(&cmd->r2t_lock); | |
397 | ||
398 | return NULL; | |
399 | } | |
400 | ||
401 | static inline int iscsit_check_received_cmdsn(struct iscsi_session *sess, u32 cmdsn) | |
402 | { | |
403 | int ret; | |
404 | ||
405 | /* | |
406 | * This is the proper method of checking received CmdSN against | |
407 | * ExpCmdSN and MaxCmdSN values, as well as accounting for out | |
408 | * or order CmdSNs due to multiple connection sessions and/or | |
409 | * CRC failures. | |
410 | */ | |
411 | if (iscsi_sna_gt(cmdsn, sess->max_cmd_sn)) { | |
412 | pr_err("Received CmdSN: 0x%08x is greater than" | |
413 | " MaxCmdSN: 0x%08x, protocol error.\n", cmdsn, | |
414 | sess->max_cmd_sn); | |
415 | ret = CMDSN_ERROR_CANNOT_RECOVER; | |
416 | ||
417 | } else if (cmdsn == sess->exp_cmd_sn) { | |
418 | sess->exp_cmd_sn++; | |
419 | pr_debug("Received CmdSN matches ExpCmdSN," | |
420 | " incremented ExpCmdSN to: 0x%08x\n", | |
421 | sess->exp_cmd_sn); | |
422 | ret = CMDSN_NORMAL_OPERATION; | |
423 | ||
424 | } else if (iscsi_sna_gt(cmdsn, sess->exp_cmd_sn)) { | |
425 | pr_debug("Received CmdSN: 0x%08x is greater" | |
426 | " than ExpCmdSN: 0x%08x, not acknowledging.\n", | |
427 | cmdsn, sess->exp_cmd_sn); | |
428 | ret = CMDSN_HIGHER_THAN_EXP; | |
429 | ||
430 | } else { | |
431 | pr_err("Received CmdSN: 0x%08x is less than" | |
432 | " ExpCmdSN: 0x%08x, ignoring.\n", cmdsn, | |
433 | sess->exp_cmd_sn); | |
434 | ret = CMDSN_LOWER_THAN_EXP; | |
435 | } | |
436 | ||
437 | return ret; | |
438 | } | |
439 | ||
440 | /* | |
441 | * Commands may be received out of order if MC/S is in use. | |
442 | * Ensure they are executed in CmdSN order. | |
443 | */ | |
444 | int iscsit_sequence_cmd( | |
445 | struct iscsi_conn *conn, | |
446 | struct iscsi_cmd *cmd, | |
447 | u32 cmdsn) | |
448 | { | |
449 | int ret; | |
450 | int cmdsn_ret; | |
451 | ||
452 | mutex_lock(&conn->sess->cmdsn_mutex); | |
453 | ||
454 | cmdsn_ret = iscsit_check_received_cmdsn(conn->sess, cmdsn); | |
455 | switch (cmdsn_ret) { | |
456 | case CMDSN_NORMAL_OPERATION: | |
457 | ret = iscsit_execute_cmd(cmd, 0); | |
458 | if ((ret >= 0) && !list_empty(&conn->sess->sess_ooo_cmdsn_list)) | |
459 | iscsit_execute_ooo_cmdsns(conn->sess); | |
460 | break; | |
461 | case CMDSN_HIGHER_THAN_EXP: | |
462 | ret = iscsit_handle_ooo_cmdsn(conn->sess, cmd, cmdsn); | |
463 | break; | |
464 | case CMDSN_LOWER_THAN_EXP: | |
465 | cmd->i_state = ISTATE_REMOVE; | |
466 | iscsit_add_cmd_to_immediate_queue(cmd, conn, cmd->i_state); | |
467 | ret = cmdsn_ret; | |
468 | break; | |
469 | default: | |
470 | ret = cmdsn_ret; | |
471 | break; | |
472 | } | |
473 | mutex_unlock(&conn->sess->cmdsn_mutex); | |
474 | ||
475 | return ret; | |
476 | } | |
477 | ||
478 | int iscsit_check_unsolicited_dataout(struct iscsi_cmd *cmd, unsigned char *buf) | |
479 | { | |
480 | struct iscsi_conn *conn = cmd->conn; | |
481 | struct se_cmd *se_cmd = &cmd->se_cmd; | |
482 | struct iscsi_data *hdr = (struct iscsi_data *) buf; | |
483 | u32 payload_length = ntoh24(hdr->dlength); | |
484 | ||
485 | if (conn->sess->sess_ops->InitialR2T) { | |
486 | pr_err("Received unexpected unsolicited data" | |
487 | " while InitialR2T=Yes, protocol error.\n"); | |
488 | transport_send_check_condition_and_sense(se_cmd, | |
489 | TCM_UNEXPECTED_UNSOLICITED_DATA, 0); | |
490 | return -1; | |
491 | } | |
492 | ||
493 | if ((cmd->first_burst_len + payload_length) > | |
494 | conn->sess->sess_ops->FirstBurstLength) { | |
495 | pr_err("Total %u bytes exceeds FirstBurstLength: %u" | |
496 | " for this Unsolicited DataOut Burst.\n", | |
497 | (cmd->first_burst_len + payload_length), | |
498 | conn->sess->sess_ops->FirstBurstLength); | |
499 | transport_send_check_condition_and_sense(se_cmd, | |
500 | TCM_INCORRECT_AMOUNT_OF_DATA, 0); | |
501 | return -1; | |
502 | } | |
503 | ||
504 | if (!(hdr->flags & ISCSI_FLAG_CMD_FINAL)) | |
505 | return 0; | |
506 | ||
507 | if (((cmd->first_burst_len + payload_length) != cmd->data_length) && | |
508 | ((cmd->first_burst_len + payload_length) != | |
509 | conn->sess->sess_ops->FirstBurstLength)) { | |
510 | pr_err("Unsolicited non-immediate data received %u" | |
511 | " does not equal FirstBurstLength: %u, and does" | |
512 | " not equal ExpXferLen %u.\n", | |
513 | (cmd->first_burst_len + payload_length), | |
514 | conn->sess->sess_ops->FirstBurstLength, cmd->data_length); | |
515 | transport_send_check_condition_and_sense(se_cmd, | |
516 | TCM_INCORRECT_AMOUNT_OF_DATA, 0); | |
517 | return -1; | |
518 | } | |
519 | return 0; | |
520 | } | |
521 | ||
522 | struct iscsi_cmd *iscsit_find_cmd_from_itt( | |
523 | struct iscsi_conn *conn, | |
524 | u32 init_task_tag) | |
525 | { | |
526 | struct iscsi_cmd *cmd; | |
527 | ||
528 | spin_lock_bh(&conn->cmd_lock); | |
529 | list_for_each_entry(cmd, &conn->conn_cmd_list, i_list) { | |
530 | if (cmd->init_task_tag == init_task_tag) { | |
531 | spin_unlock_bh(&conn->cmd_lock); | |
532 | return cmd; | |
533 | } | |
534 | } | |
535 | spin_unlock_bh(&conn->cmd_lock); | |
536 | ||
537 | pr_err("Unable to locate ITT: 0x%08x on CID: %hu", | |
538 | init_task_tag, conn->cid); | |
539 | return NULL; | |
540 | } | |
541 | ||
542 | struct iscsi_cmd *iscsit_find_cmd_from_itt_or_dump( | |
543 | struct iscsi_conn *conn, | |
544 | u32 init_task_tag, | |
545 | u32 length) | |
546 | { | |
547 | struct iscsi_cmd *cmd; | |
548 | ||
549 | spin_lock_bh(&conn->cmd_lock); | |
550 | list_for_each_entry(cmd, &conn->conn_cmd_list, i_list) { | |
551 | if (cmd->init_task_tag == init_task_tag) { | |
552 | spin_unlock_bh(&conn->cmd_lock); | |
553 | return cmd; | |
554 | } | |
555 | } | |
556 | spin_unlock_bh(&conn->cmd_lock); | |
557 | ||
558 | pr_err("Unable to locate ITT: 0x%08x on CID: %hu," | |
559 | " dumping payload\n", init_task_tag, conn->cid); | |
560 | if (length) | |
561 | iscsit_dump_data_payload(conn, length, 1); | |
562 | ||
563 | return NULL; | |
564 | } | |
565 | ||
566 | struct iscsi_cmd *iscsit_find_cmd_from_ttt( | |
567 | struct iscsi_conn *conn, | |
568 | u32 targ_xfer_tag) | |
569 | { | |
570 | struct iscsi_cmd *cmd = NULL; | |
571 | ||
572 | spin_lock_bh(&conn->cmd_lock); | |
573 | list_for_each_entry(cmd, &conn->conn_cmd_list, i_list) { | |
574 | if (cmd->targ_xfer_tag == targ_xfer_tag) { | |
575 | spin_unlock_bh(&conn->cmd_lock); | |
576 | return cmd; | |
577 | } | |
578 | } | |
579 | spin_unlock_bh(&conn->cmd_lock); | |
580 | ||
581 | pr_err("Unable to locate TTT: 0x%08x on CID: %hu\n", | |
582 | targ_xfer_tag, conn->cid); | |
583 | return NULL; | |
584 | } | |
585 | ||
586 | int iscsit_find_cmd_for_recovery( | |
587 | struct iscsi_session *sess, | |
588 | struct iscsi_cmd **cmd_ptr, | |
589 | struct iscsi_conn_recovery **cr_ptr, | |
590 | u32 init_task_tag) | |
591 | { | |
592 | struct iscsi_cmd *cmd = NULL; | |
593 | struct iscsi_conn_recovery *cr; | |
594 | /* | |
595 | * Scan through the inactive connection recovery list's command list. | |
596 | * If init_task_tag matches the command is still alligent. | |
597 | */ | |
598 | spin_lock(&sess->cr_i_lock); | |
599 | list_for_each_entry(cr, &sess->cr_inactive_list, cr_list) { | |
600 | spin_lock(&cr->conn_recovery_cmd_lock); | |
601 | list_for_each_entry(cmd, &cr->conn_recovery_cmd_list, i_list) { | |
602 | if (cmd->init_task_tag == init_task_tag) { | |
603 | spin_unlock(&cr->conn_recovery_cmd_lock); | |
604 | spin_unlock(&sess->cr_i_lock); | |
605 | ||
606 | *cr_ptr = cr; | |
607 | *cmd_ptr = cmd; | |
608 | return -2; | |
609 | } | |
610 | } | |
611 | spin_unlock(&cr->conn_recovery_cmd_lock); | |
612 | } | |
613 | spin_unlock(&sess->cr_i_lock); | |
614 | /* | |
615 | * Scan through the active connection recovery list's command list. | |
616 | * If init_task_tag matches the command is ready to be reassigned. | |
617 | */ | |
618 | spin_lock(&sess->cr_a_lock); | |
619 | list_for_each_entry(cr, &sess->cr_active_list, cr_list) { | |
620 | spin_lock(&cr->conn_recovery_cmd_lock); | |
621 | list_for_each_entry(cmd, &cr->conn_recovery_cmd_list, i_list) { | |
622 | if (cmd->init_task_tag == init_task_tag) { | |
623 | spin_unlock(&cr->conn_recovery_cmd_lock); | |
624 | spin_unlock(&sess->cr_a_lock); | |
625 | ||
626 | *cr_ptr = cr; | |
627 | *cmd_ptr = cmd; | |
628 | return 0; | |
629 | } | |
630 | } | |
631 | spin_unlock(&cr->conn_recovery_cmd_lock); | |
632 | } | |
633 | spin_unlock(&sess->cr_a_lock); | |
634 | ||
635 | return -1; | |
636 | } | |
637 | ||
638 | void iscsit_add_cmd_to_immediate_queue( | |
639 | struct iscsi_cmd *cmd, | |
640 | struct iscsi_conn *conn, | |
641 | u8 state) | |
642 | { | |
643 | struct iscsi_queue_req *qr; | |
644 | ||
645 | qr = kmem_cache_zalloc(lio_qr_cache, GFP_ATOMIC); | |
646 | if (!qr) { | |
647 | pr_err("Unable to allocate memory for" | |
648 | " struct iscsi_queue_req\n"); | |
649 | return; | |
650 | } | |
651 | INIT_LIST_HEAD(&qr->qr_list); | |
652 | qr->cmd = cmd; | |
653 | qr->state = state; | |
654 | ||
655 | spin_lock_bh(&conn->immed_queue_lock); | |
656 | list_add_tail(&qr->qr_list, &conn->immed_queue_list); | |
657 | atomic_inc(&cmd->immed_queue_count); | |
658 | atomic_set(&conn->check_immediate_queue, 1); | |
659 | spin_unlock_bh(&conn->immed_queue_lock); | |
660 | ||
661 | wake_up_process(conn->thread_set->tx_thread); | |
662 | } | |
663 | ||
664 | struct iscsi_queue_req *iscsit_get_cmd_from_immediate_queue(struct iscsi_conn *conn) | |
665 | { | |
666 | struct iscsi_queue_req *qr; | |
667 | ||
668 | spin_lock_bh(&conn->immed_queue_lock); | |
669 | if (list_empty(&conn->immed_queue_list)) { | |
670 | spin_unlock_bh(&conn->immed_queue_lock); | |
671 | return NULL; | |
672 | } | |
673 | list_for_each_entry(qr, &conn->immed_queue_list, qr_list) | |
674 | break; | |
675 | ||
676 | list_del(&qr->qr_list); | |
677 | if (qr->cmd) | |
678 | atomic_dec(&qr->cmd->immed_queue_count); | |
679 | spin_unlock_bh(&conn->immed_queue_lock); | |
680 | ||
681 | return qr; | |
682 | } | |
683 | ||
684 | static void iscsit_remove_cmd_from_immediate_queue( | |
685 | struct iscsi_cmd *cmd, | |
686 | struct iscsi_conn *conn) | |
687 | { | |
688 | struct iscsi_queue_req *qr, *qr_tmp; | |
689 | ||
690 | spin_lock_bh(&conn->immed_queue_lock); | |
691 | if (!atomic_read(&cmd->immed_queue_count)) { | |
692 | spin_unlock_bh(&conn->immed_queue_lock); | |
693 | return; | |
694 | } | |
695 | ||
696 | list_for_each_entry_safe(qr, qr_tmp, &conn->immed_queue_list, qr_list) { | |
697 | if (qr->cmd != cmd) | |
698 | continue; | |
699 | ||
700 | atomic_dec(&qr->cmd->immed_queue_count); | |
701 | list_del(&qr->qr_list); | |
702 | kmem_cache_free(lio_qr_cache, qr); | |
703 | } | |
704 | spin_unlock_bh(&conn->immed_queue_lock); | |
705 | ||
706 | if (atomic_read(&cmd->immed_queue_count)) { | |
707 | pr_err("ITT: 0x%08x immed_queue_count: %d\n", | |
708 | cmd->init_task_tag, | |
709 | atomic_read(&cmd->immed_queue_count)); | |
710 | } | |
711 | } | |
712 | ||
713 | void iscsit_add_cmd_to_response_queue( | |
714 | struct iscsi_cmd *cmd, | |
715 | struct iscsi_conn *conn, | |
716 | u8 state) | |
717 | { | |
718 | struct iscsi_queue_req *qr; | |
719 | ||
720 | qr = kmem_cache_zalloc(lio_qr_cache, GFP_ATOMIC); | |
721 | if (!qr) { | |
722 | pr_err("Unable to allocate memory for" | |
723 | " struct iscsi_queue_req\n"); | |
724 | return; | |
725 | } | |
726 | INIT_LIST_HEAD(&qr->qr_list); | |
727 | qr->cmd = cmd; | |
728 | qr->state = state; | |
729 | ||
730 | spin_lock_bh(&conn->response_queue_lock); | |
731 | list_add_tail(&qr->qr_list, &conn->response_queue_list); | |
732 | atomic_inc(&cmd->response_queue_count); | |
733 | spin_unlock_bh(&conn->response_queue_lock); | |
734 | ||
735 | wake_up_process(conn->thread_set->tx_thread); | |
736 | } | |
737 | ||
738 | struct iscsi_queue_req *iscsit_get_cmd_from_response_queue(struct iscsi_conn *conn) | |
739 | { | |
740 | struct iscsi_queue_req *qr; | |
741 | ||
742 | spin_lock_bh(&conn->response_queue_lock); | |
743 | if (list_empty(&conn->response_queue_list)) { | |
744 | spin_unlock_bh(&conn->response_queue_lock); | |
745 | return NULL; | |
746 | } | |
747 | ||
748 | list_for_each_entry(qr, &conn->response_queue_list, qr_list) | |
749 | break; | |
750 | ||
751 | list_del(&qr->qr_list); | |
752 | if (qr->cmd) | |
753 | atomic_dec(&qr->cmd->response_queue_count); | |
754 | spin_unlock_bh(&conn->response_queue_lock); | |
755 | ||
756 | return qr; | |
757 | } | |
758 | ||
759 | static void iscsit_remove_cmd_from_response_queue( | |
760 | struct iscsi_cmd *cmd, | |
761 | struct iscsi_conn *conn) | |
762 | { | |
763 | struct iscsi_queue_req *qr, *qr_tmp; | |
764 | ||
765 | spin_lock_bh(&conn->response_queue_lock); | |
766 | if (!atomic_read(&cmd->response_queue_count)) { | |
767 | spin_unlock_bh(&conn->response_queue_lock); | |
768 | return; | |
769 | } | |
770 | ||
771 | list_for_each_entry_safe(qr, qr_tmp, &conn->response_queue_list, | |
772 | qr_list) { | |
773 | if (qr->cmd != cmd) | |
774 | continue; | |
775 | ||
776 | atomic_dec(&qr->cmd->response_queue_count); | |
777 | list_del(&qr->qr_list); | |
778 | kmem_cache_free(lio_qr_cache, qr); | |
779 | } | |
780 | spin_unlock_bh(&conn->response_queue_lock); | |
781 | ||
782 | if (atomic_read(&cmd->response_queue_count)) { | |
783 | pr_err("ITT: 0x%08x response_queue_count: %d\n", | |
784 | cmd->init_task_tag, | |
785 | atomic_read(&cmd->response_queue_count)); | |
786 | } | |
787 | } | |
788 | ||
789 | void iscsit_free_queue_reqs_for_conn(struct iscsi_conn *conn) | |
790 | { | |
791 | struct iscsi_queue_req *qr, *qr_tmp; | |
792 | ||
793 | spin_lock_bh(&conn->immed_queue_lock); | |
794 | list_for_each_entry_safe(qr, qr_tmp, &conn->immed_queue_list, qr_list) { | |
795 | list_del(&qr->qr_list); | |
796 | if (qr->cmd) | |
797 | atomic_dec(&qr->cmd->immed_queue_count); | |
798 | ||
799 | kmem_cache_free(lio_qr_cache, qr); | |
800 | } | |
801 | spin_unlock_bh(&conn->immed_queue_lock); | |
802 | ||
803 | spin_lock_bh(&conn->response_queue_lock); | |
804 | list_for_each_entry_safe(qr, qr_tmp, &conn->response_queue_list, | |
805 | qr_list) { | |
806 | list_del(&qr->qr_list); | |
807 | if (qr->cmd) | |
808 | atomic_dec(&qr->cmd->response_queue_count); | |
809 | ||
810 | kmem_cache_free(lio_qr_cache, qr); | |
811 | } | |
812 | spin_unlock_bh(&conn->response_queue_lock); | |
813 | } | |
814 | ||
815 | void iscsit_release_cmd(struct iscsi_cmd *cmd) | |
816 | { | |
817 | struct iscsi_conn *conn = cmd->conn; | |
818 | int i; | |
819 | ||
820 | iscsit_free_r2ts_from_list(cmd); | |
821 | iscsit_free_all_datain_reqs(cmd); | |
822 | ||
823 | kfree(cmd->buf_ptr); | |
824 | kfree(cmd->pdu_list); | |
825 | kfree(cmd->seq_list); | |
826 | kfree(cmd->tmr_req); | |
827 | kfree(cmd->iov_data); | |
828 | ||
829 | for (i = 0; i < cmd->t_mem_sg_nents; i++) | |
830 | __free_page(sg_page(&cmd->t_mem_sg[i])); | |
831 | ||
832 | kfree(cmd->t_mem_sg); | |
833 | ||
834 | if (conn) { | |
835 | iscsit_remove_cmd_from_immediate_queue(cmd, conn); | |
836 | iscsit_remove_cmd_from_response_queue(cmd, conn); | |
837 | } | |
838 | ||
839 | kmem_cache_free(lio_cmd_cache, cmd); | |
840 | } | |
841 | ||
842 | int iscsit_check_session_usage_count(struct iscsi_session *sess) | |
843 | { | |
844 | spin_lock_bh(&sess->session_usage_lock); | |
845 | if (sess->session_usage_count != 0) { | |
846 | sess->session_waiting_on_uc = 1; | |
847 | spin_unlock_bh(&sess->session_usage_lock); | |
848 | if (in_interrupt()) | |
849 | return 2; | |
850 | ||
851 | wait_for_completion(&sess->session_waiting_on_uc_comp); | |
852 | return 1; | |
853 | } | |
854 | spin_unlock_bh(&sess->session_usage_lock); | |
855 | ||
856 | return 0; | |
857 | } | |
858 | ||
859 | void iscsit_dec_session_usage_count(struct iscsi_session *sess) | |
860 | { | |
861 | spin_lock_bh(&sess->session_usage_lock); | |
862 | sess->session_usage_count--; | |
863 | ||
864 | if (!sess->session_usage_count && sess->session_waiting_on_uc) | |
865 | complete(&sess->session_waiting_on_uc_comp); | |
866 | ||
867 | spin_unlock_bh(&sess->session_usage_lock); | |
868 | } | |
869 | ||
870 | void iscsit_inc_session_usage_count(struct iscsi_session *sess) | |
871 | { | |
872 | spin_lock_bh(&sess->session_usage_lock); | |
873 | sess->session_usage_count++; | |
874 | spin_unlock_bh(&sess->session_usage_lock); | |
875 | } | |
876 | ||
877 | /* | |
878 | * Used before iscsi_do[rx,tx]_data() to determine iov and [rx,tx]_marker | |
879 | * array counts needed for sync and steering. | |
880 | */ | |
881 | static int iscsit_determine_sync_and_steering_counts( | |
882 | struct iscsi_conn *conn, | |
883 | struct iscsi_data_count *count) | |
884 | { | |
885 | u32 length = count->data_length; | |
886 | u32 marker, markint; | |
887 | ||
888 | count->sync_and_steering = 1; | |
889 | ||
890 | marker = (count->type == ISCSI_RX_DATA) ? | |
891 | conn->of_marker : conn->if_marker; | |
892 | markint = (count->type == ISCSI_RX_DATA) ? | |
893 | (conn->conn_ops->OFMarkInt * 4) : | |
894 | (conn->conn_ops->IFMarkInt * 4); | |
895 | count->ss_iov_count = count->iov_count; | |
896 | ||
897 | while (length > 0) { | |
898 | if (length >= marker) { | |
899 | count->ss_iov_count += 3; | |
900 | count->ss_marker_count += 2; | |
901 | ||
902 | length -= marker; | |
903 | marker = markint; | |
904 | } else | |
905 | length = 0; | |
906 | } | |
907 | ||
908 | return 0; | |
909 | } | |
910 | ||
911 | /* | |
912 | * Setup conn->if_marker and conn->of_marker values based upon | |
913 | * the initial marker-less interval. (see iSCSI v19 A.2) | |
914 | */ | |
915 | int iscsit_set_sync_and_steering_values(struct iscsi_conn *conn) | |
916 | { | |
917 | int login_ifmarker_count = 0, login_ofmarker_count = 0, next_marker = 0; | |
918 | /* | |
919 | * IFMarkInt and OFMarkInt are negotiated as 32-bit words. | |
920 | */ | |
921 | u32 IFMarkInt = (conn->conn_ops->IFMarkInt * 4); | |
922 | u32 OFMarkInt = (conn->conn_ops->OFMarkInt * 4); | |
923 | ||
924 | if (conn->conn_ops->OFMarker) { | |
925 | /* | |
926 | * Account for the first Login Command received not | |
927 | * via iscsi_recv_msg(). | |
928 | */ | |
929 | conn->of_marker += ISCSI_HDR_LEN; | |
930 | if (conn->of_marker <= OFMarkInt) { | |
931 | conn->of_marker = (OFMarkInt - conn->of_marker); | |
932 | } else { | |
933 | login_ofmarker_count = (conn->of_marker / OFMarkInt); | |
934 | next_marker = (OFMarkInt * (login_ofmarker_count + 1)) + | |
935 | (login_ofmarker_count * MARKER_SIZE); | |
936 | conn->of_marker = (next_marker - conn->of_marker); | |
937 | } | |
938 | conn->of_marker_offset = 0; | |
939 | pr_debug("Setting OFMarker value to %u based on Initial" | |
940 | " Markerless Interval.\n", conn->of_marker); | |
941 | } | |
942 | ||
943 | if (conn->conn_ops->IFMarker) { | |
944 | if (conn->if_marker <= IFMarkInt) { | |
945 | conn->if_marker = (IFMarkInt - conn->if_marker); | |
946 | } else { | |
947 | login_ifmarker_count = (conn->if_marker / IFMarkInt); | |
948 | next_marker = (IFMarkInt * (login_ifmarker_count + 1)) + | |
949 | (login_ifmarker_count * MARKER_SIZE); | |
950 | conn->if_marker = (next_marker - conn->if_marker); | |
951 | } | |
952 | pr_debug("Setting IFMarker value to %u based on Initial" | |
953 | " Markerless Interval.\n", conn->if_marker); | |
954 | } | |
955 | ||
956 | return 0; | |
957 | } | |
958 | ||
959 | struct iscsi_conn *iscsit_get_conn_from_cid(struct iscsi_session *sess, u16 cid) | |
960 | { | |
961 | struct iscsi_conn *conn; | |
962 | ||
963 | spin_lock_bh(&sess->conn_lock); | |
964 | list_for_each_entry(conn, &sess->sess_conn_list, conn_list) { | |
965 | if ((conn->cid == cid) && | |
966 | (conn->conn_state == TARG_CONN_STATE_LOGGED_IN)) { | |
967 | iscsit_inc_conn_usage_count(conn); | |
968 | spin_unlock_bh(&sess->conn_lock); | |
969 | return conn; | |
970 | } | |
971 | } | |
972 | spin_unlock_bh(&sess->conn_lock); | |
973 | ||
974 | return NULL; | |
975 | } | |
976 | ||
977 | struct iscsi_conn *iscsit_get_conn_from_cid_rcfr(struct iscsi_session *sess, u16 cid) | |
978 | { | |
979 | struct iscsi_conn *conn; | |
980 | ||
981 | spin_lock_bh(&sess->conn_lock); | |
982 | list_for_each_entry(conn, &sess->sess_conn_list, conn_list) { | |
983 | if (conn->cid == cid) { | |
984 | iscsit_inc_conn_usage_count(conn); | |
985 | spin_lock(&conn->state_lock); | |
986 | atomic_set(&conn->connection_wait_rcfr, 1); | |
987 | spin_unlock(&conn->state_lock); | |
988 | spin_unlock_bh(&sess->conn_lock); | |
989 | return conn; | |
990 | } | |
991 | } | |
992 | spin_unlock_bh(&sess->conn_lock); | |
993 | ||
994 | return NULL; | |
995 | } | |
996 | ||
997 | void iscsit_check_conn_usage_count(struct iscsi_conn *conn) | |
998 | { | |
999 | spin_lock_bh(&conn->conn_usage_lock); | |
1000 | if (conn->conn_usage_count != 0) { | |
1001 | conn->conn_waiting_on_uc = 1; | |
1002 | spin_unlock_bh(&conn->conn_usage_lock); | |
1003 | ||
1004 | wait_for_completion(&conn->conn_waiting_on_uc_comp); | |
1005 | return; | |
1006 | } | |
1007 | spin_unlock_bh(&conn->conn_usage_lock); | |
1008 | } | |
1009 | ||
1010 | void iscsit_dec_conn_usage_count(struct iscsi_conn *conn) | |
1011 | { | |
1012 | spin_lock_bh(&conn->conn_usage_lock); | |
1013 | conn->conn_usage_count--; | |
1014 | ||
1015 | if (!conn->conn_usage_count && conn->conn_waiting_on_uc) | |
1016 | complete(&conn->conn_waiting_on_uc_comp); | |
1017 | ||
1018 | spin_unlock_bh(&conn->conn_usage_lock); | |
1019 | } | |
1020 | ||
1021 | void iscsit_inc_conn_usage_count(struct iscsi_conn *conn) | |
1022 | { | |
1023 | spin_lock_bh(&conn->conn_usage_lock); | |
1024 | conn->conn_usage_count++; | |
1025 | spin_unlock_bh(&conn->conn_usage_lock); | |
1026 | } | |
1027 | ||
1028 | static int iscsit_add_nopin(struct iscsi_conn *conn, int want_response) | |
1029 | { | |
1030 | u8 state; | |
1031 | struct iscsi_cmd *cmd; | |
1032 | ||
1033 | cmd = iscsit_allocate_cmd(conn, GFP_ATOMIC); | |
1034 | if (!cmd) | |
1035 | return -1; | |
1036 | ||
1037 | cmd->iscsi_opcode = ISCSI_OP_NOOP_IN; | |
1038 | state = (want_response) ? ISTATE_SEND_NOPIN_WANT_RESPONSE : | |
1039 | ISTATE_SEND_NOPIN_NO_RESPONSE; | |
1040 | cmd->init_task_tag = 0xFFFFFFFF; | |
1041 | spin_lock_bh(&conn->sess->ttt_lock); | |
1042 | cmd->targ_xfer_tag = (want_response) ? conn->sess->targ_xfer_tag++ : | |
1043 | 0xFFFFFFFF; | |
1044 | if (want_response && (cmd->targ_xfer_tag == 0xFFFFFFFF)) | |
1045 | cmd->targ_xfer_tag = conn->sess->targ_xfer_tag++; | |
1046 | spin_unlock_bh(&conn->sess->ttt_lock); | |
1047 | ||
1048 | spin_lock_bh(&conn->cmd_lock); | |
1049 | list_add_tail(&cmd->i_list, &conn->conn_cmd_list); | |
1050 | spin_unlock_bh(&conn->cmd_lock); | |
1051 | ||
1052 | if (want_response) | |
1053 | iscsit_start_nopin_response_timer(conn); | |
1054 | iscsit_add_cmd_to_immediate_queue(cmd, conn, state); | |
1055 | ||
1056 | return 0; | |
1057 | } | |
1058 | ||
1059 | static void iscsit_handle_nopin_response_timeout(unsigned long data) | |
1060 | { | |
1061 | struct iscsi_conn *conn = (struct iscsi_conn *) data; | |
1062 | ||
1063 | iscsit_inc_conn_usage_count(conn); | |
1064 | ||
1065 | spin_lock_bh(&conn->nopin_timer_lock); | |
1066 | if (conn->nopin_response_timer_flags & ISCSI_TF_STOP) { | |
1067 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1068 | iscsit_dec_conn_usage_count(conn); | |
1069 | return; | |
1070 | } | |
1071 | ||
1072 | pr_debug("Did not receive response to NOPIN on CID: %hu on" | |
1073 | " SID: %u, failing connection.\n", conn->cid, | |
1074 | conn->sess->sid); | |
1075 | conn->nopin_response_timer_flags &= ~ISCSI_TF_RUNNING; | |
1076 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1077 | ||
1078 | { | |
1079 | struct iscsi_portal_group *tpg = conn->sess->tpg; | |
1080 | struct iscsi_tiqn *tiqn = tpg->tpg_tiqn; | |
1081 | ||
1082 | if (tiqn) { | |
1083 | spin_lock_bh(&tiqn->sess_err_stats.lock); | |
1084 | strcpy(tiqn->sess_err_stats.last_sess_fail_rem_name, | |
1085 | (void *)conn->sess->sess_ops->InitiatorName); | |
1086 | tiqn->sess_err_stats.last_sess_failure_type = | |
1087 | ISCSI_SESS_ERR_CXN_TIMEOUT; | |
1088 | tiqn->sess_err_stats.cxn_timeout_errors++; | |
1089 | conn->sess->conn_timeout_errors++; | |
1090 | spin_unlock_bh(&tiqn->sess_err_stats.lock); | |
1091 | } | |
1092 | } | |
1093 | ||
1094 | iscsit_cause_connection_reinstatement(conn, 0); | |
1095 | iscsit_dec_conn_usage_count(conn); | |
1096 | } | |
1097 | ||
1098 | void iscsit_mod_nopin_response_timer(struct iscsi_conn *conn) | |
1099 | { | |
1100 | struct iscsi_session *sess = conn->sess; | |
1101 | struct iscsi_node_attrib *na = iscsit_tpg_get_node_attrib(sess); | |
1102 | ||
1103 | spin_lock_bh(&conn->nopin_timer_lock); | |
1104 | if (!(conn->nopin_response_timer_flags & ISCSI_TF_RUNNING)) { | |
1105 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1106 | return; | |
1107 | } | |
1108 | ||
1109 | mod_timer(&conn->nopin_response_timer, | |
1110 | (get_jiffies_64() + na->nopin_response_timeout * HZ)); | |
1111 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1112 | } | |
1113 | ||
1114 | /* | |
1115 | * Called with conn->nopin_timer_lock held. | |
1116 | */ | |
1117 | void iscsit_start_nopin_response_timer(struct iscsi_conn *conn) | |
1118 | { | |
1119 | struct iscsi_session *sess = conn->sess; | |
1120 | struct iscsi_node_attrib *na = iscsit_tpg_get_node_attrib(sess); | |
1121 | ||
1122 | spin_lock_bh(&conn->nopin_timer_lock); | |
1123 | if (conn->nopin_response_timer_flags & ISCSI_TF_RUNNING) { | |
1124 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1125 | return; | |
1126 | } | |
1127 | ||
1128 | init_timer(&conn->nopin_response_timer); | |
1129 | conn->nopin_response_timer.expires = | |
1130 | (get_jiffies_64() + na->nopin_response_timeout * HZ); | |
1131 | conn->nopin_response_timer.data = (unsigned long)conn; | |
1132 | conn->nopin_response_timer.function = iscsit_handle_nopin_response_timeout; | |
1133 | conn->nopin_response_timer_flags &= ~ISCSI_TF_STOP; | |
1134 | conn->nopin_response_timer_flags |= ISCSI_TF_RUNNING; | |
1135 | add_timer(&conn->nopin_response_timer); | |
1136 | ||
1137 | pr_debug("Started NOPIN Response Timer on CID: %d to %u" | |
1138 | " seconds\n", conn->cid, na->nopin_response_timeout); | |
1139 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1140 | } | |
1141 | ||
1142 | void iscsit_stop_nopin_response_timer(struct iscsi_conn *conn) | |
1143 | { | |
1144 | spin_lock_bh(&conn->nopin_timer_lock); | |
1145 | if (!(conn->nopin_response_timer_flags & ISCSI_TF_RUNNING)) { | |
1146 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1147 | return; | |
1148 | } | |
1149 | conn->nopin_response_timer_flags |= ISCSI_TF_STOP; | |
1150 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1151 | ||
1152 | del_timer_sync(&conn->nopin_response_timer); | |
1153 | ||
1154 | spin_lock_bh(&conn->nopin_timer_lock); | |
1155 | conn->nopin_response_timer_flags &= ~ISCSI_TF_RUNNING; | |
1156 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1157 | } | |
1158 | ||
1159 | static void iscsit_handle_nopin_timeout(unsigned long data) | |
1160 | { | |
1161 | struct iscsi_conn *conn = (struct iscsi_conn *) data; | |
1162 | ||
1163 | iscsit_inc_conn_usage_count(conn); | |
1164 | ||
1165 | spin_lock_bh(&conn->nopin_timer_lock); | |
1166 | if (conn->nopin_timer_flags & ISCSI_TF_STOP) { | |
1167 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1168 | iscsit_dec_conn_usage_count(conn); | |
1169 | return; | |
1170 | } | |
1171 | conn->nopin_timer_flags &= ~ISCSI_TF_RUNNING; | |
1172 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1173 | ||
1174 | iscsit_add_nopin(conn, 1); | |
1175 | iscsit_dec_conn_usage_count(conn); | |
1176 | } | |
1177 | ||
1178 | /* | |
1179 | * Called with conn->nopin_timer_lock held. | |
1180 | */ | |
1181 | void __iscsit_start_nopin_timer(struct iscsi_conn *conn) | |
1182 | { | |
1183 | struct iscsi_session *sess = conn->sess; | |
1184 | struct iscsi_node_attrib *na = iscsit_tpg_get_node_attrib(sess); | |
1185 | /* | |
1186 | * NOPIN timeout is disabled. | |
1187 | */ | |
1188 | if (!na->nopin_timeout) | |
1189 | return; | |
1190 | ||
1191 | if (conn->nopin_timer_flags & ISCSI_TF_RUNNING) | |
1192 | return; | |
1193 | ||
1194 | init_timer(&conn->nopin_timer); | |
1195 | conn->nopin_timer.expires = (get_jiffies_64() + na->nopin_timeout * HZ); | |
1196 | conn->nopin_timer.data = (unsigned long)conn; | |
1197 | conn->nopin_timer.function = iscsit_handle_nopin_timeout; | |
1198 | conn->nopin_timer_flags &= ~ISCSI_TF_STOP; | |
1199 | conn->nopin_timer_flags |= ISCSI_TF_RUNNING; | |
1200 | add_timer(&conn->nopin_timer); | |
1201 | ||
1202 | pr_debug("Started NOPIN Timer on CID: %d at %u second" | |
1203 | " interval\n", conn->cid, na->nopin_timeout); | |
1204 | } | |
1205 | ||
1206 | void iscsit_start_nopin_timer(struct iscsi_conn *conn) | |
1207 | { | |
1208 | struct iscsi_session *sess = conn->sess; | |
1209 | struct iscsi_node_attrib *na = iscsit_tpg_get_node_attrib(sess); | |
1210 | /* | |
1211 | * NOPIN timeout is disabled.. | |
1212 | */ | |
1213 | if (!na->nopin_timeout) | |
1214 | return; | |
1215 | ||
1216 | spin_lock_bh(&conn->nopin_timer_lock); | |
1217 | if (conn->nopin_timer_flags & ISCSI_TF_RUNNING) { | |
1218 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1219 | return; | |
1220 | } | |
1221 | ||
1222 | init_timer(&conn->nopin_timer); | |
1223 | conn->nopin_timer.expires = (get_jiffies_64() + na->nopin_timeout * HZ); | |
1224 | conn->nopin_timer.data = (unsigned long)conn; | |
1225 | conn->nopin_timer.function = iscsit_handle_nopin_timeout; | |
1226 | conn->nopin_timer_flags &= ~ISCSI_TF_STOP; | |
1227 | conn->nopin_timer_flags |= ISCSI_TF_RUNNING; | |
1228 | add_timer(&conn->nopin_timer); | |
1229 | ||
1230 | pr_debug("Started NOPIN Timer on CID: %d at %u second" | |
1231 | " interval\n", conn->cid, na->nopin_timeout); | |
1232 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1233 | } | |
1234 | ||
1235 | void iscsit_stop_nopin_timer(struct iscsi_conn *conn) | |
1236 | { | |
1237 | spin_lock_bh(&conn->nopin_timer_lock); | |
1238 | if (!(conn->nopin_timer_flags & ISCSI_TF_RUNNING)) { | |
1239 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1240 | return; | |
1241 | } | |
1242 | conn->nopin_timer_flags |= ISCSI_TF_STOP; | |
1243 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1244 | ||
1245 | del_timer_sync(&conn->nopin_timer); | |
1246 | ||
1247 | spin_lock_bh(&conn->nopin_timer_lock); | |
1248 | conn->nopin_timer_flags &= ~ISCSI_TF_RUNNING; | |
1249 | spin_unlock_bh(&conn->nopin_timer_lock); | |
1250 | } | |
1251 | ||
1252 | int iscsit_send_tx_data( | |
1253 | struct iscsi_cmd *cmd, | |
1254 | struct iscsi_conn *conn, | |
1255 | int use_misc) | |
1256 | { | |
1257 | int tx_sent, tx_size; | |
1258 | u32 iov_count; | |
1259 | struct kvec *iov; | |
1260 | ||
1261 | send_data: | |
1262 | tx_size = cmd->tx_size; | |
1263 | ||
1264 | if (!use_misc) { | |
1265 | iov = &cmd->iov_data[0]; | |
1266 | iov_count = cmd->iov_data_count; | |
1267 | } else { | |
1268 | iov = &cmd->iov_misc[0]; | |
1269 | iov_count = cmd->iov_misc_count; | |
1270 | } | |
1271 | ||
1272 | tx_sent = tx_data(conn, &iov[0], iov_count, tx_size); | |
1273 | if (tx_size != tx_sent) { | |
1274 | if (tx_sent == -EAGAIN) { | |
1275 | pr_err("tx_data() returned -EAGAIN\n"); | |
1276 | goto send_data; | |
1277 | } else | |
1278 | return -1; | |
1279 | } | |
1280 | cmd->tx_size = 0; | |
1281 | ||
1282 | return 0; | |
1283 | } | |
1284 | ||
1285 | int iscsit_fe_sendpage_sg( | |
1286 | struct iscsi_cmd *cmd, | |
1287 | struct iscsi_conn *conn) | |
1288 | { | |
1289 | struct scatterlist *sg = cmd->first_data_sg; | |
1290 | struct kvec iov; | |
1291 | u32 tx_hdr_size, data_len; | |
1292 | u32 offset = cmd->first_data_sg_off; | |
1293 | int tx_sent; | |
1294 | ||
1295 | send_hdr: | |
1296 | tx_hdr_size = ISCSI_HDR_LEN; | |
1297 | if (conn->conn_ops->HeaderDigest) | |
1298 | tx_hdr_size += ISCSI_CRC_LEN; | |
1299 | ||
1300 | iov.iov_base = cmd->pdu; | |
1301 | iov.iov_len = tx_hdr_size; | |
1302 | ||
1303 | tx_sent = tx_data(conn, &iov, 1, tx_hdr_size); | |
1304 | if (tx_hdr_size != tx_sent) { | |
1305 | if (tx_sent == -EAGAIN) { | |
1306 | pr_err("tx_data() returned -EAGAIN\n"); | |
1307 | goto send_hdr; | |
1308 | } | |
1309 | return -1; | |
1310 | } | |
1311 | ||
1312 | data_len = cmd->tx_size - tx_hdr_size - cmd->padding; | |
1313 | if (conn->conn_ops->DataDigest) | |
1314 | data_len -= ISCSI_CRC_LEN; | |
1315 | ||
1316 | /* | |
1317 | * Perform sendpage() for each page in the scatterlist | |
1318 | */ | |
1319 | while (data_len) { | |
1320 | u32 space = (sg->length - offset); | |
1321 | u32 sub_len = min_t(u32, data_len, space); | |
1322 | send_pg: | |
1323 | tx_sent = conn->sock->ops->sendpage(conn->sock, | |
1324 | sg_page(sg), sg->offset + offset, sub_len, 0); | |
1325 | if (tx_sent != sub_len) { | |
1326 | if (tx_sent == -EAGAIN) { | |
1327 | pr_err("tcp_sendpage() returned" | |
1328 | " -EAGAIN\n"); | |
1329 | goto send_pg; | |
1330 | } | |
1331 | ||
1332 | pr_err("tcp_sendpage() failure: %d\n", | |
1333 | tx_sent); | |
1334 | return -1; | |
1335 | } | |
1336 | ||
1337 | data_len -= sub_len; | |
1338 | offset = 0; | |
1339 | sg = sg_next(sg); | |
1340 | } | |
1341 | ||
1342 | send_padding: | |
1343 | if (cmd->padding) { | |
1344 | struct kvec *iov_p = | |
1345 | &cmd->iov_data[cmd->iov_data_count-1]; | |
1346 | ||
1347 | tx_sent = tx_data(conn, iov_p, 1, cmd->padding); | |
1348 | if (cmd->padding != tx_sent) { | |
1349 | if (tx_sent == -EAGAIN) { | |
1350 | pr_err("tx_data() returned -EAGAIN\n"); | |
1351 | goto send_padding; | |
1352 | } | |
1353 | return -1; | |
1354 | } | |
1355 | } | |
1356 | ||
1357 | send_datacrc: | |
1358 | if (conn->conn_ops->DataDigest) { | |
1359 | struct kvec *iov_d = | |
1360 | &cmd->iov_data[cmd->iov_data_count]; | |
1361 | ||
1362 | tx_sent = tx_data(conn, iov_d, 1, ISCSI_CRC_LEN); | |
1363 | if (ISCSI_CRC_LEN != tx_sent) { | |
1364 | if (tx_sent == -EAGAIN) { | |
1365 | pr_err("tx_data() returned -EAGAIN\n"); | |
1366 | goto send_datacrc; | |
1367 | } | |
1368 | return -1; | |
1369 | } | |
1370 | } | |
1371 | ||
1372 | return 0; | |
1373 | } | |
1374 | ||
1375 | /* | |
1376 | * This function is used for mainly sending a ISCSI_TARG_LOGIN_RSP PDU | |
1377 | * back to the Initiator when an expection condition occurs with the | |
1378 | * errors set in status_class and status_detail. | |
1379 | * | |
1380 | * Parameters: iSCSI Connection, Status Class, Status Detail. | |
1381 | * Returns: 0 on success, -1 on error. | |
1382 | */ | |
1383 | int iscsit_tx_login_rsp(struct iscsi_conn *conn, u8 status_class, u8 status_detail) | |
1384 | { | |
1385 | u8 iscsi_hdr[ISCSI_HDR_LEN]; | |
1386 | int err; | |
1387 | struct kvec iov; | |
1388 | struct iscsi_login_rsp *hdr; | |
1389 | ||
1390 | iscsit_collect_login_stats(conn, status_class, status_detail); | |
1391 | ||
1392 | memset(&iov, 0, sizeof(struct kvec)); | |
1393 | memset(&iscsi_hdr, 0x0, ISCSI_HDR_LEN); | |
1394 | ||
1395 | hdr = (struct iscsi_login_rsp *)&iscsi_hdr; | |
1396 | hdr->opcode = ISCSI_OP_LOGIN_RSP; | |
1397 | hdr->status_class = status_class; | |
1398 | hdr->status_detail = status_detail; | |
1399 | hdr->itt = cpu_to_be32(conn->login_itt); | |
1400 | ||
1401 | iov.iov_base = &iscsi_hdr; | |
1402 | iov.iov_len = ISCSI_HDR_LEN; | |
1403 | ||
1404 | PRINT_BUFF(iscsi_hdr, ISCSI_HDR_LEN); | |
1405 | ||
1406 | err = tx_data(conn, &iov, 1, ISCSI_HDR_LEN); | |
1407 | if (err != ISCSI_HDR_LEN) { | |
1408 | pr_err("tx_data returned less than expected\n"); | |
1409 | return -1; | |
1410 | } | |
1411 | ||
1412 | return 0; | |
1413 | } | |
1414 | ||
1415 | void iscsit_print_session_params(struct iscsi_session *sess) | |
1416 | { | |
1417 | struct iscsi_conn *conn; | |
1418 | ||
1419 | pr_debug("-----------------------------[Session Params for" | |
1420 | " SID: %u]-----------------------------\n", sess->sid); | |
1421 | spin_lock_bh(&sess->conn_lock); | |
1422 | list_for_each_entry(conn, &sess->sess_conn_list, conn_list) | |
1423 | iscsi_dump_conn_ops(conn->conn_ops); | |
1424 | spin_unlock_bh(&sess->conn_lock); | |
1425 | ||
1426 | iscsi_dump_sess_ops(sess->sess_ops); | |
1427 | } | |
1428 | ||
1429 | static int iscsit_do_rx_data( | |
1430 | struct iscsi_conn *conn, | |
1431 | struct iscsi_data_count *count) | |
1432 | { | |
1433 | int data = count->data_length, rx_loop = 0, total_rx = 0, iov_len; | |
1434 | u32 rx_marker_val[count->ss_marker_count], rx_marker_iov = 0; | |
1435 | struct kvec iov[count->ss_iov_count], *iov_p; | |
1436 | struct msghdr msg; | |
1437 | ||
1438 | if (!conn || !conn->sock || !conn->conn_ops) | |
1439 | return -1; | |
1440 | ||
1441 | memset(&msg, 0, sizeof(struct msghdr)); | |
1442 | ||
1443 | if (count->sync_and_steering) { | |
1444 | int size = 0; | |
1445 | u32 i, orig_iov_count = 0; | |
1446 | u32 orig_iov_len = 0, orig_iov_loc = 0; | |
1447 | u32 iov_count = 0, per_iov_bytes = 0; | |
1448 | u32 *rx_marker, old_rx_marker = 0; | |
1449 | struct kvec *iov_record; | |
1450 | ||
1451 | memset(&rx_marker_val, 0, | |
1452 | count->ss_marker_count * sizeof(u32)); | |
1453 | memset(&iov, 0, count->ss_iov_count * sizeof(struct kvec)); | |
1454 | ||
1455 | iov_record = count->iov; | |
1456 | orig_iov_count = count->iov_count; | |
1457 | rx_marker = &conn->of_marker; | |
1458 | ||
1459 | i = 0; | |
1460 | size = data; | |
1461 | orig_iov_len = iov_record[orig_iov_loc].iov_len; | |
1462 | while (size > 0) { | |
1463 | pr_debug("rx_data: #1 orig_iov_len %u," | |
1464 | " orig_iov_loc %u\n", orig_iov_len, orig_iov_loc); | |
1465 | pr_debug("rx_data: #2 rx_marker %u, size" | |
1466 | " %u\n", *rx_marker, size); | |
1467 | ||
1468 | if (orig_iov_len >= *rx_marker) { | |
1469 | iov[iov_count].iov_len = *rx_marker; | |
1470 | iov[iov_count++].iov_base = | |
1471 | (iov_record[orig_iov_loc].iov_base + | |
1472 | per_iov_bytes); | |
1473 | ||
1474 | iov[iov_count].iov_len = (MARKER_SIZE / 2); | |
1475 | iov[iov_count++].iov_base = | |
1476 | &rx_marker_val[rx_marker_iov++]; | |
1477 | iov[iov_count].iov_len = (MARKER_SIZE / 2); | |
1478 | iov[iov_count++].iov_base = | |
1479 | &rx_marker_val[rx_marker_iov++]; | |
1480 | old_rx_marker = *rx_marker; | |
1481 | ||
1482 | /* | |
1483 | * OFMarkInt is in 32-bit words. | |
1484 | */ | |
1485 | *rx_marker = (conn->conn_ops->OFMarkInt * 4); | |
1486 | size -= old_rx_marker; | |
1487 | orig_iov_len -= old_rx_marker; | |
1488 | per_iov_bytes += old_rx_marker; | |
1489 | ||
1490 | pr_debug("rx_data: #3 new_rx_marker" | |
1491 | " %u, size %u\n", *rx_marker, size); | |
1492 | } else { | |
1493 | iov[iov_count].iov_len = orig_iov_len; | |
1494 | iov[iov_count++].iov_base = | |
1495 | (iov_record[orig_iov_loc].iov_base + | |
1496 | per_iov_bytes); | |
1497 | ||
1498 | per_iov_bytes = 0; | |
1499 | *rx_marker -= orig_iov_len; | |
1500 | size -= orig_iov_len; | |
1501 | ||
1502 | if (size) | |
1503 | orig_iov_len = | |
1504 | iov_record[++orig_iov_loc].iov_len; | |
1505 | ||
1506 | pr_debug("rx_data: #4 new_rx_marker" | |
1507 | " %u, size %u\n", *rx_marker, size); | |
1508 | } | |
1509 | } | |
1510 | data += (rx_marker_iov * (MARKER_SIZE / 2)); | |
1511 | ||
1512 | iov_p = &iov[0]; | |
1513 | iov_len = iov_count; | |
1514 | ||
1515 | if (iov_count > count->ss_iov_count) { | |
1516 | pr_err("iov_count: %d, count->ss_iov_count:" | |
1517 | " %d\n", iov_count, count->ss_iov_count); | |
1518 | return -1; | |
1519 | } | |
1520 | if (rx_marker_iov > count->ss_marker_count) { | |
1521 | pr_err("rx_marker_iov: %d, count->ss_marker" | |
1522 | "_count: %d\n", rx_marker_iov, | |
1523 | count->ss_marker_count); | |
1524 | return -1; | |
1525 | } | |
1526 | } else { | |
1527 | iov_p = count->iov; | |
1528 | iov_len = count->iov_count; | |
1529 | } | |
1530 | ||
1531 | while (total_rx < data) { | |
1532 | rx_loop = kernel_recvmsg(conn->sock, &msg, iov_p, iov_len, | |
1533 | (data - total_rx), MSG_WAITALL); | |
1534 | if (rx_loop <= 0) { | |
1535 | pr_debug("rx_loop: %d total_rx: %d\n", | |
1536 | rx_loop, total_rx); | |
1537 | return rx_loop; | |
1538 | } | |
1539 | total_rx += rx_loop; | |
1540 | pr_debug("rx_loop: %d, total_rx: %d, data: %d\n", | |
1541 | rx_loop, total_rx, data); | |
1542 | } | |
1543 | ||
1544 | if (count->sync_and_steering) { | |
1545 | int j; | |
1546 | for (j = 0; j < rx_marker_iov; j++) { | |
1547 | pr_debug("rx_data: #5 j: %d, offset: %d\n", | |
1548 | j, rx_marker_val[j]); | |
1549 | conn->of_marker_offset = rx_marker_val[j]; | |
1550 | } | |
1551 | total_rx -= (rx_marker_iov * (MARKER_SIZE / 2)); | |
1552 | } | |
1553 | ||
1554 | return total_rx; | |
1555 | } | |
1556 | ||
1557 | static int iscsit_do_tx_data( | |
1558 | struct iscsi_conn *conn, | |
1559 | struct iscsi_data_count *count) | |
1560 | { | |
1561 | int data = count->data_length, total_tx = 0, tx_loop = 0, iov_len; | |
1562 | u32 tx_marker_val[count->ss_marker_count], tx_marker_iov = 0; | |
1563 | struct kvec iov[count->ss_iov_count], *iov_p; | |
1564 | struct msghdr msg; | |
1565 | ||
1566 | if (!conn || !conn->sock || !conn->conn_ops) | |
1567 | return -1; | |
1568 | ||
1569 | if (data <= 0) { | |
1570 | pr_err("Data length is: %d\n", data); | |
1571 | return -1; | |
1572 | } | |
1573 | ||
1574 | memset(&msg, 0, sizeof(struct msghdr)); | |
1575 | ||
1576 | if (count->sync_and_steering) { | |
1577 | int size = 0; | |
1578 | u32 i, orig_iov_count = 0; | |
1579 | u32 orig_iov_len = 0, orig_iov_loc = 0; | |
1580 | u32 iov_count = 0, per_iov_bytes = 0; | |
1581 | u32 *tx_marker, old_tx_marker = 0; | |
1582 | struct kvec *iov_record; | |
1583 | ||
1584 | memset(&tx_marker_val, 0, | |
1585 | count->ss_marker_count * sizeof(u32)); | |
1586 | memset(&iov, 0, count->ss_iov_count * sizeof(struct kvec)); | |
1587 | ||
1588 | iov_record = count->iov; | |
1589 | orig_iov_count = count->iov_count; | |
1590 | tx_marker = &conn->if_marker; | |
1591 | ||
1592 | i = 0; | |
1593 | size = data; | |
1594 | orig_iov_len = iov_record[orig_iov_loc].iov_len; | |
1595 | while (size > 0) { | |
1596 | pr_debug("tx_data: #1 orig_iov_len %u," | |
1597 | " orig_iov_loc %u\n", orig_iov_len, orig_iov_loc); | |
1598 | pr_debug("tx_data: #2 tx_marker %u, size" | |
1599 | " %u\n", *tx_marker, size); | |
1600 | ||
1601 | if (orig_iov_len >= *tx_marker) { | |
1602 | iov[iov_count].iov_len = *tx_marker; | |
1603 | iov[iov_count++].iov_base = | |
1604 | (iov_record[orig_iov_loc].iov_base + | |
1605 | per_iov_bytes); | |
1606 | ||
1607 | tx_marker_val[tx_marker_iov] = | |
1608 | (size - *tx_marker); | |
1609 | iov[iov_count].iov_len = (MARKER_SIZE / 2); | |
1610 | iov[iov_count++].iov_base = | |
1611 | &tx_marker_val[tx_marker_iov++]; | |
1612 | iov[iov_count].iov_len = (MARKER_SIZE / 2); | |
1613 | iov[iov_count++].iov_base = | |
1614 | &tx_marker_val[tx_marker_iov++]; | |
1615 | old_tx_marker = *tx_marker; | |
1616 | ||
1617 | /* | |
1618 | * IFMarkInt is in 32-bit words. | |
1619 | */ | |
1620 | *tx_marker = (conn->conn_ops->IFMarkInt * 4); | |
1621 | size -= old_tx_marker; | |
1622 | orig_iov_len -= old_tx_marker; | |
1623 | per_iov_bytes += old_tx_marker; | |
1624 | ||
1625 | pr_debug("tx_data: #3 new_tx_marker" | |
1626 | " %u, size %u\n", *tx_marker, size); | |
1627 | pr_debug("tx_data: #4 offset %u\n", | |
1628 | tx_marker_val[tx_marker_iov-1]); | |
1629 | } else { | |
1630 | iov[iov_count].iov_len = orig_iov_len; | |
1631 | iov[iov_count++].iov_base | |
1632 | = (iov_record[orig_iov_loc].iov_base + | |
1633 | per_iov_bytes); | |
1634 | ||
1635 | per_iov_bytes = 0; | |
1636 | *tx_marker -= orig_iov_len; | |
1637 | size -= orig_iov_len; | |
1638 | ||
1639 | if (size) | |
1640 | orig_iov_len = | |
1641 | iov_record[++orig_iov_loc].iov_len; | |
1642 | ||
1643 | pr_debug("tx_data: #5 new_tx_marker" | |
1644 | " %u, size %u\n", *tx_marker, size); | |
1645 | } | |
1646 | } | |
1647 | ||
1648 | data += (tx_marker_iov * (MARKER_SIZE / 2)); | |
1649 | ||
1650 | iov_p = &iov[0]; | |
1651 | iov_len = iov_count; | |
1652 | ||
1653 | if (iov_count > count->ss_iov_count) { | |
1654 | pr_err("iov_count: %d, count->ss_iov_count:" | |
1655 | " %d\n", iov_count, count->ss_iov_count); | |
1656 | return -1; | |
1657 | } | |
1658 | if (tx_marker_iov > count->ss_marker_count) { | |
1659 | pr_err("tx_marker_iov: %d, count->ss_marker" | |
1660 | "_count: %d\n", tx_marker_iov, | |
1661 | count->ss_marker_count); | |
1662 | return -1; | |
1663 | } | |
1664 | } else { | |
1665 | iov_p = count->iov; | |
1666 | iov_len = count->iov_count; | |
1667 | } | |
1668 | ||
1669 | while (total_tx < data) { | |
1670 | tx_loop = kernel_sendmsg(conn->sock, &msg, iov_p, iov_len, | |
1671 | (data - total_tx)); | |
1672 | if (tx_loop <= 0) { | |
1673 | pr_debug("tx_loop: %d total_tx %d\n", | |
1674 | tx_loop, total_tx); | |
1675 | return tx_loop; | |
1676 | } | |
1677 | total_tx += tx_loop; | |
1678 | pr_debug("tx_loop: %d, total_tx: %d, data: %d\n", | |
1679 | tx_loop, total_tx, data); | |
1680 | } | |
1681 | ||
1682 | if (count->sync_and_steering) | |
1683 | total_tx -= (tx_marker_iov * (MARKER_SIZE / 2)); | |
1684 | ||
1685 | return total_tx; | |
1686 | } | |
1687 | ||
1688 | int rx_data( | |
1689 | struct iscsi_conn *conn, | |
1690 | struct kvec *iov, | |
1691 | int iov_count, | |
1692 | int data) | |
1693 | { | |
1694 | struct iscsi_data_count c; | |
1695 | ||
1696 | if (!conn || !conn->sock || !conn->conn_ops) | |
1697 | return -1; | |
1698 | ||
1699 | memset(&c, 0, sizeof(struct iscsi_data_count)); | |
1700 | c.iov = iov; | |
1701 | c.iov_count = iov_count; | |
1702 | c.data_length = data; | |
1703 | c.type = ISCSI_RX_DATA; | |
1704 | ||
1705 | if (conn->conn_ops->OFMarker && | |
1706 | (conn->conn_state >= TARG_CONN_STATE_LOGGED_IN)) { | |
1707 | if (iscsit_determine_sync_and_steering_counts(conn, &c) < 0) | |
1708 | return -1; | |
1709 | } | |
1710 | ||
1711 | return iscsit_do_rx_data(conn, &c); | |
1712 | } | |
1713 | ||
1714 | int tx_data( | |
1715 | struct iscsi_conn *conn, | |
1716 | struct kvec *iov, | |
1717 | int iov_count, | |
1718 | int data) | |
1719 | { | |
1720 | struct iscsi_data_count c; | |
1721 | ||
1722 | if (!conn || !conn->sock || !conn->conn_ops) | |
1723 | return -1; | |
1724 | ||
1725 | memset(&c, 0, sizeof(struct iscsi_data_count)); | |
1726 | c.iov = iov; | |
1727 | c.iov_count = iov_count; | |
1728 | c.data_length = data; | |
1729 | c.type = ISCSI_TX_DATA; | |
1730 | ||
1731 | if (conn->conn_ops->IFMarker && | |
1732 | (conn->conn_state >= TARG_CONN_STATE_LOGGED_IN)) { | |
1733 | if (iscsit_determine_sync_and_steering_counts(conn, &c) < 0) | |
1734 | return -1; | |
1735 | } | |
1736 | ||
1737 | return iscsit_do_tx_data(conn, &c); | |
1738 | } | |
1739 | ||
1740 | void iscsit_collect_login_stats( | |
1741 | struct iscsi_conn *conn, | |
1742 | u8 status_class, | |
1743 | u8 status_detail) | |
1744 | { | |
1745 | struct iscsi_param *intrname = NULL; | |
1746 | struct iscsi_tiqn *tiqn; | |
1747 | struct iscsi_login_stats *ls; | |
1748 | ||
1749 | tiqn = iscsit_snmp_get_tiqn(conn); | |
1750 | if (!tiqn) | |
1751 | return; | |
1752 | ||
1753 | ls = &tiqn->login_stats; | |
1754 | ||
1755 | spin_lock(&ls->lock); | |
1756 | if (!strcmp(conn->login_ip, ls->last_intr_fail_ip_addr) && | |
1757 | ((get_jiffies_64() - ls->last_fail_time) < 10)) { | |
1758 | /* We already have the failure info for this login */ | |
1759 | spin_unlock(&ls->lock); | |
1760 | return; | |
1761 | } | |
1762 | ||
1763 | if (status_class == ISCSI_STATUS_CLS_SUCCESS) | |
1764 | ls->accepts++; | |
1765 | else if (status_class == ISCSI_STATUS_CLS_REDIRECT) { | |
1766 | ls->redirects++; | |
1767 | ls->last_fail_type = ISCSI_LOGIN_FAIL_REDIRECT; | |
1768 | } else if ((status_class == ISCSI_STATUS_CLS_INITIATOR_ERR) && | |
1769 | (status_detail == ISCSI_LOGIN_STATUS_AUTH_FAILED)) { | |
1770 | ls->authenticate_fails++; | |
1771 | ls->last_fail_type = ISCSI_LOGIN_FAIL_AUTHENTICATE; | |
1772 | } else if ((status_class == ISCSI_STATUS_CLS_INITIATOR_ERR) && | |
1773 | (status_detail == ISCSI_LOGIN_STATUS_TGT_FORBIDDEN)) { | |
1774 | ls->authorize_fails++; | |
1775 | ls->last_fail_type = ISCSI_LOGIN_FAIL_AUTHORIZE; | |
1776 | } else if ((status_class == ISCSI_STATUS_CLS_INITIATOR_ERR) && | |
1777 | (status_detail == ISCSI_LOGIN_STATUS_INIT_ERR)) { | |
1778 | ls->negotiate_fails++; | |
1779 | ls->last_fail_type = ISCSI_LOGIN_FAIL_NEGOTIATE; | |
1780 | } else { | |
1781 | ls->other_fails++; | |
1782 | ls->last_fail_type = ISCSI_LOGIN_FAIL_OTHER; | |
1783 | } | |
1784 | ||
1785 | /* Save initiator name, ip address and time, if it is a failed login */ | |
1786 | if (status_class != ISCSI_STATUS_CLS_SUCCESS) { | |
1787 | if (conn->param_list) | |
1788 | intrname = iscsi_find_param_from_key(INITIATORNAME, | |
1789 | conn->param_list); | |
1790 | strcpy(ls->last_intr_fail_name, | |
1791 | (intrname ? intrname->value : "Unknown")); | |
1792 | ||
1793 | ls->last_intr_fail_ip_family = conn->sock->sk->sk_family; | |
1794 | snprintf(ls->last_intr_fail_ip_addr, IPV6_ADDRESS_SPACE, | |
1795 | "%s", conn->login_ip); | |
1796 | ls->last_fail_time = get_jiffies_64(); | |
1797 | } | |
1798 | ||
1799 | spin_unlock(&ls->lock); | |
1800 | } | |
1801 | ||
1802 | struct iscsi_tiqn *iscsit_snmp_get_tiqn(struct iscsi_conn *conn) | |
1803 | { | |
1804 | struct iscsi_portal_group *tpg; | |
1805 | ||
1806 | if (!conn || !conn->sess) | |
1807 | return NULL; | |
1808 | ||
1809 | tpg = conn->sess->tpg; | |
1810 | if (!tpg) | |
1811 | return NULL; | |
1812 | ||
1813 | if (!tpg->tpg_tiqn) | |
1814 | return NULL; | |
1815 | ||
1816 | return tpg->tpg_tiqn; | |
1817 | } |