]>
Commit | Line | Data |
---|---|---|
1a59d1b8 | 1 | // SPDX-License-Identifier: GPL-2.0-or-later |
c66ac9db NB |
2 | /******************************************************************************* |
3 | * Filename: target_core_tpg.c | |
4 | * | |
5 | * This file contains generic Target Portal Group related functions. | |
6 | * | |
4c76251e | 7 | * (c) Copyright 2002-2013 Datera, Inc. |
c66ac9db NB |
8 | * |
9 | * Nicholas A. Bellinger <nab@kernel.org> | |
10 | * | |
c66ac9db NB |
11 | ******************************************************************************/ |
12 | ||
13 | #include <linux/net.h> | |
14 | #include <linux/string.h> | |
15 | #include <linux/timer.h> | |
16 | #include <linux/slab.h> | |
17 | #include <linux/spinlock.h> | |
c66ac9db | 18 | #include <linux/in.h> |
c53181af | 19 | #include <linux/export.h> |
c66ac9db NB |
20 | #include <net/sock.h> |
21 | #include <net/tcp.h> | |
ba929992 | 22 | #include <scsi/scsi_proto.h> |
c66ac9db NB |
23 | |
24 | #include <target/target_core_base.h> | |
c4795fb2 CH |
25 | #include <target/target_core_backend.h> |
26 | #include <target/target_core_fabric.h> | |
c66ac9db | 27 | |
e26d99ae | 28 | #include "target_core_internal.h" |
adf653f9 | 29 | #include "target_core_alua.h" |
e2480563 | 30 | #include "target_core_pr.h" |
e986a35a | 31 | #include "target_core_ua.h" |
e3d6f909 AG |
32 | |
33 | extern struct se_device *g_lun0_dev; | |
34 | ||
c66ac9db NB |
35 | /* __core_tpg_get_initiator_node_acl(): |
36 | * | |
403edd78 | 37 | * mutex_lock(&tpg->acl_node_mutex); must be held when calling |
c66ac9db NB |
38 | */ |
39 | struct se_node_acl *__core_tpg_get_initiator_node_acl( | |
40 | struct se_portal_group *tpg, | |
41 | const char *initiatorname) | |
42 | { | |
43 | struct se_node_acl *acl; | |
44 | ||
45 | list_for_each_entry(acl, &tpg->acl_node_list, acl_list) { | |
6708bb27 | 46 | if (!strcmp(acl->initiatorname, initiatorname)) |
c66ac9db NB |
47 | return acl; |
48 | } | |
49 | ||
50 | return NULL; | |
51 | } | |
52 | ||
53 | /* core_tpg_get_initiator_node_acl(): | |
54 | * | |
55 | * | |
56 | */ | |
57 | struct se_node_acl *core_tpg_get_initiator_node_acl( | |
58 | struct se_portal_group *tpg, | |
59 | unsigned char *initiatorname) | |
60 | { | |
61 | struct se_node_acl *acl; | |
21aaa23b NB |
62 | /* |
63 | * Obtain se_node_acl->acl_kref using fabric driver provided | |
64 | * initiatorname[] during node acl endpoint lookup driven by | |
65 | * new se_session login. | |
66 | * | |
67 | * The reference is held until se_session shutdown -> release | |
68 | * occurs via fabric driver invoked transport_deregister_session() | |
69 | * or transport_free_session() code. | |
70 | */ | |
403edd78 | 71 | mutex_lock(&tpg->acl_node_mutex); |
fcf29481 | 72 | acl = __core_tpg_get_initiator_node_acl(tpg, initiatorname); |
21aaa23b NB |
73 | if (acl) { |
74 | if (!kref_get_unless_zero(&acl->acl_kref)) | |
75 | acl = NULL; | |
76 | } | |
403edd78 | 77 | mutex_unlock(&tpg->acl_node_mutex); |
c66ac9db | 78 | |
fcf29481 | 79 | return acl; |
c66ac9db | 80 | } |
b3fde035 | 81 | EXPORT_SYMBOL(core_tpg_get_initiator_node_acl); |
c66ac9db | 82 | |
e986a35a HR |
83 | void core_allocate_nexus_loss_ua( |
84 | struct se_node_acl *nacl) | |
85 | { | |
86 | struct se_dev_entry *deve; | |
87 | ||
88 | if (!nacl) | |
89 | return; | |
90 | ||
91 | rcu_read_lock(); | |
92 | hlist_for_each_entry_rcu(deve, &nacl->lun_entry_hlist, link) | |
93 | core_scsi3_ua_allocate(deve, 0x29, | |
94 | ASCQ_29H_NEXUS_LOSS_OCCURRED); | |
95 | rcu_read_unlock(); | |
96 | } | |
97 | EXPORT_SYMBOL(core_allocate_nexus_loss_ua); | |
98 | ||
c66ac9db NB |
99 | /* core_tpg_add_node_to_devs(): |
100 | * | |
101 | * | |
102 | */ | |
103 | void core_tpg_add_node_to_devs( | |
104 | struct se_node_acl *acl, | |
df9766ca NB |
105 | struct se_portal_group *tpg, |
106 | struct se_lun *lun_orig) | |
c66ac9db | 107 | { |
03a68b44 | 108 | bool lun_access_ro = true; |
c66ac9db NB |
109 | struct se_lun *lun; |
110 | struct se_device *dev; | |
111 | ||
6bb82612 NB |
112 | mutex_lock(&tpg->tpg_lun_mutex); |
113 | hlist_for_each_entry_rcu(lun, &tpg->tpg_lun_hlist, link) { | |
df9766ca | 114 | if (lun_orig && lun != lun_orig) |
c66ac9db NB |
115 | continue; |
116 | ||
4cc987ea NB |
117 | dev = rcu_dereference_check(lun->lun_se_dev, |
118 | lockdep_is_held(&tpg->tpg_lun_mutex)); | |
c66ac9db NB |
119 | /* |
120 | * By default in LIO-Target $FABRIC_MOD, | |
121 | * demo_mode_write_protect is ON, or READ_ONLY; | |
122 | */ | |
6708bb27 | 123 | if (!tpg->se_tpg_tfo->tpg_check_demo_mode_write_protect(tpg)) { |
03a68b44 | 124 | lun_access_ro = false; |
c66ac9db NB |
125 | } else { |
126 | /* | |
127 | * Allow only optical drives to issue R/W in default RO | |
128 | * demo mode. | |
129 | */ | |
e3d6f909 | 130 | if (dev->transport->get_device_type(dev) == TYPE_DISK) |
03a68b44 | 131 | lun_access_ro = true; |
c66ac9db | 132 | else |
03a68b44 | 133 | lun_access_ro = false; |
c66ac9db NB |
134 | } |
135 | ||
f2d30680 | 136 | pr_debug("TARGET_CORE[%s]->TPG[%u]_LUN[%llu] - Adding %s" |
c66ac9db | 137 | " access for LUN in Demo Mode\n", |
30c7ca93 | 138 | tpg->se_tpg_tfo->fabric_name, |
e3d6f909 | 139 | tpg->se_tpg_tfo->tpg_get_tag(tpg), lun->unpacked_lun, |
03a68b44 | 140 | lun_access_ro ? "READ-ONLY" : "READ-WRITE"); |
c66ac9db | 141 | |
e80ac6c4 | 142 | core_enable_device_list_for_node(lun, NULL, lun->unpacked_lun, |
03a68b44 | 143 | lun_access_ro, acl, tpg); |
e2480563 NB |
144 | /* |
145 | * Check to see if there are any existing persistent reservation | |
146 | * APTPL pre-registrations that need to be enabled for this dynamic | |
147 | * LUN ACL now.. | |
148 | */ | |
149 | core_scsi3_check_aptpl_registration(dev, tpg, lun, acl, | |
150 | lun->unpacked_lun); | |
c66ac9db | 151 | } |
6bb82612 | 152 | mutex_unlock(&tpg->tpg_lun_mutex); |
c66ac9db NB |
153 | } |
154 | ||
d36ad77f NB |
155 | static void |
156 | target_set_nacl_queue_depth(struct se_portal_group *tpg, | |
157 | struct se_node_acl *acl, u32 queue_depth) | |
c66ac9db | 158 | { |
d36ad77f NB |
159 | acl->queue_depth = queue_depth; |
160 | ||
c66ac9db | 161 | if (!acl->queue_depth) { |
d36ad77f | 162 | pr_warn("Queue depth for %s Initiator Node: %s is 0," |
30c7ca93 | 163 | "defaulting to 1.\n", tpg->se_tpg_tfo->fabric_name, |
c66ac9db NB |
164 | acl->initiatorname); |
165 | acl->queue_depth = 1; | |
166 | } | |
c66ac9db NB |
167 | } |
168 | ||
e413f472 CH |
169 | static struct se_node_acl *target_alloc_node_acl(struct se_portal_group *tpg, |
170 | const unsigned char *initiatorname) | |
4a5a75f3 | 171 | { |
c66ac9db | 172 | struct se_node_acl *acl; |
d36ad77f | 173 | u32 queue_depth; |
4a5a75f3 | 174 | |
144bc4c2 CH |
175 | acl = kzalloc(max(sizeof(*acl), tpg->se_tpg_tfo->node_acl_size), |
176 | GFP_KERNEL); | |
6708bb27 | 177 | if (!acl) |
c66ac9db | 178 | return NULL; |
4a5a75f3 | 179 | |
c66ac9db NB |
180 | INIT_LIST_HEAD(&acl->acl_list); |
181 | INIT_LIST_HEAD(&acl->acl_sess_list); | |
29a05dee | 182 | INIT_HLIST_HEAD(&acl->lun_entry_hlist); |
afb999ff | 183 | kref_init(&acl->acl_kref); |
01468346 | 184 | init_completion(&acl->acl_free_comp); |
c66ac9db | 185 | spin_lock_init(&acl->nacl_sess_lock); |
29a05dee | 186 | mutex_init(&acl->lun_entry_mutex); |
c66ac9db | 187 | atomic_set(&acl->acl_pr_ref_count, 0); |
d36ad77f | 188 | |
e1750d20 | 189 | if (tpg->se_tpg_tfo->tpg_get_default_depth) |
d36ad77f | 190 | queue_depth = tpg->se_tpg_tfo->tpg_get_default_depth(tpg); |
e1750d20 | 191 | else |
d36ad77f NB |
192 | queue_depth = 1; |
193 | target_set_nacl_queue_depth(tpg, acl, queue_depth); | |
194 | ||
c66ac9db NB |
195 | snprintf(acl->initiatorname, TRANSPORT_IQN_LEN, "%s", initiatorname); |
196 | acl->se_tpg = tpg; | |
197 | acl->acl_index = scsi_get_new_index(SCSI_AUTH_INTR_INDEX); | |
4a5a75f3 | 198 | |
e3d6f909 | 199 | tpg->se_tpg_tfo->set_default_node_attributes(acl); |
c66ac9db | 200 | |
e413f472 | 201 | return acl; |
4a5a75f3 JE |
202 | } |
203 | ||
e413f472 | 204 | static void target_add_node_acl(struct se_node_acl *acl) |
c66ac9db | 205 | { |
e413f472 | 206 | struct se_portal_group *tpg = acl->se_tpg; |
c66ac9db | 207 | |
403edd78 | 208 | mutex_lock(&tpg->acl_node_mutex); |
e413f472 | 209 | list_add_tail(&acl->acl_list, &tpg->acl_node_list); |
403edd78 | 210 | mutex_unlock(&tpg->acl_node_mutex); |
e413f472 CH |
211 | |
212 | pr_debug("%s_TPG[%hu] - Added %s ACL with TCQ Depth: %d for %s" | |
213 | " Initiator Node: %s\n", | |
30c7ca93 | 214 | tpg->se_tpg_tfo->fabric_name, |
e413f472 CH |
215 | tpg->se_tpg_tfo->tpg_get_tag(tpg), |
216 | acl->dynamic_node_acl ? "DYNAMIC" : "", | |
217 | acl->queue_depth, | |
30c7ca93 | 218 | tpg->se_tpg_tfo->fabric_name, |
e413f472 | 219 | acl->initiatorname); |
c66ac9db NB |
220 | } |
221 | ||
21aaa23b NB |
222 | bool target_tpg_has_node_acl(struct se_portal_group *tpg, |
223 | const char *initiatorname) | |
224 | { | |
225 | struct se_node_acl *acl; | |
226 | bool found = false; | |
227 | ||
228 | mutex_lock(&tpg->acl_node_mutex); | |
229 | list_for_each_entry(acl, &tpg->acl_node_list, acl_list) { | |
230 | if (!strcmp(acl->initiatorname, initiatorname)) { | |
231 | found = true; | |
232 | break; | |
233 | } | |
234 | } | |
235 | mutex_unlock(&tpg->acl_node_mutex); | |
236 | ||
237 | return found; | |
238 | } | |
239 | EXPORT_SYMBOL(target_tpg_has_node_acl); | |
240 | ||
c66ac9db NB |
241 | struct se_node_acl *core_tpg_check_initiator_node_acl( |
242 | struct se_portal_group *tpg, | |
243 | unsigned char *initiatorname) | |
244 | { | |
245 | struct se_node_acl *acl; | |
246 | ||
247 | acl = core_tpg_get_initiator_node_acl(tpg, initiatorname); | |
6708bb27 | 248 | if (acl) |
c66ac9db NB |
249 | return acl; |
250 | ||
6708bb27 | 251 | if (!tpg->se_tpg_tfo->tpg_check_demo_mode(tpg)) |
c66ac9db NB |
252 | return NULL; |
253 | ||
e413f472 | 254 | acl = target_alloc_node_acl(tpg, initiatorname); |
6708bb27 | 255 | if (!acl) |
c66ac9db | 256 | return NULL; |
21aaa23b NB |
257 | /* |
258 | * When allocating a dynamically generated node_acl, go ahead | |
259 | * and take the extra kref now before returning to the fabric | |
260 | * driver caller. | |
261 | * | |
262 | * Note this reference will be released at session shutdown | |
263 | * time within transport_free_session() code. | |
264 | */ | |
265 | kref_get(&acl->acl_kref); | |
c66ac9db NB |
266 | acl->dynamic_node_acl = 1; |
267 | ||
052605c6 NB |
268 | /* |
269 | * Here we only create demo-mode MappedLUNs from the active | |
35d1efe8 | 270 | * TPG LUNs if the fabric is not explicitly asking for |
052605c6 NB |
271 | * tpg_check_demo_mode_login_only() == 1. |
272 | */ | |
cdf88a2f AG |
273 | if ((tpg->se_tpg_tfo->tpg_check_demo_mode_login_only == NULL) || |
274 | (tpg->se_tpg_tfo->tpg_check_demo_mode_login_only(tpg) != 1)) | |
df9766ca | 275 | core_tpg_add_node_to_devs(acl, tpg, NULL); |
c66ac9db | 276 | |
e413f472 | 277 | target_add_node_acl(acl); |
c66ac9db NB |
278 | return acl; |
279 | } | |
280 | EXPORT_SYMBOL(core_tpg_check_initiator_node_acl); | |
281 | ||
282 | void core_tpg_wait_for_nacl_pr_ref(struct se_node_acl *nacl) | |
283 | { | |
284 | while (atomic_read(&nacl->acl_pr_ref_count) != 0) | |
285 | cpu_relax(); | |
286 | } | |
287 | ||
c66ac9db NB |
288 | struct se_node_acl *core_tpg_add_initiator_node_acl( |
289 | struct se_portal_group *tpg, | |
c7d6a803 | 290 | const char *initiatorname) |
c66ac9db | 291 | { |
c7d6a803 | 292 | struct se_node_acl *acl; |
c66ac9db | 293 | |
403edd78 | 294 | mutex_lock(&tpg->acl_node_mutex); |
c66ac9db | 295 | acl = __core_tpg_get_initiator_node_acl(tpg, initiatorname); |
6708bb27 | 296 | if (acl) { |
c66ac9db NB |
297 | if (acl->dynamic_node_acl) { |
298 | acl->dynamic_node_acl = 0; | |
6708bb27 | 299 | pr_debug("%s_TPG[%u] - Replacing dynamic ACL" |
30c7ca93 | 300 | " for %s\n", tpg->se_tpg_tfo->fabric_name, |
e3d6f909 | 301 | tpg->se_tpg_tfo->tpg_get_tag(tpg), initiatorname); |
403edd78 | 302 | mutex_unlock(&tpg->acl_node_mutex); |
e413f472 | 303 | return acl; |
c66ac9db NB |
304 | } |
305 | ||
6708bb27 | 306 | pr_err("ACL entry for %s Initiator" |
c66ac9db | 307 | " Node %s already exists for TPG %u, ignoring" |
30c7ca93 | 308 | " request.\n", tpg->se_tpg_tfo->fabric_name, |
e3d6f909 | 309 | initiatorname, tpg->se_tpg_tfo->tpg_get_tag(tpg)); |
403edd78 | 310 | mutex_unlock(&tpg->acl_node_mutex); |
c66ac9db NB |
311 | return ERR_PTR(-EEXIST); |
312 | } | |
403edd78 | 313 | mutex_unlock(&tpg->acl_node_mutex); |
c66ac9db | 314 | |
e413f472 CH |
315 | acl = target_alloc_node_acl(tpg, initiatorname); |
316 | if (!acl) | |
c66ac9db | 317 | return ERR_PTR(-ENOMEM); |
c66ac9db | 318 | |
e413f472 | 319 | target_add_node_acl(acl); |
c66ac9db NB |
320 | return acl; |
321 | } | |
c66ac9db | 322 | |
fba81f88 | 323 | static void target_shutdown_sessions(struct se_node_acl *acl) |
c66ac9db | 324 | { |
bc6e6bb4 | 325 | struct se_session *sess; |
140854cb | 326 | unsigned long flags; |
c66ac9db | 327 | |
bc6e6bb4 | 328 | restart: |
fba81f88 | 329 | spin_lock_irqsave(&acl->nacl_sess_lock, flags); |
bc6e6bb4 | 330 | list_for_each_entry(sess, &acl->acl_sess_list, sess_acl_list) { |
6f55b06f | 331 | if (atomic_read(&sess->stopped)) |
c66ac9db | 332 | continue; |
337c0607 | 333 | |
fba81f88 | 334 | list_del_init(&sess->sess_acl_list); |
bc6e6bb4 | 335 | spin_unlock_irqrestore(&acl->nacl_sess_lock, flags); |
d94331fa CH |
336 | |
337 | if (acl->se_tpg->se_tpg_tfo->close_session) | |
338 | acl->se_tpg->se_tpg_tfo->close_session(sess); | |
bc6e6bb4 | 339 | goto restart; |
c66ac9db | 340 | } |
bc6e6bb4 CH |
341 | spin_unlock_irqrestore(&acl->nacl_sess_lock, flags); |
342 | } | |
343 | ||
344 | void core_tpg_del_initiator_node_acl(struct se_node_acl *acl) | |
345 | { | |
346 | struct se_portal_group *tpg = acl->se_tpg; | |
347 | ||
348 | mutex_lock(&tpg->acl_node_mutex); | |
349 | if (acl->dynamic_node_acl) | |
350 | acl->dynamic_node_acl = 0; | |
6f48655f | 351 | list_del_init(&acl->acl_list); |
bc6e6bb4 CH |
352 | mutex_unlock(&tpg->acl_node_mutex); |
353 | ||
fba81f88 | 354 | target_shutdown_sessions(acl); |
bc6e6bb4 | 355 | |
337c0607 NB |
356 | target_put_nacl(acl); |
357 | /* | |
358 | * Wait for last target_put_nacl() to complete in target_complete_nacl() | |
359 | * for active fabric session transport_deregister_session() callbacks. | |
360 | */ | |
361 | wait_for_completion(&acl->acl_free_comp); | |
c66ac9db NB |
362 | |
363 | core_tpg_wait_for_nacl_pr_ref(acl); | |
c66ac9db NB |
364 | core_free_device_list_for_node(acl, tpg); |
365 | ||
6708bb27 | 366 | pr_debug("%s_TPG[%hu] - Deleted ACL with TCQ Depth: %d for %s" |
30c7ca93 | 367 | " Initiator Node: %s\n", tpg->se_tpg_tfo->fabric_name, |
e3d6f909 | 368 | tpg->se_tpg_tfo->tpg_get_tag(tpg), acl->queue_depth, |
30c7ca93 | 369 | tpg->se_tpg_tfo->fabric_name, acl->initiatorname); |
c66ac9db | 370 | |
144bc4c2 | 371 | kfree(acl); |
c66ac9db | 372 | } |
c66ac9db NB |
373 | |
374 | /* core_tpg_set_initiator_node_queue_depth(): | |
375 | * | |
376 | * | |
377 | */ | |
378 | int core_tpg_set_initiator_node_queue_depth( | |
d36ad77f NB |
379 | struct se_node_acl *acl, |
380 | u32 queue_depth) | |
c66ac9db | 381 | { |
d36ad77f | 382 | struct se_portal_group *tpg = acl->se_tpg; |
c66ac9db | 383 | |
46861cdd NB |
384 | /* |
385 | * Allow the setting of se_node_acl queue_depth to be idempotent, | |
386 | * and not force a session shutdown event if the value is not | |
387 | * changing. | |
388 | */ | |
389 | if (acl->queue_depth == queue_depth) | |
390 | return 0; | |
c66ac9db NB |
391 | /* |
392 | * User has requested to change the queue depth for a Initiator Node. | |
393 | * Change the value in the Node's struct se_node_acl, and call | |
d36ad77f | 394 | * target_set_nacl_queue_depth() to set the new queue depth. |
c66ac9db | 395 | */ |
d36ad77f NB |
396 | target_set_nacl_queue_depth(tpg, acl, queue_depth); |
397 | ||
bc6e6bb4 CH |
398 | /* |
399 | * Shutdown all pending sessions to force session reinstatement. | |
400 | */ | |
fba81f88 | 401 | target_shutdown_sessions(acl); |
c66ac9db | 402 | |
bfb9035c | 403 | pr_debug("Successfully changed queue depth to: %d for Initiator" |
d36ad77f | 404 | " Node: %s on %s Target Portal Group: %u\n", acl->queue_depth, |
30c7ca93 | 405 | acl->initiatorname, tpg->se_tpg_tfo->fabric_name, |
e3d6f909 | 406 | tpg->se_tpg_tfo->tpg_get_tag(tpg)); |
c66ac9db | 407 | |
c66ac9db NB |
408 | return 0; |
409 | } | |
410 | EXPORT_SYMBOL(core_tpg_set_initiator_node_queue_depth); | |
411 | ||
79e62fc3 AG |
412 | /* core_tpg_set_initiator_node_tag(): |
413 | * | |
414 | * Initiator nodeacl tags are not used internally, but may be used by | |
415 | * userspace to emulate aliases or groups. | |
416 | * Returns length of newly-set tag or -EINVAL. | |
417 | */ | |
418 | int core_tpg_set_initiator_node_tag( | |
419 | struct se_portal_group *tpg, | |
420 | struct se_node_acl *acl, | |
421 | const char *new_tag) | |
422 | { | |
423 | if (strlen(new_tag) >= MAX_ACL_TAG_SIZE) | |
424 | return -EINVAL; | |
425 | ||
426 | if (!strncmp("NULL", new_tag, 4)) { | |
427 | acl->acl_tag[0] = '\0'; | |
428 | return 0; | |
429 | } | |
430 | ||
431 | return snprintf(acl->acl_tag, MAX_ACL_TAG_SIZE, "%s", new_tag); | |
432 | } | |
433 | EXPORT_SYMBOL(core_tpg_set_initiator_node_tag); | |
434 | ||
5277797d NB |
435 | static void core_tpg_lun_ref_release(struct percpu_ref *ref) |
436 | { | |
437 | struct se_lun *lun = container_of(ref, struct se_lun, lun_ref); | |
438 | ||
bd4e2d29 | 439 | complete(&lun->lun_shutdown_comp); |
5277797d NB |
440 | } |
441 | ||
2bce1a6d | 442 | /* Does not change se_wwn->priv. */ |
c66ac9db | 443 | int core_tpg_register( |
c66ac9db NB |
444 | struct se_wwn *se_wwn, |
445 | struct se_portal_group *se_tpg, | |
e4aae5af | 446 | int proto_id) |
c66ac9db | 447 | { |
adf653f9 CH |
448 | int ret; |
449 | ||
bc0c94b1 NB |
450 | if (!se_tpg) |
451 | return -EINVAL; | |
452 | /* | |
453 | * For the typical case where core_tpg_register() is called by a | |
454 | * fabric driver from target_core_fabric_ops->fabric_make_tpg() | |
455 | * configfs context, use the original tf_ops pointer already saved | |
456 | * by target-core in target_fabric_make_wwn(). | |
457 | * | |
458 | * Otherwise, for special cases like iscsi-target discovery TPGs | |
459 | * the caller is responsible for setting ->se_tpg_tfo ahead of | |
460 | * calling core_tpg_register(). | |
461 | */ | |
462 | if (se_wwn) | |
463 | se_tpg->se_tpg_tfo = se_wwn->wwn_tf->tf_ops; | |
c66ac9db | 464 | |
bc0c94b1 NB |
465 | if (!se_tpg->se_tpg_tfo) { |
466 | pr_err("Unable to locate se_tpg->se_tpg_tfo pointer\n"); | |
467 | return -EINVAL; | |
c66ac9db NB |
468 | } |
469 | ||
6bb82612 | 470 | INIT_HLIST_HEAD(&se_tpg->tpg_lun_hlist); |
e4aae5af | 471 | se_tpg->proto_id = proto_id; |
c66ac9db NB |
472 | se_tpg->se_tpg_wwn = se_wwn; |
473 | atomic_set(&se_tpg->tpg_pr_ref_count, 0); | |
474 | INIT_LIST_HEAD(&se_tpg->acl_node_list); | |
c66ac9db | 475 | INIT_LIST_HEAD(&se_tpg->tpg_sess_list); |
c66ac9db | 476 | spin_lock_init(&se_tpg->session_lock); |
6bb82612 | 477 | mutex_init(&se_tpg->tpg_lun_mutex); |
403edd78 | 478 | mutex_init(&se_tpg->acl_node_mutex); |
c66ac9db | 479 | |
e4aae5af | 480 | if (se_tpg->proto_id >= 0) { |
adf653f9 CH |
481 | se_tpg->tpg_virt_lun0 = core_tpg_alloc_lun(se_tpg, 0); |
482 | if (IS_ERR(se_tpg->tpg_virt_lun0)) | |
483 | return PTR_ERR(se_tpg->tpg_virt_lun0); | |
484 | ||
485 | ret = core_tpg_add_lun(se_tpg, se_tpg->tpg_virt_lun0, | |
03a68b44 | 486 | true, g_lun0_dev); |
adf653f9 CH |
487 | if (ret < 0) { |
488 | kfree(se_tpg->tpg_virt_lun0); | |
489 | return ret; | |
c66ac9db NB |
490 | } |
491 | } | |
492 | ||
e4aae5af | 493 | pr_debug("TARGET_CORE[%s]: Allocated portal_group for endpoint: %s, " |
30c7ca93 | 494 | "Proto: %d, Portal Tag: %u\n", se_tpg->se_tpg_tfo->fabric_name, |
bc0c94b1 NB |
495 | se_tpg->se_tpg_tfo->tpg_get_wwn(se_tpg) ? |
496 | se_tpg->se_tpg_tfo->tpg_get_wwn(se_tpg) : NULL, | |
497 | se_tpg->proto_id, se_tpg->se_tpg_tfo->tpg_get_tag(se_tpg)); | |
c66ac9db NB |
498 | |
499 | return 0; | |
500 | } | |
501 | EXPORT_SYMBOL(core_tpg_register); | |
502 | ||
503 | int core_tpg_deregister(struct se_portal_group *se_tpg) | |
504 | { | |
e4aae5af | 505 | const struct target_core_fabric_ops *tfo = se_tpg->se_tpg_tfo; |
e89d15ee | 506 | struct se_node_acl *nacl, *nacl_tmp; |
22793de5 | 507 | LIST_HEAD(node_list); |
e89d15ee | 508 | |
e4aae5af | 509 | pr_debug("TARGET_CORE[%s]: Deallocating portal_group for endpoint: %s, " |
30c7ca93 | 510 | "Proto: %d, Portal Tag: %u\n", tfo->fabric_name, |
e4aae5af CH |
511 | tfo->tpg_get_wwn(se_tpg) ? tfo->tpg_get_wwn(se_tpg) : NULL, |
512 | se_tpg->proto_id, tfo->tpg_get_tag(se_tpg)); | |
c66ac9db | 513 | |
c66ac9db NB |
514 | while (atomic_read(&se_tpg->tpg_pr_ref_count) != 0) |
515 | cpu_relax(); | |
403edd78 | 516 | |
22793de5 NB |
517 | mutex_lock(&se_tpg->acl_node_mutex); |
518 | list_splice_init(&se_tpg->acl_node_list, &node_list); | |
519 | mutex_unlock(&se_tpg->acl_node_mutex); | |
e89d15ee NB |
520 | /* |
521 | * Release any remaining demo-mode generated se_node_acl that have | |
522 | * not been released because of TFO->tpg_check_demo_mode_cache() == 1 | |
523 | * in transport_deregister_session(). | |
524 | */ | |
22793de5 | 525 | list_for_each_entry_safe(nacl, nacl_tmp, &node_list, acl_list) { |
6f48655f | 526 | list_del_init(&nacl->acl_list); |
e89d15ee NB |
527 | |
528 | core_tpg_wait_for_nacl_pr_ref(nacl); | |
529 | core_free_device_list_for_node(nacl, se_tpg); | |
144bc4c2 | 530 | kfree(nacl); |
e89d15ee | 531 | } |
c66ac9db | 532 | |
adf653f9 CH |
533 | if (se_tpg->proto_id >= 0) { |
534 | core_tpg_remove_lun(se_tpg, se_tpg->tpg_virt_lun0); | |
535 | kfree_rcu(se_tpg->tpg_virt_lun0, rcu_head); | |
536 | } | |
c66ac9db | 537 | |
c66ac9db NB |
538 | return 0; |
539 | } | |
540 | EXPORT_SYMBOL(core_tpg_deregister); | |
541 | ||
d344f8a1 | 542 | struct se_lun *core_tpg_alloc_lun( |
c66ac9db | 543 | struct se_portal_group *tpg, |
f2d30680 | 544 | u64 unpacked_lun) |
c66ac9db NB |
545 | { |
546 | struct se_lun *lun; | |
547 | ||
6bb82612 NB |
548 | lun = kzalloc(sizeof(*lun), GFP_KERNEL); |
549 | if (!lun) { | |
550 | pr_err("Unable to allocate se_lun memory\n"); | |
551 | return ERR_PTR(-ENOMEM); | |
c66ac9db | 552 | } |
6bb82612 | 553 | lun->unpacked_lun = unpacked_lun; |
6bb82612 | 554 | atomic_set(&lun->lun_acl_count, 0); |
bd4e2d29 | 555 | init_completion(&lun->lun_shutdown_comp); |
adf653f9 CH |
556 | INIT_LIST_HEAD(&lun->lun_deve_list); |
557 | INIT_LIST_HEAD(&lun->lun_dev_link); | |
558 | atomic_set(&lun->lun_tg_pt_secondary_offline, 0); | |
559 | spin_lock_init(&lun->lun_deve_lock); | |
560 | mutex_init(&lun->lun_tg_pt_md_mutex); | |
561 | INIT_LIST_HEAD(&lun->lun_tg_pt_gp_link); | |
562 | spin_lock_init(&lun->lun_tg_pt_gp_lock); | |
adf653f9 | 563 | lun->lun_tpg = tpg; |
c66ac9db NB |
564 | |
565 | return lun; | |
566 | } | |
567 | ||
d344f8a1 | 568 | int core_tpg_add_lun( |
c66ac9db NB |
569 | struct se_portal_group *tpg, |
570 | struct se_lun *lun, | |
03a68b44 | 571 | bool lun_access_ro, |
340dbf72 | 572 | struct se_device *dev) |
c66ac9db | 573 | { |
e3d6f909 AG |
574 | int ret; |
575 | ||
2aad2a86 | 576 | ret = percpu_ref_init(&lun->lun_ref, core_tpg_lun_ref_release, 0, |
a34375ef | 577 | GFP_KERNEL); |
e3d6f909 | 578 | if (ret < 0) |
adf653f9 | 579 | goto out; |
c66ac9db | 580 | |
adf653f9 CH |
581 | ret = core_alloc_rtpi(lun, dev); |
582 | if (ret) | |
583 | goto out_kill_ref; | |
584 | ||
69088a04 | 585 | if (!(dev->transport_flags & TRANSPORT_FLAG_PASSTHROUGH_ALUA) && |
adf653f9 CH |
586 | !(dev->se_hba->hba_flags & HBA_FLAGS_INTERNAL_USE)) |
587 | target_attach_tg_pt_gp(lun, dev->t10_alua.default_tg_pt_gp); | |
5277797d | 588 | |
6bb82612 | 589 | mutex_lock(&tpg->tpg_lun_mutex); |
adf653f9 | 590 | |
adf653f9 | 591 | spin_lock(&dev->se_port_lock); |
4cc987ea | 592 | lun->lun_index = dev->dev_index; |
adf653f9 CH |
593 | rcu_assign_pointer(lun->lun_se_dev, dev); |
594 | dev->export_count++; | |
595 | list_add_tail(&lun->lun_dev_link, &dev->dev_sep_list); | |
596 | spin_unlock(&dev->se_port_lock); | |
5277797d | 597 | |
eeeb9522 | 598 | if (dev->dev_flags & DF_READ_ONLY) |
03a68b44 | 599 | lun->lun_access_ro = true; |
eeeb9522 | 600 | else |
03a68b44 | 601 | lun->lun_access_ro = lun_access_ro; |
6bb82612 NB |
602 | if (!(dev->se_hba->hba_flags & HBA_FLAGS_INTERNAL_USE)) |
603 | hlist_add_head_rcu(&lun->link, &tpg->tpg_lun_hlist); | |
604 | mutex_unlock(&tpg->tpg_lun_mutex); | |
c66ac9db NB |
605 | |
606 | return 0; | |
adf653f9 CH |
607 | |
608 | out_kill_ref: | |
609 | percpu_ref_exit(&lun->lun_ref); | |
610 | out: | |
611 | return ret; | |
c66ac9db NB |
612 | } |
613 | ||
cd9d7cba | 614 | void core_tpg_remove_lun( |
c66ac9db NB |
615 | struct se_portal_group *tpg, |
616 | struct se_lun *lun) | |
617 | { | |
4cc987ea NB |
618 | /* |
619 | * rcu_dereference_raw protected by se_lun->lun_group symlink | |
620 | * reference to se_device->dev_group. | |
621 | */ | |
622 | struct se_device *dev = rcu_dereference_raw(lun->lun_se_dev); | |
6bb82612 | 623 | |
49cb77e2 NB |
624 | lun->lun_shutdown = true; |
625 | ||
4a9a6c8d | 626 | core_clear_lun_from_tpg(lun, tpg); |
9e37d042 NB |
627 | /* |
628 | * Wait for any active I/O references to percpu se_lun->lun_ref to | |
629 | * be released. Also, se_lun->lun_ref is now used by PR and ALUA | |
630 | * logic when referencing a remote target port during ALL_TGT_PT=1 | |
631 | * and generating UNIT_ATTENTIONs for ALUA access state transition. | |
632 | */ | |
4a9a6c8d | 633 | transport_clear_lun_ref(lun); |
c66ac9db | 634 | |
6bb82612 | 635 | mutex_lock(&tpg->tpg_lun_mutex); |
adf653f9 | 636 | if (lun->lun_se_dev) { |
adf653f9 | 637 | target_detach_tg_pt_gp(lun); |
c66ac9db | 638 | |
adf653f9 CH |
639 | spin_lock(&dev->se_port_lock); |
640 | list_del(&lun->lun_dev_link); | |
641 | dev->export_count--; | |
642 | rcu_assign_pointer(lun->lun_se_dev, NULL); | |
643 | spin_unlock(&dev->se_port_lock); | |
644 | } | |
6bb82612 NB |
645 | if (!(dev->se_hba->hba_flags & HBA_FLAGS_INTERNAL_USE)) |
646 | hlist_del_rcu(&lun->link); | |
49cb77e2 NB |
647 | |
648 | lun->lun_shutdown = false; | |
6bb82612 | 649 | mutex_unlock(&tpg->tpg_lun_mutex); |
c66ac9db | 650 | |
9a1049da | 651 | percpu_ref_exit(&lun->lun_ref); |
c66ac9db | 652 | } |