2 * Copyright (c) 2009, 2010, Oracle and/or its affiliates. All rights reserved.
4 * Copyright (c) 2012, Intel Corporation.
6 * Author: Zach Brown <zab@zabbo.net>
7 * Author: Peter J. Braam <braam@clusterfs.com>
8 * Author: Phil Schwan <phil@clusterfs.com>
9 * Author: Eric Barton <eric@bartonsoftware.com>
11 * This file is part of Portals, http://www.sf.net/projects/sandiaportals/
13 * Portals is free software; you can redistribute it and/or
14 * modify it under the terms of version 2 of the GNU General Public
15 * License as published by the Free Software Foundation.
17 * Portals is distributed in the hope that it will be useful,
18 * but WITHOUT ANY WARRANTY; without even the implied warranty of
19 * MERCHANTABILITY or FITNESS FOR A PARTICULAR PURPOSE. See the
20 * GNU General Public License for more details.
22 * You should have received a copy of the GNU General Public License
23 * along with Portals; if not, write to the Free Software
24 * Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
31 * pro_send_hello : send hello message
32 * pro_recv_hello : receive hello message
33 * pro_pack : pack message header
34 * pro_unpack : unpack message header
35 * pro_queue_tx_zcack() : Called holding BH lock: kss_lock
36 * return 1 if ACK is piggybacked, otherwise return 0
37 * pro_queue_tx_msg() : Called holding BH lock: kss_lock
38 * return the ACK that piggybacked by my message, or NULL
39 * pro_handle_zcreq() : handler of incoming ZC-REQ
40 * pro_handle_zcack() : handler of incoming ZC-ACK
41 * pro_match_tx() : Called holding glock
45 ksocknal_queue_tx_msg_v1(ksock_conn_t
*conn
, ksock_tx_t
*tx_msg
)
47 /* V1.x, just enqueue it */
48 list_add_tail(&tx_msg
->tx_list
, &conn
->ksnc_tx_queue
);
53 ksocknal_next_tx_carrier(ksock_conn_t
*conn
)
55 ksock_tx_t
*tx
= conn
->ksnc_tx_carrier
;
57 /* Called holding BH lock: conn->ksnc_scheduler->kss_lock */
58 LASSERT (!list_empty(&conn
->ksnc_tx_queue
));
61 /* Next TX that can carry ZC-ACK or LNet message */
62 if (tx
->tx_list
.next
== &conn
->ksnc_tx_queue
) {
63 /* no more packets queued */
64 conn
->ksnc_tx_carrier
= NULL
;
66 conn
->ksnc_tx_carrier
= list_entry(tx
->tx_list
.next
,
68 LASSERT (conn
->ksnc_tx_carrier
->tx_msg
.ksm_type
== tx
->tx_msg
.ksm_type
);
73 ksocknal_queue_tx_zcack_v2(ksock_conn_t
*conn
,
74 ksock_tx_t
*tx_ack
, __u64 cookie
)
76 ksock_tx_t
*tx
= conn
->ksnc_tx_carrier
;
78 LASSERT (tx_ack
== NULL
||
79 tx_ack
->tx_msg
.ksm_type
== KSOCK_MSG_NOOP
);
82 * Enqueue or piggyback tx_ack / cookie
83 * . no tx can piggyback cookie of tx_ack (or cookie), just
84 * enqueue the tx_ack (if tx_ack != NUL) and return NULL.
85 * . There is tx can piggyback cookie of tx_ack (or cookie),
86 * piggyback the cookie and return the tx.
90 list_add_tail(&tx_ack
->tx_list
,
91 &conn
->ksnc_tx_queue
);
92 conn
->ksnc_tx_carrier
= tx_ack
;
97 if (tx
->tx_msg
.ksm_type
== KSOCK_MSG_NOOP
) {
98 /* tx is noop zc-ack, can't piggyback zc-ack cookie */
100 list_add_tail(&tx_ack
->tx_list
,
101 &conn
->ksnc_tx_queue
);
105 LASSERT(tx
->tx_msg
.ksm_type
== KSOCK_MSG_LNET
);
106 LASSERT(tx
->tx_msg
.ksm_zc_cookies
[1] == 0);
109 cookie
= tx_ack
->tx_msg
.ksm_zc_cookies
[1];
111 /* piggyback the zc-ack cookie */
112 tx
->tx_msg
.ksm_zc_cookies
[1] = cookie
;
113 /* move on to the next TX which can carry cookie */
114 ksocknal_next_tx_carrier(conn
);
120 ksocknal_queue_tx_msg_v2(ksock_conn_t
*conn
, ksock_tx_t
*tx_msg
)
122 ksock_tx_t
*tx
= conn
->ksnc_tx_carrier
;
126 * . If there is no NOOP on the connection, just enqueue
127 * tx_msg and return NULL
128 * . If there is NOOP on the connection, piggyback the cookie
129 * and replace the NOOP tx, and return the NOOP tx.
131 if (tx
== NULL
) { /* nothing on queue */
132 list_add_tail(&tx_msg
->tx_list
, &conn
->ksnc_tx_queue
);
133 conn
->ksnc_tx_carrier
= tx_msg
;
137 if (tx
->tx_msg
.ksm_type
== KSOCK_MSG_LNET
) { /* nothing to carry */
138 list_add_tail(&tx_msg
->tx_list
, &conn
->ksnc_tx_queue
);
142 LASSERT (tx
->tx_msg
.ksm_type
== KSOCK_MSG_NOOP
);
144 /* There is a noop zc-ack can be piggybacked */
145 tx_msg
->tx_msg
.ksm_zc_cookies
[1] = tx
->tx_msg
.ksm_zc_cookies
[1];
146 ksocknal_next_tx_carrier(conn
);
148 /* use new_tx to replace the noop zc-ack packet */
149 list_add(&tx_msg
->tx_list
, &tx
->tx_list
);
150 list_del(&tx
->tx_list
);
156 ksocknal_queue_tx_zcack_v3(ksock_conn_t
*conn
,
157 ksock_tx_t
*tx_ack
, __u64 cookie
)
161 if (conn
->ksnc_type
!= SOCKLND_CONN_ACK
)
162 return ksocknal_queue_tx_zcack_v2(conn
, tx_ack
, cookie
);
164 /* non-blocking ZC-ACK (to router) */
165 LASSERT (tx_ack
== NULL
||
166 tx_ack
->tx_msg
.ksm_type
== KSOCK_MSG_NOOP
);
168 if ((tx
= conn
->ksnc_tx_carrier
) == NULL
) {
169 if (tx_ack
!= NULL
) {
170 list_add_tail(&tx_ack
->tx_list
,
171 &conn
->ksnc_tx_queue
);
172 conn
->ksnc_tx_carrier
= tx_ack
;
177 /* conn->ksnc_tx_carrier != NULL */
180 cookie
= tx_ack
->tx_msg
.ksm_zc_cookies
[1];
182 if (cookie
== SOCKNAL_KEEPALIVE_PING
) /* ignore keepalive PING */
185 if (tx
->tx_msg
.ksm_zc_cookies
[1] == SOCKNAL_KEEPALIVE_PING
) {
186 /* replace the keepalive PING with a real ACK */
187 LASSERT (tx
->tx_msg
.ksm_zc_cookies
[0] == 0);
188 tx
->tx_msg
.ksm_zc_cookies
[1] = cookie
;
192 if (cookie
== tx
->tx_msg
.ksm_zc_cookies
[0] ||
193 cookie
== tx
->tx_msg
.ksm_zc_cookies
[1]) {
194 CWARN("%s: duplicated ZC cookie: "LPU64
"\n",
195 libcfs_id2str(conn
->ksnc_peer
->ksnp_id
), cookie
);
196 return 1; /* XXX return error in the future */
199 if (tx
->tx_msg
.ksm_zc_cookies
[0] == 0) {
200 /* NOOP tx has only one ZC-ACK cookie, can carry at least one more */
201 if (tx
->tx_msg
.ksm_zc_cookies
[1] > cookie
) {
202 tx
->tx_msg
.ksm_zc_cookies
[0] = tx
->tx_msg
.ksm_zc_cookies
[1];
203 tx
->tx_msg
.ksm_zc_cookies
[1] = cookie
;
205 tx
->tx_msg
.ksm_zc_cookies
[0] = cookie
;
208 if (tx
->tx_msg
.ksm_zc_cookies
[0] - tx
->tx_msg
.ksm_zc_cookies
[1] > 2) {
209 /* not likely to carry more ACKs, skip it to simplify logic */
210 ksocknal_next_tx_carrier(conn
);
216 /* takes two or more cookies already */
218 if (tx
->tx_msg
.ksm_zc_cookies
[0] > tx
->tx_msg
.ksm_zc_cookies
[1]) {
221 /* two separated cookies: (a+2, a) or (a+1, a) */
222 LASSERT (tx
->tx_msg
.ksm_zc_cookies
[0] -
223 tx
->tx_msg
.ksm_zc_cookies
[1] <= 2);
225 if (tx
->tx_msg
.ksm_zc_cookies
[0] -
226 tx
->tx_msg
.ksm_zc_cookies
[1] == 2) {
227 if (cookie
== tx
->tx_msg
.ksm_zc_cookies
[1] + 1)
229 } else if (cookie
== tx
->tx_msg
.ksm_zc_cookies
[1] - 1) {
230 tmp
= tx
->tx_msg
.ksm_zc_cookies
[1];
231 } else if (cookie
== tx
->tx_msg
.ksm_zc_cookies
[0] + 1) {
232 tmp
= tx
->tx_msg
.ksm_zc_cookies
[0];
236 /* range of cookies */
237 tx
->tx_msg
.ksm_zc_cookies
[0] = tmp
- 1;
238 tx
->tx_msg
.ksm_zc_cookies
[1] = tmp
+ 1;
243 /* ksm_zc_cookies[0] < ksm_zc_cookies[1], it is range of cookies */
244 if (cookie
>= tx
->tx_msg
.ksm_zc_cookies
[0] &&
245 cookie
<= tx
->tx_msg
.ksm_zc_cookies
[1]) {
246 CWARN("%s: duplicated ZC cookie: "LPU64
"\n",
247 libcfs_id2str(conn
->ksnc_peer
->ksnp_id
), cookie
);
248 return 1; /* XXX: return error in the future */
251 if (cookie
== tx
->tx_msg
.ksm_zc_cookies
[1] + 1) {
252 tx
->tx_msg
.ksm_zc_cookies
[1] = cookie
;
256 if (cookie
== tx
->tx_msg
.ksm_zc_cookies
[0] - 1) {
257 tx
->tx_msg
.ksm_zc_cookies
[0] = cookie
;
262 /* failed to piggyback ZC-ACK */
263 if (tx_ack
!= NULL
) {
264 list_add_tail(&tx_ack
->tx_list
, &conn
->ksnc_tx_queue
);
265 /* the next tx can piggyback at least 1 ACK */
266 ksocknal_next_tx_carrier(conn
);
273 ksocknal_match_tx(ksock_conn_t
*conn
, ksock_tx_t
*tx
, int nonblk
)
277 #if SOCKNAL_VERSION_DEBUG
278 if (!*ksocknal_tunables
.ksnd_typed_conns
)
279 return SOCKNAL_MATCH_YES
;
282 if (tx
== NULL
|| tx
->tx_lnetmsg
== NULL
) {
284 nob
= offsetof(ksock_msg_t
, ksm_u
);
286 nob
= tx
->tx_lnetmsg
->msg_len
+
287 ((conn
->ksnc_proto
== &ksocknal_protocol_v1x
) ?
288 sizeof(lnet_hdr_t
) : sizeof(ksock_msg_t
));
291 /* default checking for typed connection */
292 switch (conn
->ksnc_type
) {
294 CERROR("ksnc_type bad: %u\n", conn
->ksnc_type
);
296 case SOCKLND_CONN_ANY
:
297 return SOCKNAL_MATCH_YES
;
299 case SOCKLND_CONN_BULK_IN
:
300 return SOCKNAL_MATCH_MAY
;
302 case SOCKLND_CONN_BULK_OUT
:
303 if (nob
< *ksocknal_tunables
.ksnd_min_bulk
)
304 return SOCKNAL_MATCH_MAY
;
306 return SOCKNAL_MATCH_YES
;
308 case SOCKLND_CONN_CONTROL
:
309 if (nob
>= *ksocknal_tunables
.ksnd_min_bulk
)
310 return SOCKNAL_MATCH_MAY
;
312 return SOCKNAL_MATCH_YES
;
317 ksocknal_match_tx_v3(ksock_conn_t
*conn
, ksock_tx_t
*tx
, int nonblk
)
321 if (tx
== NULL
|| tx
->tx_lnetmsg
== NULL
)
322 nob
= offsetof(ksock_msg_t
, ksm_u
);
324 nob
= tx
->tx_lnetmsg
->msg_len
+ sizeof(ksock_msg_t
);
326 switch (conn
->ksnc_type
) {
328 CERROR("ksnc_type bad: %u\n", conn
->ksnc_type
);
330 case SOCKLND_CONN_ANY
:
331 return SOCKNAL_MATCH_NO
;
333 case SOCKLND_CONN_ACK
:
335 return SOCKNAL_MATCH_YES
;
336 else if (tx
== NULL
|| tx
->tx_lnetmsg
== NULL
)
337 return SOCKNAL_MATCH_MAY
;
339 return SOCKNAL_MATCH_NO
;
341 case SOCKLND_CONN_BULK_OUT
:
343 return SOCKNAL_MATCH_NO
;
344 else if (nob
< *ksocknal_tunables
.ksnd_min_bulk
)
345 return SOCKNAL_MATCH_MAY
;
347 return SOCKNAL_MATCH_YES
;
349 case SOCKLND_CONN_CONTROL
:
351 return SOCKNAL_MATCH_NO
;
352 else if (nob
>= *ksocknal_tunables
.ksnd_min_bulk
)
353 return SOCKNAL_MATCH_MAY
;
355 return SOCKNAL_MATCH_YES
;
359 /* (Sink) handle incoming ZC request from sender */
361 ksocknal_handle_zcreq(ksock_conn_t
*c
, __u64 cookie
, int remote
)
363 ksock_peer_t
*peer
= c
->ksnc_peer
;
368 read_lock(&ksocknal_data
.ksnd_global_lock
);
370 conn
= ksocknal_find_conn_locked(peer
, NULL
, !!remote
);
372 ksock_sched_t
*sched
= conn
->ksnc_scheduler
;
374 LASSERT(conn
->ksnc_proto
->pro_queue_tx_zcack
!= NULL
);
376 spin_lock_bh(&sched
->kss_lock
);
378 rc
= conn
->ksnc_proto
->pro_queue_tx_zcack(conn
, NULL
, cookie
);
380 spin_unlock_bh(&sched
->kss_lock
);
382 if (rc
) { /* piggybacked */
383 read_unlock(&ksocknal_data
.ksnd_global_lock
);
388 read_unlock(&ksocknal_data
.ksnd_global_lock
);
390 /* ACK connection is not ready, or can't piggyback the ACK */
391 tx
= ksocknal_alloc_tx_noop(cookie
, !!remote
);
395 if ((rc
= ksocknal_launch_packet(peer
->ksnp_ni
, tx
, peer
->ksnp_id
)) == 0)
398 ksocknal_free_tx(tx
);
402 /* (Sender) handle ZC_ACK from sink */
404 ksocknal_handle_zcack(ksock_conn_t
*conn
, __u64 cookie1
, __u64 cookie2
)
406 ksock_peer_t
*peer
= conn
->ksnc_peer
;
415 count
= (cookie1
> cookie2
) ? 2 : (cookie2
- cookie1
+ 1);
417 if (cookie2
== SOCKNAL_KEEPALIVE_PING
&&
418 conn
->ksnc_proto
== &ksocknal_protocol_v3x
) {
419 /* keepalive PING for V3.x, just ignore it */
420 return count
== 1 ? 0 : -EPROTO
;
423 spin_lock(&peer
->ksnp_lock
);
425 list_for_each_entry_safe(tx
, tmp
,
426 &peer
->ksnp_zc_req_list
, tx_zc_list
) {
427 __u64 c
= tx
->tx_msg
.ksm_zc_cookies
[0];
429 if (c
== cookie1
|| c
== cookie2
|| (cookie1
< c
&& c
< cookie2
)) {
430 tx
->tx_msg
.ksm_zc_cookies
[0] = 0;
431 list_del(&tx
->tx_zc_list
);
432 list_add(&tx
->tx_zc_list
, &zlist
);
439 spin_unlock(&peer
->ksnp_lock
);
441 while (!list_empty(&zlist
)) {
442 tx
= list_entry(zlist
.next
, ksock_tx_t
, tx_zc_list
);
443 list_del(&tx
->tx_zc_list
);
444 ksocknal_tx_decref(tx
);
447 return count
== 0 ? 0 : -EPROTO
;
451 ksocknal_send_hello_v1 (ksock_conn_t
*conn
, ksock_hello_msg_t
*hello
)
453 socket_t
*sock
= conn
->ksnc_sock
;
455 lnet_magicversion_t
*hmv
;
459 CLASSERT(sizeof(lnet_magicversion_t
) == offsetof(lnet_hdr_t
, src_nid
));
461 LIBCFS_ALLOC(hdr
, sizeof(*hdr
));
463 CERROR("Can't allocate lnet_hdr_t\n");
467 hmv
= (lnet_magicversion_t
*)&hdr
->dest_nid
;
469 /* Re-organize V2.x message header to V1.x (lnet_hdr_t)
470 * header and send out */
471 hmv
->magic
= cpu_to_le32 (LNET_PROTO_TCP_MAGIC
);
472 hmv
->version_major
= cpu_to_le16 (KSOCK_PROTO_V1_MAJOR
);
473 hmv
->version_minor
= cpu_to_le16 (KSOCK_PROTO_V1_MINOR
);
475 if (the_lnet
.ln_testprotocompat
!= 0) {
476 /* single-shot proto check */
478 if ((the_lnet
.ln_testprotocompat
& 1) != 0) {
479 hmv
->version_major
++; /* just different! */
480 the_lnet
.ln_testprotocompat
&= ~1;
482 if ((the_lnet
.ln_testprotocompat
& 2) != 0) {
483 hmv
->magic
= LNET_PROTO_MAGIC
;
484 the_lnet
.ln_testprotocompat
&= ~2;
489 hdr
->src_nid
= cpu_to_le64 (hello
->kshm_src_nid
);
490 hdr
->src_pid
= cpu_to_le32 (hello
->kshm_src_pid
);
491 hdr
->type
= cpu_to_le32 (LNET_MSG_HELLO
);
492 hdr
->payload_length
= cpu_to_le32 (hello
->kshm_nips
* sizeof(__u32
));
493 hdr
->msg
.hello
.type
= cpu_to_le32 (hello
->kshm_ctype
);
494 hdr
->msg
.hello
.incarnation
= cpu_to_le64 (hello
->kshm_src_incarnation
);
496 rc
= libcfs_sock_write(sock
, hdr
, sizeof(*hdr
),lnet_acceptor_timeout());
499 CNETERR("Error %d sending HELLO hdr to %pI4h/%d\n",
500 rc
, &conn
->ksnc_ipaddr
, conn
->ksnc_port
);
504 if (hello
->kshm_nips
== 0)
507 for (i
= 0; i
< (int) hello
->kshm_nips
; i
++) {
508 hello
->kshm_ips
[i
] = __cpu_to_le32 (hello
->kshm_ips
[i
]);
511 rc
= libcfs_sock_write(sock
, hello
->kshm_ips
,
512 hello
->kshm_nips
* sizeof(__u32
),
513 lnet_acceptor_timeout());
515 CNETERR("Error %d sending HELLO payload (%d)"
516 " to %pI4h/%d\n", rc
, hello
->kshm_nips
,
517 &conn
->ksnc_ipaddr
, conn
->ksnc_port
);
520 LIBCFS_FREE(hdr
, sizeof(*hdr
));
526 ksocknal_send_hello_v2 (ksock_conn_t
*conn
, ksock_hello_msg_t
*hello
)
528 socket_t
*sock
= conn
->ksnc_sock
;
531 hello
->kshm_magic
= LNET_PROTO_MAGIC
;
532 hello
->kshm_version
= conn
->ksnc_proto
->pro_version
;
534 if (the_lnet
.ln_testprotocompat
!= 0) {
535 /* single-shot proto check */
537 if ((the_lnet
.ln_testprotocompat
& 1) != 0) {
538 hello
->kshm_version
++; /* just different! */
539 the_lnet
.ln_testprotocompat
&= ~1;
544 rc
= libcfs_sock_write(sock
, hello
, offsetof(ksock_hello_msg_t
, kshm_ips
),
545 lnet_acceptor_timeout());
548 CNETERR("Error %d sending HELLO hdr to %pI4h/%d\n",
549 rc
, &conn
->ksnc_ipaddr
, conn
->ksnc_port
);
553 if (hello
->kshm_nips
== 0)
556 rc
= libcfs_sock_write(sock
, hello
->kshm_ips
,
557 hello
->kshm_nips
* sizeof(__u32
),
558 lnet_acceptor_timeout());
560 CNETERR("Error %d sending HELLO payload (%d)"
561 " to %pI4h/%d\n", rc
, hello
->kshm_nips
,
562 &conn
->ksnc_ipaddr
, conn
->ksnc_port
);
569 ksocknal_recv_hello_v1(ksock_conn_t
*conn
, ksock_hello_msg_t
*hello
,int timeout
)
571 socket_t
*sock
= conn
->ksnc_sock
;
576 LIBCFS_ALLOC(hdr
, sizeof(*hdr
));
578 CERROR("Can't allocate lnet_hdr_t\n");
582 rc
= libcfs_sock_read(sock
, &hdr
->src_nid
,
583 sizeof (*hdr
) - offsetof (lnet_hdr_t
, src_nid
),
586 CERROR("Error %d reading rest of HELLO hdr from %pI4h\n",
587 rc
, &conn
->ksnc_ipaddr
);
588 LASSERT (rc
< 0 && rc
!= -EALREADY
);
592 /* ...and check we got what we expected */
593 if (hdr
->type
!= cpu_to_le32 (LNET_MSG_HELLO
)) {
594 CERROR("Expecting a HELLO hdr,"
595 " but got type %d from %pI4h\n",
596 le32_to_cpu (hdr
->type
),
602 hello
->kshm_src_nid
= le64_to_cpu (hdr
->src_nid
);
603 hello
->kshm_src_pid
= le32_to_cpu (hdr
->src_pid
);
604 hello
->kshm_src_incarnation
= le64_to_cpu (hdr
->msg
.hello
.incarnation
);
605 hello
->kshm_ctype
= le32_to_cpu (hdr
->msg
.hello
.type
);
606 hello
->kshm_nips
= le32_to_cpu (hdr
->payload_length
) /
609 if (hello
->kshm_nips
> LNET_MAX_INTERFACES
) {
610 CERROR("Bad nips %d from ip %pI4h\n",
611 hello
->kshm_nips
, &conn
->ksnc_ipaddr
);
616 if (hello
->kshm_nips
== 0)
619 rc
= libcfs_sock_read(sock
, hello
->kshm_ips
,
620 hello
->kshm_nips
* sizeof(__u32
), timeout
);
622 CERROR("Error %d reading IPs from ip %pI4h\n",
623 rc
, &conn
->ksnc_ipaddr
);
624 LASSERT(rc
< 0 && rc
!= -EALREADY
);
628 for (i
= 0; i
< (int) hello
->kshm_nips
; i
++) {
629 hello
->kshm_ips
[i
] = __le32_to_cpu(hello
->kshm_ips
[i
]);
631 if (hello
->kshm_ips
[i
] == 0) {
632 CERROR("Zero IP[%d] from ip %pI4h\n",
633 i
, &conn
->ksnc_ipaddr
);
639 LIBCFS_FREE(hdr
, sizeof(*hdr
));
645 ksocknal_recv_hello_v2 (ksock_conn_t
*conn
, ksock_hello_msg_t
*hello
, int timeout
)
647 socket_t
*sock
= conn
->ksnc_sock
;
651 if (hello
->kshm_magic
== LNET_PROTO_MAGIC
)
656 rc
= libcfs_sock_read(sock
, &hello
->kshm_src_nid
,
657 offsetof(ksock_hello_msg_t
, kshm_ips
) -
658 offsetof(ksock_hello_msg_t
, kshm_src_nid
),
661 CERROR("Error %d reading HELLO from %pI4h\n",
662 rc
, &conn
->ksnc_ipaddr
);
663 LASSERT(rc
< 0 && rc
!= -EALREADY
);
667 if (conn
->ksnc_flip
) {
668 __swab32s(&hello
->kshm_src_pid
);
669 __swab64s(&hello
->kshm_src_nid
);
670 __swab32s(&hello
->kshm_dst_pid
);
671 __swab64s(&hello
->kshm_dst_nid
);
672 __swab64s(&hello
->kshm_src_incarnation
);
673 __swab64s(&hello
->kshm_dst_incarnation
);
674 __swab32s(&hello
->kshm_ctype
);
675 __swab32s(&hello
->kshm_nips
);
678 if (hello
->kshm_nips
> LNET_MAX_INTERFACES
) {
679 CERROR("Bad nips %d from ip %pI4h\n",
680 hello
->kshm_nips
, &conn
->ksnc_ipaddr
);
684 if (hello
->kshm_nips
== 0)
687 rc
= libcfs_sock_read(sock
, hello
->kshm_ips
,
688 hello
->kshm_nips
* sizeof(__u32
), timeout
);
690 CERROR("Error %d reading IPs from ip %pI4h\n",
691 rc
, &conn
->ksnc_ipaddr
);
692 LASSERT(rc
< 0 && rc
!= -EALREADY
);
696 for (i
= 0; i
< (int) hello
->kshm_nips
; i
++) {
698 __swab32s(&hello
->kshm_ips
[i
]);
700 if (hello
->kshm_ips
[i
] == 0) {
701 CERROR("Zero IP[%d] from ip %pI4h\n",
702 i
, &conn
->ksnc_ipaddr
);
711 ksocknal_pack_msg_v1(ksock_tx_t
*tx
)
713 /* V1.x has no KSOCK_MSG_NOOP */
714 LASSERT(tx
->tx_msg
.ksm_type
!= KSOCK_MSG_NOOP
);
715 LASSERT(tx
->tx_lnetmsg
!= NULL
);
717 tx
->tx_iov
[0].iov_base
= (void *)&tx
->tx_lnetmsg
->msg_hdr
;
718 tx
->tx_iov
[0].iov_len
= sizeof(lnet_hdr_t
);
720 tx
->tx_resid
= tx
->tx_nob
= tx
->tx_lnetmsg
->msg_len
+ sizeof(lnet_hdr_t
);
724 ksocknal_pack_msg_v2(ksock_tx_t
*tx
)
726 tx
->tx_iov
[0].iov_base
= (void *)&tx
->tx_msg
;
728 if (tx
->tx_lnetmsg
!= NULL
) {
729 LASSERT(tx
->tx_msg
.ksm_type
!= KSOCK_MSG_NOOP
);
731 tx
->tx_msg
.ksm_u
.lnetmsg
.ksnm_hdr
= tx
->tx_lnetmsg
->msg_hdr
;
732 tx
->tx_iov
[0].iov_len
= sizeof(ksock_msg_t
);
733 tx
->tx_resid
= tx
->tx_nob
= sizeof(ksock_msg_t
) + tx
->tx_lnetmsg
->msg_len
;
735 LASSERT(tx
->tx_msg
.ksm_type
== KSOCK_MSG_NOOP
);
737 tx
->tx_iov
[0].iov_len
= offsetof(ksock_msg_t
, ksm_u
.lnetmsg
.ksnm_hdr
);
738 tx
->tx_resid
= tx
->tx_nob
= offsetof(ksock_msg_t
, ksm_u
.lnetmsg
.ksnm_hdr
);
740 /* Don't checksum before start sending, because packet can be piggybacked with ACK */
744 ksocknal_unpack_msg_v1(ksock_msg_t
*msg
)
747 msg
->ksm_type
= KSOCK_MSG_LNET
;
748 msg
->ksm_zc_cookies
[0] = msg
->ksm_zc_cookies
[1] = 0;
752 ksocknal_unpack_msg_v2(ksock_msg_t
*msg
)
754 return; /* Do nothing */
757 ksock_proto_t ksocknal_protocol_v1x
=
759 .pro_version
= KSOCK_PROTO_V1
,
760 .pro_send_hello
= ksocknal_send_hello_v1
,
761 .pro_recv_hello
= ksocknal_recv_hello_v1
,
762 .pro_pack
= ksocknal_pack_msg_v1
,
763 .pro_unpack
= ksocknal_unpack_msg_v1
,
764 .pro_queue_tx_msg
= ksocknal_queue_tx_msg_v1
,
765 .pro_handle_zcreq
= NULL
,
766 .pro_handle_zcack
= NULL
,
767 .pro_queue_tx_zcack
= NULL
,
768 .pro_match_tx
= ksocknal_match_tx
771 ksock_proto_t ksocknal_protocol_v2x
=
773 .pro_version
= KSOCK_PROTO_V2
,
774 .pro_send_hello
= ksocknal_send_hello_v2
,
775 .pro_recv_hello
= ksocknal_recv_hello_v2
,
776 .pro_pack
= ksocknal_pack_msg_v2
,
777 .pro_unpack
= ksocknal_unpack_msg_v2
,
778 .pro_queue_tx_msg
= ksocknal_queue_tx_msg_v2
,
779 .pro_queue_tx_zcack
= ksocknal_queue_tx_zcack_v2
,
780 .pro_handle_zcreq
= ksocknal_handle_zcreq
,
781 .pro_handle_zcack
= ksocknal_handle_zcack
,
782 .pro_match_tx
= ksocknal_match_tx
785 ksock_proto_t ksocknal_protocol_v3x
=
787 .pro_version
= KSOCK_PROTO_V3
,
788 .pro_send_hello
= ksocknal_send_hello_v2
,
789 .pro_recv_hello
= ksocknal_recv_hello_v2
,
790 .pro_pack
= ksocknal_pack_msg_v2
,
791 .pro_unpack
= ksocknal_unpack_msg_v2
,
792 .pro_queue_tx_msg
= ksocknal_queue_tx_msg_v2
,
793 .pro_queue_tx_zcack
= ksocknal_queue_tx_zcack_v3
,
794 .pro_handle_zcreq
= ksocknal_handle_zcreq
,
795 .pro_handle_zcack
= ksocknal_handle_zcack
,
796 .pro_match_tx
= ksocknal_match_tx_v3