]>
git.proxmox.com Git - mirror_ovs.git/blob - datapath/linux/compat/include/linux/skbuff.h
1 #ifndef __LINUX_SKBUFF_WRAPPER_H
2 #define __LINUX_SKBUFF_WRAPPER_H 1
4 #include_next <linux/skbuff.h>
6 #include <linux/version.h>
8 #if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,24)
9 /* In version 2.6.24 the return type of skb_headroom() changed from 'int' to
10 * 'unsigned int'. We use skb_headroom() as one arm of a min(a,b) invocation
11 * in make_writable() in actions.c, so we need the correct type. */
12 #define skb_headroom rpl_skb_headroom
13 static inline unsigned int rpl_skb_headroom(const struct sk_buff
*skb
)
15 return skb
->data
- skb
->head
;
19 #ifndef HAVE_SKB_COPY_FROM_LINEAR_DATA_OFFSET
20 static inline void skb_copy_from_linear_data_offset(const struct sk_buff
*skb
,
21 const int offset
, void *to
,
22 const unsigned int len
)
24 memcpy(to
, skb
->data
+ offset
, len
);
27 static inline void skb_copy_to_linear_data_offset(struct sk_buff
*skb
,
30 const unsigned int len
)
32 memcpy(skb
->data
+ offset
, from
, len
);
35 #endif /* !HAVE_SKB_COPY_FROM_LINEAR_DATA_OFFSET */
38 * The networking layer reserves some headroom in skb data (via
39 * dev_alloc_skb). This is used to avoid having to reallocate skb data when
40 * the header has to grow. In the default case, if the header has to grow
41 * 16 bytes or less we avoid the reallocation.
43 * Unfortunately this headroom changes the DMA alignment of the resulting
44 * network packet. As for NET_IP_ALIGN, this unaligned DMA is expensive
45 * on some architectures. An architecture can override this value,
46 * perhaps setting it to a cacheline in size (since that will maintain
47 * cacheline alignment of the DMA). It must be a power of 2.
49 * Various parts of the networking layer expect at least 16 bytes of
50 * headroom, you should not reduce this.
53 #define NET_SKB_PAD 16
56 #ifndef HAVE_SKB_COW_HEAD
57 static inline int __skb_cow(struct sk_buff
*skb
, unsigned int headroom
,
62 if (headroom
< NET_SKB_PAD
)
63 headroom
= NET_SKB_PAD
;
64 if (headroom
> skb_headroom(skb
))
65 delta
= headroom
- skb_headroom(skb
);
68 return pskb_expand_head(skb
, ALIGN(delta
, NET_SKB_PAD
), 0,
73 static inline int skb_cow_head(struct sk_buff
*skb
, unsigned int headroom
)
75 return __skb_cow(skb
, headroom
, skb_header_cloned(skb
));
77 #endif /* !HAVE_SKB_COW_HEAD */
79 #if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,23)
80 static inline int skb_clone_writable(struct sk_buff
*skb
, int len
)
86 #ifndef HAVE_SKB_DST_ACCESSOR_FUNCS
87 static inline struct dst_entry
*skb_dst(const struct sk_buff
*skb
)
89 return (struct dst_entry
*)skb
->dst
;
92 static inline void skb_dst_set(struct sk_buff
*skb
, struct dst_entry
*dst
)
97 static inline struct rtable
*skb_rtable(const struct sk_buff
*skb
)
99 return (struct rtable
*)skb
->dst
;
103 #if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,17)
104 /* Emulate Linux 2.6.17 and later behavior, in which kfree_skb silently ignores
105 * null pointer arguments. */
106 #define kfree_skb(skb) kfree_skb_maybe_null(skb)
107 static inline void kfree_skb_maybe_null(struct sk_buff
*skb
)
109 if (likely(skb
!= NULL
))
115 #ifndef CHECKSUM_PARTIAL
116 #define CHECKSUM_PARTIAL CHECKSUM_HW
118 #ifndef CHECKSUM_COMPLETE
119 #define CHECKSUM_COMPLETE CHECKSUM_HW
123 #define mac_header mac.raw
124 #define network_header nh.raw
125 #define transport_header h.raw
128 #ifndef HAVE_SKBUFF_HEADER_HELPERS
129 static inline unsigned char *skb_transport_header(const struct sk_buff
*skb
)
134 static inline void skb_reset_transport_header(struct sk_buff
*skb
)
136 skb
->h
.raw
= skb
->data
;
139 static inline void skb_set_transport_header(struct sk_buff
*skb
,
142 skb
->h
.raw
= skb
->data
+ offset
;
145 static inline unsigned char *skb_network_header(const struct sk_buff
*skb
)
150 static inline void skb_reset_network_header(struct sk_buff
*skb
)
152 skb
->nh
.raw
= skb
->data
;
155 static inline void skb_set_network_header(struct sk_buff
*skb
, const int offset
)
157 skb
->nh
.raw
= skb
->data
+ offset
;
160 static inline unsigned char *skb_mac_header(const struct sk_buff
*skb
)
165 static inline void skb_reset_mac_header(struct sk_buff
*skb
)
167 skb
->mac_header
= skb
->data
;
170 static inline void skb_set_mac_header(struct sk_buff
*skb
, const int offset
)
172 skb
->mac
.raw
= skb
->data
+ offset
;
175 static inline int skb_transport_offset(const struct sk_buff
*skb
)
177 return skb_transport_header(skb
) - skb
->data
;
180 static inline int skb_network_offset(const struct sk_buff
*skb
)
182 return skb_network_header(skb
) - skb
->data
;
185 static inline void skb_copy_to_linear_data(struct sk_buff
*skb
,
187 const unsigned int len
)
189 memcpy(skb
->data
, from
, len
);
191 #endif /* !HAVE_SKBUFF_HEADER_HELPERS */
193 #if LINUX_VERSION_CODE < KERNEL_VERSION(2,6,18)
194 #warning "TSO/UFO not supported on kernels earlier than 2.6.18"
196 static inline int skb_is_gso(const struct sk_buff
*skb
)
201 static inline struct sk_buff
*skb_gso_segment(struct sk_buff
*skb
,
206 #endif /* before 2.6.18 */
208 #ifndef HAVE_SKB_WARN_LRO
210 static inline bool skb_warn_if_lro(const struct sk_buff
*skb
)
215 extern void __skb_warn_lro_forwarding(const struct sk_buff
*skb
);
217 static inline bool skb_warn_if_lro(const struct sk_buff
*skb
)
219 /* LRO sets gso_size but not gso_type, whereas if GSO is really
220 * wanted then gso_type will be set. */
221 struct skb_shared_info
*shinfo
= skb_shinfo(skb
);
222 if (shinfo
->gso_size
!= 0 && unlikely(shinfo
->gso_type
== 0)) {
223 __skb_warn_lro_forwarding(skb
);
228 #endif /* NETIF_F_LRO */
229 #endif /* HAVE_SKB_WARN_LRO */
231 #ifndef HAVE_CONSUME_SKB
232 #define consume_skb kfree_skb
235 #if LINUX_VERSION_CODE < KERNEL_VERSION(3,2,0)
236 static inline struct page
*skb_frag_page(const skb_frag_t
*frag
)
242 #if LINUX_VERSION_CODE < KERNEL_VERSION(3,0,0)
243 static inline void skb_reset_mac_len(struct sk_buff
*skb
)
245 skb
->mac_len
= skb
->network_header
- skb
->mac_header
;