Blame view
net/ipv4/inet_fragment.c
14.4 KB
2874c5fd2 treewide: Replace... |
1 |
// SPDX-License-Identifier: GPL-2.0-or-later |
7eb95156d [INET]: Collect f... |
2 3 4 |
/* * inet fragments management * |
7eb95156d [INET]: Collect f... |
5 6 7 8 9 10 11 12 13 14 |
* Authors: Pavel Emelyanov <xemul@openvz.org> * Started as consolidation of ipv4/ip_fragment.c, * ipv6/reassembly. and ipv6 nf conntrack reassembly */ #include <linux/list.h> #include <linux/spinlock.h> #include <linux/module.h> #include <linux/timer.h> #include <linux/mm.h> |
321a3a99e [INET]: Consolida... |
15 |
#include <linux/random.h> |
1e4b82873 [INET]: Consolida... |
16 17 |
#include <linux/skbuff.h> #include <linux/rtnetlink.h> |
5a0e3ad6a include cleanup: ... |
18 |
#include <linux/slab.h> |
0eb71a9da rhashtable: split... |
19 |
#include <linux/rhashtable.h> |
7eb95156d [INET]: Collect f... |
20 |
|
5a3da1fe9 inet: limit lengt... |
21 |
#include <net/sock.h> |
7eb95156d [INET]: Collect f... |
22 |
#include <net/inet_frag.h> |
be991971d inet: generalize ... |
23 |
#include <net/inet_ecn.h> |
c23f35d19 net: IP defrag: e... |
24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 |
#include <net/ip.h> #include <net/ipv6.h> /* Use skb->cb to track consecutive/adjacent fragments coming at * the end of the queue. Nodes in the rb-tree queue will * contain "runs" of one or more adjacent fragments. * * Invariants: * - next_frag is NULL at the tail of a "run"; * - the head of a "run" has the sum of all fragment lengths in frag_run_len. */ struct ipfrag_skb_cb { union { struct inet_skb_parm h4; struct inet6_skb_parm h6; }; struct sk_buff *next_frag; int frag_run_len; }; #define FRAG_CB(skb) ((struct ipfrag_skb_cb *)((skb)->cb)) static void fragcb_clear(struct sk_buff *skb) { RB_CLEAR_NODE(&skb->rbnode); FRAG_CB(skb)->next_frag = NULL; FRAG_CB(skb)->frag_run_len = skb->len; } /* Append skb to the last "run". */ static void fragrun_append_to_last(struct inet_frag_queue *q, struct sk_buff *skb) { fragcb_clear(skb); FRAG_CB(q->last_run_head)->frag_run_len += skb->len; FRAG_CB(q->fragments_tail)->next_frag = skb; q->fragments_tail = skb; } /* Create a new "run" with the skb. */ static void fragrun_create(struct inet_frag_queue *q, struct sk_buff *skb) { BUILD_BUG_ON(sizeof(struct ipfrag_skb_cb) > sizeof(skb->cb)); fragcb_clear(skb); if (q->last_run_head) rb_link_node(&skb->rbnode, &q->last_run_head->rbnode, &q->last_run_head->rbnode.rb_right); else rb_link_node(&skb->rbnode, NULL, &q->rb_fragments.rb_node); rb_insert_color(&skb->rbnode, &q->rb_fragments); q->fragments_tail = skb; q->last_run_head = skb; } |
be991971d inet: generalize ... |
80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 |
/* Given the OR values of all fragments, apply RFC 3168 5.3 requirements * Value : 0xff if frame should be dropped. * 0 or INET_ECN_CE value, to be ORed in to final iph->tos field */ const u8 ip_frag_ecn_table[16] = { /* at least one fragment had CE, and others ECT_0 or ECT_1 */ [IPFRAG_ECN_CE | IPFRAG_ECN_ECT_0] = INET_ECN_CE, [IPFRAG_ECN_CE | IPFRAG_ECN_ECT_1] = INET_ECN_CE, [IPFRAG_ECN_CE | IPFRAG_ECN_ECT_0 | IPFRAG_ECN_ECT_1] = INET_ECN_CE, /* invalid combinations : drop frame */ [IPFRAG_ECN_NOT_ECT | IPFRAG_ECN_CE] = 0xff, [IPFRAG_ECN_NOT_ECT | IPFRAG_ECN_ECT_0] = 0xff, [IPFRAG_ECN_NOT_ECT | IPFRAG_ECN_ECT_1] = 0xff, [IPFRAG_ECN_NOT_ECT | IPFRAG_ECN_ECT_0 | IPFRAG_ECN_ECT_1] = 0xff, [IPFRAG_ECN_NOT_ECT | IPFRAG_ECN_CE | IPFRAG_ECN_ECT_0] = 0xff, [IPFRAG_ECN_NOT_ECT | IPFRAG_ECN_CE | IPFRAG_ECN_ECT_1] = 0xff, [IPFRAG_ECN_NOT_ECT | IPFRAG_ECN_CE | IPFRAG_ECN_ECT_0 | IPFRAG_ECN_ECT_1] = 0xff, }; EXPORT_SYMBOL(ip_frag_ecn_table); |
7eb95156d [INET]: Collect f... |
101 |
|
d4ad4d22e inet: frags: use ... |
102 |
int inet_frags_init(struct inet_frags *f) |
7eb95156d [INET]: Collect f... |
103 |
{ |
d4ad4d22e inet: frags: use ... |
104 105 106 107 |
f->frags_cachep = kmem_cache_create(f->frags_cache_name, f->qsize, 0, 0, NULL); if (!f->frags_cachep) return -ENOMEM; |
dc93f46bc inet: frags: fix ... |
108 109 |
refcount_set(&f->refcnt, 1); init_completion(&f->completion); |
d4ad4d22e inet: frags: use ... |
110 |
return 0; |
7eb95156d [INET]: Collect f... |
111 112 113 114 115 |
} EXPORT_SYMBOL(inet_frags_init); void inet_frags_fini(struct inet_frags *f) { |
dc93f46bc inet: frags: fix ... |
116 117 118 119 |
if (refcount_dec_and_test(&f->refcnt)) complete(&f->completion); wait_for_completion(&f->completion); |
648700f76 inet: frags: use ... |
120 |
|
d4ad4d22e inet: frags: use ... |
121 |
kmem_cache_destroy(f->frags_cachep); |
648700f76 inet: frags: use ... |
122 |
f->frags_cachep = NULL; |
7eb95156d [INET]: Collect f... |
123 124 |
} EXPORT_SYMBOL(inet_frags_fini); |
277e650dd [INET]: Consolida... |
125 |
|
3c8fc8782 inet: frags: rewo... |
126 |
/* called from rhashtable_free_and_destroy() at netns_frags dismantle */ |
648700f76 inet: frags: use ... |
127 |
static void inet_frags_free_cb(void *ptr, void *arg) |
277e650dd [INET]: Consolida... |
128 |
{ |
648700f76 inet: frags: use ... |
129 |
struct inet_frag_queue *fq = ptr; |
3c8fc8782 inet: frags: rewo... |
130 |
int count; |
ab1c724f6 inet: frag: use s... |
131 |
|
3c8fc8782 inet: frags: rewo... |
132 |
count = del_timer_sync(&fq->timer) ? 1 : 0; |
19952cc4f net: frag queue p... |
133 |
|
648700f76 inet: frags: use ... |
134 135 136 |
spin_lock_bh(&fq->lock); if (!(fq->flags & INET_FRAG_COMPLETE)) { fq->flags |= INET_FRAG_COMPLETE; |
3c8fc8782 inet: frags: rewo... |
137 138 139 |
count++; } else if (fq->flags & INET_FRAG_HASH_DEAD) { count++; |
ab1c724f6 inet: frag: use s... |
140 |
} |
648700f76 inet: frags: use ... |
141 |
spin_unlock_bh(&fq->lock); |
ab1c724f6 inet: frag: use s... |
142 |
|
3c8fc8782 inet: frags: rewo... |
143 144 |
if (refcount_sub_and_test(count, &fq->refcnt)) inet_frag_destroy(fq); |
ab1c724f6 inet: frag: use s... |
145 |
} |
d5dd88794 inet: fix various... |
146 |
static void fqdir_work_fn(struct work_struct *work) |
ab1c724f6 inet: frag: use s... |
147 |
{ |
d5dd88794 inet: fix various... |
148 |
struct fqdir *fqdir = container_of(work, struct fqdir, destroy_work); |
dc93f46bc inet: frags: fix ... |
149 |
struct inet_frags *f = fqdir->f; |
ab1c724f6 inet: frag: use s... |
150 |
|
6ce3b4dce inet: rename netn... |
151 |
rhashtable_free_and_destroy(&fqdir->rhashtable, inet_frags_free_cb, NULL); |
dc93f46bc inet: frags: fix ... |
152 153 154 155 156 157 158 159 160 |
/* We need to make sure all ongoing call_rcu(..., inet_frag_destroy_rcu) * have completed, since they need to dereference fqdir. * Would it not be nice to have kfree_rcu_barrier() ? :) */ rcu_barrier(); if (refcount_dec_and_test(&f->refcnt)) complete(&f->completion); |
4907abc60 net: dynamically ... |
161 |
kfree(fqdir); |
277e650dd [INET]: Consolida... |
162 |
} |
3c8fc8782 inet: frags: rewo... |
163 |
|
6b73d1971 inet: frags: unin... |
164 165 166 167 168 169 170 171 172 173 174 175 176 177 |
int fqdir_init(struct fqdir **fqdirp, struct inet_frags *f, struct net *net) { struct fqdir *fqdir = kzalloc(sizeof(*fqdir), GFP_KERNEL); int res; if (!fqdir) return -ENOMEM; fqdir->f = f; fqdir->net = net; res = rhashtable_init(&fqdir->rhashtable, &fqdir->f->rhash_params); if (res < 0) { kfree(fqdir); return res; } |
dc93f46bc inet: frags: fix ... |
178 |
refcount_inc(&f->refcnt); |
6b73d1971 inet: frags: unin... |
179 180 181 182 |
*fqdirp = fqdir; return 0; } EXPORT_SYMBOL(fqdir_init); |
3c8fc8782 inet: frags: rewo... |
183 184 |
void fqdir_exit(struct fqdir *fqdir) { |
d5dd88794 inet: fix various... |
185 186 |
INIT_WORK(&fqdir->destroy_work, fqdir_work_fn); queue_work(system_wq, &fqdir->destroy_work); |
3c8fc8782 inet: frags: rewo... |
187 |
} |
89fb90051 net: rename inet_... |
188 |
EXPORT_SYMBOL(fqdir_exit); |
277e650dd [INET]: Consolida... |
189 |
|
093ba7291 inet: frags: add ... |
190 |
void inet_frag_kill(struct inet_frag_queue *fq) |
277e650dd [INET]: Consolida... |
191 192 |
{ if (del_timer(&fq->timer)) |
edcb69187 net: convert inet... |
193 |
refcount_dec(&fq->refcnt); |
277e650dd [INET]: Consolida... |
194 |
|
06aa8b8a0 inet: frags: rena... |
195 |
if (!(fq->flags & INET_FRAG_COMPLETE)) { |
6ce3b4dce inet: rename netn... |
196 |
struct fqdir *fqdir = fq->fqdir; |
648700f76 inet: frags: use ... |
197 198 |
fq->flags |= INET_FRAG_COMPLETE; |
3c8fc8782 inet: frags: rewo... |
199 |
rcu_read_lock(); |
32707c4df inet: frags: Remo... |
200 201 202 203 |
/* The RCU read lock provides a memory barrier * guaranteeing that if fqdir->dead is false then * the hash table destruction will not start until * after we unlock. Paired with inet_frags_exit_net(). |
3c8fc8782 inet: frags: rewo... |
204 |
*/ |
32707c4df inet: frags: Remo... |
205 |
if (!fqdir->dead) { |
3c8fc8782 inet: frags: rewo... |
206 207 208 209 210 211 212 |
rhashtable_remove_fast(&fqdir->rhashtable, &fq->node, fqdir->f->rhash_params); refcount_dec(&fq->refcnt); } else { fq->flags |= INET_FRAG_HASH_DEAD; } rcu_read_unlock(); |
277e650dd [INET]: Consolida... |
213 214 |
} } |
277e650dd [INET]: Consolida... |
215 |
EXPORT_SYMBOL(inet_frag_kill); |
1e4b82873 [INET]: Consolida... |
216 |
|
648700f76 inet: frags: use ... |
217 218 219 220 |
static void inet_frag_destroy_rcu(struct rcu_head *head) { struct inet_frag_queue *q = container_of(head, struct inet_frag_queue, rcu); |
6ce3b4dce inet: rename netn... |
221 |
struct inet_frags *f = q->fqdir->f; |
648700f76 inet: frags: use ... |
222 223 224 225 226 |
if (f->destructor) f->destructor(q); kmem_cache_free(f->frags_cachep, q); } |
c23f35d19 net: IP defrag: e... |
227 228 229 230 231 232 233 234 235 236 237 238 239 240 241 242 243 244 245 246 247 |
unsigned int inet_frag_rbtree_purge(struct rb_root *root) { struct rb_node *p = rb_first(root); unsigned int sum = 0; while (p) { struct sk_buff *skb = rb_entry(p, struct sk_buff, rbnode); p = rb_next(p); rb_erase(&skb->rbnode, root); while (skb) { struct sk_buff *next = FRAG_CB(skb)->next_frag; sum += skb->truesize; kfree_skb(skb); skb = next; } } return sum; } EXPORT_SYMBOL(inet_frag_rbtree_purge); |
093ba7291 inet: frags: add ... |
248 |
void inet_frag_destroy(struct inet_frag_queue *q) |
1e4b82873 [INET]: Consolida... |
249 |
{ |
6ce3b4dce inet: rename netn... |
250 |
struct fqdir *fqdir; |
d433673e5 net: frag helper ... |
251 |
unsigned int sum, sum_truesize = 0; |
093ba7291 inet: frags: add ... |
252 |
struct inet_frags *f; |
1e4b82873 [INET]: Consolida... |
253 |
|
06aa8b8a0 inet: frags: rena... |
254 |
WARN_ON(!(q->flags & INET_FRAG_COMPLETE)); |
547b792ca net: convert BUG_... |
255 |
WARN_ON(del_timer(&q->timer) != 0); |
1e4b82873 [INET]: Consolida... |
256 257 |
/* Release all fragment data. */ |
6ce3b4dce inet: rename netn... |
258 259 |
fqdir = q->fqdir; f = fqdir->f; |
d8cf757fb net: remove unuse... |
260 |
sum_truesize = inet_frag_rbtree_purge(&q->rb_fragments); |
d433673e5 net: frag helper ... |
261 |
sum = sum_truesize + f->qsize; |
1e4b82873 [INET]: Consolida... |
262 |
|
648700f76 inet: frags: use ... |
263 |
call_rcu(&q->rcu, inet_frag_destroy_rcu); |
5719b296f inet: frag: don't... |
264 |
|
6ce3b4dce inet: rename netn... |
265 |
sub_frag_mem_limit(fqdir, sum); |
1e4b82873 [INET]: Consolida... |
266 267 |
} EXPORT_SYMBOL(inet_frag_destroy); |
8e7999c44 [INET]: Consolida... |
268 |
|
6ce3b4dce inet: rename netn... |
269 |
static struct inet_frag_queue *inet_frag_alloc(struct fqdir *fqdir, |
f926e2366 inet: frags: fix ... |
270 271 |
struct inet_frags *f, void *arg) |
e521db9d7 [INET]: Consolida... |
272 273 |
{ struct inet_frag_queue *q; |
d4ad4d22e inet: frags: use ... |
274 |
q = kmem_cache_zalloc(f->frags_cachep, GFP_ATOMIC); |
51456b291 ipv4: coding styl... |
275 |
if (!q) |
e521db9d7 [INET]: Consolida... |
276 |
return NULL; |
6ce3b4dce inet: rename netn... |
277 |
q->fqdir = fqdir; |
c6fda2822 [INET]: Consolida... |
278 |
f->constructor(q, arg); |
6ce3b4dce inet: rename netn... |
279 |
add_frag_mem_limit(fqdir, f->qsize); |
d433673e5 net: frag helper ... |
280 |
|
78802011f inet: frags: Conv... |
281 |
timer_setup(&q->timer, f->frag_expire, 0); |
e521db9d7 [INET]: Consolida... |
282 |
spin_lock_init(&q->lock); |
648700f76 inet: frags: use ... |
283 |
refcount_set(&q->refcnt, 3); |
e521db9d7 [INET]: Consolida... |
284 285 286 |
return q; } |
c6fda2822 [INET]: Consolida... |
287 |
|
6ce3b4dce inet: rename netn... |
288 |
static struct inet_frag_queue *inet_frag_create(struct fqdir *fqdir, |
0d5b9311b inet: frags: bett... |
289 290 |
void *arg, struct inet_frag_queue **prev) |
c6fda2822 [INET]: Consolida... |
291 |
{ |
6ce3b4dce inet: rename netn... |
292 |
struct inet_frags *f = fqdir->f; |
c6fda2822 [INET]: Consolida... |
293 |
struct inet_frag_queue *q; |
6ce3b4dce inet: rename netn... |
294 |
q = inet_frag_alloc(fqdir, f, arg); |
0d5b9311b inet: frags: bett... |
295 296 |
if (!q) { *prev = ERR_PTR(-ENOMEM); |
c6fda2822 [INET]: Consolida... |
297 |
return NULL; |
0d5b9311b inet: frags: bett... |
298 |
} |
6ce3b4dce inet: rename netn... |
299 |
mod_timer(&q->timer, jiffies + fqdir->timeout); |
648700f76 inet: frags: use ... |
300 |
|
6ce3b4dce inet: rename netn... |
301 |
*prev = rhashtable_lookup_get_insert_key(&fqdir->rhashtable, &q->key, |
0d5b9311b inet: frags: bett... |
302 303 |
&q->node, f->rhash_params); if (*prev) { |
648700f76 inet: frags: use ... |
304 305 306 307 308 309 |
q->flags |= INET_FRAG_COMPLETE; inet_frag_kill(q); inet_frag_destroy(q); return NULL; } return q; |
c6fda2822 [INET]: Consolida... |
310 |
} |
abd6523d1 [INET]: Consolida... |
311 |
|
648700f76 inet: frags: use ... |
312 |
/* TODO : call from rcu_read_lock() and no longer use refcount_inc_not_zero() */ |
6ce3b4dce inet: rename netn... |
313 |
struct inet_frag_queue *inet_frag_find(struct fqdir *fqdir, void *key) |
abd6523d1 [INET]: Consolida... |
314 |
{ |
0d5b9311b inet: frags: bett... |
315 |
struct inet_frag_queue *fq = NULL, *prev; |
abd6523d1 [INET]: Consolida... |
316 |
|
6ce3b4dce inet: rename netn... |
317 |
if (!fqdir->high_thresh || frag_mem_limit(fqdir) > fqdir->high_thresh) |
56e2c94f0 inet: frag: enfor... |
318 |
return NULL; |
648700f76 inet: frags: use ... |
319 |
rcu_read_lock(); |
e3a57d18b inet: frag: remov... |
320 |
|
6ce3b4dce inet: rename netn... |
321 |
prev = rhashtable_lookup(&fqdir->rhashtable, key, fqdir->f->rhash_params); |
0d5b9311b inet: frags: bett... |
322 |
if (!prev) |
6ce3b4dce inet: rename netn... |
323 |
fq = inet_frag_create(fqdir, key, &prev); |
c7148c03d net/ipv4: cleanup... |
324 |
if (!IS_ERR_OR_NULL(prev)) { |
0d5b9311b inet: frags: bett... |
325 |
fq = prev; |
648700f76 inet: frags: use ... |
326 327 |
if (!refcount_inc_not_zero(&fq->refcnt)) fq = NULL; |
e3a57d18b inet: frag: remov... |
328 |
} |
648700f76 inet: frags: use ... |
329 |
rcu_read_unlock(); |
0d5b9311b inet: frags: bett... |
330 |
return fq; |
abd6523d1 [INET]: Consolida... |
331 332 |
} EXPORT_SYMBOL(inet_frag_find); |
c23f35d19 net: IP defrag: e... |
333 334 335 336 337 338 339 340 341 342 343 344 345 346 347 348 349 350 351 352 353 354 355 356 357 358 359 360 361 362 363 364 365 366 367 368 369 370 371 372 373 374 375 376 377 378 379 380 381 382 383 384 385 386 387 388 389 390 391 392 393 394 395 396 397 398 399 400 401 402 403 404 405 406 407 408 409 410 411 412 413 414 415 416 417 418 419 420 421 422 423 424 425 426 427 428 429 430 431 432 433 |
int inet_frag_queue_insert(struct inet_frag_queue *q, struct sk_buff *skb, int offset, int end) { struct sk_buff *last = q->fragments_tail; /* RFC5722, Section 4, amended by Errata ID : 3089 * When reassembling an IPv6 datagram, if * one or more its constituent fragments is determined to be an * overlapping fragment, the entire datagram (and any constituent * fragments) MUST be silently discarded. * * Duplicates, however, should be ignored (i.e. skb dropped, but the * queue/fragments kept for later reassembly). */ if (!last) fragrun_create(q, skb); /* First fragment. */ else if (last->ip_defrag_offset + last->len < end) { /* This is the common case: skb goes to the end. */ /* Detect and discard overlaps. */ if (offset < last->ip_defrag_offset + last->len) return IPFRAG_OVERLAP; if (offset == last->ip_defrag_offset + last->len) fragrun_append_to_last(q, skb); else fragrun_create(q, skb); } else { /* Binary search. Note that skb can become the first fragment, * but not the last (covered above). */ struct rb_node **rbn, *parent; rbn = &q->rb_fragments.rb_node; do { struct sk_buff *curr; int curr_run_end; parent = *rbn; curr = rb_to_skb(parent); curr_run_end = curr->ip_defrag_offset + FRAG_CB(curr)->frag_run_len; if (end <= curr->ip_defrag_offset) rbn = &parent->rb_left; else if (offset >= curr_run_end) rbn = &parent->rb_right; else if (offset >= curr->ip_defrag_offset && end <= curr_run_end) return IPFRAG_DUP; else return IPFRAG_OVERLAP; } while (*rbn); /* Here we have parent properly set, and rbn pointing to * one of its NULL left/right children. Insert skb. */ fragcb_clear(skb); rb_link_node(&skb->rbnode, parent, rbn); rb_insert_color(&skb->rbnode, &q->rb_fragments); } skb->ip_defrag_offset = offset; return IPFRAG_OK; } EXPORT_SYMBOL(inet_frag_queue_insert); void *inet_frag_reasm_prepare(struct inet_frag_queue *q, struct sk_buff *skb, struct sk_buff *parent) { struct sk_buff *fp, *head = skb_rb_first(&q->rb_fragments); struct sk_buff **nextp; int delta; if (head != skb) { fp = skb_clone(skb, GFP_ATOMIC); if (!fp) return NULL; FRAG_CB(fp)->next_frag = FRAG_CB(skb)->next_frag; if (RB_EMPTY_NODE(&skb->rbnode)) FRAG_CB(parent)->next_frag = fp; else rb_replace_node(&skb->rbnode, &fp->rbnode, &q->rb_fragments); if (q->fragments_tail == skb) q->fragments_tail = fp; skb_morph(skb, head); FRAG_CB(skb)->next_frag = FRAG_CB(head)->next_frag; rb_replace_node(&head->rbnode, &skb->rbnode, &q->rb_fragments); consume_skb(head); head = skb; } WARN_ON(head->ip_defrag_offset != 0); delta = -head->truesize; /* Head of list must not be cloned. */ if (skb_unclone(head, GFP_ATOMIC)) return NULL; delta += head->truesize; if (delta) |
6ce3b4dce inet: rename netn... |
434 |
add_frag_mem_limit(q->fqdir, delta); |
c23f35d19 net: IP defrag: e... |
435 436 437 438 439 440 441 442 443 444 445 446 447 448 449 450 451 452 453 454 455 |
/* If the first fragment is fragmented itself, we split * it to two chunks: the first with data and paged part * and the second, holding only fragments. */ if (skb_has_frag_list(head)) { struct sk_buff *clone; int i, plen = 0; clone = alloc_skb(0, GFP_ATOMIC); if (!clone) return NULL; skb_shinfo(clone)->frag_list = skb_shinfo(head)->frag_list; skb_frag_list_init(head); for (i = 0; i < skb_shinfo(head)->nr_frags; i++) plen += skb_frag_size(&skb_shinfo(head)->frags[i]); clone->data_len = head->data_len - plen; clone->len = clone->data_len; head->truesize += clone->truesize; clone->csum = 0; clone->ip_summed = head->ip_summed; |
6ce3b4dce inet: rename netn... |
456 |
add_frag_mem_limit(q->fqdir, clone->truesize); |
c23f35d19 net: IP defrag: e... |
457 458 459 460 461 462 463 464 465 466 467 |
skb_shinfo(head)->frag_list = clone; nextp = &clone->next; } else { nextp = &skb_shinfo(head)->frag_list; } return nextp; } EXPORT_SYMBOL(inet_frag_reasm_prepare); void inet_frag_reasm_finish(struct inet_frag_queue *q, struct sk_buff *head, |
891584f48 inet: frags: re-i... |
468 |
void *reasm_data, bool try_coalesce) |
c23f35d19 net: IP defrag: e... |
469 470 471 472 |
{ struct sk_buff **nextp = (struct sk_buff **)reasm_data; struct rb_node *rbn; struct sk_buff *fp; |
891584f48 inet: frags: re-i... |
473 |
int sum_truesize; |
c23f35d19 net: IP defrag: e... |
474 475 476 477 478 479 480 |
skb_push(head, head->data - skb_network_header(head)); /* Traverse the tree in order, to build frag_list. */ fp = FRAG_CB(head)->next_frag; rbn = rb_next(&head->rbnode); rb_erase(&head->rbnode, &q->rb_fragments); |
891584f48 inet: frags: re-i... |
481 482 |
sum_truesize = head->truesize; |
c23f35d19 net: IP defrag: e... |
483 484 485 486 487 488 |
while (rbn || fp) { /* fp points to the next sk_buff in the current run; * rbn points to the next run. */ /* Go through the current run. */ while (fp) { |
891584f48 inet: frags: re-i... |
489 490 491 492 493 |
struct sk_buff *next_frag = FRAG_CB(fp)->next_frag; bool stolen; int delta; sum_truesize += fp->truesize; |
c23f35d19 net: IP defrag: e... |
494 495 496 497 |
if (head->ip_summed != fp->ip_summed) head->ip_summed = CHECKSUM_NONE; else if (head->ip_summed == CHECKSUM_COMPLETE) head->csum = csum_add(head->csum, fp->csum); |
891584f48 inet: frags: re-i... |
498 499 500 501 502 503 504 505 506 507 508 509 510 511 512 513 514 515 |
if (try_coalesce && skb_try_coalesce(head, fp, &stolen, &delta)) { kfree_skb_partial(fp, stolen); } else { fp->prev = NULL; memset(&fp->rbnode, 0, sizeof(fp->rbnode)); fp->sk = NULL; head->data_len += fp->len; head->len += fp->len; head->truesize += fp->truesize; *nextp = fp; nextp = &fp->next; } fp = next_frag; |
c23f35d19 net: IP defrag: e... |
516 517 518 519 520 521 522 523 524 525 |
} /* Move to the next run. */ if (rbn) { struct rb_node *rbnext = rb_next(rbn); fp = rb_to_skb(rbn); rb_erase(rbn, &q->rb_fragments); rbn = rbnext; } } |
891584f48 inet: frags: re-i... |
526 |
sub_frag_mem_limit(q->fqdir, sum_truesize); |
c23f35d19 net: IP defrag: e... |
527 528 529 530 531 532 533 534 535 536 |
*nextp = NULL; skb_mark_not_on_list(head); head->prev = NULL; head->tstamp = q->stamp; } EXPORT_SYMBOL(inet_frag_reasm_finish); struct sk_buff *inet_frag_pull_head(struct inet_frag_queue *q) { |
d8cf757fb net: remove unuse... |
537 |
struct sk_buff *head, *skb; |
c23f35d19 net: IP defrag: e... |
538 |
|
d8cf757fb net: remove unuse... |
539 540 541 542 543 544 545 546 547 548 549 |
head = skb_rb_first(&q->rb_fragments); if (!head) return NULL; skb = FRAG_CB(head)->next_frag; if (skb) rb_replace_node(&head->rbnode, &skb->rbnode, &q->rb_fragments); else rb_erase(&head->rbnode, &q->rb_fragments); memset(&head->rbnode, 0, sizeof(head->rbnode)); barrier(); |
c23f35d19 net: IP defrag: e... |
550 |
|
c23f35d19 net: IP defrag: e... |
551 552 |
if (head == q->fragments_tail) q->fragments_tail = NULL; |
6ce3b4dce inet: rename netn... |
553 |
sub_frag_mem_limit(q->fqdir, head->truesize); |
c23f35d19 net: IP defrag: e... |
554 555 556 557 |
return head; } EXPORT_SYMBOL(inet_frag_pull_head); |