Merge branches 'release' and 'hp-cid' into release
[linux-2.6] / net / xfrm / xfrm_state.c
1 /*
2  * xfrm_state.c
3  *
4  * Changes:
5  *      Mitsuru KANDA @USAGI
6  *      Kazunori MIYAZAWA @USAGI
7  *      Kunihiro Ishiguro <kunihiro@ipinfusion.com>
8  *              IPv6 support
9  *      YOSHIFUJI Hideaki @USAGI
10  *              Split up af-specific functions
11  *      Derek Atkins <derek@ihtfp.com>
12  *              Add UDP Encapsulation
13  *
14  */
15
16 #include <linux/workqueue.h>
17 #include <net/xfrm.h>
18 #include <linux/pfkeyv2.h>
19 #include <linux/ipsec.h>
20 #include <linux/module.h>
21 #include <linux/cache.h>
22 #include <linux/audit.h>
23 #include <asm/uaccess.h>
24
25 #include "xfrm_hash.h"
26
27 struct sock *xfrm_nl;
28 EXPORT_SYMBOL(xfrm_nl);
29
30 u32 sysctl_xfrm_aevent_etime __read_mostly = XFRM_AE_ETIME;
31 EXPORT_SYMBOL(sysctl_xfrm_aevent_etime);
32
33 u32 sysctl_xfrm_aevent_rseqth __read_mostly = XFRM_AE_SEQT_SIZE;
34 EXPORT_SYMBOL(sysctl_xfrm_aevent_rseqth);
35
36 u32 sysctl_xfrm_acq_expires __read_mostly = 30;
37
38 /* Each xfrm_state may be linked to two tables:
39
40    1. Hash table by (spi,daddr,ah/esp) to find SA by SPI. (input,ctl)
41    2. Hash table by (daddr,family,reqid) to find what SAs exist for given
42       destination/tunnel endpoint. (output)
43  */
44
45 static DEFINE_SPINLOCK(xfrm_state_lock);
46
47 /* Hash table to find appropriate SA towards given target (endpoint
48  * of tunnel or destination of transport mode) allowed by selector.
49  *
50  * Main use is finding SA after policy selected tunnel or transport mode.
51  * Also, it can be used by ah/esp icmp error handler to find offending SA.
52  */
53 static struct hlist_head *xfrm_state_bydst __read_mostly;
54 static struct hlist_head *xfrm_state_bysrc __read_mostly;
55 static struct hlist_head *xfrm_state_byspi __read_mostly;
56 static unsigned int xfrm_state_hmask __read_mostly;
57 static unsigned int xfrm_state_hashmax __read_mostly = 1 * 1024 * 1024;
58 static unsigned int xfrm_state_num;
59 static unsigned int xfrm_state_genid;
60
61 static struct xfrm_state_afinfo *xfrm_state_get_afinfo(unsigned int family);
62 static void xfrm_state_put_afinfo(struct xfrm_state_afinfo *afinfo);
63
64 #ifdef CONFIG_AUDITSYSCALL
65 static void xfrm_audit_state_replay(struct xfrm_state *x,
66                                     struct sk_buff *skb, __be32 net_seq);
67 #else
68 #define xfrm_audit_state_replay(x, s, sq)       do { ; } while (0)
69 #endif /* CONFIG_AUDITSYSCALL */
70
71 static inline unsigned int xfrm_dst_hash(xfrm_address_t *daddr,
72                                          xfrm_address_t *saddr,
73                                          u32 reqid,
74                                          unsigned short family)
75 {
76         return __xfrm_dst_hash(daddr, saddr, reqid, family, xfrm_state_hmask);
77 }
78
79 static inline unsigned int xfrm_src_hash(xfrm_address_t *daddr,
80                                          xfrm_address_t *saddr,
81                                          unsigned short family)
82 {
83         return __xfrm_src_hash(daddr, saddr, family, xfrm_state_hmask);
84 }
85
86 static inline unsigned int
87 xfrm_spi_hash(xfrm_address_t *daddr, __be32 spi, u8 proto, unsigned short family)
88 {
89         return __xfrm_spi_hash(daddr, spi, proto, family, xfrm_state_hmask);
90 }
91
92 static void xfrm_hash_transfer(struct hlist_head *list,
93                                struct hlist_head *ndsttable,
94                                struct hlist_head *nsrctable,
95                                struct hlist_head *nspitable,
96                                unsigned int nhashmask)
97 {
98         struct hlist_node *entry, *tmp;
99         struct xfrm_state *x;
100
101         hlist_for_each_entry_safe(x, entry, tmp, list, bydst) {
102                 unsigned int h;
103
104                 h = __xfrm_dst_hash(&x->id.daddr, &x->props.saddr,
105                                     x->props.reqid, x->props.family,
106                                     nhashmask);
107                 hlist_add_head(&x->bydst, ndsttable+h);
108
109                 h = __xfrm_src_hash(&x->id.daddr, &x->props.saddr,
110                                     x->props.family,
111                                     nhashmask);
112                 hlist_add_head(&x->bysrc, nsrctable+h);
113
114                 if (x->id.spi) {
115                         h = __xfrm_spi_hash(&x->id.daddr, x->id.spi,
116                                             x->id.proto, x->props.family,
117                                             nhashmask);
118                         hlist_add_head(&x->byspi, nspitable+h);
119                 }
120         }
121 }
122
123 static unsigned long xfrm_hash_new_size(void)
124 {
125         return ((xfrm_state_hmask + 1) << 1) *
126                 sizeof(struct hlist_head);
127 }
128
129 static DEFINE_MUTEX(hash_resize_mutex);
130
131 static void xfrm_hash_resize(struct work_struct *__unused)
132 {
133         struct hlist_head *ndst, *nsrc, *nspi, *odst, *osrc, *ospi;
134         unsigned long nsize, osize;
135         unsigned int nhashmask, ohashmask;
136         int i;
137
138         mutex_lock(&hash_resize_mutex);
139
140         nsize = xfrm_hash_new_size();
141         ndst = xfrm_hash_alloc(nsize);
142         if (!ndst)
143                 goto out_unlock;
144         nsrc = xfrm_hash_alloc(nsize);
145         if (!nsrc) {
146                 xfrm_hash_free(ndst, nsize);
147                 goto out_unlock;
148         }
149         nspi = xfrm_hash_alloc(nsize);
150         if (!nspi) {
151                 xfrm_hash_free(ndst, nsize);
152                 xfrm_hash_free(nsrc, nsize);
153                 goto out_unlock;
154         }
155
156         spin_lock_bh(&xfrm_state_lock);
157
158         nhashmask = (nsize / sizeof(struct hlist_head)) - 1U;
159         for (i = xfrm_state_hmask; i >= 0; i--)
160                 xfrm_hash_transfer(xfrm_state_bydst+i, ndst, nsrc, nspi,
161                                    nhashmask);
162
163         odst = xfrm_state_bydst;
164         osrc = xfrm_state_bysrc;
165         ospi = xfrm_state_byspi;
166         ohashmask = xfrm_state_hmask;
167
168         xfrm_state_bydst = ndst;
169         xfrm_state_bysrc = nsrc;
170         xfrm_state_byspi = nspi;
171         xfrm_state_hmask = nhashmask;
172
173         spin_unlock_bh(&xfrm_state_lock);
174
175         osize = (ohashmask + 1) * sizeof(struct hlist_head);
176         xfrm_hash_free(odst, osize);
177         xfrm_hash_free(osrc, osize);
178         xfrm_hash_free(ospi, osize);
179
180 out_unlock:
181         mutex_unlock(&hash_resize_mutex);
182 }
183
184 static DECLARE_WORK(xfrm_hash_work, xfrm_hash_resize);
185
186 DECLARE_WAIT_QUEUE_HEAD(km_waitq);
187 EXPORT_SYMBOL(km_waitq);
188
189 static DEFINE_RWLOCK(xfrm_state_afinfo_lock);
190 static struct xfrm_state_afinfo *xfrm_state_afinfo[NPROTO];
191
192 static struct work_struct xfrm_state_gc_work;
193 static HLIST_HEAD(xfrm_state_gc_list);
194 static DEFINE_SPINLOCK(xfrm_state_gc_lock);
195
196 int __xfrm_state_delete(struct xfrm_state *x);
197
198 int km_query(struct xfrm_state *x, struct xfrm_tmpl *t, struct xfrm_policy *pol);
199 void km_state_expired(struct xfrm_state *x, int hard, u32 pid);
200
201 static struct xfrm_state_afinfo *xfrm_state_lock_afinfo(unsigned int family)
202 {
203         struct xfrm_state_afinfo *afinfo;
204         if (unlikely(family >= NPROTO))
205                 return NULL;
206         write_lock_bh(&xfrm_state_afinfo_lock);
207         afinfo = xfrm_state_afinfo[family];
208         if (unlikely(!afinfo))
209                 write_unlock_bh(&xfrm_state_afinfo_lock);
210         return afinfo;
211 }
212
213 static void xfrm_state_unlock_afinfo(struct xfrm_state_afinfo *afinfo)
214         __releases(xfrm_state_afinfo_lock)
215 {
216         write_unlock_bh(&xfrm_state_afinfo_lock);
217 }
218
219 int xfrm_register_type(const struct xfrm_type *type, unsigned short family)
220 {
221         struct xfrm_state_afinfo *afinfo = xfrm_state_lock_afinfo(family);
222         const struct xfrm_type **typemap;
223         int err = 0;
224
225         if (unlikely(afinfo == NULL))
226                 return -EAFNOSUPPORT;
227         typemap = afinfo->type_map;
228
229         if (likely(typemap[type->proto] == NULL))
230                 typemap[type->proto] = type;
231         else
232                 err = -EEXIST;
233         xfrm_state_unlock_afinfo(afinfo);
234         return err;
235 }
236 EXPORT_SYMBOL(xfrm_register_type);
237
238 int xfrm_unregister_type(const struct xfrm_type *type, unsigned short family)
239 {
240         struct xfrm_state_afinfo *afinfo = xfrm_state_lock_afinfo(family);
241         const struct xfrm_type **typemap;
242         int err = 0;
243
244         if (unlikely(afinfo == NULL))
245                 return -EAFNOSUPPORT;
246         typemap = afinfo->type_map;
247
248         if (unlikely(typemap[type->proto] != type))
249                 err = -ENOENT;
250         else
251                 typemap[type->proto] = NULL;
252         xfrm_state_unlock_afinfo(afinfo);
253         return err;
254 }
255 EXPORT_SYMBOL(xfrm_unregister_type);
256
257 static const struct xfrm_type *xfrm_get_type(u8 proto, unsigned short family)
258 {
259         struct xfrm_state_afinfo *afinfo;
260         const struct xfrm_type **typemap;
261         const struct xfrm_type *type;
262         int modload_attempted = 0;
263
264 retry:
265         afinfo = xfrm_state_get_afinfo(family);
266         if (unlikely(afinfo == NULL))
267                 return NULL;
268         typemap = afinfo->type_map;
269
270         type = typemap[proto];
271         if (unlikely(type && !try_module_get(type->owner)))
272                 type = NULL;
273         if (!type && !modload_attempted) {
274                 xfrm_state_put_afinfo(afinfo);
275                 request_module("xfrm-type-%d-%d", family, proto);
276                 modload_attempted = 1;
277                 goto retry;
278         }
279
280         xfrm_state_put_afinfo(afinfo);
281         return type;
282 }
283
284 static void xfrm_put_type(const struct xfrm_type *type)
285 {
286         module_put(type->owner);
287 }
288
289 int xfrm_register_mode(struct xfrm_mode *mode, int family)
290 {
291         struct xfrm_state_afinfo *afinfo;
292         struct xfrm_mode **modemap;
293         int err;
294
295         if (unlikely(mode->encap >= XFRM_MODE_MAX))
296                 return -EINVAL;
297
298         afinfo = xfrm_state_lock_afinfo(family);
299         if (unlikely(afinfo == NULL))
300                 return -EAFNOSUPPORT;
301
302         err = -EEXIST;
303         modemap = afinfo->mode_map;
304         if (modemap[mode->encap])
305                 goto out;
306
307         err = -ENOENT;
308         if (!try_module_get(afinfo->owner))
309                 goto out;
310
311         mode->afinfo = afinfo;
312         modemap[mode->encap] = mode;
313         err = 0;
314
315 out:
316         xfrm_state_unlock_afinfo(afinfo);
317         return err;
318 }
319 EXPORT_SYMBOL(xfrm_register_mode);
320
321 int xfrm_unregister_mode(struct xfrm_mode *mode, int family)
322 {
323         struct xfrm_state_afinfo *afinfo;
324         struct xfrm_mode **modemap;
325         int err;
326
327         if (unlikely(mode->encap >= XFRM_MODE_MAX))
328                 return -EINVAL;
329
330         afinfo = xfrm_state_lock_afinfo(family);
331         if (unlikely(afinfo == NULL))
332                 return -EAFNOSUPPORT;
333
334         err = -ENOENT;
335         modemap = afinfo->mode_map;
336         if (likely(modemap[mode->encap] == mode)) {
337                 modemap[mode->encap] = NULL;
338                 module_put(mode->afinfo->owner);
339                 err = 0;
340         }
341
342         xfrm_state_unlock_afinfo(afinfo);
343         return err;
344 }
345 EXPORT_SYMBOL(xfrm_unregister_mode);
346
347 static struct xfrm_mode *xfrm_get_mode(unsigned int encap, int family)
348 {
349         struct xfrm_state_afinfo *afinfo;
350         struct xfrm_mode *mode;
351         int modload_attempted = 0;
352
353         if (unlikely(encap >= XFRM_MODE_MAX))
354                 return NULL;
355
356 retry:
357         afinfo = xfrm_state_get_afinfo(family);
358         if (unlikely(afinfo == NULL))
359                 return NULL;
360
361         mode = afinfo->mode_map[encap];
362         if (unlikely(mode && !try_module_get(mode->owner)))
363                 mode = NULL;
364         if (!mode && !modload_attempted) {
365                 xfrm_state_put_afinfo(afinfo);
366                 request_module("xfrm-mode-%d-%d", family, encap);
367                 modload_attempted = 1;
368                 goto retry;
369         }
370
371         xfrm_state_put_afinfo(afinfo);
372         return mode;
373 }
374
375 static void xfrm_put_mode(struct xfrm_mode *mode)
376 {
377         module_put(mode->owner);
378 }
379
380 static void xfrm_state_gc_destroy(struct xfrm_state *x)
381 {
382         del_timer_sync(&x->timer);
383         del_timer_sync(&x->rtimer);
384         kfree(x->aalg);
385         kfree(x->ealg);
386         kfree(x->calg);
387         kfree(x->encap);
388         kfree(x->coaddr);
389         if (x->inner_mode)
390                 xfrm_put_mode(x->inner_mode);
391         if (x->outer_mode)
392                 xfrm_put_mode(x->outer_mode);
393         if (x->type) {
394                 x->type->destructor(x);
395                 xfrm_put_type(x->type);
396         }
397         security_xfrm_state_free(x);
398         kfree(x);
399 }
400
401 static void xfrm_state_gc_task(struct work_struct *data)
402 {
403         struct xfrm_state *x;
404         struct hlist_node *entry, *tmp;
405         struct hlist_head gc_list;
406
407         spin_lock_bh(&xfrm_state_gc_lock);
408         gc_list.first = xfrm_state_gc_list.first;
409         INIT_HLIST_HEAD(&xfrm_state_gc_list);
410         spin_unlock_bh(&xfrm_state_gc_lock);
411
412         hlist_for_each_entry_safe(x, entry, tmp, &gc_list, bydst)
413                 xfrm_state_gc_destroy(x);
414
415         wake_up(&km_waitq);
416 }
417
418 static inline unsigned long make_jiffies(long secs)
419 {
420         if (secs >= (MAX_SCHEDULE_TIMEOUT-1)/HZ)
421                 return MAX_SCHEDULE_TIMEOUT-1;
422         else
423                 return secs*HZ;
424 }
425
426 static void xfrm_timer_handler(unsigned long data)
427 {
428         struct xfrm_state *x = (struct xfrm_state*)data;
429         unsigned long now = get_seconds();
430         long next = LONG_MAX;
431         int warn = 0;
432         int err = 0;
433
434         spin_lock(&x->lock);
435         if (x->km.state == XFRM_STATE_DEAD)
436                 goto out;
437         if (x->km.state == XFRM_STATE_EXPIRED)
438                 goto expired;
439         if (x->lft.hard_add_expires_seconds) {
440                 long tmo = x->lft.hard_add_expires_seconds +
441                         x->curlft.add_time - now;
442                 if (tmo <= 0)
443                         goto expired;
444                 if (tmo < next)
445                         next = tmo;
446         }
447         if (x->lft.hard_use_expires_seconds) {
448                 long tmo = x->lft.hard_use_expires_seconds +
449                         (x->curlft.use_time ? : now) - now;
450                 if (tmo <= 0)
451                         goto expired;
452                 if (tmo < next)
453                         next = tmo;
454         }
455         if (x->km.dying)
456                 goto resched;
457         if (x->lft.soft_add_expires_seconds) {
458                 long tmo = x->lft.soft_add_expires_seconds +
459                         x->curlft.add_time - now;
460                 if (tmo <= 0)
461                         warn = 1;
462                 else if (tmo < next)
463                         next = tmo;
464         }
465         if (x->lft.soft_use_expires_seconds) {
466                 long tmo = x->lft.soft_use_expires_seconds +
467                         (x->curlft.use_time ? : now) - now;
468                 if (tmo <= 0)
469                         warn = 1;
470                 else if (tmo < next)
471                         next = tmo;
472         }
473
474         x->km.dying = warn;
475         if (warn)
476                 km_state_expired(x, 0, 0);
477 resched:
478         if (next != LONG_MAX)
479                 mod_timer(&x->timer, jiffies + make_jiffies(next));
480
481         goto out;
482
483 expired:
484         if (x->km.state == XFRM_STATE_ACQ && x->id.spi == 0) {
485                 x->km.state = XFRM_STATE_EXPIRED;
486                 wake_up(&km_waitq);
487                 next = 2;
488                 goto resched;
489         }
490
491         err = __xfrm_state_delete(x);
492         if (!err && x->id.spi)
493                 km_state_expired(x, 1, 0);
494
495         xfrm_audit_state_delete(x, err ? 0 : 1,
496                                 audit_get_loginuid(current), 0);
497
498 out:
499         spin_unlock(&x->lock);
500 }
501
502 static void xfrm_replay_timer_handler(unsigned long data);
503
504 struct xfrm_state *xfrm_state_alloc(void)
505 {
506         struct xfrm_state *x;
507
508         x = kzalloc(sizeof(struct xfrm_state), GFP_ATOMIC);
509
510         if (x) {
511                 atomic_set(&x->refcnt, 1);
512                 atomic_set(&x->tunnel_users, 0);
513                 INIT_HLIST_NODE(&x->bydst);
514                 INIT_HLIST_NODE(&x->bysrc);
515                 INIT_HLIST_NODE(&x->byspi);
516                 setup_timer(&x->timer, xfrm_timer_handler, (unsigned long)x);
517                 setup_timer(&x->rtimer, xfrm_replay_timer_handler,
518                                 (unsigned long)x);
519                 x->curlft.add_time = get_seconds();
520                 x->lft.soft_byte_limit = XFRM_INF;
521                 x->lft.soft_packet_limit = XFRM_INF;
522                 x->lft.hard_byte_limit = XFRM_INF;
523                 x->lft.hard_packet_limit = XFRM_INF;
524                 x->replay_maxage = 0;
525                 x->replay_maxdiff = 0;
526                 spin_lock_init(&x->lock);
527         }
528         return x;
529 }
530 EXPORT_SYMBOL(xfrm_state_alloc);
531
532 void __xfrm_state_destroy(struct xfrm_state *x)
533 {
534         BUG_TRAP(x->km.state == XFRM_STATE_DEAD);
535
536         spin_lock_bh(&xfrm_state_gc_lock);
537         hlist_add_head(&x->bydst, &xfrm_state_gc_list);
538         spin_unlock_bh(&xfrm_state_gc_lock);
539         schedule_work(&xfrm_state_gc_work);
540 }
541 EXPORT_SYMBOL(__xfrm_state_destroy);
542
543 int __xfrm_state_delete(struct xfrm_state *x)
544 {
545         int err = -ESRCH;
546
547         if (x->km.state != XFRM_STATE_DEAD) {
548                 x->km.state = XFRM_STATE_DEAD;
549                 spin_lock(&xfrm_state_lock);
550                 hlist_del(&x->bydst);
551                 hlist_del(&x->bysrc);
552                 if (x->id.spi)
553                         hlist_del(&x->byspi);
554                 xfrm_state_num--;
555                 spin_unlock(&xfrm_state_lock);
556
557                 /* All xfrm_state objects are created by xfrm_state_alloc.
558                  * The xfrm_state_alloc call gives a reference, and that
559                  * is what we are dropping here.
560                  */
561                 xfrm_state_put(x);
562                 err = 0;
563         }
564
565         return err;
566 }
567 EXPORT_SYMBOL(__xfrm_state_delete);
568
569 int xfrm_state_delete(struct xfrm_state *x)
570 {
571         int err;
572
573         spin_lock_bh(&x->lock);
574         err = __xfrm_state_delete(x);
575         spin_unlock_bh(&x->lock);
576
577         return err;
578 }
579 EXPORT_SYMBOL(xfrm_state_delete);
580
581 #ifdef CONFIG_SECURITY_NETWORK_XFRM
582 static inline int
583 xfrm_state_flush_secctx_check(u8 proto, struct xfrm_audit *audit_info)
584 {
585         int i, err = 0;
586
587         for (i = 0; i <= xfrm_state_hmask; i++) {
588                 struct hlist_node *entry;
589                 struct xfrm_state *x;
590
591                 hlist_for_each_entry(x, entry, xfrm_state_bydst+i, bydst) {
592                         if (xfrm_id_proto_match(x->id.proto, proto) &&
593                            (err = security_xfrm_state_delete(x)) != 0) {
594                                 xfrm_audit_state_delete(x, 0,
595                                                         audit_info->loginuid,
596                                                         audit_info->secid);
597                                 return err;
598                         }
599                 }
600         }
601
602         return err;
603 }
604 #else
605 static inline int
606 xfrm_state_flush_secctx_check(u8 proto, struct xfrm_audit *audit_info)
607 {
608         return 0;
609 }
610 #endif
611
612 int xfrm_state_flush(u8 proto, struct xfrm_audit *audit_info)
613 {
614         int i, err = 0;
615
616         spin_lock_bh(&xfrm_state_lock);
617         err = xfrm_state_flush_secctx_check(proto, audit_info);
618         if (err)
619                 goto out;
620
621         for (i = 0; i <= xfrm_state_hmask; i++) {
622                 struct hlist_node *entry;
623                 struct xfrm_state *x;
624 restart:
625                 hlist_for_each_entry(x, entry, xfrm_state_bydst+i, bydst) {
626                         if (!xfrm_state_kern(x) &&
627                             xfrm_id_proto_match(x->id.proto, proto)) {
628                                 xfrm_state_hold(x);
629                                 spin_unlock_bh(&xfrm_state_lock);
630
631                                 err = xfrm_state_delete(x);
632                                 xfrm_audit_state_delete(x, err ? 0 : 1,
633                                                         audit_info->loginuid,
634                                                         audit_info->secid);
635                                 xfrm_state_put(x);
636
637                                 spin_lock_bh(&xfrm_state_lock);
638                                 goto restart;
639                         }
640                 }
641         }
642         err = 0;
643
644 out:
645         spin_unlock_bh(&xfrm_state_lock);
646         wake_up(&km_waitq);
647         return err;
648 }
649 EXPORT_SYMBOL(xfrm_state_flush);
650
651 void xfrm_sad_getinfo(struct xfrmk_sadinfo *si)
652 {
653         spin_lock_bh(&xfrm_state_lock);
654         si->sadcnt = xfrm_state_num;
655         si->sadhcnt = xfrm_state_hmask;
656         si->sadhmcnt = xfrm_state_hashmax;
657         spin_unlock_bh(&xfrm_state_lock);
658 }
659 EXPORT_SYMBOL(xfrm_sad_getinfo);
660
661 static int
662 xfrm_init_tempsel(struct xfrm_state *x, struct flowi *fl,
663                   struct xfrm_tmpl *tmpl,
664                   xfrm_address_t *daddr, xfrm_address_t *saddr,
665                   unsigned short family)
666 {
667         struct xfrm_state_afinfo *afinfo = xfrm_state_get_afinfo(family);
668         if (!afinfo)
669                 return -1;
670         afinfo->init_tempsel(x, fl, tmpl, daddr, saddr);
671         xfrm_state_put_afinfo(afinfo);
672         return 0;
673 }
674
675 static struct xfrm_state *__xfrm_state_lookup(xfrm_address_t *daddr, __be32 spi, u8 proto, unsigned short family)
676 {
677         unsigned int h = xfrm_spi_hash(daddr, spi, proto, family);
678         struct xfrm_state *x;
679         struct hlist_node *entry;
680
681         hlist_for_each_entry(x, entry, xfrm_state_byspi+h, byspi) {
682                 if (x->props.family != family ||
683                     x->id.spi       != spi ||
684                     x->id.proto     != proto)
685                         continue;
686
687                 switch (family) {
688                 case AF_INET:
689                         if (x->id.daddr.a4 != daddr->a4)
690                                 continue;
691                         break;
692                 case AF_INET6:
693                         if (!ipv6_addr_equal((struct in6_addr *)daddr,
694                                              (struct in6_addr *)
695                                              x->id.daddr.a6))
696                                 continue;
697                         break;
698                 }
699
700                 xfrm_state_hold(x);
701                 return x;
702         }
703
704         return NULL;
705 }
706
707 static struct xfrm_state *__xfrm_state_lookup_byaddr(xfrm_address_t *daddr, xfrm_address_t *saddr, u8 proto, unsigned short family)
708 {
709         unsigned int h = xfrm_src_hash(daddr, saddr, family);
710         struct xfrm_state *x;
711         struct hlist_node *entry;
712
713         hlist_for_each_entry(x, entry, xfrm_state_bysrc+h, bysrc) {
714                 if (x->props.family != family ||
715                     x->id.proto     != proto)
716                         continue;
717
718                 switch (family) {
719                 case AF_INET:
720                         if (x->id.daddr.a4 != daddr->a4 ||
721                             x->props.saddr.a4 != saddr->a4)
722                                 continue;
723                         break;
724                 case AF_INET6:
725                         if (!ipv6_addr_equal((struct in6_addr *)daddr,
726                                              (struct in6_addr *)
727                                              x->id.daddr.a6) ||
728                             !ipv6_addr_equal((struct in6_addr *)saddr,
729                                              (struct in6_addr *)
730                                              x->props.saddr.a6))
731                                 continue;
732                         break;
733                 }
734
735                 xfrm_state_hold(x);
736                 return x;
737         }
738
739         return NULL;
740 }
741
742 static inline struct xfrm_state *
743 __xfrm_state_locate(struct xfrm_state *x, int use_spi, int family)
744 {
745         if (use_spi)
746                 return __xfrm_state_lookup(&x->id.daddr, x->id.spi,
747                                            x->id.proto, family);
748         else
749                 return __xfrm_state_lookup_byaddr(&x->id.daddr,
750                                                   &x->props.saddr,
751                                                   x->id.proto, family);
752 }
753
754 static void xfrm_hash_grow_check(int have_hash_collision)
755 {
756         if (have_hash_collision &&
757             (xfrm_state_hmask + 1) < xfrm_state_hashmax &&
758             xfrm_state_num > xfrm_state_hmask)
759                 schedule_work(&xfrm_hash_work);
760 }
761
762 struct xfrm_state *
763 xfrm_state_find(xfrm_address_t *daddr, xfrm_address_t *saddr,
764                 struct flowi *fl, struct xfrm_tmpl *tmpl,
765                 struct xfrm_policy *pol, int *err,
766                 unsigned short family)
767 {
768         unsigned int h;
769         struct hlist_node *entry;
770         struct xfrm_state *x, *x0;
771         int acquire_in_progress = 0;
772         int error = 0;
773         struct xfrm_state *best = NULL;
774
775         spin_lock_bh(&xfrm_state_lock);
776         h = xfrm_dst_hash(daddr, saddr, tmpl->reqid, family);
777         hlist_for_each_entry(x, entry, xfrm_state_bydst+h, bydst) {
778                 if (x->props.family == family &&
779                     x->props.reqid == tmpl->reqid &&
780                     !(x->props.flags & XFRM_STATE_WILDRECV) &&
781                     xfrm_state_addr_check(x, daddr, saddr, family) &&
782                     tmpl->mode == x->props.mode &&
783                     tmpl->id.proto == x->id.proto &&
784                     (tmpl->id.spi == x->id.spi || !tmpl->id.spi)) {
785                         /* Resolution logic:
786                            1. There is a valid state with matching selector.
787                               Done.
788                            2. Valid state with inappropriate selector. Skip.
789
790                            Entering area of "sysdeps".
791
792                            3. If state is not valid, selector is temporary,
793                               it selects only session which triggered
794                               previous resolution. Key manager will do
795                               something to install a state with proper
796                               selector.
797                          */
798                         if (x->km.state == XFRM_STATE_VALID) {
799                                 if (!xfrm_selector_match(&x->sel, fl, x->sel.family) ||
800                                     !security_xfrm_state_pol_flow_match(x, pol, fl))
801                                         continue;
802                                 if (!best ||
803                                     best->km.dying > x->km.dying ||
804                                     (best->km.dying == x->km.dying &&
805                                      best->curlft.add_time < x->curlft.add_time))
806                                         best = x;
807                         } else if (x->km.state == XFRM_STATE_ACQ) {
808                                 acquire_in_progress = 1;
809                         } else if (x->km.state == XFRM_STATE_ERROR ||
810                                    x->km.state == XFRM_STATE_EXPIRED) {
811                                 if (xfrm_selector_match(&x->sel, fl, x->sel.family) &&
812                                     security_xfrm_state_pol_flow_match(x, pol, fl))
813                                         error = -ESRCH;
814                         }
815                 }
816         }
817
818         x = best;
819         if (!x && !error && !acquire_in_progress) {
820                 if (tmpl->id.spi &&
821                     (x0 = __xfrm_state_lookup(daddr, tmpl->id.spi,
822                                               tmpl->id.proto, family)) != NULL) {
823                         xfrm_state_put(x0);
824                         error = -EEXIST;
825                         goto out;
826                 }
827                 x = xfrm_state_alloc();
828                 if (x == NULL) {
829                         error = -ENOMEM;
830                         goto out;
831                 }
832                 /* Initialize temporary selector matching only
833                  * to current session. */
834                 xfrm_init_tempsel(x, fl, tmpl, daddr, saddr, family);
835
836                 error = security_xfrm_state_alloc_acquire(x, pol->security, fl->secid);
837                 if (error) {
838                         x->km.state = XFRM_STATE_DEAD;
839                         xfrm_state_put(x);
840                         x = NULL;
841                         goto out;
842                 }
843
844                 if (km_query(x, tmpl, pol) == 0) {
845                         x->km.state = XFRM_STATE_ACQ;
846                         hlist_add_head(&x->bydst, xfrm_state_bydst+h);
847                         h = xfrm_src_hash(daddr, saddr, family);
848                         hlist_add_head(&x->bysrc, xfrm_state_bysrc+h);
849                         if (x->id.spi) {
850                                 h = xfrm_spi_hash(&x->id.daddr, x->id.spi, x->id.proto, family);
851                                 hlist_add_head(&x->byspi, xfrm_state_byspi+h);
852                         }
853                         x->lft.hard_add_expires_seconds = sysctl_xfrm_acq_expires;
854                         x->timer.expires = jiffies + sysctl_xfrm_acq_expires*HZ;
855                         add_timer(&x->timer);
856                         xfrm_state_num++;
857                         xfrm_hash_grow_check(x->bydst.next != NULL);
858                 } else {
859                         x->km.state = XFRM_STATE_DEAD;
860                         xfrm_state_put(x);
861                         x = NULL;
862                         error = -ESRCH;
863                 }
864         }
865 out:
866         if (x)
867                 xfrm_state_hold(x);
868         else
869                 *err = acquire_in_progress ? -EAGAIN : error;
870         spin_unlock_bh(&xfrm_state_lock);
871         return x;
872 }
873
874 struct xfrm_state *
875 xfrm_stateonly_find(xfrm_address_t *daddr, xfrm_address_t *saddr,
876                     unsigned short family, u8 mode, u8 proto, u32 reqid)
877 {
878         unsigned int h;
879         struct xfrm_state *rx = NULL, *x = NULL;
880         struct hlist_node *entry;
881
882         spin_lock(&xfrm_state_lock);
883         h = xfrm_dst_hash(daddr, saddr, reqid, family);
884         hlist_for_each_entry(x, entry, xfrm_state_bydst+h, bydst) {
885                 if (x->props.family == family &&
886                     x->props.reqid == reqid &&
887                     !(x->props.flags & XFRM_STATE_WILDRECV) &&
888                     xfrm_state_addr_check(x, daddr, saddr, family) &&
889                     mode == x->props.mode &&
890                     proto == x->id.proto &&
891                     x->km.state == XFRM_STATE_VALID) {
892                         rx = x;
893                         break;
894                 }
895         }
896
897         if (rx)
898                 xfrm_state_hold(rx);
899         spin_unlock(&xfrm_state_lock);
900
901
902         return rx;
903 }
904 EXPORT_SYMBOL(xfrm_stateonly_find);
905
906 static void __xfrm_state_insert(struct xfrm_state *x)
907 {
908         unsigned int h;
909
910         x->genid = ++xfrm_state_genid;
911
912         h = xfrm_dst_hash(&x->id.daddr, &x->props.saddr,
913                           x->props.reqid, x->props.family);
914         hlist_add_head(&x->bydst, xfrm_state_bydst+h);
915
916         h = xfrm_src_hash(&x->id.daddr, &x->props.saddr, x->props.family);
917         hlist_add_head(&x->bysrc, xfrm_state_bysrc+h);
918
919         if (x->id.spi) {
920                 h = xfrm_spi_hash(&x->id.daddr, x->id.spi, x->id.proto,
921                                   x->props.family);
922
923                 hlist_add_head(&x->byspi, xfrm_state_byspi+h);
924         }
925
926         mod_timer(&x->timer, jiffies + HZ);
927         if (x->replay_maxage)
928                 mod_timer(&x->rtimer, jiffies + x->replay_maxage);
929
930         wake_up(&km_waitq);
931
932         xfrm_state_num++;
933
934         xfrm_hash_grow_check(x->bydst.next != NULL);
935 }
936
937 /* xfrm_state_lock is held */
938 static void __xfrm_state_bump_genids(struct xfrm_state *xnew)
939 {
940         unsigned short family = xnew->props.family;
941         u32 reqid = xnew->props.reqid;
942         struct xfrm_state *x;
943         struct hlist_node *entry;
944         unsigned int h;
945
946         h = xfrm_dst_hash(&xnew->id.daddr, &xnew->props.saddr, reqid, family);
947         hlist_for_each_entry(x, entry, xfrm_state_bydst+h, bydst) {
948                 if (x->props.family     == family &&
949                     x->props.reqid      == reqid &&
950                     !xfrm_addr_cmp(&x->id.daddr, &xnew->id.daddr, family) &&
951                     !xfrm_addr_cmp(&x->props.saddr, &xnew->props.saddr, family))
952                         x->genid = xfrm_state_genid;
953         }
954 }
955
956 void xfrm_state_insert(struct xfrm_state *x)
957 {
958         spin_lock_bh(&xfrm_state_lock);
959         __xfrm_state_bump_genids(x);
960         __xfrm_state_insert(x);
961         spin_unlock_bh(&xfrm_state_lock);
962 }
963 EXPORT_SYMBOL(xfrm_state_insert);
964
965 /* xfrm_state_lock is held */
966 static struct xfrm_state *__find_acq_core(unsigned short family, u8 mode, u32 reqid, u8 proto, xfrm_address_t *daddr, xfrm_address_t *saddr, int create)
967 {
968         unsigned int h = xfrm_dst_hash(daddr, saddr, reqid, family);
969         struct hlist_node *entry;
970         struct xfrm_state *x;
971
972         hlist_for_each_entry(x, entry, xfrm_state_bydst+h, bydst) {
973                 if (x->props.reqid  != reqid ||
974                     x->props.mode   != mode ||
975                     x->props.family != family ||
976                     x->km.state     != XFRM_STATE_ACQ ||
977                     x->id.spi       != 0 ||
978                     x->id.proto     != proto)
979                         continue;
980
981                 switch (family) {
982                 case AF_INET:
983                         if (x->id.daddr.a4    != daddr->a4 ||
984                             x->props.saddr.a4 != saddr->a4)
985                                 continue;
986                         break;
987                 case AF_INET6:
988                         if (!ipv6_addr_equal((struct in6_addr *)x->id.daddr.a6,
989                                              (struct in6_addr *)daddr) ||
990                             !ipv6_addr_equal((struct in6_addr *)
991                                              x->props.saddr.a6,
992                                              (struct in6_addr *)saddr))
993                                 continue;
994                         break;
995                 }
996
997                 xfrm_state_hold(x);
998                 return x;
999         }
1000
1001         if (!create)
1002                 return NULL;
1003
1004         x = xfrm_state_alloc();
1005         if (likely(x)) {
1006                 switch (family) {
1007                 case AF_INET:
1008                         x->sel.daddr.a4 = daddr->a4;
1009                         x->sel.saddr.a4 = saddr->a4;
1010                         x->sel.prefixlen_d = 32;
1011                         x->sel.prefixlen_s = 32;
1012                         x->props.saddr.a4 = saddr->a4;
1013                         x->id.daddr.a4 = daddr->a4;
1014                         break;
1015
1016                 case AF_INET6:
1017                         ipv6_addr_copy((struct in6_addr *)x->sel.daddr.a6,
1018                                        (struct in6_addr *)daddr);
1019                         ipv6_addr_copy((struct in6_addr *)x->sel.saddr.a6,
1020                                        (struct in6_addr *)saddr);
1021                         x->sel.prefixlen_d = 128;
1022                         x->sel.prefixlen_s = 128;
1023                         ipv6_addr_copy((struct in6_addr *)x->props.saddr.a6,
1024                                        (struct in6_addr *)saddr);
1025                         ipv6_addr_copy((struct in6_addr *)x->id.daddr.a6,
1026                                        (struct in6_addr *)daddr);
1027                         break;
1028                 }
1029
1030                 x->km.state = XFRM_STATE_ACQ;
1031                 x->id.proto = proto;
1032                 x->props.family = family;
1033                 x->props.mode = mode;
1034                 x->props.reqid = reqid;
1035                 x->lft.hard_add_expires_seconds = sysctl_xfrm_acq_expires;
1036                 xfrm_state_hold(x);
1037                 x->timer.expires = jiffies + sysctl_xfrm_acq_expires*HZ;
1038                 add_timer(&x->timer);
1039                 hlist_add_head(&x->bydst, xfrm_state_bydst+h);
1040                 h = xfrm_src_hash(daddr, saddr, family);
1041                 hlist_add_head(&x->bysrc, xfrm_state_bysrc+h);
1042
1043                 xfrm_state_num++;
1044
1045                 xfrm_hash_grow_check(x->bydst.next != NULL);
1046         }
1047
1048         return x;
1049 }
1050
1051 static struct xfrm_state *__xfrm_find_acq_byseq(u32 seq);
1052
1053 int xfrm_state_add(struct xfrm_state *x)
1054 {
1055         struct xfrm_state *x1;
1056         int family;
1057         int err;
1058         int use_spi = xfrm_id_proto_match(x->id.proto, IPSEC_PROTO_ANY);
1059
1060         family = x->props.family;
1061
1062         spin_lock_bh(&xfrm_state_lock);
1063
1064         x1 = __xfrm_state_locate(x, use_spi, family);
1065         if (x1) {
1066                 xfrm_state_put(x1);
1067                 x1 = NULL;
1068                 err = -EEXIST;
1069                 goto out;
1070         }
1071
1072         if (use_spi && x->km.seq) {
1073                 x1 = __xfrm_find_acq_byseq(x->km.seq);
1074                 if (x1 && ((x1->id.proto != x->id.proto) ||
1075                     xfrm_addr_cmp(&x1->id.daddr, &x->id.daddr, family))) {
1076                         xfrm_state_put(x1);
1077                         x1 = NULL;
1078                 }
1079         }
1080
1081         if (use_spi && !x1)
1082                 x1 = __find_acq_core(family, x->props.mode, x->props.reqid,
1083                                      x->id.proto,
1084                                      &x->id.daddr, &x->props.saddr, 0);
1085
1086         __xfrm_state_bump_genids(x);
1087         __xfrm_state_insert(x);
1088         err = 0;
1089
1090 out:
1091         spin_unlock_bh(&xfrm_state_lock);
1092
1093         if (x1) {
1094                 xfrm_state_delete(x1);
1095                 xfrm_state_put(x1);
1096         }
1097
1098         return err;
1099 }
1100 EXPORT_SYMBOL(xfrm_state_add);
1101
1102 #ifdef CONFIG_XFRM_MIGRATE
1103 static struct xfrm_state *xfrm_state_clone(struct xfrm_state *orig, int *errp)
1104 {
1105         int err = -ENOMEM;
1106         struct xfrm_state *x = xfrm_state_alloc();
1107         if (!x)
1108                 goto error;
1109
1110         memcpy(&x->id, &orig->id, sizeof(x->id));
1111         memcpy(&x->sel, &orig->sel, sizeof(x->sel));
1112         memcpy(&x->lft, &orig->lft, sizeof(x->lft));
1113         x->props.mode = orig->props.mode;
1114         x->props.replay_window = orig->props.replay_window;
1115         x->props.reqid = orig->props.reqid;
1116         x->props.family = orig->props.family;
1117         x->props.saddr = orig->props.saddr;
1118
1119         if (orig->aalg) {
1120                 x->aalg = xfrm_algo_clone(orig->aalg);
1121                 if (!x->aalg)
1122                         goto error;
1123         }
1124         x->props.aalgo = orig->props.aalgo;
1125
1126         if (orig->ealg) {
1127                 x->ealg = xfrm_algo_clone(orig->ealg);
1128                 if (!x->ealg)
1129                         goto error;
1130         }
1131         x->props.ealgo = orig->props.ealgo;
1132
1133         if (orig->calg) {
1134                 x->calg = xfrm_algo_clone(orig->calg);
1135                 if (!x->calg)
1136                         goto error;
1137         }
1138         x->props.calgo = orig->props.calgo;
1139
1140         if (orig->encap) {
1141                 x->encap = kmemdup(orig->encap, sizeof(*x->encap), GFP_KERNEL);
1142                 if (!x->encap)
1143                         goto error;
1144         }
1145
1146         if (orig->coaddr) {
1147                 x->coaddr = kmemdup(orig->coaddr, sizeof(*x->coaddr),
1148                                     GFP_KERNEL);
1149                 if (!x->coaddr)
1150                         goto error;
1151         }
1152
1153         err = xfrm_init_state(x);
1154         if (err)
1155                 goto error;
1156
1157         x->props.flags = orig->props.flags;
1158
1159         x->curlft.add_time = orig->curlft.add_time;
1160         x->km.state = orig->km.state;
1161         x->km.seq = orig->km.seq;
1162
1163         return x;
1164
1165  error:
1166         if (errp)
1167                 *errp = err;
1168         if (x) {
1169                 kfree(x->aalg);
1170                 kfree(x->ealg);
1171                 kfree(x->calg);
1172                 kfree(x->encap);
1173                 kfree(x->coaddr);
1174         }
1175         kfree(x);
1176         return NULL;
1177 }
1178
1179 /* xfrm_state_lock is held */
1180 struct xfrm_state * xfrm_migrate_state_find(struct xfrm_migrate *m)
1181 {
1182         unsigned int h;
1183         struct xfrm_state *x;
1184         struct hlist_node *entry;
1185
1186         if (m->reqid) {
1187                 h = xfrm_dst_hash(&m->old_daddr, &m->old_saddr,
1188                                   m->reqid, m->old_family);
1189                 hlist_for_each_entry(x, entry, xfrm_state_bydst+h, bydst) {
1190                         if (x->props.mode != m->mode ||
1191                             x->id.proto != m->proto)
1192                                 continue;
1193                         if (m->reqid && x->props.reqid != m->reqid)
1194                                 continue;
1195                         if (xfrm_addr_cmp(&x->id.daddr, &m->old_daddr,
1196                                           m->old_family) ||
1197                             xfrm_addr_cmp(&x->props.saddr, &m->old_saddr,
1198                                           m->old_family))
1199                                 continue;
1200                         xfrm_state_hold(x);
1201                         return x;
1202                 }
1203         } else {
1204                 h = xfrm_src_hash(&m->old_daddr, &m->old_saddr,
1205                                   m->old_family);
1206                 hlist_for_each_entry(x, entry, xfrm_state_bysrc+h, bysrc) {
1207                         if (x->props.mode != m->mode ||
1208                             x->id.proto != m->proto)
1209                                 continue;
1210                         if (xfrm_addr_cmp(&x->id.daddr, &m->old_daddr,
1211                                           m->old_family) ||
1212                             xfrm_addr_cmp(&x->props.saddr, &m->old_saddr,
1213                                           m->old_family))
1214                                 continue;
1215                         xfrm_state_hold(x);
1216                         return x;
1217                 }
1218         }
1219
1220         return NULL;
1221 }
1222 EXPORT_SYMBOL(xfrm_migrate_state_find);
1223
1224 struct xfrm_state * xfrm_state_migrate(struct xfrm_state *x,
1225                                        struct xfrm_migrate *m)
1226 {
1227         struct xfrm_state *xc;
1228         int err;
1229
1230         xc = xfrm_state_clone(x, &err);
1231         if (!xc)
1232                 return NULL;
1233
1234         memcpy(&xc->id.daddr, &m->new_daddr, sizeof(xc->id.daddr));
1235         memcpy(&xc->props.saddr, &m->new_saddr, sizeof(xc->props.saddr));
1236
1237         /* add state */
1238         if (!xfrm_addr_cmp(&x->id.daddr, &m->new_daddr, m->new_family)) {
1239                 /* a care is needed when the destination address of the
1240                    state is to be updated as it is a part of triplet */
1241                 xfrm_state_insert(xc);
1242         } else {
1243                 if ((err = xfrm_state_add(xc)) < 0)
1244                         goto error;
1245         }
1246
1247         return xc;
1248 error:
1249         kfree(xc);
1250         return NULL;
1251 }
1252 EXPORT_SYMBOL(xfrm_state_migrate);
1253 #endif
1254
1255 int xfrm_state_update(struct xfrm_state *x)
1256 {
1257         struct xfrm_state *x1;
1258         int err;
1259         int use_spi = xfrm_id_proto_match(x->id.proto, IPSEC_PROTO_ANY);
1260
1261         spin_lock_bh(&xfrm_state_lock);
1262         x1 = __xfrm_state_locate(x, use_spi, x->props.family);
1263
1264         err = -ESRCH;
1265         if (!x1)
1266                 goto out;
1267
1268         if (xfrm_state_kern(x1)) {
1269                 xfrm_state_put(x1);
1270                 err = -EEXIST;
1271                 goto out;
1272         }
1273
1274         if (x1->km.state == XFRM_STATE_ACQ) {
1275                 __xfrm_state_insert(x);
1276                 x = NULL;
1277         }
1278         err = 0;
1279
1280 out:
1281         spin_unlock_bh(&xfrm_state_lock);
1282
1283         if (err)
1284                 return err;
1285
1286         if (!x) {
1287                 xfrm_state_delete(x1);
1288                 xfrm_state_put(x1);
1289                 return 0;
1290         }
1291
1292         err = -EINVAL;
1293         spin_lock_bh(&x1->lock);
1294         if (likely(x1->km.state == XFRM_STATE_VALID)) {
1295                 if (x->encap && x1->encap)
1296                         memcpy(x1->encap, x->encap, sizeof(*x1->encap));
1297                 if (x->coaddr && x1->coaddr) {
1298                         memcpy(x1->coaddr, x->coaddr, sizeof(*x1->coaddr));
1299                 }
1300                 if (!use_spi && memcmp(&x1->sel, &x->sel, sizeof(x1->sel)))
1301                         memcpy(&x1->sel, &x->sel, sizeof(x1->sel));
1302                 memcpy(&x1->lft, &x->lft, sizeof(x1->lft));
1303                 x1->km.dying = 0;
1304
1305                 mod_timer(&x1->timer, jiffies + HZ);
1306                 if (x1->curlft.use_time)
1307                         xfrm_state_check_expire(x1);
1308
1309                 err = 0;
1310         }
1311         spin_unlock_bh(&x1->lock);
1312
1313         xfrm_state_put(x1);
1314
1315         return err;
1316 }
1317 EXPORT_SYMBOL(xfrm_state_update);
1318
1319 int xfrm_state_check_expire(struct xfrm_state *x)
1320 {
1321         if (!x->curlft.use_time)
1322                 x->curlft.use_time = get_seconds();
1323
1324         if (x->km.state != XFRM_STATE_VALID)
1325                 return -EINVAL;
1326
1327         if (x->curlft.bytes >= x->lft.hard_byte_limit ||
1328             x->curlft.packets >= x->lft.hard_packet_limit) {
1329                 x->km.state = XFRM_STATE_EXPIRED;
1330                 mod_timer(&x->timer, jiffies);
1331                 return -EINVAL;
1332         }
1333
1334         if (!x->km.dying &&
1335             (x->curlft.bytes >= x->lft.soft_byte_limit ||
1336              x->curlft.packets >= x->lft.soft_packet_limit)) {
1337                 x->km.dying = 1;
1338                 km_state_expired(x, 0, 0);
1339         }
1340         return 0;
1341 }
1342 EXPORT_SYMBOL(xfrm_state_check_expire);
1343
1344 struct xfrm_state *
1345 xfrm_state_lookup(xfrm_address_t *daddr, __be32 spi, u8 proto,
1346                   unsigned short family)
1347 {
1348         struct xfrm_state *x;
1349
1350         spin_lock_bh(&xfrm_state_lock);
1351         x = __xfrm_state_lookup(daddr, spi, proto, family);
1352         spin_unlock_bh(&xfrm_state_lock);
1353         return x;
1354 }
1355 EXPORT_SYMBOL(xfrm_state_lookup);
1356
1357 struct xfrm_state *
1358 xfrm_state_lookup_byaddr(xfrm_address_t *daddr, xfrm_address_t *saddr,
1359                          u8 proto, unsigned short family)
1360 {
1361         struct xfrm_state *x;
1362
1363         spin_lock_bh(&xfrm_state_lock);
1364         x = __xfrm_state_lookup_byaddr(daddr, saddr, proto, family);
1365         spin_unlock_bh(&xfrm_state_lock);
1366         return x;
1367 }
1368 EXPORT_SYMBOL(xfrm_state_lookup_byaddr);
1369
1370 struct xfrm_state *
1371 xfrm_find_acq(u8 mode, u32 reqid, u8 proto,
1372               xfrm_address_t *daddr, xfrm_address_t *saddr,
1373               int create, unsigned short family)
1374 {
1375         struct xfrm_state *x;
1376
1377         spin_lock_bh(&xfrm_state_lock);
1378         x = __find_acq_core(family, mode, reqid, proto, daddr, saddr, create);
1379         spin_unlock_bh(&xfrm_state_lock);
1380
1381         return x;
1382 }
1383 EXPORT_SYMBOL(xfrm_find_acq);
1384
1385 #ifdef CONFIG_XFRM_SUB_POLICY
1386 int
1387 xfrm_tmpl_sort(struct xfrm_tmpl **dst, struct xfrm_tmpl **src, int n,
1388                unsigned short family)
1389 {
1390         int err = 0;
1391         struct xfrm_state_afinfo *afinfo = xfrm_state_get_afinfo(family);
1392         if (!afinfo)
1393                 return -EAFNOSUPPORT;
1394
1395         spin_lock_bh(&xfrm_state_lock);
1396         if (afinfo->tmpl_sort)
1397                 err = afinfo->tmpl_sort(dst, src, n);
1398         spin_unlock_bh(&xfrm_state_lock);
1399         xfrm_state_put_afinfo(afinfo);
1400         return err;
1401 }
1402 EXPORT_SYMBOL(xfrm_tmpl_sort);
1403
1404 int
1405 xfrm_state_sort(struct xfrm_state **dst, struct xfrm_state **src, int n,
1406                 unsigned short family)
1407 {
1408         int err = 0;
1409         struct xfrm_state_afinfo *afinfo = xfrm_state_get_afinfo(family);
1410         if (!afinfo)
1411                 return -EAFNOSUPPORT;
1412
1413         spin_lock_bh(&xfrm_state_lock);
1414         if (afinfo->state_sort)
1415                 err = afinfo->state_sort(dst, src, n);
1416         spin_unlock_bh(&xfrm_state_lock);
1417         xfrm_state_put_afinfo(afinfo);
1418         return err;
1419 }
1420 EXPORT_SYMBOL(xfrm_state_sort);
1421 #endif
1422
1423 /* Silly enough, but I'm lazy to build resolution list */
1424
1425 static struct xfrm_state *__xfrm_find_acq_byseq(u32 seq)
1426 {
1427         int i;
1428
1429         for (i = 0; i <= xfrm_state_hmask; i++) {
1430                 struct hlist_node *entry;
1431                 struct xfrm_state *x;
1432
1433                 hlist_for_each_entry(x, entry, xfrm_state_bydst+i, bydst) {
1434                         if (x->km.seq == seq &&
1435                             x->km.state == XFRM_STATE_ACQ) {
1436                                 xfrm_state_hold(x);
1437                                 return x;
1438                         }
1439                 }
1440         }
1441         return NULL;
1442 }
1443
1444 struct xfrm_state *xfrm_find_acq_byseq(u32 seq)
1445 {
1446         struct xfrm_state *x;
1447
1448         spin_lock_bh(&xfrm_state_lock);
1449         x = __xfrm_find_acq_byseq(seq);
1450         spin_unlock_bh(&xfrm_state_lock);
1451         return x;
1452 }
1453 EXPORT_SYMBOL(xfrm_find_acq_byseq);
1454
1455 u32 xfrm_get_acqseq(void)
1456 {
1457         u32 res;
1458         static u32 acqseq;
1459         static DEFINE_SPINLOCK(acqseq_lock);
1460
1461         spin_lock_bh(&acqseq_lock);
1462         res = (++acqseq ? : ++acqseq);
1463         spin_unlock_bh(&acqseq_lock);
1464         return res;
1465 }
1466 EXPORT_SYMBOL(xfrm_get_acqseq);
1467
1468 int xfrm_alloc_spi(struct xfrm_state *x, u32 low, u32 high)
1469 {
1470         unsigned int h;
1471         struct xfrm_state *x0;
1472         int err = -ENOENT;
1473         __be32 minspi = htonl(low);
1474         __be32 maxspi = htonl(high);
1475
1476         spin_lock_bh(&x->lock);
1477         if (x->km.state == XFRM_STATE_DEAD)
1478                 goto unlock;
1479
1480         err = 0;
1481         if (x->id.spi)
1482                 goto unlock;
1483
1484         err = -ENOENT;
1485
1486         if (minspi == maxspi) {
1487                 x0 = xfrm_state_lookup(&x->id.daddr, minspi, x->id.proto, x->props.family);
1488                 if (x0) {
1489                         xfrm_state_put(x0);
1490                         goto unlock;
1491                 }
1492                 x->id.spi = minspi;
1493         } else {
1494                 u32 spi = 0;
1495                 for (h=0; h<high-low+1; h++) {
1496                         spi = low + net_random()%(high-low+1);
1497                         x0 = xfrm_state_lookup(&x->id.daddr, htonl(spi), x->id.proto, x->props.family);
1498                         if (x0 == NULL) {
1499                                 x->id.spi = htonl(spi);
1500                                 break;
1501                         }
1502                         xfrm_state_put(x0);
1503                 }
1504         }
1505         if (x->id.spi) {
1506                 spin_lock_bh(&xfrm_state_lock);
1507                 h = xfrm_spi_hash(&x->id.daddr, x->id.spi, x->id.proto, x->props.family);
1508                 hlist_add_head(&x->byspi, xfrm_state_byspi+h);
1509                 spin_unlock_bh(&xfrm_state_lock);
1510
1511                 err = 0;
1512         }
1513
1514 unlock:
1515         spin_unlock_bh(&x->lock);
1516
1517         return err;
1518 }
1519 EXPORT_SYMBOL(xfrm_alloc_spi);
1520
1521 int xfrm_state_walk(u8 proto, int (*func)(struct xfrm_state *, int, void*),
1522                     void *data)
1523 {
1524         int i;
1525         struct xfrm_state *x, *last = NULL;
1526         struct hlist_node *entry;
1527         int count = 0;
1528         int err = 0;
1529
1530         spin_lock_bh(&xfrm_state_lock);
1531         for (i = 0; i <= xfrm_state_hmask; i++) {
1532                 hlist_for_each_entry(x, entry, xfrm_state_bydst+i, bydst) {
1533                         if (!xfrm_id_proto_match(x->id.proto, proto))
1534                                 continue;
1535                         if (last) {
1536                                 err = func(last, count, data);
1537                                 if (err)
1538                                         goto out;
1539                         }
1540                         last = x;
1541                         count++;
1542                 }
1543         }
1544         if (count == 0) {
1545                 err = -ENOENT;
1546                 goto out;
1547         }
1548         err = func(last, 0, data);
1549 out:
1550         spin_unlock_bh(&xfrm_state_lock);
1551         return err;
1552 }
1553 EXPORT_SYMBOL(xfrm_state_walk);
1554
1555
1556 void xfrm_replay_notify(struct xfrm_state *x, int event)
1557 {
1558         struct km_event c;
1559         /* we send notify messages in case
1560          *  1. we updated on of the sequence numbers, and the seqno difference
1561          *     is at least x->replay_maxdiff, in this case we also update the
1562          *     timeout of our timer function
1563          *  2. if x->replay_maxage has elapsed since last update,
1564          *     and there were changes
1565          *
1566          *  The state structure must be locked!
1567          */
1568
1569         switch (event) {
1570         case XFRM_REPLAY_UPDATE:
1571                 if (x->replay_maxdiff &&
1572                     (x->replay.seq - x->preplay.seq < x->replay_maxdiff) &&
1573                     (x->replay.oseq - x->preplay.oseq < x->replay_maxdiff)) {
1574                         if (x->xflags & XFRM_TIME_DEFER)
1575                                 event = XFRM_REPLAY_TIMEOUT;
1576                         else
1577                                 return;
1578                 }
1579
1580                 break;
1581
1582         case XFRM_REPLAY_TIMEOUT:
1583                 if ((x->replay.seq == x->preplay.seq) &&
1584                     (x->replay.bitmap == x->preplay.bitmap) &&
1585                     (x->replay.oseq == x->preplay.oseq)) {
1586                         x->xflags |= XFRM_TIME_DEFER;
1587                         return;
1588                 }
1589
1590                 break;
1591         }
1592
1593         memcpy(&x->preplay, &x->replay, sizeof(struct xfrm_replay_state));
1594         c.event = XFRM_MSG_NEWAE;
1595         c.data.aevent = event;
1596         km_state_notify(x, &c);
1597
1598         if (x->replay_maxage &&
1599             !mod_timer(&x->rtimer, jiffies + x->replay_maxage))
1600                 x->xflags &= ~XFRM_TIME_DEFER;
1601 }
1602
1603 static void xfrm_replay_timer_handler(unsigned long data)
1604 {
1605         struct xfrm_state *x = (struct xfrm_state*)data;
1606
1607         spin_lock(&x->lock);
1608
1609         if (x->km.state == XFRM_STATE_VALID) {
1610                 if (xfrm_aevent_is_on())
1611                         xfrm_replay_notify(x, XFRM_REPLAY_TIMEOUT);
1612                 else
1613                         x->xflags |= XFRM_TIME_DEFER;
1614         }
1615
1616         spin_unlock(&x->lock);
1617 }
1618
1619 int xfrm_replay_check(struct xfrm_state *x,
1620                       struct sk_buff *skb, __be32 net_seq)
1621 {
1622         u32 diff;
1623         u32 seq = ntohl(net_seq);
1624
1625         if (unlikely(seq == 0))
1626                 goto err;
1627
1628         if (likely(seq > x->replay.seq))
1629                 return 0;
1630
1631         diff = x->replay.seq - seq;
1632         if (diff >= min_t(unsigned int, x->props.replay_window,
1633                           sizeof(x->replay.bitmap) * 8)) {
1634                 x->stats.replay_window++;
1635                 goto err;
1636         }
1637
1638         if (x->replay.bitmap & (1U << diff)) {
1639                 x->stats.replay++;
1640                 goto err;
1641         }
1642         return 0;
1643
1644 err:
1645         xfrm_audit_state_replay(x, skb, net_seq);
1646         return -EINVAL;
1647 }
1648
1649 void xfrm_replay_advance(struct xfrm_state *x, __be32 net_seq)
1650 {
1651         u32 diff;
1652         u32 seq = ntohl(net_seq);
1653
1654         if (seq > x->replay.seq) {
1655                 diff = seq - x->replay.seq;
1656                 if (diff < x->props.replay_window)
1657                         x->replay.bitmap = ((x->replay.bitmap) << diff) | 1;
1658                 else
1659                         x->replay.bitmap = 1;
1660                 x->replay.seq = seq;
1661         } else {
1662                 diff = x->replay.seq - seq;
1663                 x->replay.bitmap |= (1U << diff);
1664         }
1665
1666         if (xfrm_aevent_is_on())
1667                 xfrm_replay_notify(x, XFRM_REPLAY_UPDATE);
1668 }
1669
1670 static LIST_HEAD(xfrm_km_list);
1671 static DEFINE_RWLOCK(xfrm_km_lock);
1672
1673 void km_policy_notify(struct xfrm_policy *xp, int dir, struct km_event *c)
1674 {
1675         struct xfrm_mgr *km;
1676
1677         read_lock(&xfrm_km_lock);
1678         list_for_each_entry(km, &xfrm_km_list, list)
1679                 if (km->notify_policy)
1680                         km->notify_policy(xp, dir, c);
1681         read_unlock(&xfrm_km_lock);
1682 }
1683
1684 void km_state_notify(struct xfrm_state *x, struct km_event *c)
1685 {
1686         struct xfrm_mgr *km;
1687         read_lock(&xfrm_km_lock);
1688         list_for_each_entry(km, &xfrm_km_list, list)
1689                 if (km->notify)
1690                         km->notify(x, c);
1691         read_unlock(&xfrm_km_lock);
1692 }
1693
1694 EXPORT_SYMBOL(km_policy_notify);
1695 EXPORT_SYMBOL(km_state_notify);
1696
1697 void km_state_expired(struct xfrm_state *x, int hard, u32 pid)
1698 {
1699         struct km_event c;
1700
1701         c.data.hard = hard;
1702         c.pid = pid;
1703         c.event = XFRM_MSG_EXPIRE;
1704         km_state_notify(x, &c);
1705
1706         if (hard)
1707                 wake_up(&km_waitq);
1708 }
1709
1710 EXPORT_SYMBOL(km_state_expired);
1711 /*
1712  * We send to all registered managers regardless of failure
1713  * We are happy with one success
1714 */
1715 int km_query(struct xfrm_state *x, struct xfrm_tmpl *t, struct xfrm_policy *pol)
1716 {
1717         int err = -EINVAL, acqret;
1718         struct xfrm_mgr *km;
1719
1720         read_lock(&xfrm_km_lock);
1721         list_for_each_entry(km, &xfrm_km_list, list) {
1722                 acqret = km->acquire(x, t, pol, XFRM_POLICY_OUT);
1723                 if (!acqret)
1724                         err = acqret;
1725         }
1726         read_unlock(&xfrm_km_lock);
1727         return err;
1728 }
1729 EXPORT_SYMBOL(km_query);
1730
1731 int km_new_mapping(struct xfrm_state *x, xfrm_address_t *ipaddr, __be16 sport)
1732 {
1733         int err = -EINVAL;
1734         struct xfrm_mgr *km;
1735
1736         read_lock(&xfrm_km_lock);
1737         list_for_each_entry(km, &xfrm_km_list, list) {
1738                 if (km->new_mapping)
1739                         err = km->new_mapping(x, ipaddr, sport);
1740                 if (!err)
1741                         break;
1742         }
1743         read_unlock(&xfrm_km_lock);
1744         return err;
1745 }
1746 EXPORT_SYMBOL(km_new_mapping);
1747
1748 void km_policy_expired(struct xfrm_policy *pol, int dir, int hard, u32 pid)
1749 {
1750         struct km_event c;
1751
1752         c.data.hard = hard;
1753         c.pid = pid;
1754         c.event = XFRM_MSG_POLEXPIRE;
1755         km_policy_notify(pol, dir, &c);
1756
1757         if (hard)
1758                 wake_up(&km_waitq);
1759 }
1760 EXPORT_SYMBOL(km_policy_expired);
1761
1762 #ifdef CONFIG_XFRM_MIGRATE
1763 int km_migrate(struct xfrm_selector *sel, u8 dir, u8 type,
1764                struct xfrm_migrate *m, int num_migrate)
1765 {
1766         int err = -EINVAL;
1767         int ret;
1768         struct xfrm_mgr *km;
1769
1770         read_lock(&xfrm_km_lock);
1771         list_for_each_entry(km, &xfrm_km_list, list) {
1772                 if (km->migrate) {
1773                         ret = km->migrate(sel, dir, type, m, num_migrate);
1774                         if (!ret)
1775                                 err = ret;
1776                 }
1777         }
1778         read_unlock(&xfrm_km_lock);
1779         return err;
1780 }
1781 EXPORT_SYMBOL(km_migrate);
1782 #endif
1783
1784 int km_report(u8 proto, struct xfrm_selector *sel, xfrm_address_t *addr)
1785 {
1786         int err = -EINVAL;
1787         int ret;
1788         struct xfrm_mgr *km;
1789
1790         read_lock(&xfrm_km_lock);
1791         list_for_each_entry(km, &xfrm_km_list, list) {
1792                 if (km->report) {
1793                         ret = km->report(proto, sel, addr);
1794                         if (!ret)
1795                                 err = ret;
1796                 }
1797         }
1798         read_unlock(&xfrm_km_lock);
1799         return err;
1800 }
1801 EXPORT_SYMBOL(km_report);
1802
1803 int xfrm_user_policy(struct sock *sk, int optname, u8 __user *optval, int optlen)
1804 {
1805         int err;
1806         u8 *data;
1807         struct xfrm_mgr *km;
1808         struct xfrm_policy *pol = NULL;
1809
1810         if (optlen <= 0 || optlen > PAGE_SIZE)
1811                 return -EMSGSIZE;
1812
1813         data = kmalloc(optlen, GFP_KERNEL);
1814         if (!data)
1815                 return -ENOMEM;
1816
1817         err = -EFAULT;
1818         if (copy_from_user(data, optval, optlen))
1819                 goto out;
1820
1821         err = -EINVAL;
1822         read_lock(&xfrm_km_lock);
1823         list_for_each_entry(km, &xfrm_km_list, list) {
1824                 pol = km->compile_policy(sk, optname, data,
1825                                          optlen, &err);
1826                 if (err >= 0)
1827                         break;
1828         }
1829         read_unlock(&xfrm_km_lock);
1830
1831         if (err >= 0) {
1832                 xfrm_sk_policy_insert(sk, err, pol);
1833                 xfrm_pol_put(pol);
1834                 err = 0;
1835         }
1836
1837 out:
1838         kfree(data);
1839         return err;
1840 }
1841 EXPORT_SYMBOL(xfrm_user_policy);
1842
1843 int xfrm_register_km(struct xfrm_mgr *km)
1844 {
1845         write_lock_bh(&xfrm_km_lock);
1846         list_add_tail(&km->list, &xfrm_km_list);
1847         write_unlock_bh(&xfrm_km_lock);
1848         return 0;
1849 }
1850 EXPORT_SYMBOL(xfrm_register_km);
1851
1852 int xfrm_unregister_km(struct xfrm_mgr *km)
1853 {
1854         write_lock_bh(&xfrm_km_lock);
1855         list_del(&km->list);
1856         write_unlock_bh(&xfrm_km_lock);
1857         return 0;
1858 }
1859 EXPORT_SYMBOL(xfrm_unregister_km);
1860
1861 int xfrm_state_register_afinfo(struct xfrm_state_afinfo *afinfo)
1862 {
1863         int err = 0;
1864         if (unlikely(afinfo == NULL))
1865                 return -EINVAL;
1866         if (unlikely(afinfo->family >= NPROTO))
1867                 return -EAFNOSUPPORT;
1868         write_lock_bh(&xfrm_state_afinfo_lock);
1869         if (unlikely(xfrm_state_afinfo[afinfo->family] != NULL))
1870                 err = -ENOBUFS;
1871         else
1872                 xfrm_state_afinfo[afinfo->family] = afinfo;
1873         write_unlock_bh(&xfrm_state_afinfo_lock);
1874         return err;
1875 }
1876 EXPORT_SYMBOL(xfrm_state_register_afinfo);
1877
1878 int xfrm_state_unregister_afinfo(struct xfrm_state_afinfo *afinfo)
1879 {
1880         int err = 0;
1881         if (unlikely(afinfo == NULL))
1882                 return -EINVAL;
1883         if (unlikely(afinfo->family >= NPROTO))
1884                 return -EAFNOSUPPORT;
1885         write_lock_bh(&xfrm_state_afinfo_lock);
1886         if (likely(xfrm_state_afinfo[afinfo->family] != NULL)) {
1887                 if (unlikely(xfrm_state_afinfo[afinfo->family] != afinfo))
1888                         err = -EINVAL;
1889                 else
1890                         xfrm_state_afinfo[afinfo->family] = NULL;
1891         }
1892         write_unlock_bh(&xfrm_state_afinfo_lock);
1893         return err;
1894 }
1895 EXPORT_SYMBOL(xfrm_state_unregister_afinfo);
1896
1897 static struct xfrm_state_afinfo *xfrm_state_get_afinfo(unsigned int family)
1898 {
1899         struct xfrm_state_afinfo *afinfo;
1900         if (unlikely(family >= NPROTO))
1901                 return NULL;
1902         read_lock(&xfrm_state_afinfo_lock);
1903         afinfo = xfrm_state_afinfo[family];
1904         if (unlikely(!afinfo))
1905                 read_unlock(&xfrm_state_afinfo_lock);
1906         return afinfo;
1907 }
1908
1909 static void xfrm_state_put_afinfo(struct xfrm_state_afinfo *afinfo)
1910         __releases(xfrm_state_afinfo_lock)
1911 {
1912         read_unlock(&xfrm_state_afinfo_lock);
1913 }
1914
1915 /* Temporarily located here until net/xfrm/xfrm_tunnel.c is created */
1916 void xfrm_state_delete_tunnel(struct xfrm_state *x)
1917 {
1918         if (x->tunnel) {
1919                 struct xfrm_state *t = x->tunnel;
1920
1921                 if (atomic_read(&t->tunnel_users) == 2)
1922                         xfrm_state_delete(t);
1923                 atomic_dec(&t->tunnel_users);
1924                 xfrm_state_put(t);
1925                 x->tunnel = NULL;
1926         }
1927 }
1928 EXPORT_SYMBOL(xfrm_state_delete_tunnel);
1929
1930 int xfrm_state_mtu(struct xfrm_state *x, int mtu)
1931 {
1932         int res;
1933
1934         spin_lock_bh(&x->lock);
1935         if (x->km.state == XFRM_STATE_VALID &&
1936             x->type && x->type->get_mtu)
1937                 res = x->type->get_mtu(x, mtu);
1938         else
1939                 res = mtu - x->props.header_len;
1940         spin_unlock_bh(&x->lock);
1941         return res;
1942 }
1943
1944 int xfrm_init_state(struct xfrm_state *x)
1945 {
1946         struct xfrm_state_afinfo *afinfo;
1947         int family = x->props.family;
1948         int err;
1949
1950         err = -EAFNOSUPPORT;
1951         afinfo = xfrm_state_get_afinfo(family);
1952         if (!afinfo)
1953                 goto error;
1954
1955         err = 0;
1956         if (afinfo->init_flags)
1957                 err = afinfo->init_flags(x);
1958
1959         xfrm_state_put_afinfo(afinfo);
1960
1961         if (err)
1962                 goto error;
1963
1964         err = -EPROTONOSUPPORT;
1965         x->inner_mode = xfrm_get_mode(x->props.mode, x->sel.family);
1966         if (x->inner_mode == NULL)
1967                 goto error;
1968
1969         if (!(x->inner_mode->flags & XFRM_MODE_FLAG_TUNNEL) &&
1970             family != x->sel.family)
1971                 goto error;
1972
1973         x->type = xfrm_get_type(x->id.proto, family);
1974         if (x->type == NULL)
1975                 goto error;
1976
1977         err = x->type->init_state(x);
1978         if (err)
1979                 goto error;
1980
1981         x->outer_mode = xfrm_get_mode(x->props.mode, family);
1982         if (x->outer_mode == NULL)
1983                 goto error;
1984
1985         x->km.state = XFRM_STATE_VALID;
1986
1987 error:
1988         return err;
1989 }
1990
1991 EXPORT_SYMBOL(xfrm_init_state);
1992
1993 void __init xfrm_state_init(void)
1994 {
1995         unsigned int sz;
1996
1997         sz = sizeof(struct hlist_head) * 8;
1998
1999         xfrm_state_bydst = xfrm_hash_alloc(sz);
2000         xfrm_state_bysrc = xfrm_hash_alloc(sz);
2001         xfrm_state_byspi = xfrm_hash_alloc(sz);
2002         if (!xfrm_state_bydst || !xfrm_state_bysrc || !xfrm_state_byspi)
2003                 panic("XFRM: Cannot allocate bydst/bysrc/byspi hashes.");
2004         xfrm_state_hmask = ((sz / sizeof(struct hlist_head)) - 1);
2005
2006         INIT_WORK(&xfrm_state_gc_work, xfrm_state_gc_task);
2007 }
2008
2009 #ifdef CONFIG_AUDITSYSCALL
2010 static void xfrm_audit_helper_sainfo(struct xfrm_state *x,
2011                                      struct audit_buffer *audit_buf)
2012 {
2013         struct xfrm_sec_ctx *ctx = x->security;
2014         u32 spi = ntohl(x->id.spi);
2015
2016         if (ctx)
2017                 audit_log_format(audit_buf, " sec_alg=%u sec_doi=%u sec_obj=%s",
2018                                  ctx->ctx_alg, ctx->ctx_doi, ctx->ctx_str);
2019
2020         switch(x->props.family) {
2021         case AF_INET:
2022                 audit_log_format(audit_buf,
2023                                  " src=" NIPQUAD_FMT " dst=" NIPQUAD_FMT,
2024                                  NIPQUAD(x->props.saddr.a4),
2025                                  NIPQUAD(x->id.daddr.a4));
2026                 break;
2027         case AF_INET6:
2028                 audit_log_format(audit_buf,
2029                                  " src=" NIP6_FMT " dst=" NIP6_FMT,
2030                                  NIP6(*(struct in6_addr *)x->props.saddr.a6),
2031                                  NIP6(*(struct in6_addr *)x->id.daddr.a6));
2032                 break;
2033         }
2034
2035         audit_log_format(audit_buf, " spi=%u(0x%x)", spi, spi);
2036 }
2037
2038 static void xfrm_audit_helper_pktinfo(struct sk_buff *skb, u16 family,
2039                                       struct audit_buffer *audit_buf)
2040 {
2041         struct iphdr *iph4;
2042         struct ipv6hdr *iph6;
2043
2044         switch (family) {
2045         case AF_INET:
2046                 iph4 = ip_hdr(skb);
2047                 audit_log_format(audit_buf,
2048                                  " src=" NIPQUAD_FMT " dst=" NIPQUAD_FMT,
2049                                  NIPQUAD(iph4->saddr),
2050                                  NIPQUAD(iph4->daddr));
2051                 break;
2052         case AF_INET6:
2053                 iph6 = ipv6_hdr(skb);
2054                 audit_log_format(audit_buf,
2055                                  " src=" NIP6_FMT " dst=" NIP6_FMT
2056                                  " flowlbl=0x%x%x%x",
2057                                  NIP6(iph6->saddr),
2058                                  NIP6(iph6->daddr),
2059                                  iph6->flow_lbl[0] & 0x0f,
2060                                  iph6->flow_lbl[1],
2061                                  iph6->flow_lbl[2]);
2062                 break;
2063         }
2064 }
2065
2066 void xfrm_audit_state_add(struct xfrm_state *x, int result,
2067                           u32 auid, u32 secid)
2068 {
2069         struct audit_buffer *audit_buf;
2070
2071         audit_buf = xfrm_audit_start("SAD-add");
2072         if (audit_buf == NULL)
2073                 return;
2074         xfrm_audit_helper_usrinfo(auid, secid, audit_buf);
2075         xfrm_audit_helper_sainfo(x, audit_buf);
2076         audit_log_format(audit_buf, " res=%u", result);
2077         audit_log_end(audit_buf);
2078 }
2079 EXPORT_SYMBOL_GPL(xfrm_audit_state_add);
2080
2081 void xfrm_audit_state_delete(struct xfrm_state *x, int result,
2082                              u32 auid, u32 secid)
2083 {
2084         struct audit_buffer *audit_buf;
2085
2086         audit_buf = xfrm_audit_start("SAD-delete");
2087         if (audit_buf == NULL)
2088                 return;
2089         xfrm_audit_helper_usrinfo(auid, secid, audit_buf);
2090         xfrm_audit_helper_sainfo(x, audit_buf);
2091         audit_log_format(audit_buf, " res=%u", result);
2092         audit_log_end(audit_buf);
2093 }
2094 EXPORT_SYMBOL_GPL(xfrm_audit_state_delete);
2095
2096 void xfrm_audit_state_replay_overflow(struct xfrm_state *x,
2097                                       struct sk_buff *skb)
2098 {
2099         struct audit_buffer *audit_buf;
2100         u32 spi;
2101
2102         audit_buf = xfrm_audit_start("SA-replay-overflow");
2103         if (audit_buf == NULL)
2104                 return;
2105         xfrm_audit_helper_pktinfo(skb, x->props.family, audit_buf);
2106         /* don't record the sequence number because it's inherent in this kind
2107          * of audit message */
2108         spi = ntohl(x->id.spi);
2109         audit_log_format(audit_buf, " spi=%u(0x%x)", spi, spi);
2110         audit_log_end(audit_buf);
2111 }
2112 EXPORT_SYMBOL_GPL(xfrm_audit_state_replay_overflow);
2113
2114 static void xfrm_audit_state_replay(struct xfrm_state *x,
2115                              struct sk_buff *skb, __be32 net_seq)
2116 {
2117         struct audit_buffer *audit_buf;
2118         u32 spi;
2119
2120         audit_buf = xfrm_audit_start("SA-replayed-pkt");
2121         if (audit_buf == NULL)
2122                 return;
2123         xfrm_audit_helper_pktinfo(skb, x->props.family, audit_buf);
2124         spi = ntohl(x->id.spi);
2125         audit_log_format(audit_buf, " spi=%u(0x%x) seqno=%u",
2126                          spi, spi, ntohl(net_seq));
2127         audit_log_end(audit_buf);
2128 }
2129
2130 void xfrm_audit_state_notfound_simple(struct sk_buff *skb, u16 family)
2131 {
2132         struct audit_buffer *audit_buf;
2133
2134         audit_buf = xfrm_audit_start("SA-notfound");
2135         if (audit_buf == NULL)
2136                 return;
2137         xfrm_audit_helper_pktinfo(skb, family, audit_buf);
2138         audit_log_end(audit_buf);
2139 }
2140 EXPORT_SYMBOL_GPL(xfrm_audit_state_notfound_simple);
2141
2142 void xfrm_audit_state_notfound(struct sk_buff *skb, u16 family,
2143                                __be32 net_spi, __be32 net_seq)
2144 {
2145         struct audit_buffer *audit_buf;
2146         u32 spi;
2147
2148         audit_buf = xfrm_audit_start("SA-notfound");
2149         if (audit_buf == NULL)
2150                 return;
2151         xfrm_audit_helper_pktinfo(skb, family, audit_buf);
2152         spi = ntohl(net_spi);
2153         audit_log_format(audit_buf, " spi=%u(0x%x) seqno=%u",
2154                          spi, spi, ntohl(net_seq));
2155         audit_log_end(audit_buf);
2156 }
2157 EXPORT_SYMBOL_GPL(xfrm_audit_state_notfound);
2158
2159 void xfrm_audit_state_icvfail(struct xfrm_state *x,
2160                               struct sk_buff *skb, u8 proto)
2161 {
2162         struct audit_buffer *audit_buf;
2163         __be32 net_spi;
2164         __be32 net_seq;
2165
2166         audit_buf = xfrm_audit_start("SA-icv-failure");
2167         if (audit_buf == NULL)
2168                 return;
2169         xfrm_audit_helper_pktinfo(skb, x->props.family, audit_buf);
2170         if (xfrm_parse_spi(skb, proto, &net_spi, &net_seq) == 0) {
2171                 u32 spi = ntohl(net_spi);
2172                 audit_log_format(audit_buf, " spi=%u(0x%x) seqno=%u",
2173                                  spi, spi, ntohl(net_seq));
2174         }
2175         audit_log_end(audit_buf);
2176 }
2177 EXPORT_SYMBOL_GPL(xfrm_audit_state_icvfail);
2178 #endif /* CONFIG_AUDITSYSCALL */