Loading...
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 32 33 34 35 36 37 38 39 40 41 42 43 44 45 46 47 48 49 50 51 52 53 54 55 56 57 58 59 60 61 62 63 64 65 66 67 68 69 70 71 72 73 74 75 76 77 78 79 80 81 82 83 84 85 86 87 88 89 90 91 92 93 94 95 96 97 98 99 100 101 102 103 104 105 106 107 108 109 110 111 112 113 114 115 116 117 118 119 120 121 122 123 124 125 126 127 128 129 130 131 132 133 134 135 136 137 138 139 140 141 142 143 144 145 146 147 148 149 150 151 152 153 154 155 156 157 | /*
* net/dst.h Protocol independent destination cache definitions.
*
* Authors: Alexey Kuznetsov, <kuznet@ms2.inr.ac.ru>
*
*/
#ifndef _NET_DST_H
#define _NET_DST_H
#include <linux/config.h>
#include <net/neighbour.h>
/*
* 0 - no debugging messages
* 1 - rare events and bugs (default)
* 2 - trace mode.
*/
#ifdef NO_ANK_FIX
#define RT_CACHE_DEBUG 0
#else
#define RT_CACHE_DEBUG 1
#endif
#define DST_GC_MIN (1*HZ)
#define DST_GC_INC (5*HZ)
#define DST_GC_MAX (120*HZ)
struct sk_buff;
struct dst_entry
{
struct dst_entry *next;
atomic_t refcnt; /* tree/hash references */
atomic_t use; /* client references */
struct device *dev;
int obsolete;
unsigned long lastuse;
unsigned mxlock;
unsigned window;
unsigned pmtu;
unsigned rtt;
unsigned long rate_last; /* rate limiting for ICMP */
unsigned long rate_tokens;
int error;
struct neighbour *neighbour;
struct hh_cache *hh;
int (*input)(struct sk_buff*);
int (*output)(struct sk_buff*);
#ifdef CONFIG_NET_CLS_ROUTE
__u32 tclassid;
#endif
struct dst_ops *ops;
char info[0];
};
struct dst_ops
{
unsigned short family;
unsigned short protocol;
unsigned gc_thresh;
int (*gc)(void);
struct dst_entry * (*check)(struct dst_entry *, __u32 cookie);
struct dst_entry * (*reroute)(struct dst_entry *,
struct sk_buff *);
void (*destroy)(struct dst_entry *);
struct dst_entry * (*negative_advice)(struct dst_entry *);
void (*link_failure)(struct sk_buff *);
atomic_t entries;
};
#ifdef __KERNEL__
extern struct dst_entry * dst_garbage_list;
extern atomic_t dst_total;
extern __inline__
struct dst_entry * dst_clone(struct dst_entry * dst)
{
if (dst)
atomic_inc(&dst->use);
return dst;
}
extern __inline__
void dst_release(struct dst_entry * dst)
{
if (dst)
atomic_dec(&dst->use);
}
extern __inline__
struct dst_entry * dst_check(struct dst_entry ** dst_p, u32 cookie)
{
struct dst_entry * dst = *dst_p;
if (dst && dst->obsolete)
dst = dst->ops->check(dst, cookie);
return (*dst_p = dst);
}
extern __inline__
struct dst_entry * dst_reroute(struct dst_entry ** dst_p, struct sk_buff *skb)
{
struct dst_entry * dst = *dst_p;
if (dst && dst->obsolete)
dst = dst->ops->reroute(dst, skb);
return (*dst_p = dst);
}
extern void * dst_alloc(int size, struct dst_ops * ops);
extern void __dst_free(struct dst_entry * dst);
extern void dst_destroy(struct dst_entry * dst);
extern __inline__
void dst_free(struct dst_entry * dst)
{
if (dst->obsolete > 1)
return;
if (!atomic_read(&dst->use)) {
dst_destroy(dst);
return;
}
__dst_free(dst);
}
extern __inline__ void dst_confirm(struct dst_entry *dst)
{
if (dst)
neigh_confirm(dst->neighbour);
}
extern __inline__ void dst_negative_advice(struct dst_entry **dst_p)
{
struct dst_entry * dst = *dst_p;
if (dst && dst->ops->negative_advice)
*dst_p = dst->ops->negative_advice(dst);
}
extern __inline__ void dst_link_failure(struct sk_buff *skb)
{
struct dst_entry * dst = skb->dst;
if (dst && dst->ops && dst->ops->link_failure)
dst->ops->link_failure(skb);
}
#endif
#endif /* _NET_DST_H */
|