62c4f0a2d5
Signed-off-by: David Woodhouse <dwmw2@infradead.org>
274 lines
5.1 KiB
C
274 lines
5.1 KiB
C
#ifndef __NET_PKT_ACT_H
|
|
#define __NET_PKT_ACT_H
|
|
|
|
#include <asm/uaccess.h>
|
|
#include <asm/system.h>
|
|
#include <linux/bitops.h>
|
|
#include <linux/types.h>
|
|
#include <linux/kernel.h>
|
|
#include <linux/sched.h>
|
|
#include <linux/string.h>
|
|
#include <linux/mm.h>
|
|
#include <linux/socket.h>
|
|
#include <linux/sockios.h>
|
|
#include <linux/in.h>
|
|
#include <linux/errno.h>
|
|
#include <linux/interrupt.h>
|
|
#include <linux/skbuff.h>
|
|
#include <linux/rtnetlink.h>
|
|
#include <linux/module.h>
|
|
#include <linux/init.h>
|
|
#include <linux/proc_fs.h>
|
|
#include <net/sock.h>
|
|
#include <net/pkt_sched.h>
|
|
|
|
#define tca_st(val) (struct tcf_##val *)
|
|
#define PRIV(a,name) ( tca_st(name) (a)->priv)
|
|
|
|
#if 0 /* control */
|
|
#define DPRINTK(format,args...) printk(KERN_DEBUG format,##args)
|
|
#else
|
|
#define DPRINTK(format,args...)
|
|
#endif
|
|
|
|
#if 0 /* data */
|
|
#define D2PRINTK(format,args...) printk(KERN_DEBUG format,##args)
|
|
#else
|
|
#define D2PRINTK(format,args...)
|
|
#endif
|
|
|
|
static __inline__ unsigned
|
|
tcf_hash(u32 index)
|
|
{
|
|
return index & MY_TAB_MASK;
|
|
}
|
|
|
|
/* probably move this from being inline
|
|
* and put into act_generic
|
|
*/
|
|
static inline void
|
|
tcf_hash_destroy(struct tcf_st *p)
|
|
{
|
|
unsigned h = tcf_hash(p->index);
|
|
struct tcf_st **p1p;
|
|
|
|
for (p1p = &tcf_ht[h]; *p1p; p1p = &(*p1p)->next) {
|
|
if (*p1p == p) {
|
|
write_lock_bh(&tcf_t_lock);
|
|
*p1p = p->next;
|
|
write_unlock_bh(&tcf_t_lock);
|
|
#ifdef CONFIG_NET_ESTIMATOR
|
|
gen_kill_estimator(&p->bstats, &p->rate_est);
|
|
#endif
|
|
kfree(p);
|
|
return;
|
|
}
|
|
}
|
|
BUG_TRAP(0);
|
|
}
|
|
|
|
static inline int
|
|
tcf_hash_release(struct tcf_st *p, int bind )
|
|
{
|
|
int ret = 0;
|
|
if (p) {
|
|
if (bind) {
|
|
p->bindcnt--;
|
|
}
|
|
p->refcnt--;
|
|
if(p->bindcnt <=0 && p->refcnt <= 0) {
|
|
tcf_hash_destroy(p);
|
|
ret = 1;
|
|
}
|
|
}
|
|
return ret;
|
|
}
|
|
|
|
static __inline__ int
|
|
tcf_dump_walker(struct sk_buff *skb, struct netlink_callback *cb,
|
|
struct tc_action *a)
|
|
{
|
|
struct tcf_st *p;
|
|
int err =0, index = -1,i= 0, s_i = 0, n_i = 0;
|
|
struct rtattr *r ;
|
|
|
|
read_lock(&tcf_t_lock);
|
|
|
|
s_i = cb->args[0];
|
|
|
|
for (i = 0; i < MY_TAB_SIZE; i++) {
|
|
p = tcf_ht[tcf_hash(i)];
|
|
|
|
for (; p; p = p->next) {
|
|
index++;
|
|
if (index < s_i)
|
|
continue;
|
|
a->priv = p;
|
|
a->order = n_i;
|
|
r = (struct rtattr*) skb->tail;
|
|
RTA_PUT(skb, a->order, 0, NULL);
|
|
err = tcf_action_dump_1(skb, a, 0, 0);
|
|
if (0 > err) {
|
|
index--;
|
|
skb_trim(skb, (u8*)r - skb->data);
|
|
goto done;
|
|
}
|
|
r->rta_len = skb->tail - (u8*)r;
|
|
n_i++;
|
|
if (n_i >= TCA_ACT_MAX_PRIO) {
|
|
goto done;
|
|
}
|
|
}
|
|
}
|
|
done:
|
|
read_unlock(&tcf_t_lock);
|
|
if (n_i)
|
|
cb->args[0] += n_i;
|
|
return n_i;
|
|
|
|
rtattr_failure:
|
|
skb_trim(skb, (u8*)r - skb->data);
|
|
goto done;
|
|
}
|
|
|
|
static __inline__ int
|
|
tcf_del_walker(struct sk_buff *skb, struct tc_action *a)
|
|
{
|
|
struct tcf_st *p, *s_p;
|
|
struct rtattr *r ;
|
|
int i= 0, n_i = 0;
|
|
|
|
r = (struct rtattr*) skb->tail;
|
|
RTA_PUT(skb, a->order, 0, NULL);
|
|
RTA_PUT(skb, TCA_KIND, IFNAMSIZ, a->ops->kind);
|
|
for (i = 0; i < MY_TAB_SIZE; i++) {
|
|
p = tcf_ht[tcf_hash(i)];
|
|
|
|
while (p != NULL) {
|
|
s_p = p->next;
|
|
if (ACT_P_DELETED == tcf_hash_release(p, 0)) {
|
|
module_put(a->ops->owner);
|
|
}
|
|
n_i++;
|
|
p = s_p;
|
|
}
|
|
}
|
|
RTA_PUT(skb, TCA_FCNT, 4, &n_i);
|
|
r->rta_len = skb->tail - (u8*)r;
|
|
|
|
return n_i;
|
|
rtattr_failure:
|
|
skb_trim(skb, (u8*)r - skb->data);
|
|
return -EINVAL;
|
|
}
|
|
|
|
static __inline__ int
|
|
tcf_generic_walker(struct sk_buff *skb, struct netlink_callback *cb, int type,
|
|
struct tc_action *a)
|
|
{
|
|
if (type == RTM_DELACTION) {
|
|
return tcf_del_walker(skb,a);
|
|
} else if (type == RTM_GETACTION) {
|
|
return tcf_dump_walker(skb,cb,a);
|
|
} else {
|
|
printk("tcf_generic_walker: unknown action %d\n",type);
|
|
return -EINVAL;
|
|
}
|
|
}
|
|
|
|
static __inline__ struct tcf_st *
|
|
tcf_hash_lookup(u32 index)
|
|
{
|
|
struct tcf_st *p;
|
|
|
|
read_lock(&tcf_t_lock);
|
|
for (p = tcf_ht[tcf_hash(index)]; p; p = p->next) {
|
|
if (p->index == index)
|
|
break;
|
|
}
|
|
read_unlock(&tcf_t_lock);
|
|
return p;
|
|
}
|
|
|
|
static __inline__ u32
|
|
tcf_hash_new_index(void)
|
|
{
|
|
do {
|
|
if (++idx_gen == 0)
|
|
idx_gen = 1;
|
|
} while (tcf_hash_lookup(idx_gen));
|
|
|
|
return idx_gen;
|
|
}
|
|
|
|
|
|
static inline int
|
|
tcf_hash_search(struct tc_action *a, u32 index)
|
|
{
|
|
struct tcf_st *p = tcf_hash_lookup(index);
|
|
|
|
if (p != NULL) {
|
|
a->priv = p;
|
|
return 1;
|
|
}
|
|
return 0;
|
|
}
|
|
|
|
#ifdef CONFIG_NET_ACT_INIT
|
|
static inline struct tcf_st *
|
|
tcf_hash_check(u32 index, struct tc_action *a, int ovr, int bind)
|
|
{
|
|
struct tcf_st *p = NULL;
|
|
if (index && (p = tcf_hash_lookup(index)) != NULL) {
|
|
if (bind) {
|
|
p->bindcnt++;
|
|
p->refcnt++;
|
|
}
|
|
a->priv = p;
|
|
}
|
|
return p;
|
|
}
|
|
|
|
static inline struct tcf_st *
|
|
tcf_hash_create(u32 index, struct rtattr *est, struct tc_action *a, int size, int ovr, int bind)
|
|
{
|
|
struct tcf_st *p = NULL;
|
|
|
|
p = kmalloc(size, GFP_KERNEL);
|
|
if (p == NULL)
|
|
return p;
|
|
|
|
memset(p, 0, size);
|
|
p->refcnt = 1;
|
|
|
|
if (bind) {
|
|
p->bindcnt = 1;
|
|
}
|
|
|
|
spin_lock_init(&p->lock);
|
|
p->stats_lock = &p->lock;
|
|
p->index = index ? : tcf_hash_new_index();
|
|
p->tm.install = jiffies;
|
|
p->tm.lastuse = jiffies;
|
|
#ifdef CONFIG_NET_ESTIMATOR
|
|
if (est)
|
|
gen_new_estimator(&p->bstats, &p->rate_est, p->stats_lock, est);
|
|
#endif
|
|
a->priv = (void *) p;
|
|
return p;
|
|
}
|
|
|
|
static inline void tcf_hash_insert(struct tcf_st *p)
|
|
{
|
|
unsigned h = tcf_hash(p->index);
|
|
|
|
write_lock_bh(&tcf_t_lock);
|
|
p->next = tcf_ht[h];
|
|
tcf_ht[h] = p;
|
|
write_unlock_bh(&tcf_t_lock);
|
|
}
|
|
|
|
#endif
|
|
|
|
#endif
|