3 * SNMP MIB entries for the IP subsystem.
5 * Alan Cox <gw4pts@gw4pts.ampr.org>
7 * We don't chose to implement SNMP in the kernel (this would
8 * be silly as SNMP is a pain in the backside in places). We do
9 * however need to collect the MIB statistics and export them
10 * out of /proc (eventually)
12 * This program is free software; you can redistribute it and/or
13 * modify it under the terms of the GNU General Public License
14 * as published by the Free Software Foundation; either version
15 * 2 of the License, or (at your option) any later version.
17 * $Id: snmp.h,v 1.19 2001/06/14 13:40:46 davem Exp $
24 #include <linux/cache.h>
25 #include <linux/snmp.h>
28 * Mibs are stored in array of unsigned long.
32 * - list of entries for particular API (such as /proc/net/snmp)
40 #define SNMP_MIB_ITEM(_name,_entry) { \
45 #define SNMP_MIB_SENTINEL { \
51 * We use all unsigned longs. Linux will soon be so reliable that even
52 * these will rapidly get too small 8-). Seriously consider the IpInReceives
53 * count on the 20Gb/s + networks people expect in a few years time!
57 * The rule for padding:
58 * Best is power of two because then the right structure can be found by a
59 * simple shift. The structure should be always cache line aligned.
60 * gcc needs n=alignto(cachelinesize, popcnt(sizeof(bla_mib))) shift/add
61 * instructions to emulate multiply in case it is not power-of-two.
62 * Currently n is always <=3 for all sizes so simple cache line alignment
65 * The best solution would be a global CPU local area , especially on 64
66 * and 128byte cacheline machine it makes a *lot* of sense -AK
69 #define __SNMP_MIB_ALIGN__ ____cacheline_aligned
72 #define IPSTATS_MIB_MAX __IPSTATS_MIB_MAX
74 unsigned long mibs
[IPSTATS_MIB_MAX
];
78 #define ICMP_MIB_DUMMY __ICMP_MIB_MAX
79 #define ICMP_MIB_MAX (__ICMP_MIB_MAX + 1)
82 unsigned long mibs
[ICMP_MIB_MAX
];
85 #define ICMPMSG_MIB_MAX __ICMPMSG_MIB_MAX
87 unsigned long mibs
[ICMPMSG_MIB_MAX
];
90 /* ICMP6 (IPv6-ICMP) */
91 #define ICMP6_MIB_MAX __ICMP6_MIB_MAX
93 unsigned long mibs
[ICMP6_MIB_MAX
];
96 #define ICMP6MSG_MIB_MAX __ICMP6MSG_MIB_MAX
97 struct icmpv6msg_mib
{
98 unsigned long mibs
[ICMP6MSG_MIB_MAX
];
103 #define TCP_MIB_MAX __TCP_MIB_MAX
105 unsigned long mibs
[TCP_MIB_MAX
];
106 } __SNMP_MIB_ALIGN__
;
109 #define UDP_MIB_MAX __UDP_MIB_MAX
111 unsigned long mibs
[UDP_MIB_MAX
];
112 } __SNMP_MIB_ALIGN__
;
115 #define LINUX_MIB_MAX __LINUX_MIB_MAX
117 unsigned long mibs
[LINUX_MIB_MAX
];
122 * FIXME: On x86 and some other CPUs the split into user and softirq parts
123 * is not needed because addl $1,memory is atomic against interrupts (but
124 * atomic_inc would be overkill because of the lock cycles). Wants new
125 * nonlocked_atomic_inc() primitives -AK
127 #define DEFINE_SNMP_STAT(type, name) \
128 __typeof__(type) *name[2]
129 #define DECLARE_SNMP_STAT(type, name) \
130 extern __typeof__(type) *name[2]
132 #define SNMP_STAT_BHPTR(name) (name[0])
133 #define SNMP_STAT_USRPTR(name) (name[1])
135 #define SNMP_INC_STATS_BH(mib, field) \
136 (per_cpu_ptr(mib[0], raw_smp_processor_id())->mibs[field]++)
137 #define SNMP_INC_STATS_OFFSET_BH(mib, field, offset) \
138 (per_cpu_ptr(mib[0], raw_smp_processor_id())->mibs[field + (offset)]++)
139 #define SNMP_INC_STATS_USER(mib, field) \
140 (per_cpu_ptr(mib[1], raw_smp_processor_id())->mibs[field]++)
141 #define SNMP_INC_STATS(mib, field) \
142 (per_cpu_ptr(mib[!in_softirq()], raw_smp_processor_id())->mibs[field]++)
143 #define SNMP_DEC_STATS(mib, field) \
144 (per_cpu_ptr(mib[!in_softirq()], raw_smp_processor_id())->mibs[field]--)
145 #define SNMP_ADD_STATS_BH(mib, field, addend) \
146 (per_cpu_ptr(mib[0], raw_smp_processor_id())->mibs[field] += addend)
147 #define SNMP_ADD_STATS_USER(mib, field, addend) \
148 (per_cpu_ptr(mib[1], raw_smp_processor_id())->mibs[field] += addend)