blob: 5ccaa4be7d9646765d1ab59803756a3dcbce91fc [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001#ifndef __NET_PKT_CLS_H
2#define __NET_PKT_CLS_H
3
4#include <linux/pkt_cls.h>
5#include <net/sch_generic.h>
6#include <net/act_api.h>
7
8/* Basic packet classifier frontend definitions. */
9
Eric Dumazetfd2c3ef2009-11-03 03:26:03 +000010struct tcf_walker {
Linus Torvalds1da177e2005-04-16 15:20:36 -070011 int stop;
12 int skip;
13 int count;
14 int (*fn)(struct tcf_proto *, unsigned long node, struct tcf_walker *);
15};
16
Joe Perches5c152572013-07-30 22:47:13 -070017int register_tcf_proto_ops(struct tcf_proto_ops *ops);
18int unregister_tcf_proto_ops(struct tcf_proto_ops *ops);
Linus Torvalds1da177e2005-04-16 15:20:36 -070019
20static inline unsigned long
21__cls_set_class(unsigned long *clp, unsigned long cl)
22{
WANG Conga0efb802014-09-30 16:07:24 -070023 return xchg(clp, cl);
Linus Torvalds1da177e2005-04-16 15:20:36 -070024}
25
26static inline unsigned long
27cls_set_class(struct tcf_proto *tp, unsigned long *clp,
28 unsigned long cl)
29{
30 unsigned long old_cl;
31
32 tcf_tree_lock(tp);
33 old_cl = __cls_set_class(clp, cl);
34 tcf_tree_unlock(tp);
35
36 return old_cl;
37}
38
39static inline void
40tcf_bind_filter(struct tcf_proto *tp, struct tcf_result *r, unsigned long base)
41{
42 unsigned long cl;
43
44 cl = tp->q->ops->cl_ops->bind_tcf(tp->q, base, r->classid);
45 cl = cls_set_class(tp, &r->class, cl);
46 if (cl)
47 tp->q->ops->cl_ops->unbind_tcf(tp->q, cl);
48}
49
50static inline void
51tcf_unbind_filter(struct tcf_proto *tp, struct tcf_result *r)
52{
53 unsigned long cl;
54
55 if ((cl = __cls_set_class(&r->class, 0)) != 0)
56 tp->q->ops->cl_ops->unbind_tcf(tp->q, cl);
57}
58
Eric Dumazetfd2c3ef2009-11-03 03:26:03 +000059struct tcf_exts {
Linus Torvalds1da177e2005-04-16 15:20:36 -070060#ifdef CONFIG_NET_CLS_ACT
WANG Cong33be6272013-12-15 20:15:05 -080061 __u32 type; /* for backward compat(TCA_OLD_COMPAT) */
WANG Cong22dc13c2016-08-13 22:35:00 -070062 int nr_actions;
63 struct tc_action **actions;
Linus Torvalds1da177e2005-04-16 15:20:36 -070064#endif
WANG Cong5da57f42013-12-15 20:15:07 -080065 /* Map to export classifier specific extension TLV types to the
66 * generic extensions API. Unsupported extensions must be set to 0.
67 */
Linus Torvalds1da177e2005-04-16 15:20:36 -070068 int action;
69 int police;
70};
71
WANG Congb9a24bb2016-08-19 12:36:54 -070072static inline int tcf_exts_init(struct tcf_exts *exts, int action, int police)
WANG Cong33be6272013-12-15 20:15:05 -080073{
74#ifdef CONFIG_NET_CLS_ACT
WANG Cong5da57f42013-12-15 20:15:07 -080075 exts->type = 0;
WANG Cong22dc13c2016-08-13 22:35:00 -070076 exts->nr_actions = 0;
77 exts->actions = kcalloc(TCA_ACT_MAX_PRIO, sizeof(struct tc_action *),
78 GFP_KERNEL);
WANG Congb9a24bb2016-08-19 12:36:54 -070079 if (!exts->actions)
80 return -ENOMEM;
WANG Cong33be6272013-12-15 20:15:05 -080081#endif
WANG Cong5da57f42013-12-15 20:15:07 -080082 exts->action = action;
83 exts->police = police;
WANG Congb9a24bb2016-08-19 12:36:54 -070084 return 0;
WANG Cong33be6272013-12-15 20:15:05 -080085}
86
Linus Torvalds1da177e2005-04-16 15:20:36 -070087/**
88 * tcf_exts_is_predicative - check if a predicative extension is present
89 * @exts: tc filter extensions handle
90 *
91 * Returns 1 if a predicative extension is present, i.e. an extension which
92 * might cause further actions and thus overrule the regular tcf_result.
93 */
94static inline int
95tcf_exts_is_predicative(struct tcf_exts *exts)
96{
97#ifdef CONFIG_NET_CLS_ACT
WANG Cong22dc13c2016-08-13 22:35:00 -070098 return exts->nr_actions;
Linus Torvalds1da177e2005-04-16 15:20:36 -070099#else
100 return 0;
101#endif
102}
103
104/**
105 * tcf_exts_is_available - check if at least one extension is present
106 * @exts: tc filter extensions handle
107 *
108 * Returns 1 if at least one extension is present.
109 */
110static inline int
111tcf_exts_is_available(struct tcf_exts *exts)
112{
113 /* All non-predicative extensions must be added here. */
114 return tcf_exts_is_predicative(exts);
115}
116
WANG Cong22dc13c2016-08-13 22:35:00 -0700117static inline void tcf_exts_to_list(const struct tcf_exts *exts,
118 struct list_head *actions)
119{
120#ifdef CONFIG_NET_CLS_ACT
121 int i;
122
123 for (i = 0; i < exts->nr_actions; i++) {
124 struct tc_action *a = exts->actions[i];
125
126 list_add(&a->list, actions);
127 }
128#endif
129}
130
Linus Torvalds1da177e2005-04-16 15:20:36 -0700131/**
132 * tcf_exts_exec - execute tc filter extensions
133 * @skb: socket buffer
134 * @exts: tc filter extensions handle
135 * @res: desired result
136 *
137 * Executes all configured extensions. Returns 0 on a normal execution,
138 * a negative number if the filter must be considered unmatched or
139 * a positive action code (TC_ACT_*) which must be returned to the
140 * underlying layer.
141 */
142static inline int
143tcf_exts_exec(struct sk_buff *skb, struct tcf_exts *exts,
144 struct tcf_result *res)
145{
146#ifdef CONFIG_NET_CLS_ACT
WANG Cong22dc13c2016-08-13 22:35:00 -0700147 if (exts->nr_actions)
148 return tcf_action_exec(skb, exts->actions, exts->nr_actions,
149 res);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700150#endif
Linus Torvalds1da177e2005-04-16 15:20:36 -0700151 return 0;
152}
153
WANG Cong2734437e2016-08-13 22:34:59 -0700154#ifdef CONFIG_NET_CLS_ACT
155
WANG Cong22dc13c2016-08-13 22:35:00 -0700156#define tc_no_actions(_exts) ((_exts)->nr_actions == 0)
157#define tc_single_action(_exts) ((_exts)->nr_actions == 1)
WANG Cong2734437e2016-08-13 22:34:59 -0700158
159#else /* CONFIG_NET_CLS_ACT */
160
161#define tc_no_actions(_exts) true
WANG Cong2734437e2016-08-13 22:34:59 -0700162#define tc_single_action(_exts) false
163
164#endif /* CONFIG_NET_CLS_ACT */
165
Joe Perches5c152572013-07-30 22:47:13 -0700166int tcf_exts_validate(struct net *net, struct tcf_proto *tp,
167 struct nlattr **tb, struct nlattr *rate_tlv,
Cong Wang2f7ef2f2014-04-25 13:54:06 -0700168 struct tcf_exts *exts, bool ovr);
WANG Cong18d02642014-09-25 10:26:37 -0700169void tcf_exts_destroy(struct tcf_exts *exts);
Joe Perches5c152572013-07-30 22:47:13 -0700170void tcf_exts_change(struct tcf_proto *tp, struct tcf_exts *dst,
171 struct tcf_exts *src);
WANG Cong5da57f42013-12-15 20:15:07 -0800172int tcf_exts_dump(struct sk_buff *skb, struct tcf_exts *exts);
173int tcf_exts_dump_stats(struct sk_buff *skb, struct tcf_exts *exts);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700174
175/**
176 * struct tcf_pkt_info - packet information
177 */
Eric Dumazetfd2c3ef2009-11-03 03:26:03 +0000178struct tcf_pkt_info {
Linus Torvalds1da177e2005-04-16 15:20:36 -0700179 unsigned char * ptr;
180 int nexthdr;
181};
182
183#ifdef CONFIG_NET_EMATCH
184
185struct tcf_ematch_ops;
186
187/**
188 * struct tcf_ematch - extended match (ematch)
189 *
190 * @matchid: identifier to allow userspace to reidentify a match
191 * @flags: flags specifying attributes and the relation to other matches
192 * @ops: the operations lookup table of the corresponding ematch module
193 * @datalen: length of the ematch specific configuration data
194 * @data: ematch specific data
195 */
Eric Dumazetfd2c3ef2009-11-03 03:26:03 +0000196struct tcf_ematch {
Linus Torvalds1da177e2005-04-16 15:20:36 -0700197 struct tcf_ematch_ops * ops;
198 unsigned long data;
199 unsigned int datalen;
200 u16 matchid;
201 u16 flags;
John Fastabend82a470f2014-10-05 21:27:53 -0700202 struct net *net;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700203};
204
205static inline int tcf_em_is_container(struct tcf_ematch *em)
206{
207 return !em->ops;
208}
209
210static inline int tcf_em_is_simple(struct tcf_ematch *em)
211{
212 return em->flags & TCF_EM_SIMPLE;
213}
214
215static inline int tcf_em_is_inverted(struct tcf_ematch *em)
216{
217 return em->flags & TCF_EM_INVERT;
218}
219
220static inline int tcf_em_last_match(struct tcf_ematch *em)
221{
222 return (em->flags & TCF_EM_REL_MASK) == TCF_EM_REL_END;
223}
224
225static inline int tcf_em_early_end(struct tcf_ematch *em, int result)
226{
227 if (tcf_em_last_match(em))
228 return 1;
229
230 if (result == 0 && em->flags & TCF_EM_REL_AND)
231 return 1;
232
233 if (result != 0 && em->flags & TCF_EM_REL_OR)
234 return 1;
235
236 return 0;
237}
238
239/**
240 * struct tcf_ematch_tree - ematch tree handle
241 *
242 * @hdr: ematch tree header supplied by userspace
243 * @matches: array of ematches
244 */
Eric Dumazetfd2c3ef2009-11-03 03:26:03 +0000245struct tcf_ematch_tree {
Linus Torvalds1da177e2005-04-16 15:20:36 -0700246 struct tcf_ematch_tree_hdr hdr;
247 struct tcf_ematch * matches;
248
249};
250
251/**
252 * struct tcf_ematch_ops - ematch module operations
253 *
254 * @kind: identifier (kind) of this ematch module
255 * @datalen: length of expected configuration data (optional)
256 * @change: called during validation (optional)
257 * @match: called during ematch tree evaluation, must return 1/0
258 * @destroy: called during destroyage (optional)
259 * @dump: called during dumping process (optional)
260 * @owner: owner, must be set to THIS_MODULE
261 * @link: link to previous/next ematch module (internal use)
262 */
Eric Dumazetfd2c3ef2009-11-03 03:26:03 +0000263struct tcf_ematch_ops {
Linus Torvalds1da177e2005-04-16 15:20:36 -0700264 int kind;
265 int datalen;
John Fastabend82a470f2014-10-05 21:27:53 -0700266 int (*change)(struct net *net, void *,
Linus Torvalds1da177e2005-04-16 15:20:36 -0700267 int, struct tcf_ematch *);
268 int (*match)(struct sk_buff *, struct tcf_ematch *,
269 struct tcf_pkt_info *);
John Fastabend82a470f2014-10-05 21:27:53 -0700270 void (*destroy)(struct tcf_ematch *);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700271 int (*dump)(struct sk_buff *, struct tcf_ematch *);
272 struct module *owner;
273 struct list_head link;
274};
275
Joe Perches5c152572013-07-30 22:47:13 -0700276int tcf_em_register(struct tcf_ematch_ops *);
277void tcf_em_unregister(struct tcf_ematch_ops *);
278int tcf_em_tree_validate(struct tcf_proto *, struct nlattr *,
279 struct tcf_ematch_tree *);
John Fastabend82a470f2014-10-05 21:27:53 -0700280void tcf_em_tree_destroy(struct tcf_ematch_tree *);
Joe Perches5c152572013-07-30 22:47:13 -0700281int tcf_em_tree_dump(struct sk_buff *, struct tcf_ematch_tree *, int);
282int __tcf_em_tree_match(struct sk_buff *, struct tcf_ematch_tree *,
283 struct tcf_pkt_info *);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700284
285/**
286 * tcf_em_tree_change - replace ematch tree of a running classifier
287 *
288 * @tp: classifier kind handle
289 * @dst: destination ematch tree variable
290 * @src: source ematch tree (temporary tree from tcf_em_tree_validate)
291 *
292 * This functions replaces the ematch tree in @dst with the ematch
293 * tree in @src. The classifier in charge of the ematch tree may be
294 * running.
295 */
296static inline void tcf_em_tree_change(struct tcf_proto *tp,
297 struct tcf_ematch_tree *dst,
298 struct tcf_ematch_tree *src)
299{
300 tcf_tree_lock(tp);
301 memcpy(dst, src, sizeof(*dst));
302 tcf_tree_unlock(tp);
303}
304
305/**
306 * tcf_em_tree_match - evaulate an ematch tree
307 *
308 * @skb: socket buffer of the packet in question
309 * @tree: ematch tree to be used for evaluation
310 * @info: packet information examined by classifier
311 *
312 * This function matches @skb against the ematch tree in @tree by going
313 * through all ematches respecting their logic relations returning
314 * as soon as the result is obvious.
315 *
316 * Returns 1 if the ematch tree as-one matches, no ematches are configured
317 * or ematch is not enabled in the kernel, otherwise 0 is returned.
318 */
319static inline int tcf_em_tree_match(struct sk_buff *skb,
320 struct tcf_ematch_tree *tree,
321 struct tcf_pkt_info *info)
322{
323 if (tree->hdr.nmatches)
324 return __tcf_em_tree_match(skb, tree, info);
325 else
326 return 1;
327}
328
Patrick McHardydb3d99c2007-07-11 19:46:26 -0700329#define MODULE_ALIAS_TCF_EMATCH(kind) MODULE_ALIAS("ematch-kind-" __stringify(kind))
330
Linus Torvalds1da177e2005-04-16 15:20:36 -0700331#else /* CONFIG_NET_EMATCH */
332
Eric Dumazetfd2c3ef2009-11-03 03:26:03 +0000333struct tcf_ematch_tree {
Linus Torvalds1da177e2005-04-16 15:20:36 -0700334};
335
336#define tcf_em_tree_validate(tp, tb, t) ((void)(t), 0)
John Fastabend82a470f2014-10-05 21:27:53 -0700337#define tcf_em_tree_destroy(t) do { (void)(t); } while(0)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700338#define tcf_em_tree_dump(skb, t, tlv) (0)
339#define tcf_em_tree_change(tp, dst, src) do { } while(0)
340#define tcf_em_tree_match(skb, t, info) ((void)(info), 1)
341
342#endif /* CONFIG_NET_EMATCH */
343
344static inline unsigned char * tcf_get_base_ptr(struct sk_buff *skb, int layer)
345{
346 switch (layer) {
347 case TCF_LAYER_LINK:
348 return skb->data;
349 case TCF_LAYER_NETWORK:
Arnaldo Carvalho de Melod56f90a2007-04-10 20:50:43 -0700350 return skb_network_header(skb);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700351 case TCF_LAYER_TRANSPORT:
Arnaldo Carvalho de Melo9c702202007-04-25 18:04:18 -0700352 return skb_transport_header(skb);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700353 }
354
355 return NULL;
356}
357
Arnaldo Carvalho de Meloeddc9ec2007-04-20 22:47:35 -0700358static inline int tcf_valid_offset(const struct sk_buff *skb,
359 const unsigned char *ptr, const int len)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700360{
David S. Millerda521b22010-12-21 12:43:16 -0800361 return likely((ptr + len) <= skb_tail_pointer(skb) &&
362 ptr >= skb->head &&
363 (ptr <= (ptr + len)));
Linus Torvalds1da177e2005-04-16 15:20:36 -0700364}
365
366#ifdef CONFIG_NET_CLS_IND
Denis V. Lunev0eeb8ff2007-12-04 01:15:45 -0800367#include <net/net_namespace.h>
368
Linus Torvalds1da177e2005-04-16 15:20:36 -0700369static inline int
WANG Cong2519a602014-01-09 16:14:02 -0800370tcf_change_indev(struct net *net, struct nlattr *indev_tlv)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700371{
WANG Cong2519a602014-01-09 16:14:02 -0800372 char indev[IFNAMSIZ];
Patrick McHardyc01003c2007-03-29 11:46:52 -0700373 struct net_device *dev;
374
WANG Cong2519a602014-01-09 16:14:02 -0800375 if (nla_strlcpy(indev, indev_tlv, IFNAMSIZ) >= IFNAMSIZ)
376 return -EINVAL;
377 dev = __dev_get_by_name(net, indev);
378 if (!dev)
379 return -ENODEV;
380 return dev->ifindex;
381}
Linus Torvalds1da177e2005-04-16 15:20:36 -0700382
WANG Cong2519a602014-01-09 16:14:02 -0800383static inline bool
384tcf_match_indev(struct sk_buff *skb, int ifindex)
385{
386 if (!ifindex)
387 return true;
388 if (!skb->skb_iif)
389 return false;
390 return ifindex == skb->skb_iif;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700391}
392#endif /* CONFIG_NET_CLS_IND */
393
John Fastabenda1b7c5f2016-02-16 21:17:09 -0800394struct tc_cls_u32_knode {
395 struct tcf_exts *exts;
John Fastabende0148602016-02-17 14:59:30 -0800396 struct tc_u32_sel *sel;
John Fastabenda1b7c5f2016-02-16 21:17:09 -0800397 u32 handle;
398 u32 val;
399 u32 mask;
400 u32 link_handle;
John Fastabende0148602016-02-17 14:59:30 -0800401 u8 fshift;
John Fastabenda1b7c5f2016-02-16 21:17:09 -0800402};
403
404struct tc_cls_u32_hnode {
405 u32 handle;
406 u32 prio;
407 unsigned int divisor;
408};
409
410enum tc_clsu32_command {
411 TC_CLSU32_NEW_KNODE,
412 TC_CLSU32_REPLACE_KNODE,
413 TC_CLSU32_DELETE_KNODE,
414 TC_CLSU32_NEW_HNODE,
415 TC_CLSU32_REPLACE_HNODE,
416 TC_CLSU32_DELETE_HNODE,
417};
418
419struct tc_cls_u32_offload {
420 /* knode values */
421 enum tc_clsu32_command command;
422 union {
423 struct tc_cls_u32_knode knode;
424 struct tc_cls_u32_hnode hnode;
425 };
426};
427
Daniel Borkmann92c075d2016-06-06 22:50:39 +0200428static inline bool tc_should_offload(const struct net_device *dev,
429 const struct tcf_proto *tp, u32 flags)
John Fastabend6843e7a2016-02-26 07:53:49 -0800430{
Daniel Borkmann92c075d2016-06-06 22:50:39 +0200431 const struct Qdisc *sch = tp->q;
432 const struct Qdisc_class_ops *cops = sch->ops->cl_ops;
433
John Fastabend2b6ab0d2016-02-26 07:54:13 -0800434 if (!(dev->features & NETIF_F_HW_TC))
435 return false;
John Fastabend9e8ce792016-02-26 07:54:39 -0800436 if (flags & TCA_CLS_FLAGS_SKIP_HW)
437 return false;
John Fastabend9e8ce792016-02-26 07:54:39 -0800438 if (!dev->netdev_ops->ndo_setup_tc)
439 return false;
Daniel Borkmann92c075d2016-06-06 22:50:39 +0200440 if (cops && cops->tcf_cl_offload)
441 return cops->tcf_cl_offload(tp->classid);
John Fastabend9e8ce792016-02-26 07:54:39 -0800442
443 return true;
John Fastabend6843e7a2016-02-26 07:53:49 -0800444}
445
Samudrala, Sridhard34e3e12016-05-12 17:08:23 -0700446static inline bool tc_skip_sw(u32 flags)
447{
448 return (flags & TCA_CLS_FLAGS_SKIP_SW) ? true : false;
449}
450
451/* SKIP_HW and SKIP_SW are mutually exclusive flags. */
452static inline bool tc_flags_valid(u32 flags)
453{
454 if (flags & ~(TCA_CLS_FLAGS_SKIP_HW | TCA_CLS_FLAGS_SKIP_SW))
455 return false;
456
457 if (!(flags ^ (TCA_CLS_FLAGS_SKIP_HW | TCA_CLS_FLAGS_SKIP_SW)))
458 return false;
459
460 return true;
461}
462
Amir Vadai5b33f482016-03-08 12:42:29 +0200463enum tc_fl_command {
464 TC_CLSFLOWER_REPLACE,
465 TC_CLSFLOWER_DESTROY,
Amir Vadai10cbc682016-05-13 12:55:37 +0000466 TC_CLSFLOWER_STATS,
Amir Vadai5b33f482016-03-08 12:42:29 +0200467};
468
469struct tc_cls_flower_offload {
470 enum tc_fl_command command;
Amir Vadai8208d212016-03-11 11:08:45 +0200471 unsigned long cookie;
Amir Vadai5b33f482016-03-08 12:42:29 +0200472 struct flow_dissector *dissector;
473 struct fl_flow_key *mask;
474 struct fl_flow_key *key;
475 struct tcf_exts *exts;
476};
477
Yotam Gigib87f7932016-07-21 12:03:12 +0200478enum tc_matchall_command {
479 TC_CLSMATCHALL_REPLACE,
480 TC_CLSMATCHALL_DESTROY,
481};
482
483struct tc_cls_matchall_offload {
484 enum tc_matchall_command command;
485 struct tcf_exts *exts;
486 unsigned long cookie;
487};
488
Jakub Kicinski332ae8e2016-09-21 11:43:53 +0100489enum tc_clsbpf_command {
490 TC_CLSBPF_ADD,
491 TC_CLSBPF_REPLACE,
492 TC_CLSBPF_DESTROY,
Jakub Kicinski68d64062016-09-21 11:44:02 +0100493 TC_CLSBPF_STATS,
Jakub Kicinski332ae8e2016-09-21 11:43:53 +0100494};
495
496struct tc_cls_bpf_offload {
497 enum tc_clsbpf_command command;
498 struct tcf_exts *exts;
499 struct bpf_prog *prog;
500 const char *name;
501 bool exts_integrated;
Jakub Kicinski0d01d452016-09-21 11:43:54 +0100502 u32 gen_flags;
Jakub Kicinski332ae8e2016-09-21 11:43:53 +0100503};
504
Linus Torvalds1da177e2005-04-16 15:20:36 -0700505#endif