blob: f3c7c24bec1ca99c89270611d87f9c264c3eb786 [file] [log] [blame]
Linus Torvalds1da177e2005-04-16 15:20:36 -07001/*
2 * Common code for low-level network console, dump, and debugger code
3 *
4 * Derived from netconsole, kgdb-over-ethernet, and netdump patches
5 */
6
7#ifndef _LINUX_NETPOLL_H
8#define _LINUX_NETPOLL_H
9
10#include <linux/netdevice.h>
11#include <linux/interrupt.h>
Matt Mackall53fb95d2005-08-11 19:27:43 -070012#include <linux/rcupdate.h>
Linus Torvalds1da177e2005-04-16 15:20:36 -070013#include <linux/list.h>
14
Cong Wangb7394d22013-01-07 20:52:39 +000015union inet_addr {
16 __u32 all[4];
17 __be32 ip;
18 __be32 ip6[4];
19 struct in_addr in;
20 struct in6_addr in6;
21};
22
Linus Torvalds1da177e2005-04-16 15:20:36 -070023struct netpoll {
24 struct net_device *dev;
Stephen Hemmingerbf6bce72006-10-26 15:46:56 -070025 char dev_name[IFNAMSIZ];
26 const char *name;
Linus Torvalds1da177e2005-04-16 15:20:36 -070027 void (*rx_hook)(struct netpoll *, int, char *, int);
Stephen Hemminger5de4a472006-10-26 15:46:55 -070028
Cong Wangb7394d22013-01-07 20:52:39 +000029 union inet_addr local_ip, remote_ip;
30 bool ipv6;
Linus Torvalds1da177e2005-04-16 15:20:36 -070031 u16 local_port, remote_port;
Stephen Hemminger09538642007-11-19 19:23:29 -080032 u8 remote_mac[ETH_ALEN];
Daniel Borkmann508e14b2010-01-12 14:27:30 +000033
34 struct list_head rx; /* rx_np list element */
Neil Horman2cde6ac2013-02-11 10:25:30 +000035 struct work_struct cleanup_work;
Jeff Moyer115c1d62005-06-22 22:05:31 -070036};
37
38struct netpoll_info {
Stephen Hemminger93ec2c72006-10-26 15:46:50 -070039 atomic_t refcnt;
Daniel Borkmann508e14b2010-01-12 14:27:30 +000040
Neil Hormanca99ca12013-02-05 08:05:43 +000041 unsigned long rx_flags;
Jeff Moyerfbeec2e2005-06-22 22:05:59 -070042 spinlock_t rx_lock;
Neil Hormanbd7c4b62013-04-30 05:35:05 +000043 struct semaphore dev_lock;
Daniel Borkmann508e14b2010-01-12 14:27:30 +000044 struct list_head rx_np; /* netpolls that registered an rx_hook */
45
Cong Wangb7394d22013-01-07 20:52:39 +000046 struct sk_buff_head neigh_tx; /* list of neigh requests to reply to */
Stephen Hemmingerb6cd27e2006-10-26 15:46:51 -070047 struct sk_buff_head txq;
Daniel Borkmann508e14b2010-01-12 14:27:30 +000048
David Howells6d5aefb2006-12-05 19:36:26 +000049 struct delayed_work tx_work;
WANG Cong0e34e932010-05-06 00:47:21 -070050
51 struct netpoll *netpoll;
Amerigo Wang38e6bc12012-08-10 01:24:38 +000052 struct rcu_head rcu;
Linus Torvalds1da177e2005-04-16 15:20:36 -070053};
54
Neil Hormanca99ca12013-02-05 08:05:43 +000055#ifdef CONFIG_NETPOLL
dingtianhongda6e3782013-05-27 19:53:31 +000056extern void netpoll_rx_disable(struct net_device *dev);
Neil Hormanca99ca12013-02-05 08:05:43 +000057extern void netpoll_rx_enable(struct net_device *dev);
58#else
dingtianhongda6e3782013-05-27 19:53:31 +000059static inline void netpoll_rx_disable(struct net_device *dev) { return; }
Neil Hormanca99ca12013-02-05 08:05:43 +000060static inline void netpoll_rx_enable(struct net_device *dev) { return; }
61#endif
62
Linus Torvalds1da177e2005-04-16 15:20:36 -070063void netpoll_send_udp(struct netpoll *np, const char *msg, int len);
Satyam Sharma0bcc1812007-08-10 15:35:05 -070064void netpoll_print_options(struct netpoll *np);
Linus Torvalds1da177e2005-04-16 15:20:36 -070065int netpoll_parse_options(struct netpoll *np, char *opt);
Amerigo Wang47be03a22012-08-10 01:24:37 +000066int __netpoll_setup(struct netpoll *np, struct net_device *ndev, gfp_t gfp);
Linus Torvalds1da177e2005-04-16 15:20:36 -070067int netpoll_setup(struct netpoll *np);
68int netpoll_trap(void);
69void netpoll_set_trap(int trap);
Herbert Xu8fdd95e2010-06-10 16:12:48 +000070void __netpoll_cleanup(struct netpoll *np);
Neil Horman2cde6ac2013-02-11 10:25:30 +000071void __netpoll_free_async(struct netpoll *np);
Linus Torvalds1da177e2005-04-16 15:20:36 -070072void netpoll_cleanup(struct netpoll *np);
Amerigo Wang57c5d462012-08-10 01:24:40 +000073int __netpoll_rx(struct sk_buff *skb, struct netpoll_info *npinfo);
Neil Hormanc2355e12010-10-13 16:01:49 +000074void netpoll_send_skb_on_dev(struct netpoll *np, struct sk_buff *skb,
75 struct net_device *dev);
76static inline void netpoll_send_skb(struct netpoll *np, struct sk_buff *skb)
77{
Amerigo Wang28996562012-08-10 01:24:42 +000078 unsigned long flags;
79 local_irq_save(flags);
Neil Hormanc2355e12010-10-13 16:01:49 +000080 netpoll_send_skb_on_dev(np, skb, np->dev);
Amerigo Wang28996562012-08-10 01:24:42 +000081 local_irq_restore(flags);
Neil Hormanc2355e12010-10-13 16:01:49 +000082}
83
Stephen Hemminger5de4a472006-10-26 15:46:55 -070084
Linus Torvalds1da177e2005-04-16 15:20:36 -070085
86#ifdef CONFIG_NETPOLL
Amerigo Wang77ab8a52012-08-10 01:24:46 +000087static inline bool netpoll_rx_on(struct sk_buff *skb)
Amerigo Wang91fe4a42012-08-10 01:24:41 +000088{
89 struct netpoll_info *npinfo = rcu_dereference_bh(skb->dev->npinfo);
90
91 return npinfo && (!list_empty(&npinfo->rx_np) || npinfo->rx_flags);
92}
93
David S. Millerffb27362010-05-06 01:20:10 -070094static inline bool netpoll_rx(struct sk_buff *skb)
Linus Torvalds1da177e2005-04-16 15:20:36 -070095{
Herbert Xude85d992010-06-10 16:12:44 +000096 struct netpoll_info *npinfo;
Jeff Moyerfbeec2e2005-06-22 22:05:59 -070097 unsigned long flags;
David S. Millerffb27362010-05-06 01:20:10 -070098 bool ret = false;
Jeff Moyer115c1d62005-06-22 22:05:31 -070099
Herbert Xuf0f9dea2010-09-17 16:55:03 -0700100 local_irq_save(flags);
Herbert Xude85d992010-06-10 16:12:44 +0000101
Amerigo Wang91fe4a42012-08-10 01:24:41 +0000102 if (!netpoll_rx_on(skb))
Herbert Xude85d992010-06-10 16:12:44 +0000103 goto out;
Jeff Moyer115c1d62005-06-22 22:05:31 -0700104
Amerigo Wang91fe4a42012-08-10 01:24:41 +0000105 npinfo = rcu_dereference_bh(skb->dev->npinfo);
Herbert Xuf0f9dea2010-09-17 16:55:03 -0700106 spin_lock(&npinfo->rx_lock);
David S. Millerd9452e92008-03-04 12:28:49 -0800107 /* check rx_flags again with the lock held */
Amerigo Wang57c5d462012-08-10 01:24:40 +0000108 if (npinfo->rx_flags && __netpoll_rx(skb, npinfo))
David S. Millerffb27362010-05-06 01:20:10 -0700109 ret = true;
Herbert Xuf0f9dea2010-09-17 16:55:03 -0700110 spin_unlock(&npinfo->rx_lock);
Jeff Moyerfbeec2e2005-06-22 22:05:59 -0700111
Herbert Xude85d992010-06-10 16:12:44 +0000112out:
Herbert Xuf0f9dea2010-09-17 16:55:03 -0700113 local_irq_restore(flags);
Jeff Moyerfbeec2e2005-06-22 22:05:59 -0700114 return ret;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700115}
116
Stephen Hemmingerbea33482007-10-03 16:41:36 -0700117static inline int netpoll_receive_skb(struct sk_buff *skb)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700118{
Stephen Hemmingerbea33482007-10-03 16:41:36 -0700119 if (!list_empty(&skb->dev->napi_list))
120 return netpoll_rx(skb);
121 return 0;
122}
123
124static inline void *netpoll_poll_lock(struct napi_struct *napi)
125{
126 struct net_device *dev = napi->dev;
127
Stephen Hemmingerbea33482007-10-03 16:41:36 -0700128 if (dev && dev->npinfo) {
129 spin_lock(&napi->poll_lock);
130 napi->poll_owner = smp_processor_id();
131 return napi;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700132 }
Matt Mackall53fb95d2005-08-11 19:27:43 -0700133 return NULL;
Linus Torvalds1da177e2005-04-16 15:20:36 -0700134}
135
Matt Mackall53fb95d2005-08-11 19:27:43 -0700136static inline void netpoll_poll_unlock(void *have)
Linus Torvalds1da177e2005-04-16 15:20:36 -0700137{
Stephen Hemmingerbea33482007-10-03 16:41:36 -0700138 struct napi_struct *napi = have;
Matt Mackall53fb95d2005-08-11 19:27:43 -0700139
Stephen Hemmingerbea33482007-10-03 16:41:36 -0700140 if (napi) {
141 napi->poll_owner = -1;
142 spin_unlock(&napi->poll_lock);
Linus Torvalds1da177e2005-04-16 15:20:36 -0700143 }
144}
145
Amerigo Wang77ab8a52012-08-10 01:24:46 +0000146static inline bool netpoll_tx_running(struct net_device *dev)
Herbert Xuc18370f2010-06-10 16:12:49 +0000147{
148 return irqs_disabled();
149}
150
Linus Torvalds1da177e2005-04-16 15:20:36 -0700151#else
John W. Linville969a6e52010-08-10 16:24:41 -0700152static inline bool netpoll_rx(struct sk_buff *skb)
Stephen Hemmingerbea33482007-10-03 16:41:36 -0700153{
Amerigo Wang77ab8a52012-08-10 01:24:46 +0000154 return false;
Stephen Hemmingerbea33482007-10-03 16:41:36 -0700155}
Amerigo Wang77ab8a52012-08-10 01:24:46 +0000156static inline bool netpoll_rx_on(struct sk_buff *skb)
Herbert Xud1c76af2009-03-16 10:50:02 -0700157{
Amerigo Wang77ab8a52012-08-10 01:24:46 +0000158 return false;
Herbert Xud1c76af2009-03-16 10:50:02 -0700159}
Stephen Hemmingerbea33482007-10-03 16:41:36 -0700160static inline int netpoll_receive_skb(struct sk_buff *skb)
161{
162 return 0;
163}
164static inline void *netpoll_poll_lock(struct napi_struct *napi)
165{
166 return NULL;
167}
168static inline void netpoll_poll_unlock(void *have)
169{
170}
171static inline void netpoll_netdev_init(struct net_device *dev)
172{
173}
Amerigo Wang77ab8a52012-08-10 01:24:46 +0000174static inline bool netpoll_tx_running(struct net_device *dev)
Herbert Xuc18370f2010-06-10 16:12:49 +0000175{
Amerigo Wang77ab8a52012-08-10 01:24:46 +0000176 return false;
Herbert Xuc18370f2010-06-10 16:12:49 +0000177}
Linus Torvalds1da177e2005-04-16 15:20:36 -0700178#endif
179
180#endif