mirror of https://gitee.com/openkylin/linux.git
netlink: add flow control for memory mapped I/O
Add flow control for memory mapped RX. Since user-space usually doesn't invoke recvmsg() when using memory mapped I/O, flow control is performed in netlink_poll(). Dumps are allowed to continue if at least half of the ring frames are unused. Signed-off-by: Patrick McHardy <kaber@trash.net> Signed-off-by: David S. Miller <davem@davemloft.net>
This commit is contained in:
parent
f9c2288837
commit
cd1df525da
|
@ -3,6 +3,7 @@
|
||||||
*
|
*
|
||||||
* Authors: Alan Cox <alan@lxorguk.ukuu.org.uk>
|
* Authors: Alan Cox <alan@lxorguk.ukuu.org.uk>
|
||||||
* Alexey Kuznetsov <kuznet@ms2.inr.ac.ru>
|
* Alexey Kuznetsov <kuznet@ms2.inr.ac.ru>
|
||||||
|
* Patrick McHardy <kaber@trash.net>
|
||||||
*
|
*
|
||||||
* This program is free software; you can redistribute it and/or
|
* This program is free software; you can redistribute it and/or
|
||||||
* modify it under the terms of the GNU General Public License
|
* modify it under the terms of the GNU General Public License
|
||||||
|
@ -110,6 +111,29 @@ static inline struct hlist_head *nl_portid_hashfn(struct nl_portid_hash *hash, u
|
||||||
return &hash->table[jhash_1word(portid, hash->rnd) & hash->mask];
|
return &hash->table[jhash_1word(portid, hash->rnd) & hash->mask];
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static void netlink_overrun(struct sock *sk)
|
||||||
|
{
|
||||||
|
struct netlink_sock *nlk = nlk_sk(sk);
|
||||||
|
|
||||||
|
if (!(nlk->flags & NETLINK_RECV_NO_ENOBUFS)) {
|
||||||
|
if (!test_and_set_bit(NETLINK_CONGESTED, &nlk_sk(sk)->state)) {
|
||||||
|
sk->sk_err = ENOBUFS;
|
||||||
|
sk->sk_error_report(sk);
|
||||||
|
}
|
||||||
|
}
|
||||||
|
atomic_inc(&sk->sk_drops);
|
||||||
|
}
|
||||||
|
|
||||||
|
static void netlink_rcv_wake(struct sock *sk)
|
||||||
|
{
|
||||||
|
struct netlink_sock *nlk = nlk_sk(sk);
|
||||||
|
|
||||||
|
if (skb_queue_empty(&sk->sk_receive_queue))
|
||||||
|
clear_bit(NETLINK_CONGESTED, &nlk->state);
|
||||||
|
if (!test_bit(NETLINK_CONGESTED, &nlk->state))
|
||||||
|
wake_up_interruptible(&nlk->wait);
|
||||||
|
}
|
||||||
|
|
||||||
#ifdef CONFIG_NETLINK_MMAP
|
#ifdef CONFIG_NETLINK_MMAP
|
||||||
static bool netlink_skb_is_mmaped(const struct sk_buff *skb)
|
static bool netlink_skb_is_mmaped(const struct sk_buff *skb)
|
||||||
{
|
{
|
||||||
|
@ -441,15 +465,48 @@ static void netlink_forward_ring(struct netlink_ring *ring)
|
||||||
} while (ring->head != head);
|
} while (ring->head != head);
|
||||||
}
|
}
|
||||||
|
|
||||||
|
static bool netlink_dump_space(struct netlink_sock *nlk)
|
||||||
|
{
|
||||||
|
struct netlink_ring *ring = &nlk->rx_ring;
|
||||||
|
struct nl_mmap_hdr *hdr;
|
||||||
|
unsigned int n;
|
||||||
|
|
||||||
|
hdr = netlink_current_frame(ring, NL_MMAP_STATUS_UNUSED);
|
||||||
|
if (hdr == NULL)
|
||||||
|
return false;
|
||||||
|
|
||||||
|
n = ring->head + ring->frame_max / 2;
|
||||||
|
if (n > ring->frame_max)
|
||||||
|
n -= ring->frame_max;
|
||||||
|
|
||||||
|
hdr = __netlink_lookup_frame(ring, n);
|
||||||
|
|
||||||
|
return hdr->nm_status == NL_MMAP_STATUS_UNUSED;
|
||||||
|
}
|
||||||
|
|
||||||
static unsigned int netlink_poll(struct file *file, struct socket *sock,
|
static unsigned int netlink_poll(struct file *file, struct socket *sock,
|
||||||
poll_table *wait)
|
poll_table *wait)
|
||||||
{
|
{
|
||||||
struct sock *sk = sock->sk;
|
struct sock *sk = sock->sk;
|
||||||
struct netlink_sock *nlk = nlk_sk(sk);
|
struct netlink_sock *nlk = nlk_sk(sk);
|
||||||
unsigned int mask;
|
unsigned int mask;
|
||||||
|
int err;
|
||||||
|
|
||||||
if (nlk->cb != NULL && nlk->rx_ring.pg_vec != NULL)
|
if (nlk->rx_ring.pg_vec != NULL) {
|
||||||
netlink_dump(sk);
|
/* Memory mapped sockets don't call recvmsg(), so flow control
|
||||||
|
* for dumps is performed here. A dump is allowed to continue
|
||||||
|
* if at least half the ring is unused.
|
||||||
|
*/
|
||||||
|
while (nlk->cb != NULL && netlink_dump_space(nlk)) {
|
||||||
|
err = netlink_dump(sk);
|
||||||
|
if (err < 0) {
|
||||||
|
sk->sk_err = err;
|
||||||
|
sk->sk_error_report(sk);
|
||||||
|
break;
|
||||||
|
}
|
||||||
|
}
|
||||||
|
netlink_rcv_wake(sk);
|
||||||
|
}
|
||||||
|
|
||||||
mask = datagram_poll(file, sock, wait);
|
mask = datagram_poll(file, sock, wait);
|
||||||
|
|
||||||
|
@ -623,8 +680,7 @@ static void netlink_ring_set_copied(struct sock *sk, struct sk_buff *skb)
|
||||||
if (hdr == NULL) {
|
if (hdr == NULL) {
|
||||||
spin_unlock_bh(&sk->sk_receive_queue.lock);
|
spin_unlock_bh(&sk->sk_receive_queue.lock);
|
||||||
kfree_skb(skb);
|
kfree_skb(skb);
|
||||||
sk->sk_err = ENOBUFS;
|
netlink_overrun(sk);
|
||||||
sk->sk_error_report(sk);
|
|
||||||
return;
|
return;
|
||||||
}
|
}
|
||||||
netlink_increment_head(ring);
|
netlink_increment_head(ring);
|
||||||
|
@ -1329,19 +1385,6 @@ static int netlink_getname(struct socket *sock, struct sockaddr *addr,
|
||||||
return 0;
|
return 0;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void netlink_overrun(struct sock *sk)
|
|
||||||
{
|
|
||||||
struct netlink_sock *nlk = nlk_sk(sk);
|
|
||||||
|
|
||||||
if (!(nlk->flags & NETLINK_RECV_NO_ENOBUFS)) {
|
|
||||||
if (!test_and_set_bit(NETLINK_CONGESTED, &nlk_sk(sk)->state)) {
|
|
||||||
sk->sk_err = ENOBUFS;
|
|
||||||
sk->sk_error_report(sk);
|
|
||||||
}
|
|
||||||
}
|
|
||||||
atomic_inc(&sk->sk_drops);
|
|
||||||
}
|
|
||||||
|
|
||||||
static struct sock *netlink_getsockbyportid(struct sock *ssk, u32 portid)
|
static struct sock *netlink_getsockbyportid(struct sock *ssk, u32 portid)
|
||||||
{
|
{
|
||||||
struct sock *sock;
|
struct sock *sock;
|
||||||
|
@ -1484,16 +1527,6 @@ static struct sk_buff *netlink_trim(struct sk_buff *skb, gfp_t allocation)
|
||||||
return skb;
|
return skb;
|
||||||
}
|
}
|
||||||
|
|
||||||
static void netlink_rcv_wake(struct sock *sk)
|
|
||||||
{
|
|
||||||
struct netlink_sock *nlk = nlk_sk(sk);
|
|
||||||
|
|
||||||
if (skb_queue_empty(&sk->sk_receive_queue))
|
|
||||||
clear_bit(NETLINK_CONGESTED, &nlk->state);
|
|
||||||
if (!test_bit(NETLINK_CONGESTED, &nlk->state))
|
|
||||||
wake_up_interruptible(&nlk->wait);
|
|
||||||
}
|
|
||||||
|
|
||||||
static int netlink_unicast_kernel(struct sock *sk, struct sk_buff *skb,
|
static int netlink_unicast_kernel(struct sock *sk, struct sk_buff *skb,
|
||||||
struct sock *ssk)
|
struct sock *ssk)
|
||||||
{
|
{
|
||||||
|
@ -1597,6 +1630,7 @@ struct sk_buff *netlink_alloc_skb(struct sock *ssk, unsigned int size,
|
||||||
err2:
|
err2:
|
||||||
kfree_skb(skb);
|
kfree_skb(skb);
|
||||||
spin_unlock_bh(&sk->sk_receive_queue.lock);
|
spin_unlock_bh(&sk->sk_receive_queue.lock);
|
||||||
|
netlink_overrun(sk);
|
||||||
err1:
|
err1:
|
||||||
sock_put(sk);
|
sock_put(sk);
|
||||||
return NULL;
|
return NULL;
|
||||||
|
|
Loading…
Reference in New Issue