This patch introduces memory usage accounting for UDP.

signed-off-by: Satoshi Oshima <[EMAIL PROTECTED]>
signed-off-by: Hideo Aoki <[EMAIL PROTECTED]>

Index: 2.6.23-rc8-udp_limit/net/ipv4/ip_output.c
===================================================================
--- 2.6.23-rc8-udp_limit.orig/net/ipv4/ip_output.c
+++ 2.6.23-rc8-udp_limit/net/ipv4/ip_output.c
@@ -743,6 +743,8 @@ static inline int ip_ufo_append_data(str
                /* specify the length of each IP datagram fragment*/
                skb_shinfo(skb)->gso_size = mtu - fragheaderlen;
                skb_shinfo(skb)->gso_type = SKB_GSO_UDP;
+               atomic_add(sk_datagram_pages(skb->truesize),
+                          sk->sk_prot->memory_allocated);
                __skb_queue_tail(&sk->sk_write_queue, skb);
 
                return 0;
@@ -924,6 +926,9 @@ alloc_new_skb:
                        }
                        if (skb == NULL)
                                goto error;
+                       if (sk->sk_prot->memory_allocated)
+                               atomic_add(sk_datagram_pages(skb->truesize),
+                                          sk->sk_prot->memory_allocated);
 
                        /*
                         *      Fill in the control structures
@@ -1023,6 +1028,8 @@ alloc_new_skb:
                                frag = &skb_shinfo(skb)->frags[i];
                                skb->truesize += PAGE_SIZE;
                                atomic_add(PAGE_SIZE, &sk->sk_wmem_alloc);
+                               if (sk->sk_prot->memory_allocated)
+                                       
atomic_inc(sk->sk_prot->memory_allocated);
                        } else {
                                err = -EMSGSIZE;
                                goto error;
@@ -1123,7 +1130,9 @@ ssize_t   ip_append_page(struct sock *sk, 
                        if (unlikely(!skb)) {
                                err = -ENOBUFS;
                                goto error;
-                       }
+                       } else if (sk->sk_prot->memory_allocated)
+                               atomic_add(sk_datagram_pages(skb->truesize),
+                                          sk->sk_prot->memory_allocated);
 
                        /*
                         *      Fill in the control structures
@@ -1202,13 +1211,14 @@ int ip_push_pending_frames(struct sock *
        struct iphdr *iph;
        __be16 df = 0;
        __u8 ttl;
-       int err = 0;
+       int err = 0, send_page_size;
 
        if ((skb = __skb_dequeue(&sk->sk_write_queue)) == NULL)
                goto out;
        tail_skb = &(skb_shinfo(skb)->frag_list);
 
        /* move skb->data to ip header from ext header */
+       send_page_size = sk_datagram_pages(skb->truesize);
        if (skb->data < skb_network_header(skb))
                __skb_pull(skb, skb_network_offset(skb));
        while ((tmp_skb = __skb_dequeue(&sk->sk_write_queue)) != NULL) {
@@ -1218,6 +1228,7 @@ int ip_push_pending_frames(struct sock *
                skb->len += tmp_skb->len;
                skb->data_len += tmp_skb->len;
                skb->truesize += tmp_skb->truesize;
+               send_page_size += sk_datagram_pages(tmp_skb->truesize);
                __sock_put(tmp_skb->sk);
                tmp_skb->destructor = NULL;
                tmp_skb->sk = NULL;
@@ -1269,6 +1280,8 @@ int ip_push_pending_frames(struct sock *
        /* Netfilter gets whole the not fragmented skb. */
        err = NF_HOOK(PF_INET, NF_IP_LOCAL_OUT, skb, NULL,
                      skb->dst->dev, dst_output);
+       if (sk->sk_prot->memory_allocated)
+               atomic_sub(send_page_size, sk->sk_prot->memory_allocated);
        if (err) {
                if (err > 0)
                        err = inet->recverr ? net_xmit_errno(err) : 0;
@@ -1298,9 +1311,15 @@ void ip_flush_pending_frames(struct sock
 {
        struct inet_sock *inet = inet_sk(sk);
        struct sk_buff *skb;
+       int num_flush_mem = 0;
 
-       while ((skb = __skb_dequeue_tail(&sk->sk_write_queue)) != NULL)
+       while ((skb = __skb_dequeue_tail(&sk->sk_write_queue)) != NULL) {
+               num_flush_mem += sk_datagram_pages(skb->truesize);
                kfree_skb(skb);
+       }
+
+       if (sk->sk_prot->memory_allocated)
+               atomic_sub(num_flush_mem, sk->sk_prot->memory_allocated);
 
        inet->cork.flags &= ~IPCORK_OPT;
        kfree(inet->cork.opt);
Index: 2.6.23-rc8-udp_limit/net/ipv4/udp.c
===================================================================
--- 2.6.23-rc8-udp_limit.orig/net/ipv4/udp.c
+++ 2.6.23-rc8-udp_limit/net/ipv4/udp.c
@@ -887,6 +887,9 @@ try_again:
                err = ulen;
 
 out_free:
+       atomic_sub(sk_datagram_pages(skb->truesize),
+                  sk->sk_prot->memory_allocated);
+
        skb_free_datagram(sk, skb);
 out:
        return err;
@@ -894,6 +897,9 @@ out:
 csum_copy_err:
        UDP_INC_STATS_BH(UDP_MIB_INERRORS, is_udplite);
 
+       atomic_sub(sk_datagram_pages(skb->truesize),
+                  sk->sk_prot->memory_allocated);
+
        skb_kill_datagram(sk, skb, flags);
 
        if (noblock)
@@ -1019,6 +1025,9 @@ int udp_queue_rcv_skb(struct sock * sk, 
                goto drop;
        }
 
+       atomic_add(sk_datagram_pages(skb->truesize),
+                  sk->sk_prot->memory_allocated);
+
        UDP_INC_STATS_BH(UDP_MIB_INDATAGRAMS, up->pcflag);
        return 0;
 
@@ -1443,6 +1452,7 @@ struct proto udp_prot = {
        .hash              = udp_lib_hash,
        .unhash            = udp_lib_unhash,
        .get_port          = udp_v4_get_port,
+       .memory_allocated  = &udp_memory_allocated,
        .obj_size          = sizeof(struct udp_sock),
 #ifdef CONFIG_COMPAT
        .compat_setsockopt = compat_udp_setsockopt,
Index: 2.6.23-rc8-udp_limit/net/ipv4/af_inet.c
===================================================================
--- 2.6.23-rc8-udp_limit.orig/net/ipv4/af_inet.c
+++ 2.6.23-rc8-udp_limit/net/ipv4/af_inet.c
@@ -126,13 +126,41 @@ extern void ip_mc_drop_socket(struct soc
 static struct list_head inetsw[SOCK_MAX];
 static DEFINE_SPINLOCK(inetsw_lock);
 
+/**
+ *     __skb_queue_purge_and_sub_memory_allocated
+ *             - empty a list and subtruct memory allocation counter
+ *     @sk:   sk
+ *     @list: list to empty
+ *     Delete all buffers on an &sk_buff list and subtruct the
+ *     turesize of the sk_buff for memory accounting. Each buffer
+ *     is removed from the list and one reference dropped. This
+ *     function does not take the list lock and the caller must
+ *     hold the relevant locks to use it.
+ */
+void __skb_queue_purge_and_sub_memory_allocated(struct sock *sk,
+                                       struct sk_buff_head *list)
+{
+       struct sk_buff *skb;
+       int purged_skb_size = 0;
+       while ((skb = __skb_dequeue(list)) != NULL) {
+               purged_skb_size += sk_datagram_pages(skb->truesize);
+               kfree_skb(skb);
+       }
+       atomic_sub(purged_skb_size, sk->sk_prot->memory_allocated);
+}
+
 /* New destruction routine */
 
 void inet_sock_destruct(struct sock *sk)
 {
        struct inet_sock *inet = inet_sk(sk);
 
-       __skb_queue_purge(&sk->sk_receive_queue);
+       if (sk->sk_prot->memory_allocated && sk->sk_type != SOCK_STREAM)
+               __skb_queue_purge_and_sub_memory_allocated(sk,
+                               &sk->sk_receive_queue);
+       else
+               __skb_queue_purge(&sk->sk_receive_queue);
+
        __skb_queue_purge(&sk->sk_error_queue);
 
        if (sk->sk_type == SOCK_STREAM && sk->sk_state != TCP_CLOSE) {

-
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to [EMAIL PROTECTED]
More majordomo info at  http://vger.kernel.org/majordomo-info.html

Reply via email to