Using __get_free_pages() has some drawbacks :

1) Not NUMA aware

2) 2^X page granularity :
On arches with big PAGE_SIZE, we waste some ram for each cpu.

(We currently use 1024 pointers, that is at most 8192 bytes, but
PAGE_SIZE can be 65536 for example : With say 64 possible cpus,
thats about 56*64 Kbytes that are wasted)

Using kmalloc_node() can help to solve these two points.

Signed-off-by: Eric Dumazet <[EMAIL PROTECTED]>

 net/core/flow.c |   15 +++++----------
 1 files changed, 5 insertions(+), 10 deletions(-)

diff --git a/net/core/flow.c b/net/core/flow.c
index a618f89..bcbbe2d 100644
--- a/net/core/flow.c
+++ b/net/core/flow.c
@@ -311,18 +311,13 @@ void flow_cache_flush(void)
 static void __devinit flow_cache_cpu_prepare(int cpu)
 {
        struct tasklet_struct *tasklet;
-       unsigned long order;
 
-       for (order = 0;
-            (PAGE_SIZE << order) <
-                    (sizeof(struct flow_cache_entry *)*flow_hash_size);
-            order++)
-               /* NOTHING */;
-
-       flow_table(cpu) = (struct flow_cache_entry **)
-               __get_free_pages(GFP_KERNEL|__GFP_ZERO, order);
+       flow_table(cpu) = kmalloc_node(
+               sizeof(struct flow_cache_entry *) * flow_hash_size,
+               GFP_KERNEL | __GFP_ZERO,
+               cpu_to_node(cpu));
        if (!flow_table(cpu))
-               panic("NET: failed to allocate flow cache order %lu\n", order);
+               panic("NET: failed to allocate flow cache for cpu %d\n", cpu);
 
        flow_hash_rnd_recalc(cpu) = 1;
        flow_count(cpu) = 0;

Reply via email to