[<prev] [next>] [<thread-prev] [thread-next>] [day] [month] [year] [list]
Message-ID: <1272293707.19143.51.camel@edumazet-laptop>
Date: Mon, 26 Apr 2010 16:55:07 +0200
From: Eric Dumazet <eric.dumazet@...il.com>
To: hadi@...erus.ca
Cc: Changli Gao <xiaosuo@...il.com>,
"David S. Miller" <davem@...emloft.net>,
Tom Herbert <therbert@...gle.com>,
Stephen Hemminger <shemminger@...tta.com>,
netdev@...r.kernel.org, Andi Kleen <andi@...stfloor.org>
Subject: Re: [PATCH v6] net: batch skb dequeueing from softnet
input_pkt_queue
Le lundi 26 avril 2010 à 16:03 +0200, Eric Dumazet a écrit :
> Le samedi 24 avril 2010 à 10:10 -0400, jamal a écrit :
> > On Fri, 2010-04-23 at 18:02 -0400, jamal wrote:
> >
> > > Ive done a setup with the last patch from Changli + net-next - I will
> > > post test results tomorrow AM.
> >
> > ok, annotated results attached.
> >
> > cheers,
> > jamal
>
> Jamal, I have a Nehalem setup now, and I can see
> _raw_spin_lock_irqsave() abuse is not coming from network tree, but from
> clockevents_notify()
>
Another interesting finding:
- if all packets are received on a single queue, max speed seems to be
1.200.000 packets per second on my machine :-(
And on profile of receiving cpu (RPS enabled, pakets sent to 15 other
cpus), we can see default_send_IPI_mask_sequence_phys() is the slow
thing...
Andi, what do you think of this one ?
Dont we have a function to send an IPI to an individual cpu instead ?
void default_send_IPI_mask_sequence_phys(const struct cpumask *mask, int
vector)
{
unsigned long query_cpu;
unsigned long flags;
/*
* Hack. The clustered APIC addressing mode doesn't allow us to
send
* to an arbitrary mask, so I do a unicast to each CPU instead.
* - mbligh
*/
local_irq_save(flags);
for_each_cpu(query_cpu, mask) {
__default_send_IPI_dest_field(per_cpu(x86_cpu_to_apicid,
query_cpu), vector, APIC_DEST_PHYSICAL);
}
local_irq_restore(flags);
}
-----------------------------------------------------------------------------------------------------------------------------------------
PerfTop: 1000 irqs/sec kernel:100.0% [1000Hz cycles], (all, cpu:
7)
-----------------------------------------------------------------------------------------------------------------------------------------
samples pcnt function DSO
_______ _____ ___________________________________ _______
668.00 17.7% default_send_IPI_mask_sequence_phys vmlinux
363.00 9.6% bnx2x_rx_int vmlinux
354.00 9.4% eth_type_trans vmlinux
332.00 8.8% kmem_cache_alloc_node vmlinux
285.00 7.6% __kmalloc_node_track_caller vmlinux
278.00 7.4% _raw_spin_lock vmlinux
166.00 4.4% __slab_alloc vmlinux
147.00 3.9% __memset vmlinux
136.00 3.6% list_del vmlinux
132.00 3.5% get_partial_node vmlinux
131.00 3.5% get_rps_cpu vmlinux
102.00 2.7% enqueue_to_backlog vmlinux
95.00 2.5% unmap_single vmlinux
94.00 2.5% __alloc_skb vmlinux
74.00 2.0% vlan_gro_common vmlinux
52.00 1.4% __phys_addr vmlinux
48.00 1.3% dev_gro_receive vmlinux
39.00 1.0% swiotlb_dma_mapping_error vmlinux
36.00 1.0% swiotlb_map_page vmlinux
34.00 0.9% skb_put vmlinux
27.00 0.7% is_swiotlb_buffer vmlinux
23.00 0.6% deactivate_slab vmlinux
20.00 0.5% vlan_gro_receive vmlinux
17.00 0.5% __skb_bond_should_drop vmlinux
14.00 0.4% netif_receive_skb vmlinux
14.00 0.4% __netdev_alloc_skb vmlinux
12.00 0.3% skb_gro_reset_offset vmlinux
12.00 0.3% get_slab vmlinux
11.00 0.3% napi_skb_finish vmlinux
--
To unsubscribe from this list: send the line "unsubscribe netdev" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at http://vger.kernel.org/majordomo-info.html
Powered by blists - more mailing lists