lists.openwall.net   lists  /  announce  owl-users  owl-dev  john-users  john-dev  passwdqc-users  yescrypt  popa3d-users  /  oss-security  kernel-hardening  musl  sabotage  tlsify  passwords  /  crypt-dev  xvendor  /  Bugtraq  Full-Disclosure  linux-kernel  linux-netdev  linux-ext4  linux-hardening  linux-cve-announce  PHC 
Open Source and information security mailing list archives
 
Hash Suite: Windows password security audit tool. GUI, reports in PDF.
[<prev] [next>] [day] [month] [year] [list]
Date:	Thu, 14 Sep 2006 23:56:02 +0800
From:	Limeng [李萌] <avlimeng@...gsoft.net>
To:	<linux-kernel@...r.kernel.org>
Cc:	'金山郭小花' <guoxiaohua@...gsoft.com>,
	<limeng@...gsoft.com>
Subject: UDP Questions

Hi:
	I have a very puzzled question. When I test the limits of my network
adapters, I send many small UDP packets and compute the average packets sent
per second. Use those codes, I get the result: 75000 packets per second.
(All my sockets has set to unblock)
	//////
	socket = socket(AF_INET, SOCK_DGRAM, .....)
	bind....
	time = time(NULL);
	while (1)
	{
	sendto(socket, "", 1, 0, dstaddr, addrlen);
	count++;
}
time = time(NULL) - time;
avg = count / time; // here, I get avg = 75000
//////

This result can't satisfy me, so I add another network adapter, and try the
code blow:

//////
	socket1 = socket(AF_INET, SOCK_DGRAM, .....)
	socket2 = socket(AF_INET, SOCK_DGRAM, .....)
	bind socket1.network adapter1...
	bind socket2 network adapter2
	time = time(NULL);
	while (1)
	{
	sendto(socket1, "", 1, 0, dstaddr1, addrlen);
	sendto(socket2, "", 1, 0, dstaddr2, addrlen);
	count += 2;
}
time = time(NULL) - time;
avg = count / time; 
///////////////////

But I get the result is also 75000 packet per second, WHY?

Then I find a technique name bond, and then I configure my server's two
network adapters into a dev bond0. And then I test again with the first
paragraph of the code, then I get result 150000 packets per second. So I
want to know how bond can increate the speed of network adapters (use my own
code to send with the two network adapters at the same time is not helpful),
then I open the kernel code, and then find such code:
bond_main.c  L3861:
		if (IS_UP(slave->dev) &&
		    (slave->link == BOND_LINK_UP) &&
		    (slave->state == BOND_STATE_ACTIVE)) {
			res = bond_dev_queue_xmit(bond, skb, slave->dev);

			write_lock(&bond->curr_slave_lock);
			bond->curr_active_slave = slave->next;
			write_unlock(&bond->curr_slave_lock);

			break;
		}
It look like that when send a packet to bond dev, bond use current slave and
send packet, then change current slave to next. What is the essence
different between the bond and my code (use two network adapters)?

Any suggestions?

xixi

-
To unsubscribe from this list: send the line "unsubscribe linux-kernel" in
the body of a message to majordomo@...r.kernel.org
More majordomo info at  http://vger.kernel.org/majordomo-info.html
Please read the FAQ at  http://www.tux.org/lkml/

Powered by blists - more mailing lists

Powered by Openwall GNU/*/Linux Powered by OpenVZ